| /Linux-v5.4/drivers/gpu/drm/amd/amdkfd/ | 
| D | kfd_kernel_queue.c | 38 		enum kfd_queue_type type, unsigned int queue_size)  in initialize()  argument48 			queue_size);  in initialize()
 81 	retval = kfd_gtt_sa_allocate(dev, queue_size, &kq->pq);  in initialize()
 83 		pr_err("Failed to init pq queues size %d\n", queue_size);  in initialize()
 90 	retval = kq->ops_asic_specific.initialize(kq, dev, type, queue_size);  in initialize()
 112 	memset(kq->pq_kernel_addr, 0, queue_size);  in initialize()
 116 	prop.queue_size = queue_size;  in initialize()
 228 	queue_size_dwords = kq->queue->properties.queue_size / 4;  in acquire_packet_buffer()
 292 			(kq->queue->properties.queue_size / 4);  in rollback_packet()
 
 | 
| D | kfd_kernel_queue_cik.c | 27 			enum kfd_queue_type type, unsigned int queue_size);39 			enum kfd_queue_type type, unsigned int queue_size)  in initialize_cik()  argument
 
 | 
| D | kfd_queue.c | 34 	pr_debug("Queue Size: %llu\n", q->queue_size);  in print_queue_properties()51 	pr_debug("Queue Size: %llu\n", q->properties.queue_size);  in print_queue()
 
 | 
| D | kfd_mqd_manager_cik.c | 171 	uint32_t wptr_mask = (uint32_t)((p->queue_size / 4) - 1);  in load_mqd()205 	m->cp_hqd_pq_control |= order_base_2(q->queue_size / 4) - 1;  in __update_mqd()
 241 	m->sdma_rlc_rb_cntl = order_base_2(q->queue_size / 4)  in update_mqd_sdma()
 328 	m->cp_hqd_pq_control |= order_base_2(q->queue_size / 4) - 1;  in update_mqd_hiq()
 
 | 
| D | kfd_mqd_manager_v10.c | 184 			ffs(q->queue_size / sizeof(unsigned int)) - 1 - 1;  in update_mqd()234 	q->is_active = (q->queue_size > 0 &&  in update_mqd()
 355 	m->sdmax_rlcx_rb_cntl = (ffs(q->queue_size / sizeof(unsigned int)) - 1)  in update_mqd_sdma()
 373 	q->is_active = (q->queue_size > 0 &&  in update_mqd_sdma()
 
 | 
| D | kfd_kernel_queue.h | 57 			enum kfd_queue_type type, unsigned int queue_size);
 | 
| D | kfd_mqd_manager_vi.c | 162 	uint32_t wptr_mask = (uint32_t)((p->queue_size / 4) - 1);  in load_mqd()180 	m->cp_hqd_pq_control |=	order_base_2(q->queue_size / 4) - 1;  in __update_mqd()
 354 	m->sdmax_rlcx_rb_cntl = order_base_2(q->queue_size / 4)  in update_mqd_sdma()
 
 | 
| D | kfd_kernel_queue_v10.c | 31 			enum kfd_queue_type type, unsigned int queue_size);43 			enum kfd_queue_type type, unsigned int queue_size)  in initialize_v10()  argument
 
 | 
| /Linux-v5.4/drivers/firmware/tegra/ | 
| D | ivc.c | 549 unsigned tegra_ivc_total_queue_size(unsigned queue_size)  in tegra_ivc_total_queue_size()  argument551 	if (!IS_ALIGNED(queue_size, TEGRA_IVC_ALIGN)) {  in tegra_ivc_total_queue_size()
 553 		       __func__, queue_size, TEGRA_IVC_ALIGN);  in tegra_ivc_total_queue_size()
 557 	return queue_size + sizeof(struct tegra_ivc_header);  in tegra_ivc_total_queue_size()
 618 	size_t queue_size;  in tegra_ivc_init()  local
 636 	queue_size = tegra_ivc_total_queue_size(num_frames * frame_size);  in tegra_ivc_init()
 639 		ivc->rx.phys = dma_map_single(peer, rx, queue_size,  in tegra_ivc_init()
 644 		ivc->tx.phys = dma_map_single(peer, tx, queue_size,  in tegra_ivc_init()
 647 			dma_unmap_single(peer, ivc->rx.phys, queue_size,  in tegra_ivc_init()
 
 | 
| D | bpmp-tegra186.c | 111 	size_t message_size, queue_size;  in tegra186_bpmp_channel_init()  local121 	queue_size = tegra_ivc_total_queue_size(message_size);  in tegra186_bpmp_channel_init()
 122 	offset = queue_size * index;  in tegra186_bpmp_channel_init()
 
 | 
| /Linux-v5.4/drivers/net/wireless/intel/iwlwifi/pcie/ | 
| D | rx.c | 179 	WARN_ON(rxq->queue_size & (rxq->queue_size - 1));  in iwl_rxq_space()187 	return (rxq->read - rxq->write - 1) & (rxq->queue_size - 1);  in iwl_rxq_space()
 699 				  free_size * rxq->queue_size,  in iwl_pcie_free_rxq_dma()
 710 				   sizeof(__le32)) * rxq->queue_size,  in iwl_pcie_free_rxq_dma()
 745 		rxq->queue_size = MQ_RX_TABLE_SIZE;  in iwl_pcie_alloc_rxq_dma()
 747 		rxq->queue_size = RX_QUEUE_SIZE;  in iwl_pcie_alloc_rxq_dma()
 755 	rxq->bd = dma_alloc_coherent(dev, free_size * rxq->queue_size,  in iwl_pcie_alloc_rxq_dma()
 762 						  (use_rx_td ? sizeof(*rxq->cd) : sizeof(__le32)) * rxq->queue_size,  in iwl_pcie_alloc_rxq_dma()
 1041 	int i, err, queue_size, allocator_pool_size, num_alloc;  in _iwl_pcie_rx_init()  local
 1091 	queue_size = trans->trans_cfg->mq_rx_supported ?  in _iwl_pcie_rx_init()
 [all …]
 
 | 
| D | trans-gen2.c | 259 	int queue_size = max_t(u32, IWL_CMD_QUEUE_SIZE,  in iwl_pcie_gen2_nic_init()  local274 	if (iwl_pcie_gen2_tx_init(trans, trans_pcie->cmd_queue, queue_size))  in iwl_pcie_gen2_nic_init()
 
 | 
| /Linux-v5.4/include/linux/ | 
| D | vmw_vmci_defs.h | 804 				u64 queue_size)  in vmci_q_header_add_producer_tail()  argument806 	vmci_qp_add_pointer(&q_header->producer_tail, add, queue_size);  in vmci_q_header_add_producer_tail()
 816 				u64 queue_size)  in vmci_q_header_add_consumer_head()  argument
 818 	vmci_qp_add_pointer(&q_header->consumer_head, add, queue_size);  in vmci_q_header_add_consumer_head()
 
 | 
| /Linux-v5.4/kernel/bpf/ | 
| D | queue_stack_maps.c | 72 	u64 size, queue_size, cost;  in queue_stack_map_alloc()  local75 	cost = queue_size = sizeof(*qs) + size * attr->value_size;  in queue_stack_map_alloc()
 81 	qs = bpf_map_area_alloc(queue_size, numa_node);  in queue_stack_map_alloc()
 
 | 
| /Linux-v5.4/sound/core/ | 
| D | timer.c | 55 	int queue_size;  member1292 		prev = tu->qtail == 0 ? tu->queue_size - 1 : tu->qtail - 1;  in snd_timer_user_interrupt()
 1299 	if (tu->qused >= tu->queue_size) {  in snd_timer_user_interrupt()
 1303 		tu->qtail %= tu->queue_size;  in snd_timer_user_interrupt()
 1317 	if (tu->qused >= tu->queue_size) {  in snd_timer_user_append_to_tqueue()
 1321 		tu->qtail %= tu->queue_size;  in snd_timer_user_append_to_tqueue()
 1396 		prev = tu->qtail == 0 ? tu->queue_size - 1 : tu->qtail - 1;  in snd_timer_user_tinterrupt()
 1436 	tu->queue_size = size;  in realloc_user_queue()
 1805 	if (params.queue_size > 0 &&  in snd_timer_user_params()
 1806 	    (params.queue_size < 32 || params.queue_size > 1024)) {  in snd_timer_user_params()
 [all …]
 
 | 
| /Linux-v5.4/drivers/nvme/host/ | 
| D | rdma.c | 77 	int			queue_size;  member263 	init_attr.cap.max_send_wr = factor * queue->queue_size + 1;  in nvme_rdma_create_qp()
 265 	init_attr.cap.max_recv_wr = queue->queue_size + 1;  in nvme_rdma_create_qp()
 421 	nvme_rdma_free_ring(ibdev, queue->rsp_ring, queue->queue_size,  in nvme_rdma_destroy_queue_ib()
 464 				cq_factor * queue->queue_size + 1,  in nvme_rdma_create_queue_ib()
 475 	queue->rsp_ring = nvme_rdma_alloc_ring(ibdev, queue->queue_size,  in nvme_rdma_create_queue_ib()
 489 			      queue->queue_size,  in nvme_rdma_create_queue_ib()
 495 			queue->queue_size, idx);  in nvme_rdma_create_queue_ib()
 504 	nvme_rdma_free_ring(ibdev, queue->rsp_ring, queue->queue_size,  in nvme_rdma_create_queue_ib()
 516 		int idx, size_t queue_size)  in nvme_rdma_alloc_queue()  argument
 [all …]
 
 | 
| /Linux-v5.4/drivers/usb/host/ | 
| D | u132-hcd.c | 163 	u16 queue_size;  member517 	if (ENDP_QUEUE_SIZE > --endp->queue_size) {  in u132_hcd_giveback_urb()
 556 	if (ENDP_QUEUE_SIZE > --endp->queue_size) {  in u132_hcd_abandon_urb()
 1915 	endp->queue_size = 1;  in create_endpoint_and_queue_int()
 1932 	if (endp->queue_size++ < ENDP_QUEUE_SIZE) {  in queue_int_on_old_endpoint()
 1938 			endp->queue_size -= 1;  in queue_int_on_old_endpoint()
 2014 	endp->queue_size = 1;  in create_endpoint_and_queue_bulk()
 2029 	if (endp->queue_size++ < ENDP_QUEUE_SIZE) {  in queue_bulk_on_old_endpoint()
 2035 			endp->queue_size -= 1;  in queue_bulk_on_old_endpoint()
 2103 		endp->queue_size = 1;  in create_endpoint_and_queue_control()
 [all …]
 
 | 
| /Linux-v5.4/drivers/misc/genwqe/ | 
| D | card_ddcb.c | 1019 	unsigned int queue_size;  in setup_ddcb_queue()  local1025 	queue_size = roundup(GENWQE_DDCB_MAX * sizeof(struct ddcb), PAGE_SIZE);  in setup_ddcb_queue()
 1035 	queue->ddcb_vaddr = __genwqe_alloc_consistent(cd, queue_size,  in setup_ddcb_queue()
 1083 	__genwqe_free_consistent(cd, queue_size, queue->ddcb_vaddr,  in setup_ddcb_queue()
 1098 	unsigned int queue_size;  in free_ddcb_queue()  local
 1100 	queue_size = roundup(queue->ddcb_max * sizeof(struct ddcb), PAGE_SIZE);  in free_ddcb_queue()
 1106 		__genwqe_free_consistent(cd, queue_size, queue->ddcb_vaddr,  in free_ddcb_queue()
 
 | 
| /Linux-v5.4/include/soc/tegra/ | 
| D | ivc.h | 93 unsigned tegra_ivc_total_queue_size(unsigned queue_size);
 | 
| /Linux-v5.4/drivers/nvme/target/ | 
| D | loop.c | 515 	ctrl->tag_set.queue_depth = ctrl->ctrl.opts->queue_size;  in nvme_loop_create_io_queues()590 	ctrl->ctrl.sqsize = opts->queue_size - 1;  in nvme_loop_create_ctrl()
 603 	if (opts->queue_size > ctrl->ctrl.maxcmd) {  in nvme_loop_create_ctrl()
 607 			opts->queue_size, ctrl->ctrl.maxcmd);  in nvme_loop_create_ctrl()
 608 		opts->queue_size = ctrl->ctrl.maxcmd;  in nvme_loop_create_ctrl()
 
 | 
| /Linux-v5.4/drivers/platform/chrome/wilco_ec/ | 
| D | event.c | 103 static int queue_size = 64;  variable104 module_param(queue_size, int, 0644);
 470 	dev_data->events = event_queue_new(queue_size);  in event_device_add()
 
 | 
| /Linux-v5.4/drivers/gpu/drm/amd/amdgpu/ | 
| D | amdgpu_amdkfd_gfx_v9.c | 319 		uint32_t queue_size =  in kgd_gfx_v9_hqd_load()  local322 		uint64_t guessed_wptr = m->cp_hqd_pq_rptr & (queue_size - 1);  in kgd_gfx_v9_hqd_load()
 324 		if ((m->cp_hqd_pq_wptr_lo & (queue_size - 1)) < guessed_wptr)  in kgd_gfx_v9_hqd_load()
 325 			guessed_wptr += queue_size;  in kgd_gfx_v9_hqd_load()
 326 		guessed_wptr += m->cp_hqd_pq_wptr_lo & ~(queue_size - 1);  in kgd_gfx_v9_hqd_load()
 
 | 
| D | amdgpu_amdkfd_gfx_v10.c | 418 		uint32_t queue_size =  in kgd_hqd_load()  local421 		uint64_t guessed_wptr = m->cp_hqd_pq_rptr & (queue_size - 1);  in kgd_hqd_load()
 423 		if ((m->cp_hqd_pq_wptr_lo & (queue_size - 1)) < guessed_wptr)  in kgd_hqd_load()
 424 			guessed_wptr += queue_size;  in kgd_hqd_load()
 425 		guessed_wptr += m->cp_hqd_pq_wptr_lo & ~(queue_size - 1);  in kgd_hqd_load()
 
 | 
| /Linux-v5.4/include/uapi/linux/ | 
| D | virtio_pci.h | 148 	__le16 queue_size;		/* read-write, power of 2. */  member
 | 
| /Linux-v5.4/drivers/dma/ | 
| D | fsl-qdma.c | 475 	unsigned int queue_size[FSL_QDMA_QUEUE_MAX];  in fsl_qdma_alloc_queue_resources()  local489 					     queue_size, queue_num);  in fsl_qdma_alloc_queue_resources()
 496 			if (queue_size[i] > FSL_QDMA_CIRCULAR_DESC_SIZE_MAX ||  in fsl_qdma_alloc_queue_resources()
 497 			    queue_size[i] < FSL_QDMA_CIRCULAR_DESC_SIZE_MIN) {  in fsl_qdma_alloc_queue_resources()
 507 					   queue_size[i],  in fsl_qdma_alloc_queue_resources()
 514 			queue_temp->n_cq = queue_size[i];  in fsl_qdma_alloc_queue_resources()
 
 |