/Linux-v5.10/drivers/mailbox/ |
D | ti-msgmgr.c | 73 u8 queue_count; member 594 if (qinst->queue_id > d->queue_count) { in ti_msgmgr_queue_setup() 596 idx, qinst->queue_id, d->queue_count); in ti_msgmgr_queue_setup() 670 .queue_count = 64, 685 .queue_count = 190, 717 int queue_count; in ti_msgmgr_probe() local 765 queue_count = desc->num_valid_queues; in ti_msgmgr_probe() 766 if (!queue_count || queue_count > desc->queue_count) { in ti_msgmgr_probe() 768 queue_count, desc->queue_count); in ti_msgmgr_probe() 771 inst->num_valid_queues = queue_count; in ti_msgmgr_probe() [all …]
|
/Linux-v5.10/drivers/s390/crypto/ |
D | ap_queue.c | 143 aq->queue_count--; in ap_sm_recv() 144 if (aq->queue_count > 0) in ap_sm_recv() 157 if (!status.queue_empty || aq->queue_count <= 0) in ap_sm_recv() 160 aq->queue_count = 0; in ap_sm_recv() 186 if (aq->queue_count > 0) { in ap_sm_read() 193 if (aq->queue_count > 0) in ap_sm_read() 235 aq->queue_count++; in ap_sm_write() 236 if (aq->queue_count == 1) in ap_sm_write() 241 if (aq->queue_count < aq->card->queue_depth) { in ap_sm_write() 321 if (aq->queue_count > 0 && aq->reply) in ap_sm_reset_wait() [all …]
|
D | ap_bus.h | 186 int queue_count; /* # messages currently on AP queue. */ member
|
/Linux-v5.10/drivers/nvme/target/ |
D | loop.c | 220 BUG_ON(hctx_idx >= ctrl->ctrl.queue_count); in nvme_loop_init_hctx() 286 for (i = 1; i < ctrl->ctrl.queue_count; i++) { in nvme_loop_destroy_io_queues() 311 ctrl->ctrl.queue_count++; in nvme_loop_init_io_queues() 325 for (i = 1; i < ctrl->ctrl.queue_count; i++) { in nvme_loop_connect_io_queues() 355 ctrl->ctrl.queue_count = 1; in nvme_loop_configure_admin_queue() 408 if (ctrl->ctrl.queue_count > 1) { in nvme_loop_shutdown_ctrl() 471 ctrl->ctrl.queue_count - 1); in nvme_loop_reset_ctrl_work() 519 ctrl->tag_set.nr_hw_queues = ctrl->ctrl.queue_count - 1; in nvme_loop_create_io_queues()
|
/Linux-v5.10/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_packet_manager.c | 44 unsigned int process_count, queue_count, compute_queue_count, gws_queue_count; in pm_calc_rlib_size() local 50 queue_count = pm->dqm->active_queue_count; in pm_calc_rlib_size() 74 queue_count * map_queue_size; in pm_calc_rlib_size()
|
D | kfd_device_queue_manager.c | 378 qpd->queue_count++; in create_queue_nocpsch() 507 qpd->queue_count--; in destroy_queue_nocpsch_locked() 1309 qpd->queue_count++; in create_queue_cpsch() 1495 qpd->queue_count--; in destroy_queue_cpsch() 1768 qpd->queue_count--; in process_termination_cpsch()
|
D | kfd_packet_manager_v9.c | 49 packet->bitfields14.num_queues = (qpd->is_debug) ? 0 : qpd->queue_count; in pm_map_process_v9()
|
D | kfd_packet_manager_vi.c | 59 packet->bitfields10.num_queues = (qpd->is_debug) ? 0 : qpd->queue_count; in pm_map_process_vi()
|
/Linux-v5.10/drivers/nvme/host/ |
D | rdma.c | 326 BUG_ON(hctx_idx >= ctrl->ctrl.queue_count); in nvme_rdma_init_hctx() 660 for (i = 1; i < ctrl->ctrl.queue_count; i++) in nvme_rdma_free_io_queues() 668 for (i = 1; i < ctrl->ctrl.queue_count; i++) in nvme_rdma_stop_io_queues() 698 for (i = 1; i < ctrl->ctrl.queue_count; i++) { in nvme_rdma_start_io_queues() 731 ctrl->ctrl.queue_count = nr_io_queues + 1; in nvme_rdma_alloc_io_queues() 732 if (ctrl->ctrl.queue_count < 2) in nvme_rdma_alloc_io_queues() 766 for (i = 1; i < ctrl->ctrl.queue_count; i++) { in nvme_rdma_alloc_io_queues() 816 set->nr_hw_queues = nctrl->queue_count - 1; in nvme_rdma_alloc_tagset() 989 ctrl->ctrl.queue_count - 1); in nvme_rdma_configure_io_queues() 1028 if (ctrl->ctrl.queue_count > 1) { in nvme_rdma_teardown_io_queues() [all …]
|
D | tcp.c | 1582 set->nr_hw_queues = nctrl->queue_count - 1; in nvme_tcp_alloc_tagset() 1609 for (i = 1; i < ctrl->queue_count; i++) in nvme_tcp_free_io_queues() 1617 for (i = 1; i < ctrl->queue_count; i++) in nvme_tcp_stop_io_queues() 1625 for (i = 1; i < ctrl->queue_count; i++) { in nvme_tcp_start_io_queues() 1662 for (i = 1; i < ctrl->queue_count; i++) { in __nvme_tcp_alloc_io_queues() 1734 ctrl->queue_count = nr_io_queues + 1; in nvme_tcp_alloc_io_queues() 1735 if (ctrl->queue_count < 2) in nvme_tcp_alloc_io_queues() 1794 ctrl->queue_count - 1); in nvme_tcp_configure_io_queues() 1904 if (ctrl->queue_count <= 1) in nvme_tcp_teardown_io_queues() 1967 if (ctrl->queue_count > 1) { in nvme_tcp_setup_ctrl() [all …]
|
D | fc.c | 2280 for (i = 1; i < ctrl->ctrl.queue_count; i++) in nvme_fc_free_io_queues() 2301 struct nvme_fc_queue *queue = &ctrl->queues[ctrl->ctrl.queue_count - 1]; in nvme_fc_delete_hw_io_queues() 2304 for (i = ctrl->ctrl.queue_count - 1; i >= 1; i--, queue--) in nvme_fc_delete_hw_io_queues() 2314 for (i = 1; i < ctrl->ctrl.queue_count; i++, queue++) { in nvme_fc_create_hw_io_queues() 2333 for (i = 1; i < ctrl->ctrl.queue_count; i++) { in nvme_fc_connect_io_queues() 2353 for (i = 1; i < ctrl->ctrl.queue_count; i++) in nvme_fc_init_io_queues() 2464 if (ctrl->ctrl.queue_count > 1) { in __nvme_fc_abort_outstanding_ios() 2846 ctrl->ctrl.queue_count = nr_io_queues + 1; in nvme_fc_create_io_queues() 2862 ctrl->tag_set.nr_hw_queues = ctrl->ctrl.queue_count - 1; in nvme_fc_create_io_queues() 2907 u32 prior_ioq_cnt = ctrl->ctrl.queue_count - 1; in nvme_fc_recreate_io_queues() [all …]
|
D | pci.c | 1360 for (i = dev->ctrl.queue_count - 1; i >= lowest; i--) { in nvme_free_queues() 1361 dev->ctrl.queue_count--; in nvme_free_queues() 1390 for (i = dev->ctrl.queue_count - 1; i > 0; i--) in nvme_suspend_io_queues() 1416 for (i = dev->ctrl.queue_count - 1; i > 0; i--) { in nvme_reap_pending_cqes() 1478 if (dev->ctrl.queue_count > qid) in nvme_alloc_queue() 1498 dev->ctrl.queue_count++; in nvme_alloc_queue() 1733 for (i = dev->ctrl.queue_count; i <= dev->max_qid; i++) { in nvme_create_io_queues() 1740 max = min(dev->max_qid, dev->ctrl.queue_count - 1); in nvme_create_io_queues() 2453 if (!dead && dev->ctrl.queue_count > 0) { in nvme_dev_disable()
|
/Linux-v5.10/drivers/net/ethernet/pensando/ionic/ |
D | ionic_debugfs.c | 70 (u32 *)&ionic->ident.lif.eth.config.queue_count[IONIC_QTYPE_TXQ]); in ionic_debugfs_add_sizes() 72 (u32 *)&ionic->ident.lif.eth.config.queue_count[IONIC_QTYPE_RXQ]); in ionic_debugfs_add_sizes()
|
/Linux-v5.10/drivers/infiniband/sw/rxe/ |
D | rxe_queue.h | 136 static inline unsigned int queue_count(const struct rxe_queue *q) in queue_count() function
|
D | rxe_cq.c | 28 count = queue_count(cq->queue); in rxe_cq_chk_attr()
|
D | rxe_queue.c | 114 if (!queue_empty(q) && (num_elem < queue_count(q))) in resize_finish()
|
/Linux-v5.10/drivers/net/ethernet/intel/fm10k/ |
D | fm10k_pf.c | 501 u16 glort, queue_count, vsi_count, pc_count; in fm10k_configure_dglort_map_pf() local 516 queue_count = BIT(dglort->rss_l + dglort->pc_l); in fm10k_configure_dglort_map_pf() 523 for (queue = 0; queue < queue_count; queue++, q_idx++) { in fm10k_configure_dglort_map_pf() 533 queue_count = BIT(dglort->queue_l + dglort->rss_l + dglort->vsi_l); in fm10k_configure_dglort_map_pf() 539 for (queue = 0; queue < queue_count; queue++) { in fm10k_configure_dglort_map_pf()
|
/Linux-v5.10/drivers/net/wireless/microchip/wilc1000/ |
D | netdev.c | 683 int queue_count; in wilc_mac_xmit() local 703 queue_count = wilc_wlan_txq_add_net_pkt(ndev, (void *)tx_data, in wilc_mac_xmit() 707 if (queue_count > FLOW_CONTROL_UPPER_THRESHOLD) { in wilc_mac_xmit()
|
/Linux-v5.10/drivers/net/wireless/intel/iwlwifi/dvm/ |
D | rs.h | 314 u8 queue_count; /* number of queues that has member
|
D | rs.c | 254 while (tl->queue_count && in rs_tl_rm_old_stats() 259 tl->queue_count--; in rs_tl_rm_old_stats() 293 if (!(tl->queue_count)) { in rs_tl_add_packet() 296 tl->queue_count = 1; in rs_tl_add_packet() 314 if ((index + 1) > tl->queue_count) in rs_tl_add_packet() 315 tl->queue_count = index + 1; in rs_tl_add_packet() 367 if (!(tl->queue_count)) in rs_tl_get_load()
|
/Linux-v5.10/drivers/scsi/hisi_sas/ |
D | hisi_sas_v1_hw.c | 657 (u32)((1ULL << hisi_hba->queue_count) - 1)); in init_reg_v1_hw() 699 for (i = 0; i < hisi_hba->queue_count; i++) { in init_reg_v1_hw() 1665 for (i = 0; i < hisi_hba->queue_count; i++, idx++) { in interrupt_init_v1_hw() 1682 idx = (hisi_hba->n_phy * HISI_SAS_PHY_INT_NR) + hisi_hba->queue_count; in interrupt_init_v1_hw() 1700 hisi_hba->cq_nvecs = hisi_hba->queue_count; in interrupt_init_v1_hw()
|
D | hisi_sas_main.c | 679 int queue = i % hisi_hba->queue_count; in hisi_sas_alloc_dev() 2286 for (i = 0; i < hisi_hba->queue_count; i++) { in hisi_sas_init_mem() 2338 for (i = 0; i < hisi_hba->queue_count; i++) { in hisi_sas_alloc() 2554 &hisi_hba->queue_count)) { in hisi_sas_get_fw_info() 2714 for (i = 0; i < hisi_hba->queue_count; i++) in hisi_sas_debugfs_snapshot_cq_reg() 2726 for (i = 0; i < hisi_hba->queue_count; i++) { in hisi_sas_debugfs_snapshot_dq_reg() 3250 for (c = 0; c < hisi_hba->queue_count; c++) { in hisi_sas_debugfs_create_files() 3260 for (d = 0; d < hisi_hba->queue_count; d++) { in hisi_sas_debugfs_create_files() 3832 for (i = 0; i < hisi_hba->queue_count; i++) in hisi_sas_debugfs_release() 3835 for (i = 0; i < hisi_hba->queue_count; i++) in hisi_sas_debugfs_release() [all …]
|
D | hisi_sas_v2_hw.c | 859 int queue = i % hisi_hba->queue_count; in alloc_dev_quirk_v2_hw() 1179 (u32)((1ULL << hisi_hba->queue_count) - 1)); in init_reg_v2_hw() 1204 for (i = 0; i < hisi_hba->queue_count; i++) in init_reg_v2_hw() 1270 for (i = 0; i < hisi_hba->queue_count; i++) { in init_reg_v2_hw() 3357 for (queue_no = 0; queue_no < hisi_hba->queue_count; queue_no++) { in interrupt_init_v2_hw() 3373 hisi_hba->cq_nvecs = hisi_hba->queue_count; in interrupt_init_v2_hw() 3401 for (i = 0; i < hisi_hba->queue_count; i++) in interrupt_disable_v2_hw()
|
/Linux-v5.10/drivers/net/wireless/intel/iwlegacy/ |
D | 4965-rs.c | 240 while (tl->queue_count && tl->time_stamp < oldest_time) { in il4965_rs_tl_rm_old_stats() 244 tl->queue_count--; in il4965_rs_tl_rm_old_stats() 278 if (!(tl->queue_count)) { in il4965_rs_tl_add_packet() 281 tl->queue_count = 1; in il4965_rs_tl_add_packet() 299 if ((idx + 1) > tl->queue_count) in il4965_rs_tl_add_packet() 300 tl->queue_count = idx + 1; in il4965_rs_tl_add_packet() 323 if (!(tl->queue_count)) in il4965_rs_tl_get_load()
|
/Linux-v5.10/drivers/net/ethernet/amd/xgbe/ |
D | xgbe-dev.c | 2378 unsigned int queue_count, in xgbe_calculate_equal_fifo() argument 2385 q_fifo_size = fifo_size / queue_count; in xgbe_calculate_equal_fifo() 2396 for (i = 0; i < queue_count; i++) in xgbe_calculate_equal_fifo() 2401 unsigned int queue_count, in xgbe_set_nonprio_fifos() argument 2408 if (queue_count <= IEEE_8021QAZ_MAX_TCS) in xgbe_set_nonprio_fifos() 2415 for (i = IEEE_8021QAZ_MAX_TCS; i < queue_count; i++) { in xgbe_set_nonprio_fifos()
|