/Linux-v5.4/arch/s390/include/asm/ |
D | ap.h | 78 static inline struct ap_queue_status ap_tapq(ap_qid_t qid, unsigned long *info) in ap_tapq() argument 80 register unsigned long reg0 asm ("0") = qid; in ap_tapq() 101 static inline struct ap_queue_status ap_test_queue(ap_qid_t qid, in ap_test_queue() argument 106 qid |= 1UL << 23; /* set T bit*/ in ap_test_queue() 107 return ap_tapq(qid, info); in ap_test_queue() 116 static inline struct ap_queue_status ap_rapq(ap_qid_t qid) in ap_rapq() argument 118 register unsigned long reg0 asm ("0") = qid | (1UL << 24); in ap_rapq() 135 static inline struct ap_queue_status ap_zapq(ap_qid_t qid) in ap_zapq() argument 137 register unsigned long reg0 asm ("0") = qid | (2UL << 24); in ap_zapq() 219 static inline struct ap_queue_status ap_aqic(ap_qid_t qid, in ap_aqic() argument [all …]
|
/Linux-v5.4/drivers/infiniband/hw/cxgb4/ |
D | resource.c | 111 u32 qid; in c4iw_get_cqid() local 119 qid = entry->qid; in c4iw_get_cqid() 122 qid = c4iw_get_resource(&rdev->resource.qid_table); in c4iw_get_cqid() 123 if (!qid) in c4iw_get_cqid() 126 rdev->stats.qid.cur += rdev->qpmask + 1; in c4iw_get_cqid() 128 for (i = qid+1; i & rdev->qpmask; i++) { in c4iw_get_cqid() 132 entry->qid = i; in c4iw_get_cqid() 143 entry->qid = qid; in c4iw_get_cqid() 145 for (i = qid+1; i & rdev->qpmask; i++) { in c4iw_get_cqid() 149 entry->qid = i; in c4iw_get_cqid() [all …]
|
/Linux-v5.4/drivers/scsi/lpfc/ |
D | lpfc_debugfs.h | 562 lpfc_debug_dump_wq_by_id(struct lpfc_hba *phba, int qid) in lpfc_debug_dump_wq_by_id() argument 567 if (phba->sli4_hba.hdwq[wq_idx].io_wq->queue_id == qid) in lpfc_debug_dump_wq_by_id() 570 pr_err("IO WQ[Idx:%d|Qid:%d]\n", wq_idx, qid); in lpfc_debug_dump_wq_by_id() 575 if (phba->sli4_hba.els_wq->queue_id == qid) { in lpfc_debug_dump_wq_by_id() 576 pr_err("ELS WQ[Qid:%d]\n", qid); in lpfc_debug_dump_wq_by_id() 581 if (phba->sli4_hba.nvmels_wq->queue_id == qid) { in lpfc_debug_dump_wq_by_id() 582 pr_err("NVME LS WQ[Qid:%d]\n", qid); in lpfc_debug_dump_wq_by_id() 596 lpfc_debug_dump_mq_by_id(struct lpfc_hba *phba, int qid) in lpfc_debug_dump_mq_by_id() argument 598 if (phba->sli4_hba.mbx_wq->queue_id == qid) { in lpfc_debug_dump_mq_by_id() 599 printk(KERN_ERR "MBX WQ[Qid:%d]\n", qid); in lpfc_debug_dump_mq_by_id() [all …]
|
/Linux-v5.4/drivers/nvme/target/ |
D | trace.h | 28 #define parse_nvme_cmd(qid, opcode, fctype, cdw10) \ argument 31 (qid ? \ 55 if ((init && req->sq->qid) || (!init && req->cq->qid)) { in __assign_disk_name() 72 __field(int, qid) 85 __entry->qid = req->sq->qid; 99 __entry->qid, __entry->cid, __entry->nsid, 101 show_opcode_name(__entry->qid, __entry->opcode, 103 parse_nvme_cmd(__entry->qid, __entry->opcode, 113 __field(int, qid) 120 __entry->qid = req->cq->qid; [all …]
|
D | fabrics-cmd.c | 105 u16 qid = le16_to_cpu(c->qid); in nvmet_install_queue() local 122 nvmet_cq_setup(ctrl, req->cq, qid, sqsize + 1); in nvmet_install_queue() 123 nvmet_sq_setup(ctrl, req->sq, qid, sqsize + 1); in nvmet_install_queue() 135 qid, ret, ctrl->cntlid); in nvmet_install_queue() 211 u16 qid = le16_to_cpu(c->qid); in nvmet_execute_io_connect() local 240 if (unlikely(qid > ctrl->subsys->max_qid)) { in nvmet_execute_io_connect() 241 pr_warn("invalid queue id (%d)\n", qid); in nvmet_execute_io_connect() 243 req->cqe->result.u32 = IPO_IATTR_CONNECT_SQE(qid); in nvmet_execute_io_connect() 254 pr_debug("adding queue %d to ctrl %d.\n", qid, ctrl->cntlid); in nvmet_execute_io_connect() 285 if (cmd->connect.qid == 0) in nvmet_parse_connect_cmd()
|
/Linux-v5.4/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_process_queue_manager.c | 32 struct process_queue_manager *pqm, unsigned int qid) in get_queue_by_qid() argument 37 if ((pqn->q && pqn->q->properties.queue_id == qid) || in get_queue_by_qid() 38 (pqn->kq && pqn->kq->queue->properties.queue_id == qid)) in get_queue_by_qid() 46 unsigned int *qid) in find_available_queue_slot() argument 62 *qid = found; in find_available_queue_slot() 78 int pqm_set_gws(struct process_queue_manager *pqm, unsigned int qid, in pqm_set_gws() argument 87 pqn = get_queue_by_qid(pqm, qid); in pqm_set_gws() 168 struct file *f, unsigned int qid) in create_cp_queue() argument 177 q_properties->queue_id = qid; in create_cp_queue() 195 unsigned int *qid) in pqm_create_queue() argument [all …]
|
/Linux-v5.4/drivers/nvme/host/ |
D | trace.h | 26 #define parse_nvme_cmd(qid, opcode, fctype, cdw10) \ argument 29 ((qid) ? \ 53 __field(int, qid) 64 __entry->qid = nvme_req_qid(req); 77 __entry->qid, __entry->cid, __entry->nsid, 79 show_opcode_name(__entry->qid, __entry->opcode, 81 parse_nvme_cmd(__entry->qid, __entry->opcode, 91 __field(int, qid) 100 __entry->qid = nvme_req_qid(req); 110 __entry->qid, __entry->cid, __entry->result, [all …]
|
/Linux-v5.4/fs/9p/ |
D | cache.c | 86 if (buflen != sizeof(v9inode->qid.version)) in v9fs_cache_inode_check_aux() 89 if (memcmp(buffer, &v9inode->qid.version, in v9fs_cache_inode_check_aux() 90 sizeof(v9inode->qid.version))) in v9fs_cache_inode_check_aux() 117 &v9inode->qid.path, in v9fs_cache_inode_get_cookie() 118 sizeof(v9inode->qid.path), in v9fs_cache_inode_get_cookie() 119 &v9inode->qid.version, in v9fs_cache_inode_get_cookie() 120 sizeof(v9inode->qid.version), in v9fs_cache_inode_get_cookie() 138 fscache_relinquish_cookie(v9inode->fscache, &v9inode->qid.version, in v9fs_cache_inode_put_cookie() 190 &v9inode->qid.path, in v9fs_cache_inode_reset_cookie() 191 sizeof(v9inode->qid.path), in v9fs_cache_inode_reset_cookie() [all …]
|
D | vfs_inode_dotl.c | 69 if (memcmp(&v9inode->qid.version, in v9fs_test_inode_dotl() 70 &st->qid.version, sizeof(v9inode->qid.version))) in v9fs_test_inode_dotl() 73 if (v9inode->qid.type != st->qid.type) in v9fs_test_inode_dotl() 76 if (v9inode->qid.path != st->qid.path) in v9fs_test_inode_dotl() 92 memcpy(&v9inode->qid, &st->qid, sizeof(st->qid)); in v9fs_set_inode_dotl() 98 struct p9_qid *qid, in v9fs_qid_iget_dotl() argument 114 i_ino = v9fs_qid2ino(qid); in v9fs_qid_iget_dotl() 156 inode = v9fs_qid_iget_dotl(sb, &st->qid, fid, st, new); in v9fs_inode_from_fid_dotl() 235 struct p9_qid qid; in v9fs_vfs_atomic_open_dotl() local 289 mode, gid, &qid); in v9fs_vfs_atomic_open_dotl() [all …]
|
/Linux-v5.4/include/linux/ |
D | quota.h | 79 extern qid_t from_kqid(struct user_namespace *to, struct kqid qid); 80 extern qid_t from_kqid_munged(struct user_namespace *to, struct kqid qid); 81 extern bool qid_valid(struct kqid qid); 98 enum quota_type type, qid_t qid) in make_kqid() argument 105 kqid.uid = make_kuid(from, qid); in make_kqid() 108 kqid.gid = make_kgid(from, qid); in make_kqid() 111 kqid.projid = make_kprojid(from, qid); in make_kqid() 187 static inline bool qid_has_mapping(struct user_namespace *ns, struct kqid qid) in qid_has_mapping() argument 189 return from_kqid(ns, qid) != (qid_t) -1; in qid_has_mapping() 318 …int (*get_next_id)(struct super_block *sb, struct kqid *qid); /* Get next ID with existing structu… [all …]
|
/Linux-v5.4/drivers/s390/crypto/ |
D | zcrypt_api.c | 638 int qid = 0, rc = -ENODEV; in zcrypt_rsa_modexpo() local 684 AP_QID_QUEUE(zq->queue->qid))) in zcrypt_rsa_modexpo() 702 qid = pref_zq->queue->qid; in zcrypt_rsa_modexpo() 711 AP_QID_CARD(qid), AP_QID_QUEUE(qid)); in zcrypt_rsa_modexpo() 722 int qid = 0, rc = -ENODEV; in zcrypt_rsa_crt() local 768 AP_QID_QUEUE(zq->queue->qid))) in zcrypt_rsa_crt() 786 qid = pref_zq->queue->qid; in zcrypt_rsa_crt() 795 AP_QID_CARD(qid), AP_QID_QUEUE(qid)); in zcrypt_rsa_crt() 808 int qid = 0, rc = -ENODEV; in _zcrypt_send_cprb() local 853 tdom != AP_QID_QUEUE(zq->queue->qid))) in _zcrypt_send_cprb() [all …]
|
D | ap_queue.c | 37 status = ap_aqic(aq->qid, qirqctrl, ind); in ap_queue_enable_interruption() 47 AP_QID_CARD(aq->qid), in ap_queue_enable_interruption() 48 AP_QID_QUEUE(aq->qid)); in ap_queue_enable_interruption() 71 __ap_send(ap_qid_t qid, unsigned long long psmid, void *msg, size_t length, in __ap_send() argument 75 qid |= 0x400000UL; in __ap_send() 76 return ap_nqap(qid, psmid, msg, length); in __ap_send() 79 int ap_send(ap_qid_t qid, unsigned long long psmid, void *msg, size_t length) in ap_send() argument 83 status = __ap_send(qid, psmid, msg, length, 0); in ap_send() 98 int ap_recv(ap_qid_t qid, unsigned long long *psmid, void *msg, size_t length) in ap_recv() argument 104 status = ap_dqap(qid, psmid, msg, length); in ap_recv() [all …]
|
D | zcrypt_msgtype6.c | 297 msg->cprbx.domain = AP_QID_QUEUE(zq->queue->qid); in ICAMEX_msg_to_type6MEX_msgX() 367 msg->cprbx.domain = AP_QID_QUEUE(zq->queue->qid); in ICACRT_msg_to_type6CRT_msgX() 657 AP_QID_CARD(zq->queue->qid), in convert_type86_ica() 658 AP_QID_QUEUE(zq->queue->qid), in convert_type86_ica() 664 AP_QID_CARD(zq->queue->qid), in convert_type86_ica() 665 AP_QID_QUEUE(zq->queue->qid)); in convert_type86_ica() 668 AP_QID_CARD(zq->queue->qid), in convert_type86_ica() 669 AP_QID_QUEUE(zq->queue->qid), in convert_type86_ica() 808 AP_QID_CARD(zq->queue->qid), in convert_response_ica() 809 AP_QID_QUEUE(zq->queue->qid)); in convert_response_ica() [all …]
|
D | zcrypt_queue.c | 64 AP_QID_CARD(zq->queue->qid), in online_store() 65 AP_QID_QUEUE(zq->queue->qid), in online_store() 168 AP_QID_CARD(zq->queue->qid), AP_QID_QUEUE(zq->queue->qid)); in zcrypt_queue_register() 211 AP_QID_CARD(zq->queue->qid), AP_QID_QUEUE(zq->queue->qid)); in zcrypt_queue_unregister()
|
D | ap_bus.c | 295 static int ap_query_queue(ap_qid_t qid, int *queue_depth, int *device_type, in ap_query_queue() argument 302 if (!ap_test_config_card_id(AP_QID_CARD(qid))) in ap_query_queue() 305 status = ap_test_queue(qid, ap_apft_available(), &info); in ap_query_queue() 638 AP_QID_CARD(to_ap_queue(dev)->qid) == (int)(long) data) in __ap_queue_devices_with_id_unregister() 711 card = AP_QID_CARD(to_ap_queue(dev)->qid); in __ap_revise_reserved() 712 queue = AP_QID_QUEUE(to_ap_queue(dev)->qid); in __ap_revise_reserved() 787 card = AP_QID_CARD(to_ap_queue(dev)->qid); in ap_device_probe() 788 queue = AP_QID_QUEUE(to_ap_queue(dev)->qid); in ap_device_probe() 1318 static int ap_get_compatible_type(ap_qid_t qid, int rawtype, unsigned int func) in ap_get_compatible_type() argument 1339 status = ap_qact(qid, 0, &apinfo); in ap_get_compatible_type() [all …]
|
/Linux-v5.4/fs/xfs/ |
D | xfs_quotaops.c | 217 struct kqid qid, in xfs_fs_get_dqblk() argument 228 id = from_kqid(&init_user_ns, qid); in xfs_fs_get_dqblk() 229 return xfs_qm_scall_getquota(mp, id, xfs_quota_type(qid.type), qdq); in xfs_fs_get_dqblk() 236 struct kqid *qid, in xfs_fs_get_nextdqblk() argument 248 id = from_kqid(&init_user_ns, *qid); in xfs_fs_get_nextdqblk() 249 ret = xfs_qm_scall_getquota_next(mp, &id, xfs_quota_type(qid->type), in xfs_fs_get_nextdqblk() 255 *qid = make_kqid(current_user_ns(), qid->type, id); in xfs_fs_get_nextdqblk() 262 struct kqid qid, in xfs_fs_set_dqblk() argument 274 return xfs_qm_scall_setqlim(mp, from_kqid(&init_user_ns, qid), in xfs_fs_set_dqblk() 275 xfs_quota_type(qid.type), qdq); in xfs_fs_set_dqblk()
|
/Linux-v5.4/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_xsk.c | 15 int qid = ring->ring_idx; in ixgbe_xsk_umem() local 17 if (!xdp_on || !test_bit(qid, adapter->af_xdp_zc_qps)) in ixgbe_xsk_umem() 20 return xdp_get_umem_from_qid(adapter->netdev, qid); in ixgbe_xsk_umem() 67 u16 qid) in ixgbe_xsk_umem_enable() argument 74 if (qid >= adapter->num_rx_queues) in ixgbe_xsk_umem_enable() 77 if (qid >= netdev->real_num_rx_queues || in ixgbe_xsk_umem_enable() 78 qid >= netdev->real_num_tx_queues) in ixgbe_xsk_umem_enable() 95 ixgbe_txrx_ring_disable(adapter, qid); in ixgbe_xsk_umem_enable() 97 set_bit(qid, adapter->af_xdp_zc_qps); in ixgbe_xsk_umem_enable() 100 ixgbe_txrx_ring_enable(adapter, qid); in ixgbe_xsk_umem_enable() [all …]
|
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | params.h | 44 u16 qid, in mlx5e_qid_get_ch_if_in_group() argument 49 int ch = qid - nch * group; in mlx5e_qid_get_ch_if_in_group() 59 u16 qid, in mlx5e_qid_get_ch_and_group() argument 65 *ix = qid % nch; in mlx5e_qid_get_ch_and_group() 66 *group = qid / nch; in mlx5e_qid_get_ch_and_group() 70 struct mlx5e_params *params, u64 qid) in mlx5e_qid_validate() argument 72 return qid < params->num_channels * profile->rq_groups; in mlx5e_qid_validate()
|
/Linux-v5.4/drivers/net/wireless/mediatek/mt76/ |
D | tx.c | 254 int qid = skb_get_queue_mapping(skb); in mt76_tx() local 256 if (WARN_ON(qid >= MT_TXQ_PSD)) { in mt76_tx() 257 qid = MT_TXQ_BE; in mt76_tx() 258 skb_set_queue_mapping(skb, qid); in mt76_tx() 278 q = dev->q_tx[qid].q; in mt76_tx() 281 dev->queue_ops->tx_queue_skb(dev, qid, skb, wcid, sta); in mt76_tx() 384 enum mt76_txq_id qid = mt76_txq_get_qid(txq); in mt76_txq_send_burst() local 419 idx = dev->queue_ops->tx_queue_skb(dev, qid, skb, wcid, txq->sta); in mt76_txq_send_burst() 453 idx = dev->queue_ops->tx_queue_skb(dev, qid, skb, wcid, in mt76_txq_send_burst() 462 hwq->entry[idx].qid = sq - dev->q_tx; in mt76_txq_send_burst() [all …]
|
D | dma.c | 139 mt76_dma_tx_cleanup(struct mt76_dev *dev, enum mt76_txq_id qid, bool flush) in mt76_dma_tx_cleanup() argument 141 struct mt76_sw_queue *sq = &dev->q_tx[qid]; in mt76_dma_tx_cleanup() 160 n_swq_queued[entry.qid]++; in mt76_dma_tx_cleanup() 166 dev->drv->tx_complete_skb(dev, qid, &entry); in mt76_dma_tx_cleanup() 192 qid < IEEE80211_NUM_ACS && q->queued < q->ndesc - 8; in mt76_dma_tx_cleanup() 202 ieee80211_wake_queue(dev->hw, qid); in mt76_dma_tx_cleanup() 257 mt76_dma_tx_queue_skb_raw(struct mt76_dev *dev, enum mt76_txq_id qid, in mt76_dma_tx_queue_skb_raw() argument 260 struct mt76_queue *q = dev->q_tx[qid].q; in mt76_dma_tx_queue_skb_raw() 281 mt76_dma_tx_queue_skb(struct mt76_dev *dev, enum mt76_txq_id qid, in mt76_dma_tx_queue_skb() argument 285 struct mt76_queue *q = dev->q_tx[qid].q; in mt76_dma_tx_queue_skb() [all …]
|
/Linux-v5.4/fs/quota/ |
D | kqid.c | 120 bool qid_valid(struct kqid qid) in qid_valid() argument 122 switch (qid.type) { in qid_valid() 124 return uid_valid(qid.uid); in qid_valid() 126 return gid_valid(qid.gid); in qid_valid() 128 return projid_valid(qid.projid); in qid_valid()
|
D | quota.c | 202 struct kqid qid; in quota_getquota() local 209 qid = make_kqid(current_user_ns(), type, id); in quota_getquota() 210 if (!qid_has_mapping(sb->s_user_ns, qid)) in quota_getquota() 212 ret = sb->s_qcop->get_dqblk(sb, qid, &fdq); in quota_getquota() 228 struct kqid qid; in quota_getnextquota() local 235 qid = make_kqid(current_user_ns(), type, id); in quota_getnextquota() 236 if (!qid_has_mapping(sb->s_user_ns, qid)) in quota_getnextquota() 238 ret = sb->s_qcop->get_nextdqblk(sb, &qid, &fdq); in quota_getnextquota() 243 idq.dqb_id = from_kqid(current_user_ns(), qid); in quota_getnextquota() 280 struct kqid qid; in quota_setquota() local [all …]
|
/Linux-v5.4/net/9p/ |
D | client.c | 898 memset(&fid->qid, 0, sizeof(struct p9_qid)); in p9_fid_create() 1122 struct p9_qid qid; in p9_client_attach() local 1141 err = p9pdu_readf(&req->rc, clnt->proto_version, "Q", &qid); in p9_client_attach() 1149 qid.type, (unsigned long long)qid.path, qid.version); in p9_client_attach() 1151 memmove(&fid->qid, &qid, sizeof(struct p9_qid)); in p9_client_attach() 1220 memmove(&fid->qid, &wqids[nwqids - 1], sizeof(struct p9_qid)); in p9_client_walk() 1222 fid->qid = oldfid->qid; in p9_client_walk() 1245 struct p9_qid qid; in p9_client_open() local 1265 err = p9pdu_readf(&req->rc, clnt->proto_version, "Qd", &qid, &iounit); in p9_client_open() 1272 p9_is_proto_dotl(clnt) ? "RLOPEN" : "ROPEN", qid.type, in p9_client_open() [all …]
|
/Linux-v5.4/drivers/net/ethernet/intel/i40e/ |
D | i40e_xsk.c | 78 u16 qid) in i40e_xsk_umem_enable() argument 88 if (qid >= vsi->num_queue_pairs) in i40e_xsk_umem_enable() 91 if (qid >= netdev->real_num_rx_queues || in i40e_xsk_umem_enable() 92 qid >= netdev->real_num_tx_queues) in i40e_xsk_umem_enable() 105 set_bit(qid, vsi->af_xdp_zc_qps); in i40e_xsk_umem_enable() 110 err = i40e_queue_pair_disable(vsi, qid); in i40e_xsk_umem_enable() 114 err = i40e_queue_pair_enable(vsi, qid); in i40e_xsk_umem_enable() 119 err = i40e_xsk_wakeup(vsi->netdev, qid, XDP_WAKEUP_RX); in i40e_xsk_umem_enable() 134 static int i40e_xsk_umem_disable(struct i40e_vsi *vsi, u16 qid) in i40e_xsk_umem_disable() argument 141 umem = xdp_get_umem_from_qid(netdev, qid); in i40e_xsk_umem_disable() [all …]
|
/Linux-v5.4/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_sriov.c | 235 struct bnx2x_virtf *vf, int qid, in bnx2x_vf_queue_create() argument 241 DP(BNX2X_MSG_IOV, "vf[%d:%d]\n", vf->abs_vfid, qid); in bnx2x_vf_queue_create() 245 q_params->q_obj = &bnx2x_vfq(vf, qid, sp_obj); in bnx2x_vf_queue_create() 268 bnx2x_vf_igu_ack_sb(bp, vf, vf_igu_sb(vf, bnx2x_vfq(vf, qid, sb_idx)), in bnx2x_vf_queue_create() 275 int qid) in bnx2x_vf_queue_destroy() argument 287 q_params.q_obj = &bnx2x_vfq(vf, qid, sp_obj); in bnx2x_vf_queue_destroy() 307 if (bnx2x_vfq(vf, qid, cxt)) { in bnx2x_vf_queue_destroy() 308 bnx2x_vfq(vf, qid, cxt)->ustorm_ag_context.cdu_usage = 0; in bnx2x_vf_queue_destroy() 309 bnx2x_vfq(vf, qid, cxt)->xstorm_ag_context.cdu_reserved = 0; in bnx2x_vf_queue_destroy() 356 int qid, bool drv_only, int type) in bnx2x_vf_vlan_mac_clear() argument [all …]
|