/Linux-v5.15/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | qos.c | 44 u16 qid; member 51 mlx5e_sw_node_create_leaf(struct mlx5e_priv *priv, u16 classid, u16 qid, in mlx5e_sw_node_create_leaf() argument 62 node->qid = qid; in mlx5e_sw_node_create_leaf() 63 __set_bit(qid, priv->htb.qos_used_qids); in mlx5e_sw_node_create_leaf() 81 node->qid = MLX5E_QOS_QID_INNER; in mlx5e_sw_node_create_root() 115 if (node->qid != MLX5E_QOS_QID_INNER) { in mlx5e_sw_node_delete() 116 __clear_bit(node->qid, priv->htb.qos_used_qids); in mlx5e_sw_node_delete() 124 static u16 mlx5e_qid_from_qos(struct mlx5e_channels *chs, u16 qid) in mlx5e_qid_from_qos() argument 135 return (chs->params.num_channels + is_ptp) * mlx5e_get_dcb_num_tc(&chs->params) + qid; in mlx5e_qid_from_qos() 141 u16 qid; in mlx5e_get_txq_by_classid() local [all …]
|
D | params.h | 60 u16 qid, in mlx5e_qid_get_ch_if_in_group() argument 65 int ch = qid - nch * group; in mlx5e_qid_get_ch_if_in_group() 75 u16 qid, in mlx5e_qid_get_ch_and_group() argument 81 *ix = qid % nch; in mlx5e_qid_get_ch_and_group() 82 *group = qid / nch; in mlx5e_qid_get_ch_and_group() 86 struct mlx5e_params *params, u64 qid) in mlx5e_qid_validate() argument 88 return qid < params->num_channels * profile->rq_groups; in mlx5e_qid_validate()
|
/Linux-v5.15/arch/s390/include/asm/ |
D | ap.h | 80 static inline struct ap_queue_status ap_tapq(ap_qid_t qid, unsigned long *info) in ap_tapq() argument 92 : [qid] "d" (qid) in ap_tapq() 107 static inline struct ap_queue_status ap_test_queue(ap_qid_t qid, in ap_test_queue() argument 112 qid |= 1UL << 23; /* set T bit*/ in ap_test_queue() 113 return ap_tapq(qid, info); in ap_test_queue() 122 static inline struct ap_queue_status ap_rapq(ap_qid_t qid) in ap_rapq() argument 124 unsigned long reg0 = qid | (1UL << 24); /* fc 1UL is RAPQ */ in ap_rapq() 143 static inline struct ap_queue_status ap_zapq(ap_qid_t qid) in ap_zapq() argument 145 unsigned long reg0 = qid | (2UL << 24); /* fc 2UL is ZAPQ */ in ap_zapq() 231 static inline struct ap_queue_status ap_aqic(ap_qid_t qid, in ap_aqic() argument [all …]
|
/Linux-v5.15/drivers/infiniband/hw/cxgb4/ |
D | resource.c | 111 u32 qid; in c4iw_get_cqid() local 119 qid = entry->qid; in c4iw_get_cqid() 122 qid = c4iw_get_resource(&rdev->resource.qid_table); in c4iw_get_cqid() 123 if (!qid) in c4iw_get_cqid() 126 rdev->stats.qid.cur += rdev->qpmask + 1; in c4iw_get_cqid() 128 for (i = qid+1; i & rdev->qpmask; i++) { in c4iw_get_cqid() 132 entry->qid = i; in c4iw_get_cqid() 143 entry->qid = qid; in c4iw_get_cqid() 145 for (i = qid+1; i & rdev->qpmask; i++) { in c4iw_get_cqid() 149 entry->qid = i; in c4iw_get_cqid() [all …]
|
/Linux-v5.15/drivers/scsi/lpfc/ |
D | lpfc_debugfs.h | 570 lpfc_debug_dump_wq_by_id(struct lpfc_hba *phba, int qid) in lpfc_debug_dump_wq_by_id() argument 575 if (phba->sli4_hba.hdwq[wq_idx].io_wq->queue_id == qid) in lpfc_debug_dump_wq_by_id() 578 pr_err("IO WQ[Idx:%d|Qid:%d]\n", wq_idx, qid); in lpfc_debug_dump_wq_by_id() 583 if (phba->sli4_hba.els_wq->queue_id == qid) { in lpfc_debug_dump_wq_by_id() 584 pr_err("ELS WQ[Qid:%d]\n", qid); in lpfc_debug_dump_wq_by_id() 589 if (phba->sli4_hba.nvmels_wq->queue_id == qid) { in lpfc_debug_dump_wq_by_id() 590 pr_err("NVME LS WQ[Qid:%d]\n", qid); in lpfc_debug_dump_wq_by_id() 604 lpfc_debug_dump_mq_by_id(struct lpfc_hba *phba, int qid) in lpfc_debug_dump_mq_by_id() argument 606 if (phba->sli4_hba.mbx_wq->queue_id == qid) { in lpfc_debug_dump_mq_by_id() 607 printk(KERN_ERR "MBX WQ[Qid:%d]\n", qid); in lpfc_debug_dump_mq_by_id() [all …]
|
/Linux-v5.15/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_process_queue_manager.c | 32 struct process_queue_manager *pqm, unsigned int qid) in get_queue_by_qid() argument 37 if ((pqn->q && pqn->q->properties.queue_id == qid) || in get_queue_by_qid() 38 (pqn->kq && pqn->kq->queue->properties.queue_id == qid)) in get_queue_by_qid() 46 unsigned int *qid) in find_available_queue_slot() argument 62 *qid = found; in find_available_queue_slot() 78 int pqm_set_gws(struct process_queue_manager *pqm, unsigned int qid, in pqm_set_gws() argument 87 pqn = get_queue_by_qid(pqm, qid); in pqm_set_gws() 169 struct file *f, unsigned int qid) in init_user_queue() argument 178 q_properties->queue_id = qid; in init_user_queue() 196 unsigned int *qid, in pqm_create_queue() argument [all …]
|
/Linux-v5.15/drivers/net/ethernet/marvell/prestera/ |
D | prestera_pci.c | 208 static u32 prestera_fw_evtq_len(struct prestera_fw *fw, u8 qid) in prestera_fw_evtq_len() argument 210 return fw->evt_queue[qid].len; in prestera_fw_evtq_len() 213 static u32 prestera_fw_evtq_avail(struct prestera_fw *fw, u8 qid) in prestera_fw_evtq_avail() argument 215 u32 wr_idx = prestera_fw_read(fw, PRESTERA_EVTQ_WR_IDX_REG(qid)); in prestera_fw_evtq_avail() 216 u32 rd_idx = prestera_fw_read(fw, PRESTERA_EVTQ_RD_IDX_REG(qid)); in prestera_fw_evtq_avail() 218 return CIRC_CNT(wr_idx, rd_idx, prestera_fw_evtq_len(fw, qid)); in prestera_fw_evtq_avail() 222 u8 qid, u32 idx) in prestera_fw_evtq_rd_set() argument 224 u32 rd_idx = idx & (prestera_fw_evtq_len(fw, qid) - 1); in prestera_fw_evtq_rd_set() 226 prestera_fw_write(fw, PRESTERA_EVTQ_RD_IDX_REG(qid), rd_idx); in prestera_fw_evtq_rd_set() 229 static u8 __iomem *prestera_fw_evtq_buf(struct prestera_fw *fw, u8 qid) in prestera_fw_evtq_buf() argument [all …]
|
/Linux-v5.15/drivers/s390/crypto/ |
D | ap_queue.c | 37 status = ap_aqic(aq->qid, qirqctrl, ind); in ap_queue_enable_irq() 47 AP_QID_CARD(aq->qid), in ap_queue_enable_irq() 48 AP_QID_QUEUE(aq->qid)); in ap_queue_enable_irq() 71 __ap_send(ap_qid_t qid, unsigned long long psmid, void *msg, size_t length, in __ap_send() argument 75 qid |= 0x400000UL; in __ap_send() 76 return ap_nqap(qid, psmid, msg, length); in __ap_send() 79 int ap_send(ap_qid_t qid, unsigned long long psmid, void *msg, size_t length) in ap_send() argument 83 status = __ap_send(qid, psmid, msg, length, 0); in ap_send() 98 int ap_recv(ap_qid_t qid, unsigned long long *psmid, void *msg, size_t length) in ap_recv() argument 104 status = ap_dqap(qid, psmid, msg, length, NULL, NULL); in ap_recv() [all …]
|
D | zcrypt_api.c | 641 int cpen, qpen, qid = 0, rc = -ENODEV; in zcrypt_rsa_modexpo() local 700 AP_QID_QUEUE(zq->queue->qid))) in zcrypt_rsa_modexpo() 704 tr->last_qid == zq->queue->qid) ? in zcrypt_rsa_modexpo() 722 qid = pref_zq->queue->qid; in zcrypt_rsa_modexpo() 733 tr->last_qid = qid; in zcrypt_rsa_modexpo() 736 AP_QID_CARD(qid), AP_QID_QUEUE(qid)); in zcrypt_rsa_modexpo() 749 int cpen, qpen, qid = 0, rc = -ENODEV; in zcrypt_rsa_crt() local 808 AP_QID_QUEUE(zq->queue->qid))) in zcrypt_rsa_crt() 812 tr->last_qid == zq->queue->qid) ? in zcrypt_rsa_crt() 830 qid = pref_zq->queue->qid; in zcrypt_rsa_crt() [all …]
|
D | zcrypt_msgtype6.c | 297 msg->cprbx.domain = AP_QID_QUEUE(zq->queue->qid); in ICAMEX_msg_to_type6MEX_msgX() 367 msg->cprbx.domain = AP_QID_QUEUE(zq->queue->qid); in ICACRT_msg_to_type6CRT_msgX() 653 AP_QID_CARD(zq->queue->qid), in convert_type86_ica() 654 AP_QID_QUEUE(zq->queue->qid), in convert_type86_ica() 660 AP_QID_CARD(zq->queue->qid), in convert_type86_ica() 661 AP_QID_QUEUE(zq->queue->qid), in convert_type86_ica() 664 AP_QID_CARD(zq->queue->qid), in convert_type86_ica() 665 AP_QID_QUEUE(zq->queue->qid), in convert_type86_ica() 806 AP_QID_CARD(zq->queue->qid), in convert_response_ica() 807 AP_QID_QUEUE(zq->queue->qid), in convert_response_ica() [all …]
|
D | ap_bus.c | 312 static bool ap_queue_info(ap_qid_t qid, int *q_type, unsigned int *q_fac, in ap_queue_info() argument 332 if (AP_QID_CARD(qid) > ap_max_adapter_id || in ap_queue_info() 333 AP_QID_QUEUE(qid) > ap_max_domain_id) in ap_queue_info() 337 status = ap_test_queue(qid, ap_apft_available(), &tapq_info.value); in ap_queue_info() 770 AP_QID_CARD(to_ap_queue(dev)->qid) == (int)(long) data) in __ap_queue_devices_with_id_unregister() 780 card = AP_QID_CARD(to_ap_queue(dev)->qid); in __ap_revise_reserved() 781 queue = AP_QID_QUEUE(to_ap_queue(dev)->qid); in __ap_revise_reserved() 859 card = AP_QID_CARD(to_ap_queue(dev)->qid); in ap_device_probe() 860 queue = AP_QID_QUEUE(to_ap_queue(dev)->qid); in ap_device_probe() 874 to_ap_queue(dev)->qid); in ap_device_probe() [all …]
|
/Linux-v5.15/drivers/nvme/host/ |
D | trace.h | 26 #define parse_nvme_cmd(qid, opcode, fctype, cdw10) \ argument 29 ((qid) ? \ 53 __field(int, qid) 64 __entry->qid = nvme_req_qid(req); 77 __entry->qid, __entry->cid, __entry->nsid, 79 show_opcode_name(__entry->qid, __entry->opcode, 81 parse_nvme_cmd(__entry->qid, __entry->opcode, 91 __field(int, qid) 100 __entry->qid = nvme_req_qid(req); 110 __entry->qid, __entry->cid, __entry->result, [all …]
|
/Linux-v5.15/drivers/nvme/target/ |
D | trace.h | 28 #define parse_nvme_cmd(qid, opcode, fctype, cdw10) \ argument 31 (qid ? \ 68 __field(int, qid) 81 __entry->qid = req->sq->qid; 95 __entry->qid, __entry->cid, __entry->nsid, 97 show_opcode_name(__entry->qid, __entry->opcode, 99 parse_nvme_cmd(__entry->qid, __entry->opcode, 109 __field(int, qid) 116 __entry->qid = req->cq->qid; 125 __entry->qid, __entry->cid, __entry->result, __entry->status)
|
D | fabrics-cmd.c | 109 u16 qid = le16_to_cpu(c->qid); in nvmet_install_queue() local 123 if (ctrl->sqs[qid] != NULL) { in nvmet_install_queue() 124 pr_warn("qid %u has already been created\n", qid); in nvmet_install_queue() 125 req->error_loc = offsetof(struct nvmf_connect_command, qid); in nvmet_install_queue() 145 nvmet_cq_setup(ctrl, req->cq, qid, sqsize + 1); in nvmet_install_queue() 146 nvmet_sq_setup(ctrl, req->sq, qid, sqsize + 1); in nvmet_install_queue() 157 qid, ctrl->cntlid, ret); in nvmet_install_queue() 158 ctrl->sqs[qid] = NULL; in nvmet_install_queue() 240 u16 qid = le16_to_cpu(c->qid); in nvmet_execute_io_connect() local 273 if (unlikely(qid > ctrl->subsys->max_qid)) { in nvmet_execute_io_connect() [all …]
|
/Linux-v5.15/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_xsk.c | 15 int qid = ring->ring_idx; in ixgbe_xsk_pool() local 17 if (!xdp_on || !test_bit(qid, adapter->af_xdp_zc_qps)) in ixgbe_xsk_pool() 20 return xsk_get_pool_from_qid(adapter->netdev, qid); in ixgbe_xsk_pool() 25 u16 qid) in ixgbe_xsk_pool_enable() argument 31 if (qid >= adapter->num_rx_queues) in ixgbe_xsk_pool_enable() 34 if (qid >= netdev->real_num_rx_queues || in ixgbe_xsk_pool_enable() 35 qid >= netdev->real_num_tx_queues) in ixgbe_xsk_pool_enable() 46 ixgbe_txrx_ring_disable(adapter, qid); in ixgbe_xsk_pool_enable() 48 set_bit(qid, adapter->af_xdp_zc_qps); in ixgbe_xsk_pool_enable() 51 ixgbe_txrx_ring_enable(adapter, qid); in ixgbe_xsk_pool_enable() [all …]
|
/Linux-v5.15/fs/9p/ |
D | cache.c | 86 if (buflen != sizeof(v9inode->qid.version)) in v9fs_cache_inode_check_aux() 89 if (memcmp(buffer, &v9inode->qid.version, in v9fs_cache_inode_check_aux() 90 sizeof(v9inode->qid.version))) in v9fs_cache_inode_check_aux() 117 &v9inode->qid.path, in v9fs_cache_inode_get_cookie() 118 sizeof(v9inode->qid.path), in v9fs_cache_inode_get_cookie() 119 &v9inode->qid.version, in v9fs_cache_inode_get_cookie() 120 sizeof(v9inode->qid.version), in v9fs_cache_inode_get_cookie() 138 fscache_relinquish_cookie(v9inode->fscache, &v9inode->qid.version, in v9fs_cache_inode_put_cookie() 190 &v9inode->qid.path, in v9fs_cache_inode_reset_cookie() 191 sizeof(v9inode->qid.path), in v9fs_cache_inode_reset_cookie() [all …]
|
/Linux-v5.15/include/linux/ |
D | quota.h | 79 extern qid_t from_kqid(struct user_namespace *to, struct kqid qid); 80 extern qid_t from_kqid_munged(struct user_namespace *to, struct kqid qid); 81 extern bool qid_valid(struct kqid qid); 98 enum quota_type type, qid_t qid) in make_kqid() argument 105 kqid.uid = make_kuid(from, qid); in make_kqid() 108 kqid.gid = make_kgid(from, qid); in make_kqid() 111 kqid.projid = make_kprojid(from, qid); in make_kqid() 187 static inline bool qid_has_mapping(struct user_namespace *ns, struct kqid qid) in qid_has_mapping() argument 189 return from_kqid(ns, qid) != (qid_t) -1; in qid_has_mapping() 318 …int (*get_next_id)(struct super_block *sb, struct kqid *qid); /* Get next ID with existing structu… [all …]
|
/Linux-v5.15/drivers/net/wireless/mediatek/mt76/mt7615/ |
D | sdio_txrx.c | 81 static int mt7663s_rx_run_queue(struct mt76_dev *dev, enum mt76_rxq_id qid, in mt7663s_rx_run_queue() argument 84 struct mt76_queue *q = &dev->q_rx[qid]; in mt7663s_rx_run_queue() 90 for (i = 0; i < intr->rx.num[qid]; i++) in mt7663s_rx_run_queue() 91 len += round_up(intr->rx.len[qid][i] + 4, 4); in mt7663s_rx_run_queue() 105 err = sdio_readsb(sdio->func, buf, MCR_WRDR(qid), len); in mt7663s_rx_run_queue() 112 for (i = 0; i < intr->rx.num[qid]; i++) { in mt7663s_rx_run_queue() 116 len = intr->rx.len[qid][i]; in mt7663s_rx_run_queue() 218 int qid, err, nframes = 0, len = 0, pse_sz = 0, ple_sz = 0; in mt7663s_tx_run_queue() local 223 qid = mcu ? ARRAY_SIZE(sdio->xmit_buf) - 1 : q->qid; in mt7663s_tx_run_queue() 248 memcpy(sdio->xmit_buf[qid] + len, e->skb->data, in mt7663s_tx_run_queue() [all …]
|
/Linux-v5.15/drivers/vdpa/virtio_pci/ |
D | vp_vdpa.c | 213 static int vp_vdpa_get_vq_state(struct vdpa_device *vdpa, u16 qid, in vp_vdpa_get_vq_state() argument 248 static int vp_vdpa_set_vq_state(struct vdpa_device *vdpa, u16 qid, in vp_vdpa_set_vq_state() argument 258 !vp_modern_get_queue_enable(mdev, qid)) { in vp_vdpa_set_vq_state() 269 static void vp_vdpa_set_vq_cb(struct vdpa_device *vdpa, u16 qid, in vp_vdpa_set_vq_cb() argument 274 vp_vdpa->vring[qid].cb = *cb; in vp_vdpa_set_vq_cb() 278 u16 qid, bool ready) in vp_vdpa_set_vq_ready() argument 282 vp_modern_set_queue_enable(mdev, qid, ready); in vp_vdpa_set_vq_ready() 285 static bool vp_vdpa_get_vq_ready(struct vdpa_device *vdpa, u16 qid) in vp_vdpa_get_vq_ready() argument 289 return vp_modern_get_queue_enable(mdev, qid); in vp_vdpa_get_vq_ready() 292 static void vp_vdpa_set_vq_num(struct vdpa_device *vdpa, u16 qid, in vp_vdpa_set_vq_num() argument [all …]
|
/Linux-v5.15/fs/xfs/ |
D | xfs_quotaops.c | 216 struct kqid qid, in xfs_fs_get_dqblk() argument 225 id = from_kqid(&init_user_ns, qid); in xfs_fs_get_dqblk() 226 return xfs_qm_scall_getquota(mp, id, xfs_quota_type(qid.type), qdq); in xfs_fs_get_dqblk() 233 struct kqid *qid, in xfs_fs_get_nextdqblk() argument 243 id = from_kqid(&init_user_ns, *qid); in xfs_fs_get_nextdqblk() 244 ret = xfs_qm_scall_getquota_next(mp, &id, xfs_quota_type(qid->type), in xfs_fs_get_nextdqblk() 250 *qid = make_kqid(current_user_ns(), qid->type, id); in xfs_fs_get_nextdqblk() 257 struct kqid qid, in xfs_fs_set_dqblk() argument 267 return xfs_qm_scall_setqlim(mp, from_kqid(&init_user_ns, qid), in xfs_fs_set_dqblk() 268 xfs_quota_type(qid.type), qdq); in xfs_fs_set_dqblk()
|
/Linux-v5.15/drivers/vdpa/ifcvf/ |
D | ifcvf_main.c | 272 static int ifcvf_vdpa_get_vq_state(struct vdpa_device *vdpa_dev, u16 qid, in ifcvf_vdpa_get_vq_state() argument 277 state->split.avail_index = ifcvf_get_vq_state(vf, qid); in ifcvf_vdpa_get_vq_state() 281 static int ifcvf_vdpa_set_vq_state(struct vdpa_device *vdpa_dev, u16 qid, in ifcvf_vdpa_set_vq_state() argument 286 return ifcvf_set_vq_state(vf, qid, state->split.avail_index); in ifcvf_vdpa_set_vq_state() 289 static void ifcvf_vdpa_set_vq_cb(struct vdpa_device *vdpa_dev, u16 qid, in ifcvf_vdpa_set_vq_cb() argument 294 vf->vring[qid].cb = *cb; in ifcvf_vdpa_set_vq_cb() 298 u16 qid, bool ready) in ifcvf_vdpa_set_vq_ready() argument 302 vf->vring[qid].ready = ready; in ifcvf_vdpa_set_vq_ready() 305 static bool ifcvf_vdpa_get_vq_ready(struct vdpa_device *vdpa_dev, u16 qid) in ifcvf_vdpa_get_vq_ready() argument 309 return vf->vring[qid].ready; in ifcvf_vdpa_get_vq_ready() [all …]
|
/Linux-v5.15/fs/quota/ |
D | kqid.c | 120 bool qid_valid(struct kqid qid) in qid_valid() argument 122 switch (qid.type) { in qid_valid() 124 return uid_valid(qid.uid); in qid_valid() 126 return gid_valid(qid.gid); in qid_valid() 128 return projid_valid(qid.projid); in qid_valid()
|
/Linux-v5.15/drivers/net/wireless/mediatek/mt76/ |
D | tx.c | 246 __mt76_tx_queue_skb(struct mt76_phy *phy, int qid, struct sk_buff *skb, in __mt76_tx_queue_skb() argument 251 struct mt76_queue *q = phy->q_tx[qid]; in __mt76_tx_queue_skb() 283 int qid = skb_get_queue_mapping(skb); in mt76_tx() local 291 if (WARN_ON(qid >= MT_TXQ_PSD)) { in mt76_tx() 292 qid = MT_TXQ_BE; in mt76_tx() 293 skb_set_queue_mapping(skb, qid); in mt76_tx() 300 qid = MT_TXQ_PSD; in mt76_tx() 301 skb_set_queue_mapping(skb, qid); in mt76_tx() 311 q = phy->q_tx[qid]; in mt76_tx() 314 __mt76_tx_queue_skb(phy, qid, skb, wcid, sta, NULL); in mt76_tx() [all …]
|
/Linux-v5.15/net/9p/ |
D | client.c | 902 memset(&fid->qid, 0, sizeof(struct p9_qid)); in p9_fid_create() 1126 struct p9_qid qid; in p9_client_attach() local 1145 err = p9pdu_readf(&req->rc, clnt->proto_version, "Q", &qid); in p9_client_attach() 1153 qid.type, (unsigned long long)qid.path, qid.version); in p9_client_attach() 1155 memmove(&fid->qid, &qid, sizeof(struct p9_qid)); in p9_client_attach() 1223 memmove(&fid->qid, &wqids[nwqids - 1], sizeof(struct p9_qid)); in p9_client_walk() 1225 memmove(&fid->qid, &oldfid->qid, sizeof(struct p9_qid)); in p9_client_walk() 1248 struct p9_qid qid; in p9_client_open() local 1268 err = p9pdu_readf(&req->rc, clnt->proto_version, "Qd", &qid, &iounit); in p9_client_open() 1275 p9_is_proto_dotl(clnt) ? "RLOPEN" : "ROPEN", qid.type, in p9_client_open() [all …]
|
/Linux-v5.15/drivers/net/ethernet/intel/i40e/ |
D | i40e_xsk.c | 43 u16 qid) in i40e_xsk_pool_enable() argument 52 if (qid >= vsi->num_queue_pairs) in i40e_xsk_pool_enable() 55 if (qid >= netdev->real_num_rx_queues || in i40e_xsk_pool_enable() 56 qid >= netdev->real_num_tx_queues) in i40e_xsk_pool_enable() 63 set_bit(qid, vsi->af_xdp_zc_qps); in i40e_xsk_pool_enable() 68 err = i40e_queue_pair_disable(vsi, qid); in i40e_xsk_pool_enable() 72 err = i40e_queue_pair_enable(vsi, qid); in i40e_xsk_pool_enable() 77 err = i40e_xsk_wakeup(vsi->netdev, qid, XDP_WAKEUP_RX); in i40e_xsk_pool_enable() 93 static int i40e_xsk_pool_disable(struct i40e_vsi *vsi, u16 qid) in i40e_xsk_pool_disable() argument 100 pool = xsk_get_pool_from_qid(netdev, qid); in i40e_xsk_pool_disable() [all …]
|