Home
last modified time | relevance | path

Searched refs:eqe (Results 1 – 25 of 41) sorted by relevance

12

/Linux-v5.4/drivers/net/ethernet/mellanox/mlx4/
Deq.c123 struct mlx4_eqe *eqe = get_eqe(eq, eq->cons_index, eqe_factor, size); in next_eqe_sw() local
124 return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe; in next_eqe_sw()
129 struct mlx4_eqe *eqe = in next_slave_event_eqe() local
131 return (!!(eqe->owner & 0x80) ^ in next_slave_event_eqe()
133 eqe : NULL; in next_slave_event_eqe()
146 struct mlx4_eqe *eqe; in mlx4_gen_slave_eqe() local
150 for (eqe = next_slave_event_eqe(slave_eq); eqe; in mlx4_gen_slave_eqe()
151 eqe = next_slave_event_eqe(slave_eq)) { in mlx4_gen_slave_eqe()
152 slave = eqe->slave_id; in mlx4_gen_slave_eqe()
154 if (eqe->type == MLX4_EVENT_TYPE_PORT_CHANGE && in mlx4_gen_slave_eqe()
[all …]
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/
Devents.c142 struct mlx5_eqe *eqe = data; in any_notifier() local
145 eqe_type_str(eqe->type), eqe->sub_type); in any_notifier()
154 struct mlx5_eqe *eqe = data; in temp_warn() local
158 value_lsb = be64_to_cpu(eqe->data.temp_warning.sensor_warning_lsb); in temp_warn()
159 value_msb = be64_to_cpu(eqe->data.temp_warning.sensor_warning_msb); in temp_warn()
216 struct mlx5_eqe *eqe = data; in port_module() local
224 module_event_eqe = &eqe->data.port_module; in port_module()
307 struct mlx5_eqe *eqe = data; in pcie_core() local
309 switch (eqe->sub_type) { in pcie_core()
330 struct mlx5_eqe *eqe = data; in forward_event() local
[all …]
Deq.c135 struct mlx5_eqe *eqe; in mlx5_eq_comp_int() local
139 eqe = next_eqe_sw(eq); in mlx5_eq_comp_int()
140 if (!eqe) in mlx5_eq_comp_int()
151 cqn = be32_to_cpu(eqe->data.comp.cqn) & 0xffffff; in mlx5_eq_comp_int()
156 cq->comp(cq, eqe); in mlx5_eq_comp_int()
164 } while ((++num_eqes < MLX5_EQ_POLLING_BUDGET) && (eqe = next_eqe_sw(eq))); in mlx5_eq_comp_int()
201 struct mlx5_eqe *eqe; in mlx5_eq_async_int() local
207 eqe = next_eqe_sw(eq); in mlx5_eq_async_int()
208 if (!eqe) in mlx5_eq_async_int()
218 atomic_notifier_call_chain(&eqt->nh[eqe->type], eqe->type, eqe); in mlx5_eq_async_int()
[all …]
Dqp.c132 struct mlx5_eqe *eqe; in rsc_event_notifier() local
137 eqe = data; in rsc_event_notifier()
138 rsn = be32_to_cpu(eqe->data.dct.dctn) & 0xffffff; in rsc_event_notifier()
149 eqe = data; in rsc_event_notifier()
150 rsn = be32_to_cpu(eqe->data.qp_srq.qp_srq_n) & 0xffffff; in rsc_event_notifier()
151 rsn |= (eqe->data.qp_srq.type << MLX5_USER_INDEX_LEN); in rsc_event_notifier()
161 mlx5_core_dbg(dev, "event (%d) arrived on resource 0x%x\n", eqe->type, rsn); in rsc_event_notifier()
Dpagealloc.c458 struct mlx5_eqe *eqe; in req_pages_handler() local
465 eqe = data; in req_pages_handler()
467 func_id = be16_to_cpu(eqe->data.req_pages.func_id); in req_pages_handler()
468 npages = be32_to_cpu(eqe->data.req_pages.num_pages); in req_pages_handler()
469 ec_function = be16_to_cpu(eqe->data.req_pages.ec_function) & EC_FUNCTION_MASK; in req_pages_handler()
Den_txrx.c194 void mlx5e_completion_event(struct mlx5_core_cq *mcq, struct mlx5_eqe *eqe) in mlx5e_completion_event() argument
Dcq.c72 struct mlx5_eqe *eqe) in mlx5_add_cq_to_tasklet() argument
/Linux-v5.4/drivers/infiniband/hw/mthca/
Dmthca_eq.c236 struct mthca_eqe *eqe; in next_eqe_sw() local
237 eqe = get_eqe(eq, eq->cons_index); in next_eqe_sw()
238 return (MTHCA_EQ_ENTRY_OWNER_HW & eqe->owner) ? NULL : eqe; in next_eqe_sw()
241 static inline void set_eqe_hw(struct mthca_eqe *eqe) in set_eqe_hw() argument
243 eqe->owner = MTHCA_EQ_ENTRY_OWNER_HW; in set_eqe_hw()
262 struct mthca_eqe *eqe; in mthca_eq_int() local
267 while ((eqe = next_eqe_sw(eq))) { in mthca_eq_int()
274 switch (eqe->type) { in mthca_eq_int()
276 disarm_cqn = be32_to_cpu(eqe->event.comp.cqn) & 0xffffff; in mthca_eq_int()
282 mthca_qp_event(dev, be32_to_cpu(eqe->event.qp.qpn) & 0xffffff, in mthca_eq_int()
[all …]
/Linux-v5.4/drivers/net/ethernet/mellanox/mlxsw/
Dpci_hw.h218 MLXSW_ITEM32(pci, eqe, event_type, 0x0C, 24, 8);
225 MLXSW_ITEM32(pci, eqe, event_sub_type, 0x0C, 16, 8);
230 MLXSW_ITEM32(pci, eqe, cqn, 0x0C, 8, 7);
235 MLXSW_ITEM32(pci, eqe, owner, 0x0C, 0, 1);
240 MLXSW_ITEM32(pci, eqe, cmd_token, 0x00, 16, 16);
245 MLXSW_ITEM32(pci, eqe, cmd_status, 0x00, 0, 8);
250 MLXSW_ITEM32(pci, eqe, cmd_out_param_h, 0x04, 0, 32);
255 MLXSW_ITEM32(pci, eqe, cmd_out_param_l, 0x08, 0, 32);
Dpci.c694 static void mlxsw_pci_eq_cmd_event(struct mlxsw_pci *mlxsw_pci, char *eqe) in mlxsw_pci_eq_cmd_event() argument
696 mlxsw_pci->cmd.comp.status = mlxsw_pci_eqe_cmd_status_get(eqe); in mlxsw_pci_eq_cmd_event()
698 ((u64) mlxsw_pci_eqe_cmd_out_param_h_get(eqe)) << 32 | in mlxsw_pci_eq_cmd_event()
699 mlxsw_pci_eqe_cmd_out_param_l_get(eqe); in mlxsw_pci_eq_cmd_event()
726 char *eqe; in mlxsw_pci_eq_tasklet() local
734 while ((eqe = mlxsw_pci_eq_sw_eqe_get(q))) { in mlxsw_pci_eq_tasklet()
742 mlxsw_pci_eq_cmd_event(mlxsw_pci, eqe); in mlxsw_pci_eq_tasklet()
746 cqn = mlxsw_pci_eqe_cqn_get(eqe); in mlxsw_pci_eq_tasklet()
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/fpga/
Dcore.c151 static int fpga_err_event(struct notifier_block *nb, unsigned long event, void *eqe) in fpga_err_event() argument
155 return mlx5_fpga_event(fdev, event, eqe); in fpga_err_event()
158 static int fpga_qp_err_event(struct notifier_block *nb, unsigned long event, void *eqe) in fpga_qp_err_event() argument
162 return mlx5_fpga_event(fdev, event, eqe); in fpga_qp_err_event()
314 unsigned long event, void *eqe) in mlx5_fpga_event() argument
316 void *data = ((struct mlx5_eqe *)eqe)->data.raw; in mlx5_fpga_event()
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/lib/
Deq.h57 struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & (eq->nent - 1)); in next_eqe_sw() local
59 return ((eqe->owner & 1) ^ !!(eq->cons_index & eq->nent)) ? NULL : eqe; in next_eqe_sw()
Dclock.c472 struct mlx5_eqe *eqe = data; in mlx5_pps_event() local
473 int pin = eqe->data.pps.pin; in mlx5_pps_event()
481 be64_to_cpu(eqe->data.pps.time_stamp)); in mlx5_pps_event()
/Linux-v5.4/drivers/infiniband/hw/vmw_pvrdma/
Dpvrdma_main.c440 struct pvrdma_eqe *eqe; in pvrdma_intr1_handler() local
442 eqe = get_eqe(dev, head); in pvrdma_intr1_handler()
444 switch (eqe->type) { in pvrdma_intr1_handler()
453 pvrdma_qp_event(dev, eqe->info, eqe->type); in pvrdma_intr1_handler()
457 pvrdma_cq_event(dev, eqe->info, eqe->type); in pvrdma_intr1_handler()
462 pvrdma_srq_event(dev, eqe->info, eqe->type); in pvrdma_intr1_handler()
472 pvrdma_dev_event(dev, eqe->info, eqe->type); in pvrdma_intr1_handler()
476 pvrdma_dev_event(dev, 1, eqe->type); in pvrdma_intr1_handler()
/Linux-v5.4/include/linux/mlx5/
Dcq.h50 void (*comp)(struct mlx5_core_cq *cq, struct mlx5_eqe *eqe);
58 void (*comp)(struct mlx5_core_cq *cq, struct mlx5_eqe *eqe);
/Linux-v5.4/drivers/infiniband/hw/mlx4/
Dmad.c63 #define GET_BLK_PTR_FROM_EQE(eqe) be32_to_cpu(eqe->event.port_mgmt_change.params.tbl_change_info.bl… argument
64 #define GET_MASK_FROM_EQE(eqe) be32_to_cpu(eqe->event.port_mgmt_change.params.tbl_change_info.tbl_e… argument
1128 struct mlx4_eqe *eqe) in propagate_pkey_ev() argument
1130 __propagate_pkey_ev(dev, port_num, GET_BLK_PTR_FROM_EQE(eqe), in propagate_pkey_ev()
1131 GET_MASK_FROM_EQE(eqe)); in propagate_pkey_ev()
1189 struct mlx4_eqe *eqe = &(ew->ib_eqe); in handle_port_mgmt_change_event() local
1190 u8 port = eqe->event.port_mgmt_change.port; in handle_port_mgmt_change_event()
1195 switch (eqe->subtype) { in handle_port_mgmt_change_event()
1197 changed_attr = be32_to_cpu(eqe->event.port_mgmt_change.params.port_info.changed_attr); in handle_port_mgmt_change_event()
1202 u16 lid = be16_to_cpu(eqe->event.port_mgmt_change.params.port_info.mstr_sm_lid); in handle_port_mgmt_change_event()
[all …]
/Linux-v5.4/drivers/crypto/hisilicon/
Dqm.c69 #define QM_EQE_PHASE(eqe) (((eqe)->dw0 >> 16) & 0x1) argument
443 static struct hisi_qp *qm_to_hisi_qp(struct hisi_qm *qm, struct qm_eqe *eqe) in qm_to_hisi_qp() argument
445 u16 cqn = eqe->dw0 & QM_EQE_CQN_MASK; in qm_to_hisi_qp()
492 struct qm_eqe *eqe = qm->eqe + qm->status.eq_head; in qm_irq_handler() local
496 while (QM_EQE_PHASE(eqe) == qm->status.eqc_phase) { in qm_irq_handler()
498 qp = qm_to_hisi_qp(qm, eqe); in qm_irq_handler()
504 eqe = qm->eqe; in qm_irq_handler()
507 eqe++; in qm_irq_handler()
1648 QM_INIT_BUF(qm, eqe, QM_Q_DEPTH); in __hisi_qm_start()
1658 qm->eqe, (unsigned long)qm->eqe_dma, in __hisi_qm_start()
Dqm.h140 struct qm_eqe *eqe; member
/Linux-v5.4/drivers/net/ethernet/ibm/ehea/
Dehea_main.c921 struct ehea_eqe *eqe; in ehea_qp_aff_irq_handler() local
927 eqe = ehea_poll_eq(port->qp_eq); in ehea_qp_aff_irq_handler()
929 while (eqe) { in ehea_qp_aff_irq_handler()
930 qp_token = EHEA_BMASK_GET(EHEA_EQE_QP_TOKEN, eqe->entry); in ehea_qp_aff_irq_handler()
932 eqe->entry, qp_token); in ehea_qp_aff_irq_handler()
946 eqe = ehea_poll_eq(port->qp_eq); in ehea_qp_aff_irq_handler()
1134 static void ehea_parse_eqe(struct ehea_adapter *adapter, u64 eqe) in ehea_parse_eqe() argument
1142 ec = EHEA_BMASK_GET(NEQE_EVENT_CODE, eqe); in ehea_parse_eqe()
1143 portnum = EHEA_BMASK_GET(NEQE_PORTNUM, eqe); in ehea_parse_eqe()
1154 if (EHEA_BMASK_GET(NEQE_PORT_UP, eqe)) { in ehea_parse_eqe()
[all …]
Dehea_qmr.c305 struct ehea_eqe *eqe; in ehea_poll_eq() local
309 eqe = hw_eqit_eq_get_inc_valid(&eq->hw_queue); in ehea_poll_eq()
312 return eqe; in ehea_poll_eq()
/Linux-v5.4/drivers/infiniband/hw/mlx5/
Dodp.c1365 struct mlx5_eqe *eqe; in mlx5_ib_eq_pf_process() local
1368 while ((eqe = mlx5_eq_get_eqe(eq->core, cc))) { in mlx5_ib_eq_pf_process()
1375 pf_eqe = &eqe->data.page_fault; in mlx5_ib_eq_pf_process()
1376 pfault->event_subtype = eqe->sub_type; in mlx5_ib_eq_pf_process()
1381 eqe->sub_type, pfault->bytes_committed); in mlx5_ib_eq_pf_process()
1383 switch (eqe->sub_type) { in mlx5_ib_eq_pf_process()
1432 eqe->sub_type); in mlx5_ib_eq_pf_process()
Ddevx.c305 static u16 get_event_obj_type(unsigned long event_type, struct mlx5_eqe *eqe) in get_event_obj_type() argument
318 return eqe->data.qp_srq.type; in get_event_obj_type()
326 return MLX5_GET(affiliated_event_header, &eqe->data, obj_type); in get_event_obj_type()
1374 static void devx_cq_comp(struct mlx5_core_cq *mcq, struct mlx5_eqe *eqe) in devx_cq_comp() argument
1392 dispatch_event_fd(&obj_event->obj_sub_list, eqe); in devx_cq_comp()
2269 struct mlx5_eqe *eqe = data; in devx_get_obj_id_from_event() local
2283 obj_id = be32_to_cpu(eqe->data.qp_srq.qp_srq_n) & 0xffffff; in devx_get_obj_id_from_event()
2286 obj_id = be32_to_cpu(eqe->data.xrq_err.type_xrqn) & 0xffffff; in devx_get_obj_id_from_event()
2290 obj_id = be32_to_cpu(eqe->data.dct.dctn) & 0xffffff; in devx_get_obj_id_from_event()
2293 obj_id = be32_to_cpu(eqe->data.cq_err.cqn) & 0xffffff; in devx_get_obj_id_from_event()
[all …]
Dsrq_cmd.c663 struct mlx5_eqe *eqe; in srq_event_notifier() local
672 eqe = data; in srq_event_notifier()
673 srqn = be32_to_cpu(eqe->data.qp_srq.qp_srq_n) & 0xffffff; in srq_event_notifier()
684 srq->event(srq, eqe->type); in srq_event_notifier()
/Linux-v5.4/drivers/scsi/be2iscsi/
Dbe_main.c668 struct be_eq_entry *eqe; in be_isr_mcc() local
678 eqe = queue_tail_node(eq); in be_isr_mcc()
681 while (eqe->dw[offsetof(struct amap_eq_entry, valid) / 32] in be_isr_mcc()
683 if (((eqe->dw[offsetof(struct amap_eq_entry, in be_isr_mcc()
688 AMAP_SET_BITS(struct amap_eq_entry, valid, eqe, 0); in be_isr_mcc()
690 eqe = queue_tail_node(eq); in be_isr_mcc()
732 struct be_eq_entry *eqe; in be_isr() local
753 eqe = queue_tail_node(eq); in be_isr()
757 while (eqe->dw[offsetof(struct amap_eq_entry, valid) / 32] in be_isr()
759 if (((eqe->dw[offsetof(struct amap_eq_entry, in be_isr()
[all …]
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/en/
Dmonitor_stats.c61 unsigned long event, void *eqe) in mlx5e_monitor_event_handler() argument

12