Lines Matching refs:iwdev

225 	struct irdma_device *iwdev = rf->iwdev;  in irdma_process_aeq()  local
240 ibdev_dbg(&iwdev->ibdev, in irdma_process_aeq()
252 atomic_dec(&iwdev->vsi.qp_suspend_reqs); in irdma_process_aeq()
253 wake_up(&iwdev->suspend_wq); in irdma_process_aeq()
256 ibdev_dbg(&iwdev->ibdev, "AEQ: qp_id %d is already freed\n", in irdma_process_aeq()
323 if (iwqp->iwdev->vsi.tc_change_pending) { in irdma_process_aeq()
325 wake_up(&iwqp->iwdev->suspend_wq); in irdma_process_aeq()
335 ibdev_err(&iwdev->ibdev, in irdma_process_aeq()
389 ibdev_err(&iwdev->ibdev, "abnormal ae_id = 0x%x bool qp=%d qp_id = %d, ae_src=%d\n", in irdma_process_aeq()
391 if (rdma_protocol_roce(&iwdev->ibdev, 1)) { in irdma_process_aeq()
1076 static int irdma_alloc_set_mac(struct irdma_device *iwdev) in irdma_alloc_set_mac() argument
1080 status = irdma_alloc_local_mac_entry(iwdev->rf, in irdma_alloc_set_mac()
1081 &iwdev->mac_ip_table_idx); in irdma_alloc_set_mac()
1083 status = irdma_add_local_mac_entry(iwdev->rf, in irdma_alloc_set_mac()
1084 (const u8 *)iwdev->netdev->dev_addr, in irdma_alloc_set_mac()
1085 (u8)iwdev->mac_ip_table_idx); in irdma_alloc_set_mac()
1087 irdma_del_local_mac_entry(iwdev->rf, in irdma_alloc_set_mac()
1088 (u8)iwdev->mac_ip_table_idx); in irdma_alloc_set_mac()
1128 ibdev_dbg(&rf->iwdev->ibdev, "ERR: ceq irq config fail\n"); in irdma_cfg_ceq_vector()
1158 ibdev_dbg(&rf->iwdev->ibdev, "ERR: aeq irq config fail\n"); in irdma_cfg_aeq_vector()
1249 ibdev_dbg(&rf->iwdev->ibdev, "ERR: create ceq status = %d\n", in irdma_setup_ceq_0()
1303 ibdev_dbg(&rf->iwdev->ibdev, in irdma_setup_ceqs()
1463 static int irdma_initialize_ilq(struct irdma_device *iwdev) in irdma_initialize_ilq() argument
1474 info.sq_size = min(iwdev->rf->max_qp / 2, (u32)32768); in irdma_initialize_ilq()
1480 status = irdma_puda_create_rsrc(&iwdev->vsi, &info); in irdma_initialize_ilq()
1482 ibdev_dbg(&iwdev->ibdev, "ERR: ilq create fail\n"); in irdma_initialize_ilq()
1493 static int irdma_initialize_ieq(struct irdma_device *iwdev) in irdma_initialize_ieq() argument
1500 info.qp_id = iwdev->vsi.exception_lan_q; in irdma_initialize_ieq()
1504 info.sq_size = min(iwdev->rf->max_qp / 2, (u32)32768); in irdma_initialize_ieq()
1506 info.buf_size = iwdev->vsi.mtu + IRDMA_IPV4_PAD; in irdma_initialize_ieq()
1508 status = irdma_puda_create_rsrc(&iwdev->vsi, &info); in irdma_initialize_ieq()
1510 ibdev_dbg(&iwdev->ibdev, "ERR: ieq create fail\n"); in irdma_initialize_ieq()
1521 struct irdma_device *iwdev = vsi->back_vsi; in irdma_reinitialize_ieq() local
1522 struct irdma_pci_f *rf = iwdev->rf; in irdma_reinitialize_ieq()
1525 if (irdma_initialize_ieq(iwdev)) { in irdma_reinitialize_ieq()
1526 iwdev->rf->reset = true; in irdma_reinitialize_ieq()
1650 void irdma_rt_deinit_hw(struct irdma_device *iwdev) in irdma_rt_deinit_hw() argument
1652 ibdev_dbg(&iwdev->ibdev, "INIT: state = %d\n", iwdev->init_state); in irdma_rt_deinit_hw()
1654 switch (iwdev->init_state) { in irdma_rt_deinit_hw()
1656 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_rt_deinit_hw()
1657 irdma_del_local_mac_entry(iwdev->rf, in irdma_rt_deinit_hw()
1658 (u8)iwdev->mac_ip_table_idx); in irdma_rt_deinit_hw()
1664 if (!iwdev->roce_mode) in irdma_rt_deinit_hw()
1665 irdma_puda_dele_rsrc(&iwdev->vsi, IRDMA_PUDA_RSRC_TYPE_IEQ, in irdma_rt_deinit_hw()
1666 iwdev->rf->reset); in irdma_rt_deinit_hw()
1669 if (!iwdev->roce_mode) in irdma_rt_deinit_hw()
1670 irdma_puda_dele_rsrc(&iwdev->vsi, in irdma_rt_deinit_hw()
1672 iwdev->rf->reset); in irdma_rt_deinit_hw()
1675 ibdev_warn(&iwdev->ibdev, "bad init_state = %d\n", iwdev->init_state); in irdma_rt_deinit_hw()
1679 irdma_cleanup_cm_core(&iwdev->cm_core); in irdma_rt_deinit_hw()
1680 if (iwdev->vsi.pestat) { in irdma_rt_deinit_hw()
1681 irdma_vsi_stats_free(&iwdev->vsi); in irdma_rt_deinit_hw()
1682 kfree(iwdev->vsi.pestat); in irdma_rt_deinit_hw()
1684 if (iwdev->cleanup_wq) in irdma_rt_deinit_hw()
1685 destroy_workqueue(iwdev->cleanup_wq); in irdma_rt_deinit_hw()
1728 static void irdma_get_used_rsrc(struct irdma_device *iwdev) in irdma_get_used_rsrc() argument
1730 iwdev->rf->used_pds = find_first_zero_bit(iwdev->rf->allocated_pds, in irdma_get_used_rsrc()
1731 iwdev->rf->max_pd); in irdma_get_used_rsrc()
1732 iwdev->rf->used_qps = find_first_zero_bit(iwdev->rf->allocated_qps, in irdma_get_used_rsrc()
1733 iwdev->rf->max_qp); in irdma_get_used_rsrc()
1734 iwdev->rf->used_cqs = find_first_zero_bit(iwdev->rf->allocated_cqs, in irdma_get_used_rsrc()
1735 iwdev->rf->max_cq); in irdma_get_used_rsrc()
1736 iwdev->rf->used_mrs = find_first_zero_bit(iwdev->rf->allocated_mrs, in irdma_get_used_rsrc()
1737 iwdev->rf->max_mr); in irdma_get_used_rsrc()
1771 ibdev_warn(&rf->iwdev->ibdev, "bad init_state = %d\n", rf->init_state); in irdma_ctrl_deinit_hw()
1784 int irdma_rt_init_hw(struct irdma_device *iwdev, in irdma_rt_init_hw() argument
1787 struct irdma_pci_f *rf = iwdev->rf; in irdma_rt_init_hw()
1794 vsi_info.back_vsi = iwdev; in irdma_rt_init_hw()
1796 vsi_info.pf_data_vsi_num = iwdev->vsi_num; in irdma_rt_init_hw()
1800 irdma_sc_vsi_init(&iwdev->vsi, &vsi_info); in irdma_rt_init_hw()
1802 status = irdma_setup_cm_core(iwdev, rf->rdma_ver); in irdma_rt_init_hw()
1808 irdma_cleanup_cm_core(&iwdev->cm_core); in irdma_rt_init_hw()
1812 status = irdma_vsi_stats_init(&iwdev->vsi, &stats_info); in irdma_rt_init_hw()
1814 irdma_cleanup_cm_core(&iwdev->cm_core); in irdma_rt_init_hw()
1820 if (!iwdev->roce_mode) { in irdma_rt_init_hw()
1821 status = irdma_initialize_ilq(iwdev); in irdma_rt_init_hw()
1824 iwdev->init_state = ILQ_CREATED; in irdma_rt_init_hw()
1825 status = irdma_initialize_ieq(iwdev); in irdma_rt_init_hw()
1828 iwdev->init_state = IEQ_CREATED; in irdma_rt_init_hw()
1831 status = irdma_setup_ceqs(rf, &iwdev->vsi); in irdma_rt_init_hw()
1835 iwdev->init_state = CEQS_CREATED; in irdma_rt_init_hw()
1844 iwdev->init_state = PBLE_CHUNK_MEM; in irdma_rt_init_hw()
1852 iwdev->init_state = AEQ_CREATED; in irdma_rt_init_hw()
1856 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_rt_init_hw()
1857 irdma_alloc_set_mac(iwdev); in irdma_rt_init_hw()
1858 irdma_add_ip(iwdev); in irdma_rt_init_hw()
1859 iwdev->init_state = IP_ADDR_REGISTERED; in irdma_rt_init_hw()
1864 iwdev->cleanup_wq = alloc_workqueue("irdma-cleanup-wq", in irdma_rt_init_hw()
1866 if (!iwdev->cleanup_wq) in irdma_rt_init_hw()
1868 irdma_get_used_rsrc(iwdev); in irdma_rt_init_hw()
1869 init_waitqueue_head(&iwdev->suspend_wq); in irdma_rt_init_hw()
1875 status, iwdev->init_state); in irdma_rt_init_hw()
1876 irdma_rt_deinit_hw(iwdev); in irdma_rt_init_hw()
2087 ibdev_err(&rf->iwdev->ibdev, "cqp opcode = 0x%x maj_err_code = 0x%x min_err_code = 0x%x\n", in irdma_cqp_ce_handler()
2176 irdma_hw_modify_qp(iwqp->iwdev, iwqp, &info, 0); in irdma_next_iw_state()
2283 static int irdma_cqp_manage_apbvt_cmd(struct irdma_device *iwdev, in irdma_cqp_manage_apbvt_cmd() argument
2291 cqp_request = irdma_alloc_and_get_cqp_request(&iwdev->rf->cqp, add_port); in irdma_cqp_manage_apbvt_cmd()
2302 cqp_info->in.u.manage_apbvt_entry.cqp = &iwdev->rf->cqp.sc_cqp; in irdma_cqp_manage_apbvt_cmd()
2304 ibdev_dbg(&iwdev->ibdev, "DEV: %s: port=0x%04x\n", in irdma_cqp_manage_apbvt_cmd()
2307 status = irdma_handle_cqp_op(iwdev->rf, cqp_request); in irdma_cqp_manage_apbvt_cmd()
2308 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_cqp_manage_apbvt_cmd()
2318 struct irdma_apbvt_entry *irdma_add_apbvt(struct irdma_device *iwdev, u16 port) in irdma_add_apbvt() argument
2320 struct irdma_cm_core *cm_core = &iwdev->cm_core; in irdma_add_apbvt()
2342 if (irdma_cqp_manage_apbvt_cmd(iwdev, port, true)) { in irdma_add_apbvt()
2355 void irdma_del_apbvt(struct irdma_device *iwdev, in irdma_del_apbvt() argument
2358 struct irdma_cm_core *cm_core = &iwdev->cm_core; in irdma_del_apbvt()
2372 irdma_cqp_manage_apbvt_cmd(iwdev, entry->port, false); in irdma_del_apbvt()
2447 int irdma_manage_qhash(struct irdma_device *iwdev, struct irdma_cm_info *cminfo, in irdma_manage_qhash() argument
2453 struct irdma_cqp *iwcqp = &iwdev->rf->cqp; in irdma_manage_qhash()
2466 info->vsi = &iwdev->vsi; in irdma_manage_qhash()
2477 ether_addr_copy(info->mac_addr, iwdev->netdev->dev_addr); in irdma_manage_qhash()
2502 ibdev_dbg(&iwdev->ibdev, in irdma_manage_qhash()
2510 ibdev_dbg(&iwdev->ibdev, in irdma_manage_qhash()
2518 cqp_info->in.u.manage_qhash_table_entry.cqp = &iwdev->rf->cqp.sc_cqp; in irdma_manage_qhash()
2522 status = irdma_handle_cqp_op(iwdev->rf, cqp_request); in irdma_manage_qhash()
2560 ibdev_err(&iwqp->iwdev->ibdev, "Flush QP[%d] failed, SQ has more work", in irdma_hw_flush_wqes_callback()
2649 ibdev_err(&iwqp->iwdev->ibdev, "fatal QP event: SQ in error but not flushed, qp: %d", in irdma_hw_flush_wqes()
2665 ibdev_dbg(&rf->iwdev->ibdev, in irdma_hw_flush_wqes()
2710 struct irdma_pci_f *rf = iwqp->iwdev->rf; in irdma_flush_wqes()
2740 queue_delayed_work(iwqp->iwdev->cleanup_wq, in irdma_flush_wqes()