Home
last modified time | relevance | path

Searched refs:p_hwfn (Results 1 – 25 of 46) sorted by relevance

12

/Linux-v4.19/drivers/net/ethernet/qlogic/qed/
Dqed_dev.c72 static u32 qed_hw_bar_size(struct qed_hwfn *p_hwfn, in qed_hw_bar_size() argument
79 if (IS_VF(p_hwfn->cdev)) in qed_hw_bar_size()
80 return qed_vf_hw_bar_size(p_hwfn, bar_id); in qed_hw_bar_size()
82 val = qed_rd(p_hwfn, p_ptt, bar_reg); in qed_hw_bar_size()
87 if (p_hwfn->cdev->num_hwfns > 1) { in qed_hw_bar_size()
88 DP_INFO(p_hwfn, in qed_hw_bar_size()
92 DP_INFO(p_hwfn, in qed_hw_bar_size()
105 struct qed_hwfn *p_hwfn = &cdev->hwfns[i]; in qed_init_dp() local
107 p_hwfn->dp_level = dp_level; in qed_init_dp()
108 p_hwfn->dp_module = dp_module; in qed_init_dp()
[all …]
Dqed_ptp.c48 static enum qed_resc_lock qed_ptcdev_to_resc(struct qed_hwfn *p_hwfn) in qed_ptcdev_to_resc() argument
50 switch (qed_device_get_port_id(p_hwfn->cdev)) { in qed_ptcdev_to_resc()
64 static int qed_ptp_res_lock(struct qed_hwfn *p_hwfn, struct qed_ptt *p_ptt) in qed_ptp_res_lock() argument
70 resource = qed_ptcdev_to_resc(p_hwfn); in qed_ptp_res_lock()
76 rc = qed_mcp_resc_lock(p_hwfn, p_ptt, &params); in qed_ptp_res_lock()
83 if (p_hwfn->abs_pf_id < p_hwfn->cdev->num_ports_in_engine) in qed_ptp_res_lock()
86 DP_INFO(p_hwfn, "PF doesn't have lock ownership\n"); in qed_ptp_res_lock()
89 DP_INFO(p_hwfn, "Failed to acquire ptp resource lock\n"); in qed_ptp_res_lock()
96 static int qed_ptp_res_unlock(struct qed_hwfn *p_hwfn, struct qed_ptt *p_ptt) in qed_ptp_res_unlock() argument
102 resource = qed_ptcdev_to_resc(p_hwfn); in qed_ptp_res_unlock()
[all …]
Dqed_rdma.c62 int qed_rdma_bmap_alloc(struct qed_hwfn *p_hwfn, in qed_rdma_bmap_alloc() argument
65 DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "max_count = %08x\n", max_count); in qed_rdma_bmap_alloc()
76 DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "0\n"); in qed_rdma_bmap_alloc()
80 int qed_rdma_bmap_alloc_id(struct qed_hwfn *p_hwfn, in qed_rdma_bmap_alloc_id() argument
89 DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "%s bitmap: allocated id %d\n", in qed_rdma_bmap_alloc_id()
95 void qed_bmap_set_id(struct qed_hwfn *p_hwfn, in qed_bmap_set_id() argument
104 void qed_bmap_release_id(struct qed_hwfn *p_hwfn, in qed_bmap_release_id() argument
114 DP_NOTICE(p_hwfn, "%s bitmap: id %d already released\n", in qed_bmap_release_id()
119 DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "%s bitmap: released id %d\n", in qed_bmap_release_id()
123 int qed_bmap_test_id(struct qed_hwfn *p_hwfn, in qed_bmap_test_id() argument
[all …]
Dqed_int.c73 #define SB_ATTN_ALIGNED_SIZE(p_hwfn) \ argument
74 ALIGNED_TYPE_SIZE(struct atten_status_block, p_hwfn)
102 int (*cb)(struct qed_hwfn *p_hwfn);
115 static int qed_mcp_attn_cb(struct qed_hwfn *p_hwfn) in qed_mcp_attn_cb() argument
117 u32 tmp = qed_rd(p_hwfn, p_hwfn->p_dpc_ptt, MCP_REG_CPU_STATE); in qed_mcp_attn_cb()
120 DP_INFO(p_hwfn->cdev, "MCP_REG_CPU_STATE: %08x - Masking...\n", in qed_mcp_attn_cb()
122 qed_wr(p_hwfn, p_hwfn->p_dpc_ptt, MCP_REG_CPU_EVENT_MASK, in qed_mcp_attn_cb()
141 static int qed_pswhst_attn_cb(struct qed_hwfn *p_hwfn) in qed_pswhst_attn_cb() argument
143 u32 tmp = qed_rd(p_hwfn, p_hwfn->p_dpc_ptt, in qed_pswhst_attn_cb()
149 addr = qed_rd(p_hwfn, p_hwfn->p_dpc_ptt, in qed_pswhst_attn_cb()
[all …]
Dqed_mcp.c64 DRV_INNER_WR(p_hwfn, _p_ptt, drv_mb_addr, \
76 bool qed_mcp_is_init(struct qed_hwfn *p_hwfn) in qed_mcp_is_init() argument
78 if (!p_hwfn->mcp_info || !p_hwfn->mcp_info->public_base) in qed_mcp_is_init()
83 void qed_mcp_cmd_port_init(struct qed_hwfn *p_hwfn, struct qed_ptt *p_ptt) in qed_mcp_cmd_port_init() argument
85 u32 addr = SECTION_OFFSIZE_ADDR(p_hwfn->mcp_info->public_base, in qed_mcp_cmd_port_init()
87 u32 mfw_mb_offsize = qed_rd(p_hwfn, p_ptt, addr); in qed_mcp_cmd_port_init()
89 p_hwfn->mcp_info->port_addr = SECTION_ADDR(mfw_mb_offsize, in qed_mcp_cmd_port_init()
90 MFW_PORT(p_hwfn)); in qed_mcp_cmd_port_init()
91 DP_VERBOSE(p_hwfn, QED_MSG_SP, in qed_mcp_cmd_port_init()
93 p_hwfn->mcp_info->port_addr, MFW_PORT(p_hwfn)); in qed_mcp_cmd_port_init()
[all …]
Dqed_hw.c70 int qed_ptt_pool_alloc(struct qed_hwfn *p_hwfn) in qed_ptt_pool_alloc() argument
83 p_pool->ptts[i].hwfn_id = p_hwfn->my_id; in qed_ptt_pool_alloc()
89 p_hwfn->p_ptt_pool = p_pool; in qed_ptt_pool_alloc()
95 void qed_ptt_invalidate(struct qed_hwfn *p_hwfn) in qed_ptt_invalidate() argument
101 p_ptt = &p_hwfn->p_ptt_pool->ptts[i]; in qed_ptt_invalidate()
106 void qed_ptt_pool_free(struct qed_hwfn *p_hwfn) in qed_ptt_pool_free() argument
108 kfree(p_hwfn->p_ptt_pool); in qed_ptt_pool_free()
109 p_hwfn->p_ptt_pool = NULL; in qed_ptt_pool_free()
112 struct qed_ptt *qed_ptt_acquire(struct qed_hwfn *p_hwfn) in qed_ptt_acquire() argument
119 spin_lock_bh(&p_hwfn->p_ptt_pool->lock); in qed_ptt_acquire()
[all …]
Dqed_spq.c73 static void qed_spq_blocking_cb(struct qed_hwfn *p_hwfn, in qed_spq_blocking_cb() argument
87 static int __qed_spq_block(struct qed_hwfn *p_hwfn, in __qed_spq_block() argument
115 static int qed_spq_block(struct qed_hwfn *p_hwfn, in qed_spq_block() argument
127 rc = __qed_spq_block(p_hwfn, p_ent, p_fw_ret, false); in qed_spq_block()
133 rc = __qed_spq_block(p_hwfn, p_ent, p_fw_ret, true); in qed_spq_block()
137 p_ptt = qed_ptt_acquire(p_hwfn); in qed_spq_block()
139 DP_NOTICE(p_hwfn, "ptt, failed to acquire\n"); in qed_spq_block()
143 DP_INFO(p_hwfn, "Ramrod is stuck, requesting MCP drain\n"); in qed_spq_block()
144 rc = qed_mcp_drain(p_hwfn, p_ptt); in qed_spq_block()
146 DP_NOTICE(p_hwfn, "MCP drain failed\n"); in qed_spq_block()
[all …]
Dqed_vf.c39 static void *qed_vf_pf_prep(struct qed_hwfn *p_hwfn, u16 type, u16 length) in qed_vf_pf_prep() argument
41 struct qed_vf_iov *p_iov = p_hwfn->vf_iov_info; in qed_vf_pf_prep()
51 DP_VERBOSE(p_hwfn, in qed_vf_pf_prep()
64 p_tlv = qed_add_tlv(p_hwfn, &p_iov->offset, type, length); in qed_vf_pf_prep()
73 static void qed_vf_pf_req_end(struct qed_hwfn *p_hwfn, int req_status) in qed_vf_pf_req_end() argument
75 union pfvf_tlvs *resp = p_hwfn->vf_iov_info->pf2vf_reply; in qed_vf_pf_req_end()
77 DP_VERBOSE(p_hwfn, QED_MSG_IOV, in qed_vf_pf_req_end()
81 mutex_unlock(&(p_hwfn->vf_iov_info->mutex)); in qed_vf_pf_req_end()
84 static int qed_send_msg2pf(struct qed_hwfn *p_hwfn, u8 *done, u32 resp_size) in qed_send_msg2pf() argument
86 union vfpf_tlvs *p_req = p_hwfn->vf_iov_info->vf2pf_request; in qed_send_msg2pf()
[all …]
Dqed_iwarp.c82 static int qed_iwarp_async_event(struct qed_hwfn *p_hwfn,
88 void qed_iwarp_init_devinfo(struct qed_hwfn *p_hwfn) in qed_iwarp_init_devinfo() argument
90 struct qed_rdma_device *dev = p_hwfn->p_rdma_info->dev; in qed_iwarp_init_devinfo()
95 p_hwfn->p_rdma_info->num_qps) - in qed_iwarp_init_devinfo()
104 void qed_iwarp_init_hw(struct qed_hwfn *p_hwfn, struct qed_ptt *p_ptt) in qed_iwarp_init_hw() argument
106 p_hwfn->rdma_prs_search_reg = PRS_REG_SEARCH_TCP; in qed_iwarp_init_hw()
107 qed_wr(p_hwfn, p_ptt, p_hwfn->rdma_prs_search_reg, 1); in qed_iwarp_init_hw()
108 p_hwfn->b_rdma_enabled_in_prs = true; in qed_iwarp_init_hw()
115 static void qed_iwarp_cid_cleaned(struct qed_hwfn *p_hwfn, u32 cid) in qed_iwarp_cid_cleaned() argument
117 cid -= qed_cxt_get_proto_cid_start(p_hwfn, p_hwfn->p_rdma_info->proto); in qed_iwarp_cid_cleaned()
[all …]
Dqed_cxt.c116 #define CONN_CXT_SIZE(p_hwfn) \ argument
117 ALIGNED_TYPE_SIZE(union conn_context, p_hwfn)
121 #define TYPE0_TASK_CXT_SIZE(p_hwfn) \ argument
122 ALIGNED_TYPE_SIZE(union type0_task_context, p_hwfn)
125 #define TYPE1_TASK_CXT_SIZE(p_hwfn) sizeof(union type1_task_context) argument
316 static void qed_cxt_tm_iids(struct qed_hwfn *p_hwfn, in qed_cxt_tm_iids() argument
374 static void qed_cxt_qm_iids(struct qed_hwfn *p_hwfn, in qed_cxt_qm_iids() argument
377 struct qed_cxt_mngr *p_mngr = p_hwfn->p_cxt_mngr; in qed_cxt_qm_iids()
403 DP_VERBOSE(p_hwfn, QED_MSG_ILT, in qed_cxt_qm_iids()
408 static struct qed_tid_seg *qed_cxt_tid_seg_info(struct qed_hwfn *p_hwfn, in qed_cxt_tid_seg_info() argument
[all …]
Dqed_init_fw_funcs.c181 #define QM_INIT_TX_PQ_MAP(p_hwfn, map, chip, pq_id, rl_valid, vp_pq_id, rl_id, \ argument
195 STORE_RT_REG(p_hwfn, QM_REG_TXPQMAP_RT_OFFSET + (pq_id), \
210 static u8 qed_get_ext_voq(struct qed_hwfn *p_hwfn, in qed_get_ext_voq() argument
220 static void qed_enable_pf_rl(struct qed_hwfn *p_hwfn, bool pf_rl_en) in qed_enable_pf_rl() argument
222 STORE_RT_REG(p_hwfn, QM_REG_RLPFENABLE_RT_OFFSET, pf_rl_en ? 1 : 0); in qed_enable_pf_rl()
228 STORE_RT_REG(p_hwfn, in qed_enable_pf_rl()
232 STORE_RT_REG(p_hwfn, QM_REG_RLPFVOQENABLE_MSB_RT_OFFSET, in qed_enable_pf_rl()
236 STORE_RT_REG(p_hwfn, in qed_enable_pf_rl()
238 STORE_RT_REG(p_hwfn, in qed_enable_pf_rl()
244 STORE_RT_REG(p_hwfn, in qed_enable_pf_rl()
[all …]
Dqed_roce.c63 static void qed_roce_free_real_icid(struct qed_hwfn *p_hwfn, u16 icid);
66 qed_roce_async_event(struct qed_hwfn *p_hwfn, in qed_roce_async_event() argument
70 struct qed_rdma_events events = p_hwfn->p_rdma_info->events; in qed_roce_async_event()
80 qed_roce_free_real_icid(p_hwfn, icid); in qed_roce_async_event()
99 void qed_roce_stop(struct qed_hwfn *p_hwfn) in qed_roce_stop() argument
101 struct qed_bmap *rcid_map = &p_hwfn->p_rdma_info->real_cid_map; in qed_roce_stop()
112 DP_NOTICE(p_hwfn, "cid bitmap wait timed out\n"); in qed_roce_stop()
116 qed_spq_unregister_async_cb(p_hwfn, PROTOCOLID_ROCE); in qed_roce_stop()
155 static void qed_roce_free_cid_pair(struct qed_hwfn *p_hwfn, u16 cid) in qed_roce_free_cid_pair() argument
157 spin_lock_bh(&p_hwfn->p_rdma_info->lock); in qed_roce_free_cid_pair()
[all …]
Dqed_sriov.c47 static int qed_sriov_eqe_event(struct qed_hwfn *p_hwfn,
51 static int qed_iov_bulletin_set_mac(struct qed_hwfn *p_hwfn, u8 *mac, int vfid);
69 static int qed_sp_vf_start(struct qed_hwfn *p_hwfn, struct qed_vf_info *p_vf) in qed_sp_vf_start() argument
79 init_data.cid = qed_spq_get_cid(p_hwfn); in qed_sp_vf_start()
83 rc = qed_sp_init_request(p_hwfn, &p_ent, in qed_sp_vf_start()
94 switch (p_hwfn->hw_info.personality) { in qed_sp_vf_start()
102 DP_NOTICE(p_hwfn, "Unknown VF personality %d\n", in qed_sp_vf_start()
103 p_hwfn->hw_info.personality); in qed_sp_vf_start()
110 DP_VERBOSE(p_hwfn, in qed_sp_vf_start()
122 DP_VERBOSE(p_hwfn, QED_MSG_IOV, in qed_sp_vf_start()
[all …]
Dqed_sp_commands.c50 int qed_sp_init_request(struct qed_hwfn *p_hwfn, in qed_sp_init_request() argument
61 rc = qed_spq_get_entry(p_hwfn, pp_ent); in qed_sp_init_request()
96 DP_NOTICE(p_hwfn, "Unknown SPQE completion mode %d\n", in qed_sp_init_request()
101 DP_VERBOSE(p_hwfn, QED_MSG_SPQ, in qed_sp_init_request()
209 qed_tunn_set_pf_update_params(struct qed_hwfn *p_hwfn, in qed_tunn_set_pf_update_params() argument
213 struct qed_tunnel_info *p_tun = &p_hwfn->cdev->tunnel; in qed_tunn_set_pf_update_params()
243 static void qed_set_hw_tunn_mode(struct qed_hwfn *p_hwfn, in qed_set_hw_tunn_mode() argument
247 qed_set_gre_enable(p_hwfn, p_ptt, p_tun->l2_gre.b_mode_enabled, in qed_set_hw_tunn_mode()
249 qed_set_vxlan_enable(p_hwfn, p_ptt, p_tun->vxlan.b_mode_enabled); in qed_set_hw_tunn_mode()
251 qed_set_geneve_enable(p_hwfn, p_ptt, p_tun->l2_geneve.b_mode_enabled, in qed_set_hw_tunn_mode()
[all …]
Dqed_init_ops.c78 void qed_init_clear_rt_data(struct qed_hwfn *p_hwfn) in qed_init_clear_rt_data() argument
83 p_hwfn->rt_data.b_valid[i] = false; in qed_init_clear_rt_data()
86 void qed_init_store_rt_reg(struct qed_hwfn *p_hwfn, u32 rt_offset, u32 val) in qed_init_store_rt_reg() argument
88 p_hwfn->rt_data.init_val[rt_offset] = val; in qed_init_store_rt_reg()
89 p_hwfn->rt_data.b_valid[rt_offset] = true; in qed_init_store_rt_reg()
92 void qed_init_store_rt_agg(struct qed_hwfn *p_hwfn, in qed_init_store_rt_agg() argument
98 p_hwfn->rt_data.init_val[rt_offset + i] = p_val[i]; in qed_init_store_rt_agg()
99 p_hwfn->rt_data.b_valid[rt_offset + i] = true; in qed_init_store_rt_agg()
103 static int qed_init_rt(struct qed_hwfn *p_hwfn, in qed_init_rt() argument
107 u32 *p_init_val = &p_hwfn->rt_data.init_val[rt_offset]; in qed_init_rt()
[all …]
Dqed_cxt.h65 int qed_cxt_get_cid_info(struct qed_hwfn *p_hwfn,
76 int qed_cxt_get_tid_mem_info(struct qed_hwfn *p_hwfn,
88 u32 qed_cxt_get_proto_cid_count(struct qed_hwfn *p_hwfn,
98 int qed_cxt_set_pf_params(struct qed_hwfn *p_hwfn, u32 rdma_tasks);
108 int qed_cxt_cfg_ilt_compute(struct qed_hwfn *p_hwfn, u32 *last_line);
116 u32 qed_cxt_cfg_ilt_compute_excess(struct qed_hwfn *p_hwfn, u32 used_lines);
125 int qed_cxt_mngr_alloc(struct qed_hwfn *p_hwfn);
132 void qed_cxt_mngr_free(struct qed_hwfn *p_hwfn);
141 int qed_cxt_tables_alloc(struct qed_hwfn *p_hwfn);
148 void qed_cxt_mngr_setup(struct qed_hwfn *p_hwfn);
[all …]
Dqed_int.h88 void qed_int_igu_enable_int(struct qed_hwfn *p_hwfn,
98 void qed_int_igu_disable_int(struct qed_hwfn *p_hwfn,
109 u64 qed_int_igu_read_sisr_reg(struct qed_hwfn *p_hwfn);
128 int qed_int_sb_init(struct qed_hwfn *p_hwfn,
141 void qed_int_sb_setup(struct qed_hwfn *p_hwfn,
158 int qed_int_sb_release(struct qed_hwfn *p_hwfn,
180 void qed_int_get_num_sbs(struct qed_hwfn *p_hwfn,
199 #define SB_ALIGNED_SIZE(p_hwfn) \ argument
200 ALIGNED_TYPE_SIZE(struct status_block_e4, p_hwfn)
236 int qed_int_igu_reset_cam(struct qed_hwfn *p_hwfn, struct qed_ptt *p_ptt);
[all …]
Dqed_mcp.h280 *qed_mcp_get_link_capabilities(struct qed_hwfn *p_hwfn);
291 int qed_mcp_set_link(struct qed_hwfn *p_hwfn,
305 int qed_mcp_get_mfw_ver(struct qed_hwfn *p_hwfn,
318 int qed_mcp_get_mbi_ver(struct qed_hwfn *p_hwfn,
351 int qed_mcp_cmd(struct qed_hwfn *p_hwfn,
365 int qed_mcp_drain(struct qed_hwfn *p_hwfn,
377 int qed_mcp_get_flash_size(struct qed_hwfn *p_hwfn,
392 qed_mcp_send_drv_version(struct qed_hwfn *p_hwfn,
405 int qed_mcp_ov_update_current_config(struct qed_hwfn *p_hwfn,
418 int qed_mcp_ov_update_driver_state(struct qed_hwfn *p_hwfn,
[all …]
Dqed_ooo.h87 void qed_ooo_save_history_entry(struct qed_hwfn *p_hwfn,
91 int qed_ooo_alloc(struct qed_hwfn *p_hwfn);
93 void qed_ooo_setup(struct qed_hwfn *p_hwfn);
95 void qed_ooo_free(struct qed_hwfn *p_hwfn);
97 void qed_ooo_release_connection_isles(struct qed_hwfn *p_hwfn,
101 void qed_ooo_release_all_isles(struct qed_hwfn *p_hwfn,
104 void qed_ooo_put_free_buffer(struct qed_hwfn *p_hwfn,
109 qed_ooo_get_free_buffer(struct qed_hwfn *p_hwfn,
112 void qed_ooo_put_ready_buffer(struct qed_hwfn *p_hwfn,
117 qed_ooo_get_ready_buffer(struct qed_hwfn *p_hwfn,
[all …]
Dqed_l2.c76 int qed_l2_alloc(struct qed_hwfn *p_hwfn) in qed_l2_alloc() argument
82 if (!QED_IS_L2_PERSONALITY(p_hwfn)) in qed_l2_alloc()
88 p_hwfn->p_l2_info = p_l2_info; in qed_l2_alloc()
90 if (IS_PF(p_hwfn->cdev)) { in qed_l2_alloc()
91 p_l2_info->queues = RESC_NUM(p_hwfn, QED_L2_QUEUE); in qed_l2_alloc()
95 qed_vf_get_num_rxqs(p_hwfn, &rx); in qed_l2_alloc()
96 qed_vf_get_num_txqs(p_hwfn, &tx); in qed_l2_alloc()
116 void qed_l2_setup(struct qed_hwfn *p_hwfn) in qed_l2_setup() argument
118 if (!QED_IS_L2_PERSONALITY(p_hwfn)) in qed_l2_setup()
121 mutex_init(&p_hwfn->p_l2_info->lock); in qed_l2_setup()
[all …]
Dqed_selftest.c68 struct qed_hwfn *p_hwfn; in qed_selftest_register() local
74 p_hwfn = &cdev->hwfns[i]; in qed_selftest_register()
75 p_ptt = qed_ptt_acquire(p_hwfn); in qed_selftest_register()
77 DP_ERR(p_hwfn, "failed to acquire ptt\n"); in qed_selftest_register()
80 rc = qed_mcp_bist_register_test(p_hwfn, p_ptt); in qed_selftest_register()
81 qed_ptt_release(p_hwfn, p_ptt); in qed_selftest_register()
91 struct qed_hwfn *p_hwfn; in qed_selftest_clock() local
97 p_hwfn = &cdev->hwfns[i]; in qed_selftest_clock()
98 p_ptt = qed_ptt_acquire(p_hwfn); in qed_selftest_clock()
100 DP_ERR(p_hwfn, "failed to acquire ptt\n"); in qed_selftest_clock()
[all …]
Dqed_ooo.c47 *qed_ooo_seek_archipelago(struct qed_hwfn *p_hwfn, in qed_ooo_seek_archipelago() argument
66 static struct qed_ooo_isle *qed_ooo_seek_isle(struct qed_hwfn *p_hwfn, in qed_ooo_seek_isle() argument
74 p_archipelago = qed_ooo_seek_archipelago(p_hwfn, p_ooo_info, cid); in qed_ooo_seek_isle()
76 DP_NOTICE(p_hwfn, in qed_ooo_seek_isle()
90 void qed_ooo_save_history_entry(struct qed_hwfn *p_hwfn, in qed_ooo_save_history_entry() argument
102 int qed_ooo_alloc(struct qed_hwfn *p_hwfn) in qed_ooo_alloc() argument
110 switch (p_hwfn->hw_info.personality) { in qed_ooo_alloc()
119 DP_NOTICE(p_hwfn, in qed_ooo_alloc()
124 max_num_archipelagos = (u16)qed_cxt_get_proto_cid_count(p_hwfn, proto, in qed_ooo_alloc()
127 cid_base = (u16)qed_cxt_get_proto_cid_start(p_hwfn, proto); in qed_ooo_alloc()
[all …]
Dqed_ll2.c98 struct qed_hwfn *p_hwfn = cxt; in qed_ll2b_complete_tx_packet() local
99 struct qed_dev *cdev = p_hwfn->cdev; in qed_ll2b_complete_tx_packet()
103 dma_unmap_single(&p_hwfn->cdev->pdev->dev, first_frag_addr, in qed_ll2b_complete_tx_packet()
164 struct qed_hwfn *p_hwfn = cxt; in qed_ll2b_complete_rx_packet() local
166 struct qed_dev *cdev = p_hwfn->cdev; in qed_ll2b_complete_rx_packet()
173 DP_VERBOSE(p_hwfn, in qed_ll2b_complete_rx_packet()
194 rc = qed_ll2_alloc_buffer(p_hwfn->cdev, &new_data, in qed_ll2b_complete_rx_packet()
230 DP_VERBOSE(p_hwfn, (NETIF_MSG_RX_STATUS | NETIF_MSG_PKTDATA | in qed_ll2b_complete_rx_packet()
249 static struct qed_ll2_info *__qed_ll2_handle_sanity(struct qed_hwfn *p_hwfn, in __qed_ll2_handle_sanity() argument
259 if (!p_hwfn->p_ll2_info) in __qed_ll2_handle_sanity()
[all …]
Dqed_sp.h68 int qed_eth_cqe_completion(struct qed_hwfn *p_hwfn,
183 (*qed_spq_async_comp_cb)(struct qed_hwfn *p_hwfn,
190 qed_spq_register_async_cb(struct qed_hwfn *p_hwfn,
195 qed_spq_unregister_async_cb(struct qed_hwfn *p_hwfn,
239 int qed_spq_post(struct qed_hwfn *p_hwfn,
250 int qed_spq_alloc(struct qed_hwfn *p_hwfn);
257 void qed_spq_setup(struct qed_hwfn *p_hwfn);
264 void qed_spq_free(struct qed_hwfn *p_hwfn);
278 qed_spq_get_entry(struct qed_hwfn *p_hwfn,
288 void qed_spq_return_entry(struct qed_hwfn *p_hwfn,
[all …]
Dqed_hw.h86 void qed_gtt_init(struct qed_hwfn *p_hwfn);
93 void qed_ptt_invalidate(struct qed_hwfn *p_hwfn);
102 int qed_ptt_pool_alloc(struct qed_hwfn *p_hwfn);
109 void qed_ptt_pool_free(struct qed_hwfn *p_hwfn);
119 u32 qed_ptt_get_hw_addr(struct qed_hwfn *p_hwfn,
139 void qed_ptt_set_win(struct qed_hwfn *p_hwfn,
151 struct qed_ptt *qed_get_reserved_ptt(struct qed_hwfn *p_hwfn,
162 void qed_wr(struct qed_hwfn *p_hwfn,
175 u32 qed_rd(struct qed_hwfn *p_hwfn,
189 void qed_memcpy_from(struct qed_hwfn *p_hwfn,
[all …]

12