/Linux-v5.4/drivers/net/ethernet/qlogic/qed/ |
D | qed_dev.c | 87 static void qed_db_recovery_dp_entry(struct qed_hwfn *p_hwfn, in qed_db_recovery_dp_entry() argument 91 DP_VERBOSE(p_hwfn, in qed_db_recovery_dp_entry() 136 struct qed_hwfn *p_hwfn; in qed_db_rec_find_hwfn() local 140 p_hwfn = db_addr < cdev->hwfns[1].doorbells ? in qed_db_rec_find_hwfn() 143 p_hwfn = QED_LEADING_HWFN(cdev); in qed_db_rec_find_hwfn() 145 return p_hwfn; in qed_db_rec_find_hwfn() 156 struct qed_hwfn *p_hwfn; in qed_db_recovery_add() local 170 p_hwfn = qed_db_rec_find_hwfn(cdev, db_addr); in qed_db_recovery_add() 184 db_entry->hwfn_idx = p_hwfn->my_id; in qed_db_recovery_add() 187 qed_db_recovery_dp_entry(p_hwfn, db_entry, "Adding"); in qed_db_recovery_add() [all …]
|
D | qed_ptp.c | 50 static enum qed_resc_lock qed_ptcdev_to_resc(struct qed_hwfn *p_hwfn) in qed_ptcdev_to_resc() argument 52 switch (MFW_PORT(p_hwfn)) { in qed_ptcdev_to_resc() 66 static int qed_ptp_res_lock(struct qed_hwfn *p_hwfn, struct qed_ptt *p_ptt) in qed_ptp_res_lock() argument 72 resource = qed_ptcdev_to_resc(p_hwfn); in qed_ptp_res_lock() 78 rc = qed_mcp_resc_lock(p_hwfn, p_ptt, ¶ms); in qed_ptp_res_lock() 85 if (p_hwfn->abs_pf_id < p_hwfn->cdev->num_ports_in_engine) in qed_ptp_res_lock() 88 DP_INFO(p_hwfn, "PF doesn't have lock ownership\n"); in qed_ptp_res_lock() 91 DP_INFO(p_hwfn, "Failed to acquire ptp resource lock\n"); in qed_ptp_res_lock() 98 static int qed_ptp_res_unlock(struct qed_hwfn *p_hwfn, struct qed_ptt *p_ptt) in qed_ptp_res_unlock() argument 104 resource = qed_ptcdev_to_resc(p_hwfn); in qed_ptp_res_unlock() [all …]
|
D | qed_mcp.c | 64 DRV_INNER_WR(p_hwfn, _p_ptt, drv_mb_addr, \ 76 bool qed_mcp_is_init(struct qed_hwfn *p_hwfn) in qed_mcp_is_init() argument 78 if (!p_hwfn->mcp_info || !p_hwfn->mcp_info->public_base) in qed_mcp_is_init() 83 void qed_mcp_cmd_port_init(struct qed_hwfn *p_hwfn, struct qed_ptt *p_ptt) in qed_mcp_cmd_port_init() argument 85 u32 addr = SECTION_OFFSIZE_ADDR(p_hwfn->mcp_info->public_base, in qed_mcp_cmd_port_init() 87 u32 mfw_mb_offsize = qed_rd(p_hwfn, p_ptt, addr); in qed_mcp_cmd_port_init() 89 p_hwfn->mcp_info->port_addr = SECTION_ADDR(mfw_mb_offsize, in qed_mcp_cmd_port_init() 90 MFW_PORT(p_hwfn)); in qed_mcp_cmd_port_init() 91 DP_VERBOSE(p_hwfn, QED_MSG_SP, in qed_mcp_cmd_port_init() 93 p_hwfn->mcp_info->port_addr, MFW_PORT(p_hwfn)); in qed_mcp_cmd_port_init() [all …]
|
D | qed_int.c | 73 #define SB_ATTN_ALIGNED_SIZE(p_hwfn) \ argument 74 ALIGNED_TYPE_SIZE(struct atten_status_block, p_hwfn) 102 int (*cb)(struct qed_hwfn *p_hwfn); 115 static int qed_mcp_attn_cb(struct qed_hwfn *p_hwfn) in qed_mcp_attn_cb() argument 117 u32 tmp = qed_rd(p_hwfn, p_hwfn->p_dpc_ptt, MCP_REG_CPU_STATE); in qed_mcp_attn_cb() 120 DP_INFO(p_hwfn->cdev, "MCP_REG_CPU_STATE: %08x - Masking...\n", in qed_mcp_attn_cb() 122 qed_wr(p_hwfn, p_hwfn->p_dpc_ptt, MCP_REG_CPU_EVENT_MASK, in qed_mcp_attn_cb() 141 static int qed_pswhst_attn_cb(struct qed_hwfn *p_hwfn) in qed_pswhst_attn_cb() argument 143 u32 tmp = qed_rd(p_hwfn, p_hwfn->p_dpc_ptt, in qed_pswhst_attn_cb() 149 addr = qed_rd(p_hwfn, p_hwfn->p_dpc_ptt, in qed_pswhst_attn_cb() [all …]
|
D | qed_rdma.c | 62 int qed_rdma_bmap_alloc(struct qed_hwfn *p_hwfn, in qed_rdma_bmap_alloc() argument 65 DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "max_count = %08x\n", max_count); in qed_rdma_bmap_alloc() 76 DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "0\n"); in qed_rdma_bmap_alloc() 80 int qed_rdma_bmap_alloc_id(struct qed_hwfn *p_hwfn, in qed_rdma_bmap_alloc_id() argument 89 DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "%s bitmap: allocated id %d\n", in qed_rdma_bmap_alloc_id() 95 void qed_bmap_set_id(struct qed_hwfn *p_hwfn, in qed_bmap_set_id() argument 104 void qed_bmap_release_id(struct qed_hwfn *p_hwfn, in qed_bmap_release_id() argument 114 DP_NOTICE(p_hwfn, "%s bitmap: id %d already released\n", in qed_bmap_release_id() 119 DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "%s bitmap: released id %d\n", in qed_bmap_release_id() 123 int qed_bmap_test_id(struct qed_hwfn *p_hwfn, in qed_bmap_test_id() argument [all …]
|
D | qed_spq.c | 73 static void qed_spq_blocking_cb(struct qed_hwfn *p_hwfn, in qed_spq_blocking_cb() argument 87 static int __qed_spq_block(struct qed_hwfn *p_hwfn, in __qed_spq_block() argument 115 static int qed_spq_block(struct qed_hwfn *p_hwfn, in qed_spq_block() argument 127 rc = __qed_spq_block(p_hwfn, p_ent, p_fw_ret, false); in qed_spq_block() 133 rc = __qed_spq_block(p_hwfn, p_ent, p_fw_ret, true); in qed_spq_block() 137 p_ptt = qed_ptt_acquire(p_hwfn); in qed_spq_block() 139 DP_NOTICE(p_hwfn, "ptt, failed to acquire\n"); in qed_spq_block() 143 DP_INFO(p_hwfn, "Ramrod is stuck, requesting MCP drain\n"); in qed_spq_block() 144 rc = qed_mcp_drain(p_hwfn, p_ptt); in qed_spq_block() 145 qed_ptt_release(p_hwfn, p_ptt); in qed_spq_block() [all …]
|
D | qed_hw.c | 70 int qed_ptt_pool_alloc(struct qed_hwfn *p_hwfn) in qed_ptt_pool_alloc() argument 83 p_pool->ptts[i].hwfn_id = p_hwfn->my_id; in qed_ptt_pool_alloc() 89 p_hwfn->p_ptt_pool = p_pool; in qed_ptt_pool_alloc() 95 void qed_ptt_invalidate(struct qed_hwfn *p_hwfn) in qed_ptt_invalidate() argument 101 p_ptt = &p_hwfn->p_ptt_pool->ptts[i]; in qed_ptt_invalidate() 106 void qed_ptt_pool_free(struct qed_hwfn *p_hwfn) in qed_ptt_pool_free() argument 108 kfree(p_hwfn->p_ptt_pool); in qed_ptt_pool_free() 109 p_hwfn->p_ptt_pool = NULL; in qed_ptt_pool_free() 112 struct qed_ptt *qed_ptt_acquire(struct qed_hwfn *p_hwfn) in qed_ptt_acquire() argument 119 spin_lock_bh(&p_hwfn->p_ptt_pool->lock); in qed_ptt_acquire() [all …]
|
D | qed_vf.c | 39 static void *qed_vf_pf_prep(struct qed_hwfn *p_hwfn, u16 type, u16 length) in qed_vf_pf_prep() argument 41 struct qed_vf_iov *p_iov = p_hwfn->vf_iov_info; in qed_vf_pf_prep() 51 DP_VERBOSE(p_hwfn, in qed_vf_pf_prep() 64 p_tlv = qed_add_tlv(p_hwfn, &p_iov->offset, type, length); in qed_vf_pf_prep() 73 static void qed_vf_pf_req_end(struct qed_hwfn *p_hwfn, int req_status) in qed_vf_pf_req_end() argument 75 union pfvf_tlvs *resp = p_hwfn->vf_iov_info->pf2vf_reply; in qed_vf_pf_req_end() 77 DP_VERBOSE(p_hwfn, QED_MSG_IOV, in qed_vf_pf_req_end() 81 mutex_unlock(&(p_hwfn->vf_iov_info->mutex)); in qed_vf_pf_req_end() 84 static int qed_send_msg2pf(struct qed_hwfn *p_hwfn, u8 *done, u32 resp_size) in qed_send_msg2pf() argument 86 union vfpf_tlvs *p_req = p_hwfn->vf_iov_info->vf2pf_request; in qed_send_msg2pf() [all …]
|
D | qed_cxt.c | 115 #define CONN_CXT_SIZE(p_hwfn) \ argument 116 ALIGNED_TYPE_SIZE(union conn_context, p_hwfn) 120 #define TYPE0_TASK_CXT_SIZE(p_hwfn) \ argument 121 ALIGNED_TYPE_SIZE(union type0_task_context, p_hwfn) 124 #define TYPE1_TASK_CXT_SIZE(p_hwfn) sizeof(union type1_task_context) argument 315 static void qed_cxt_tm_iids(struct qed_hwfn *p_hwfn, in qed_cxt_tm_iids() argument 373 static void qed_cxt_qm_iids(struct qed_hwfn *p_hwfn, in qed_cxt_qm_iids() argument 376 struct qed_cxt_mngr *p_mngr = p_hwfn->p_cxt_mngr; in qed_cxt_qm_iids() 402 DP_VERBOSE(p_hwfn, QED_MSG_ILT, in qed_cxt_qm_iids() 407 static struct qed_tid_seg *qed_cxt_tid_seg_info(struct qed_hwfn *p_hwfn, in qed_cxt_tid_seg_info() argument [all …]
|
D | qed_init_fw_funcs.c | 181 #define QM_INIT_TX_PQ_MAP(p_hwfn, map, chip, pq_id, rl_valid, vp_pq_id, rl_id, \ argument 195 STORE_RT_REG(p_hwfn, QM_REG_TXPQMAP_RT_OFFSET + (pq_id), \ 210 static u8 qed_get_ext_voq(struct qed_hwfn *p_hwfn, in qed_get_ext_voq() argument 220 static void qed_enable_pf_rl(struct qed_hwfn *p_hwfn, bool pf_rl_en) in qed_enable_pf_rl() argument 222 STORE_RT_REG(p_hwfn, QM_REG_RLPFENABLE_RT_OFFSET, pf_rl_en ? 1 : 0); in qed_enable_pf_rl() 228 STORE_RT_REG(p_hwfn, in qed_enable_pf_rl() 232 STORE_RT_REG(p_hwfn, QM_REG_RLPFVOQENABLE_MSB_RT_OFFSET, in qed_enable_pf_rl() 236 STORE_RT_REG(p_hwfn, in qed_enable_pf_rl() 238 STORE_RT_REG(p_hwfn, in qed_enable_pf_rl() 244 STORE_RT_REG(p_hwfn, in qed_enable_pf_rl() [all …]
|
D | qed_iwarp.c | 87 static int qed_iwarp_async_event(struct qed_hwfn *p_hwfn, 93 void qed_iwarp_init_devinfo(struct qed_hwfn *p_hwfn) in qed_iwarp_init_devinfo() argument 95 struct qed_rdma_device *dev = p_hwfn->p_rdma_info->dev; in qed_iwarp_init_devinfo() 100 p_hwfn->p_rdma_info->num_qps) - in qed_iwarp_init_devinfo() 109 void qed_iwarp_init_hw(struct qed_hwfn *p_hwfn, struct qed_ptt *p_ptt) in qed_iwarp_init_hw() argument 111 p_hwfn->rdma_prs_search_reg = PRS_REG_SEARCH_TCP; in qed_iwarp_init_hw() 112 qed_wr(p_hwfn, p_ptt, p_hwfn->rdma_prs_search_reg, 1); in qed_iwarp_init_hw() 113 p_hwfn->b_rdma_enabled_in_prs = true; in qed_iwarp_init_hw() 120 static void qed_iwarp_cid_cleaned(struct qed_hwfn *p_hwfn, u32 cid) in qed_iwarp_cid_cleaned() argument 122 cid -= qed_cxt_get_proto_cid_start(p_hwfn, p_hwfn->p_rdma_info->proto); in qed_iwarp_cid_cleaned() [all …]
|
D | qed_roce.c | 63 static void qed_roce_free_real_icid(struct qed_hwfn *p_hwfn, u16 icid); 66 qed_roce_async_event(struct qed_hwfn *p_hwfn, in qed_roce_async_event() argument 70 struct qed_rdma_events events = p_hwfn->p_rdma_info->events; in qed_roce_async_event() 80 qed_roce_free_real_icid(p_hwfn, icid); in qed_roce_async_event() 99 void qed_roce_stop(struct qed_hwfn *p_hwfn) in qed_roce_stop() argument 101 struct qed_bmap *rcid_map = &p_hwfn->p_rdma_info->real_cid_map; in qed_roce_stop() 112 DP_NOTICE(p_hwfn, "cid bitmap wait timed out\n"); in qed_roce_stop() 116 qed_spq_unregister_async_cb(p_hwfn, PROTOCOLID_ROCE); in qed_roce_stop() 155 static void qed_roce_free_cid_pair(struct qed_hwfn *p_hwfn, u16 cid) in qed_roce_free_cid_pair() argument 157 spin_lock_bh(&p_hwfn->p_rdma_info->lock); in qed_roce_free_cid_pair() [all …]
|
D | qed_sp_commands.c | 50 void qed_sp_destroy_request(struct qed_hwfn *p_hwfn, in qed_sp_destroy_request() argument 57 if (p_ent->queue == &p_hwfn->p_spq->unlimited_pending) in qed_sp_destroy_request() 60 qed_spq_return_entry(p_hwfn, p_ent); in qed_sp_destroy_request() 63 int qed_sp_init_request(struct qed_hwfn *p_hwfn, in qed_sp_init_request() argument 74 rc = qed_spq_get_entry(p_hwfn, pp_ent); in qed_sp_init_request() 109 DP_NOTICE(p_hwfn, "Unknown SPQE completion mode %d\n", in qed_sp_init_request() 114 DP_VERBOSE(p_hwfn, QED_MSG_SPQ, in qed_sp_init_request() 127 qed_sp_destroy_request(p_hwfn, p_ent); in qed_sp_init_request() 227 qed_tunn_set_pf_update_params(struct qed_hwfn *p_hwfn, in qed_tunn_set_pf_update_params() argument 231 struct qed_tunnel_info *p_tun = &p_hwfn->cdev->tunnel; in qed_tunn_set_pf_update_params() [all …]
|
D | qed_sriov.c | 47 static int qed_sriov_eqe_event(struct qed_hwfn *p_hwfn, 51 static int qed_iov_bulletin_set_mac(struct qed_hwfn *p_hwfn, u8 *mac, int vfid); 69 static int qed_sp_vf_start(struct qed_hwfn *p_hwfn, struct qed_vf_info *p_vf) in qed_sp_vf_start() argument 79 init_data.cid = qed_spq_get_cid(p_hwfn); in qed_sp_vf_start() 83 rc = qed_sp_init_request(p_hwfn, &p_ent, in qed_sp_vf_start() 94 switch (p_hwfn->hw_info.personality) { in qed_sp_vf_start() 102 DP_NOTICE(p_hwfn, "Unknown VF personality %d\n", in qed_sp_vf_start() 103 p_hwfn->hw_info.personality); in qed_sp_vf_start() 104 qed_sp_destroy_request(p_hwfn, p_ent); in qed_sp_vf_start() 111 DP_VERBOSE(p_hwfn, in qed_sp_vf_start() [all …]
|
D | qed_init_ops.c | 78 void qed_init_clear_rt_data(struct qed_hwfn *p_hwfn) in qed_init_clear_rt_data() argument 83 p_hwfn->rt_data.b_valid[i] = false; in qed_init_clear_rt_data() 86 void qed_init_store_rt_reg(struct qed_hwfn *p_hwfn, u32 rt_offset, u32 val) in qed_init_store_rt_reg() argument 88 p_hwfn->rt_data.init_val[rt_offset] = val; in qed_init_store_rt_reg() 89 p_hwfn->rt_data.b_valid[rt_offset] = true; in qed_init_store_rt_reg() 92 void qed_init_store_rt_agg(struct qed_hwfn *p_hwfn, in qed_init_store_rt_agg() argument 98 p_hwfn->rt_data.init_val[rt_offset + i] = p_val[i]; in qed_init_store_rt_agg() 99 p_hwfn->rt_data.b_valid[rt_offset + i] = true; in qed_init_store_rt_agg() 103 static int qed_init_rt(struct qed_hwfn *p_hwfn, in qed_init_rt() argument 107 u32 *p_init_val = &p_hwfn->rt_data.init_val[rt_offset]; in qed_init_rt() [all …]
|
D | qed_cxt.h | 65 int qed_cxt_get_cid_info(struct qed_hwfn *p_hwfn, 76 int qed_cxt_get_tid_mem_info(struct qed_hwfn *p_hwfn, 88 u32 qed_cxt_get_proto_cid_count(struct qed_hwfn *p_hwfn, 98 int qed_cxt_set_pf_params(struct qed_hwfn *p_hwfn, u32 rdma_tasks); 108 int qed_cxt_cfg_ilt_compute(struct qed_hwfn *p_hwfn, u32 *last_line); 116 u32 qed_cxt_cfg_ilt_compute_excess(struct qed_hwfn *p_hwfn, u32 used_lines); 125 int qed_cxt_mngr_alloc(struct qed_hwfn *p_hwfn); 132 void qed_cxt_mngr_free(struct qed_hwfn *p_hwfn); 141 int qed_cxt_tables_alloc(struct qed_hwfn *p_hwfn); 148 void qed_cxt_mngr_setup(struct qed_hwfn *p_hwfn); [all …]
|
D | qed_mcp.h | 286 *qed_mcp_get_link_capabilities(struct qed_hwfn *p_hwfn); 297 int qed_mcp_set_link(struct qed_hwfn *p_hwfn, 311 int qed_mcp_get_mfw_ver(struct qed_hwfn *p_hwfn, 324 int qed_mcp_get_mbi_ver(struct qed_hwfn *p_hwfn, 338 int qed_mcp_get_media_type(struct qed_hwfn *p_hwfn, 353 int qed_mcp_get_transceiver_data(struct qed_hwfn *p_hwfn, 370 int qed_mcp_trans_speed_mask(struct qed_hwfn *p_hwfn, 384 int qed_mcp_get_board_config(struct qed_hwfn *p_hwfn, 404 int qed_mcp_cmd(struct qed_hwfn *p_hwfn, 418 int qed_mcp_drain(struct qed_hwfn *p_hwfn, [all …]
|
D | qed_int.h | 88 void qed_int_igu_enable_int(struct qed_hwfn *p_hwfn, 98 void qed_int_igu_disable_int(struct qed_hwfn *p_hwfn, 109 u64 qed_int_igu_read_sisr_reg(struct qed_hwfn *p_hwfn); 128 int qed_int_sb_init(struct qed_hwfn *p_hwfn, 141 void qed_int_sb_setup(struct qed_hwfn *p_hwfn, 158 int qed_int_sb_release(struct qed_hwfn *p_hwfn, 180 void qed_int_get_num_sbs(struct qed_hwfn *p_hwfn, 201 int qed_db_rec_handler(struct qed_hwfn *p_hwfn, struct qed_ptt *p_ptt); 209 #define SB_ALIGNED_SIZE(p_hwfn) \ argument 210 ALIGNED_TYPE_SIZE(struct status_block_e4, p_hwfn) [all …]
|
D | qed_ll2.c | 98 struct qed_hwfn *p_hwfn = cxt; in qed_ll2b_complete_tx_packet() local 99 struct qed_dev *cdev = p_hwfn->cdev; in qed_ll2b_complete_tx_packet() 103 dma_unmap_single(&p_hwfn->cdev->pdev->dev, first_frag_addr, in qed_ll2b_complete_tx_packet() 164 struct qed_hwfn *p_hwfn = cxt; in qed_ll2b_complete_rx_packet() local 166 struct qed_dev *cdev = p_hwfn->cdev; in qed_ll2b_complete_rx_packet() 173 DP_VERBOSE(p_hwfn, in qed_ll2b_complete_rx_packet() 194 rc = qed_ll2_alloc_buffer(p_hwfn->cdev, &new_data, in qed_ll2b_complete_rx_packet() 230 DP_VERBOSE(p_hwfn, (NETIF_MSG_RX_STATUS | NETIF_MSG_PKTDATA | in qed_ll2b_complete_rx_packet() 242 rc = qed_ll2_post_rx_buffer(p_hwfn, cdev->ll2->handle, in qed_ll2b_complete_rx_packet() 248 static struct qed_ll2_info *__qed_ll2_handle_sanity(struct qed_hwfn *p_hwfn, in __qed_ll2_handle_sanity() argument [all …]
|
D | qed_ooo.h | 87 void qed_ooo_save_history_entry(struct qed_hwfn *p_hwfn, 91 int qed_ooo_alloc(struct qed_hwfn *p_hwfn); 93 void qed_ooo_setup(struct qed_hwfn *p_hwfn); 95 void qed_ooo_free(struct qed_hwfn *p_hwfn); 97 void qed_ooo_release_connection_isles(struct qed_hwfn *p_hwfn, 101 void qed_ooo_release_all_isles(struct qed_hwfn *p_hwfn, 104 void qed_ooo_put_free_buffer(struct qed_hwfn *p_hwfn, 109 qed_ooo_get_free_buffer(struct qed_hwfn *p_hwfn, 112 void qed_ooo_put_ready_buffer(struct qed_hwfn *p_hwfn, 117 qed_ooo_get_ready_buffer(struct qed_hwfn *p_hwfn, [all …]
|
D | qed_l2.c | 76 int qed_l2_alloc(struct qed_hwfn *p_hwfn) in qed_l2_alloc() argument 82 if (!QED_IS_L2_PERSONALITY(p_hwfn)) in qed_l2_alloc() 88 p_hwfn->p_l2_info = p_l2_info; in qed_l2_alloc() 90 if (IS_PF(p_hwfn->cdev)) { in qed_l2_alloc() 91 p_l2_info->queues = RESC_NUM(p_hwfn, QED_L2_QUEUE); in qed_l2_alloc() 95 qed_vf_get_num_rxqs(p_hwfn, &rx); in qed_l2_alloc() 96 qed_vf_get_num_txqs(p_hwfn, &tx); in qed_l2_alloc() 116 void qed_l2_setup(struct qed_hwfn *p_hwfn) in qed_l2_setup() argument 118 if (!QED_IS_L2_PERSONALITY(p_hwfn)) in qed_l2_setup() 121 mutex_init(&p_hwfn->p_l2_info->lock); in qed_l2_setup() [all …]
|
D | qed_selftest.c | 68 struct qed_hwfn *p_hwfn; in qed_selftest_register() local 74 p_hwfn = &cdev->hwfns[i]; in qed_selftest_register() 75 p_ptt = qed_ptt_acquire(p_hwfn); in qed_selftest_register() 77 DP_ERR(p_hwfn, "failed to acquire ptt\n"); in qed_selftest_register() 80 rc = qed_mcp_bist_register_test(p_hwfn, p_ptt); in qed_selftest_register() 81 qed_ptt_release(p_hwfn, p_ptt); in qed_selftest_register() 91 struct qed_hwfn *p_hwfn; in qed_selftest_clock() local 97 p_hwfn = &cdev->hwfns[i]; in qed_selftest_clock() 98 p_ptt = qed_ptt_acquire(p_hwfn); in qed_selftest_clock() 100 DP_ERR(p_hwfn, "failed to acquire ptt\n"); in qed_selftest_clock() [all …]
|
D | qed_sp.h | 68 int qed_eth_cqe_completion(struct qed_hwfn *p_hwfn, 186 (*qed_spq_async_comp_cb)(struct qed_hwfn *p_hwfn, 193 qed_spq_register_async_cb(struct qed_hwfn *p_hwfn, 198 qed_spq_unregister_async_cb(struct qed_hwfn *p_hwfn, 244 int qed_spq_post(struct qed_hwfn *p_hwfn, 255 int qed_spq_alloc(struct qed_hwfn *p_hwfn); 262 void qed_spq_setup(struct qed_hwfn *p_hwfn); 269 void qed_spq_free(struct qed_hwfn *p_hwfn); 283 qed_spq_get_entry(struct qed_hwfn *p_hwfn, 293 void qed_spq_return_entry(struct qed_hwfn *p_hwfn, [all …]
|
D | qed_ooo.c | 47 *qed_ooo_seek_archipelago(struct qed_hwfn *p_hwfn, in qed_ooo_seek_archipelago() argument 66 static struct qed_ooo_isle *qed_ooo_seek_isle(struct qed_hwfn *p_hwfn, in qed_ooo_seek_isle() argument 74 p_archipelago = qed_ooo_seek_archipelago(p_hwfn, p_ooo_info, cid); in qed_ooo_seek_isle() 76 DP_NOTICE(p_hwfn, in qed_ooo_seek_isle() 90 void qed_ooo_save_history_entry(struct qed_hwfn *p_hwfn, in qed_ooo_save_history_entry() argument 102 int qed_ooo_alloc(struct qed_hwfn *p_hwfn) in qed_ooo_alloc() argument 110 switch (p_hwfn->hw_info.personality) { in qed_ooo_alloc() 119 DP_NOTICE(p_hwfn, in qed_ooo_alloc() 124 max_num_archipelagos = (u16)qed_cxt_get_proto_cid_count(p_hwfn, proto, in qed_ooo_alloc() 127 cid_base = (u16)qed_cxt_get_proto_cid_start(p_hwfn, proto); in qed_ooo_alloc() [all …]
|
D | qed_hw.h | 86 void qed_gtt_init(struct qed_hwfn *p_hwfn); 93 void qed_ptt_invalidate(struct qed_hwfn *p_hwfn); 102 int qed_ptt_pool_alloc(struct qed_hwfn *p_hwfn); 109 void qed_ptt_pool_free(struct qed_hwfn *p_hwfn); 119 u32 qed_ptt_get_hw_addr(struct qed_hwfn *p_hwfn, 139 void qed_ptt_set_win(struct qed_hwfn *p_hwfn, 151 struct qed_ptt *qed_get_reserved_ptt(struct qed_hwfn *p_hwfn, 162 void qed_wr(struct qed_hwfn *p_hwfn, 175 u32 qed_rd(struct qed_hwfn *p_hwfn, 189 void qed_memcpy_from(struct qed_hwfn *p_hwfn, [all …]
|