Lines Matching refs:aq
273 struct nix_aq_enq_req *aq; in otx2_set_rss_table() local
279 aq = otx2_mbox_alloc_msg_nix_aq_enq(mbox); in otx2_set_rss_table()
280 if (!aq) { in otx2_set_rss_table()
289 aq = otx2_mbox_alloc_msg_nix_aq_enq(mbox); in otx2_set_rss_table()
290 if (!aq) { in otx2_set_rss_table()
296 aq->rss.rq = rss->ind_tbl[idx]; in otx2_set_rss_table()
299 aq->qidx = idx; in otx2_set_rss_table()
300 aq->ctype = NIX_AQ_CTYPE_RSS; in otx2_set_rss_table()
301 aq->op = NIX_AQ_INSTOP_INIT; in otx2_set_rss_table()
706 struct nix_aq_enq_req *aq; in otx2_rq_init() local
709 aq = otx2_mbox_alloc_msg_nix_aq_enq(&pfvf->mbox); in otx2_rq_init()
710 if (!aq) in otx2_rq_init()
713 aq->rq.cq = qidx; in otx2_rq_init()
714 aq->rq.ena = 1; in otx2_rq_init()
715 aq->rq.pb_caching = 1; in otx2_rq_init()
716 aq->rq.lpb_aura = lpb_aura; /* Use large packet buffer aura */ in otx2_rq_init()
717 aq->rq.lpb_sizem1 = (DMA_BUFFER_LEN(pfvf->rbsize) / 8) - 1; in otx2_rq_init()
718 aq->rq.xqe_imm_size = 0; /* Copying of packet to CQE not needed */ in otx2_rq_init()
719 aq->rq.flow_tagw = 32; /* Copy full 32bit flow_tag to CQE header */ in otx2_rq_init()
720 aq->rq.qint_idx = 0; in otx2_rq_init()
721 aq->rq.lpb_drop_ena = 1; /* Enable RED dropping for AURA */ in otx2_rq_init()
722 aq->rq.xqe_drop_ena = 1; /* Enable RED dropping for CQ/SSO */ in otx2_rq_init()
723 aq->rq.xqe_pass = RQ_PASS_LVL_CQ(pfvf->hw.rq_skid, qset->rqe_cnt); in otx2_rq_init()
724 aq->rq.xqe_drop = RQ_DROP_LVL_CQ(pfvf->hw.rq_skid, qset->rqe_cnt); in otx2_rq_init()
725 aq->rq.lpb_aura_pass = RQ_PASS_LVL_AURA; in otx2_rq_init()
726 aq->rq.lpb_aura_drop = RQ_DROP_LVL_AURA; in otx2_rq_init()
729 aq->qidx = qidx; in otx2_rq_init()
730 aq->ctype = NIX_AQ_CTYPE_RQ; in otx2_rq_init()
731 aq->op = NIX_AQ_INSTOP_INIT; in otx2_rq_init()
740 struct nix_aq_enq_req *aq; in otx2_sq_init() local
784 aq = otx2_mbox_alloc_msg_nix_aq_enq(&pfvf->mbox); in otx2_sq_init()
785 if (!aq) in otx2_sq_init()
788 aq->sq.cq = pfvf->hw.rx_queues + qidx; in otx2_sq_init()
789 aq->sq.max_sqe_size = NIX_MAXSQESZ_W16; /* 128 byte */ in otx2_sq_init()
790 aq->sq.cq_ena = 1; in otx2_sq_init()
791 aq->sq.ena = 1; in otx2_sq_init()
793 aq->sq.smq = pfvf->hw.txschq_list[NIX_TXSCH_LVL_SMQ][0]; in otx2_sq_init()
794 aq->sq.smq_rr_quantum = DFLT_RR_QTM; in otx2_sq_init()
795 aq->sq.default_chan = pfvf->hw.tx_chan_base; in otx2_sq_init()
796 aq->sq.sqe_stype = NIX_STYPE_STF; /* Cache SQB */ in otx2_sq_init()
797 aq->sq.sqb_aura = sqb_aura; in otx2_sq_init()
798 aq->sq.sq_int_ena = NIX_SQINT_BITS; in otx2_sq_init()
799 aq->sq.qint_idx = 0; in otx2_sq_init()
803 aq->sq.cq_limit = ((SEND_CQ_SKID * 256) / (sq->sqe_cnt)); in otx2_sq_init()
806 aq->qidx = qidx; in otx2_sq_init()
807 aq->ctype = NIX_AQ_CTYPE_SQ; in otx2_sq_init()
808 aq->op = NIX_AQ_INSTOP_INIT; in otx2_sq_init()
816 struct nix_aq_enq_req *aq; in otx2_cq_init() local
849 aq = otx2_mbox_alloc_msg_nix_aq_enq(&pfvf->mbox); in otx2_cq_init()
850 if (!aq) in otx2_cq_init()
853 aq->cq.ena = 1; in otx2_cq_init()
854 aq->cq.qsize = Q_SIZE(cq->cqe_cnt, 4); in otx2_cq_init()
855 aq->cq.caching = 1; in otx2_cq_init()
856 aq->cq.base = cq->cqe->iova; in otx2_cq_init()
857 aq->cq.cint_idx = cq->cint_idx; in otx2_cq_init()
858 aq->cq.cq_err_int_ena = NIX_CQERRINT_BITS; in otx2_cq_init()
859 aq->cq.qint_idx = 0; in otx2_cq_init()
860 aq->cq.avg_level = 255; in otx2_cq_init()
863 aq->cq.drop = RQ_DROP_LVL_CQ(pfvf->hw.rq_skid, cq->cqe_cnt); in otx2_cq_init()
864 aq->cq.drop_ena = 1; in otx2_cq_init()
867 aq->cq.bp_ena = 1; in otx2_cq_init()
868 aq->cq.bpid = pfvf->bpid[0]; in otx2_cq_init()
871 aq->cq.bp = RQ_PASS_LVL_CQ(pfvf->hw.rq_skid, qset->rqe_cnt); in otx2_cq_init()
875 aq->qidx = qidx; in otx2_cq_init()
876 aq->ctype = NIX_AQ_CTYPE_CQ; in otx2_cq_init()
877 aq->op = NIX_AQ_INSTOP_INIT; in otx2_cq_init()
1088 struct npa_aq_enq_req *aq; in otx2_aura_init() local
1104 aq = otx2_mbox_alloc_msg_npa_aq_enq(&pfvf->mbox); in otx2_aura_init()
1105 if (!aq) { in otx2_aura_init()
1110 aq = otx2_mbox_alloc_msg_npa_aq_enq(&pfvf->mbox); in otx2_aura_init()
1111 if (!aq) in otx2_aura_init()
1115 aq->aura_id = aura_id; in otx2_aura_init()
1117 aq->aura.pool_addr = pool_id; in otx2_aura_init()
1118 aq->aura.pool_caching = 1; in otx2_aura_init()
1119 aq->aura.shift = ilog2(numptrs) - 8; in otx2_aura_init()
1120 aq->aura.count = numptrs; in otx2_aura_init()
1121 aq->aura.limit = numptrs; in otx2_aura_init()
1122 aq->aura.avg_level = 255; in otx2_aura_init()
1123 aq->aura.ena = 1; in otx2_aura_init()
1124 aq->aura.fc_ena = 1; in otx2_aura_init()
1125 aq->aura.fc_addr = pool->fc_addr->iova; in otx2_aura_init()
1126 aq->aura.fc_hyst_bits = 0; /* Store count on all updates */ in otx2_aura_init()
1130 aq->aura.bp_ena = 0; in otx2_aura_init()
1131 aq->aura.nix0_bpid = pfvf->bpid[0]; in otx2_aura_init()
1133 aq->aura.bp = RQ_BP_LVL_AURA; in otx2_aura_init()
1137 aq->ctype = NPA_AQ_CTYPE_AURA; in otx2_aura_init()
1138 aq->op = NPA_AQ_INSTOP_INIT; in otx2_aura_init()
1146 struct npa_aq_enq_req *aq; in otx2_pool_init() local
1160 aq = otx2_mbox_alloc_msg_npa_aq_enq(&pfvf->mbox); in otx2_pool_init()
1161 if (!aq) { in otx2_pool_init()
1168 aq = otx2_mbox_alloc_msg_npa_aq_enq(&pfvf->mbox); in otx2_pool_init()
1169 if (!aq) { in otx2_pool_init()
1175 aq->aura_id = pool_id; in otx2_pool_init()
1176 aq->pool.stack_base = pool->stack->iova; in otx2_pool_init()
1177 aq->pool.stack_caching = 1; in otx2_pool_init()
1178 aq->pool.ena = 1; in otx2_pool_init()
1179 aq->pool.buf_size = buf_size / 128; in otx2_pool_init()
1180 aq->pool.stack_max_pages = stack_pages; in otx2_pool_init()
1181 aq->pool.shift = ilog2(numptrs) - 8; in otx2_pool_init()
1182 aq->pool.ptr_start = 0; in otx2_pool_init()
1183 aq->pool.ptr_end = ~0ULL; in otx2_pool_init()
1186 aq->ctype = NPA_AQ_CTYPE_POOL; in otx2_pool_init()
1187 aq->op = NPA_AQ_INSTOP_INIT; in otx2_pool_init()