/Linux-v4.19/drivers/infiniband/hw/bnxt_re/ |
D | qplib_fp.c | 73 dev_dbg(&scq->hwq.pdev->dev, in __bnxt_qplib_add_flush_qp() 82 dev_dbg(&rcq->hwq.pdev->dev, in __bnxt_qplib_add_flush_qp() 142 qp->sq.hwq.prod = 0; in bnxt_qplib_clean_qp() 143 qp->sq.hwq.cons = 0; in bnxt_qplib_clean_qp() 145 qp->rq.hwq.prod = 0; in bnxt_qplib_clean_qp() 146 qp->rq.hwq.cons = 0; in bnxt_qplib_clean_qp() 181 rq->hwq.max_elements * qp->rq_hdr_buf_size, in bnxt_qplib_free_qp_hdr_buf() 185 sq->hwq.max_elements * qp->sq_hdr_buf_size, in bnxt_qplib_free_qp_hdr_buf() 202 if (qp->sq_hdr_buf_size && sq->hwq.max_elements) { in bnxt_qplib_alloc_qp_hdr_buf() 204 sq->hwq.max_elements * in bnxt_qplib_alloc_qp_hdr_buf() [all …]
|
D | qplib_res.c | 136 void bnxt_qplib_free_hwq(struct pci_dev *pdev, struct bnxt_qplib_hwq *hwq) in bnxt_qplib_free_hwq() argument 140 if (!hwq->max_elements) in bnxt_qplib_free_hwq() 142 if (hwq->level >= PBL_LVL_MAX) in bnxt_qplib_free_hwq() 145 for (i = 0; i < hwq->level + 1; i++) { in bnxt_qplib_free_hwq() 146 if (i == hwq->level) in bnxt_qplib_free_hwq() 147 __free_pbl(pdev, &hwq->pbl[i], hwq->is_user); in bnxt_qplib_free_hwq() 149 __free_pbl(pdev, &hwq->pbl[i], false); in bnxt_qplib_free_hwq() 152 hwq->level = PBL_LVL_MAX; in bnxt_qplib_free_hwq() 153 hwq->max_elements = 0; in bnxt_qplib_free_hwq() 154 hwq->element_size = 0; in bnxt_qplib_free_hwq() [all …]
|
D | qplib_res.h | 49 #define HWQ_CMP(idx, hwq) ((idx) & ((hwq)->max_elements - 1)) argument 51 #define HWQ_FREE_SLOTS(hwq) (hwq->max_elements - \ argument 52 ((HWQ_CMP(hwq->prod, hwq)\ 53 - HWQ_CMP(hwq->cons, hwq))\ 54 & (hwq->max_elements - 1))) 201 void bnxt_qplib_free_hwq(struct pci_dev *pdev, struct bnxt_qplib_hwq *hwq); 202 int bnxt_qplib_alloc_init_hwq(struct pci_dev *pdev, struct bnxt_qplib_hwq *hwq,
|
D | qplib_sp.c | 580 if (mrw->hwq.max_elements) in bnxt_qplib_free_mrw() 581 bnxt_qplib_free_hwq(res->pdev, &mrw->hwq); in bnxt_qplib_free_mrw() 639 if (mrw->hwq.max_elements) { in bnxt_qplib_dereg_mrw() 642 bnxt_qplib_free_hwq(res->pdev, &mrw->hwq); in bnxt_qplib_dereg_mrw() 676 if (mr->hwq.max_elements) in bnxt_qplib_reg_mr() 677 bnxt_qplib_free_hwq(res->pdev, &mr->hwq); in bnxt_qplib_reg_mr() 679 mr->hwq.max_elements = pages; in bnxt_qplib_reg_mr() 681 rc = bnxt_qplib_alloc_init_hwq(res->pdev, &mr->hwq, NULL, 0, in bnxt_qplib_reg_mr() 682 &mr->hwq.max_elements, in bnxt_qplib_reg_mr() 691 pbl_ptr = (dma_addr_t **)mr->hwq.pbl_ptr; in bnxt_qplib_reg_mr() [all …]
|
D | qplib_fp.h | 53 struct bnxt_qplib_hwq hwq; member 237 struct bnxt_qplib_hwq hwq; member 340 return HWQ_CMP((qplib_q->hwq.prod + qplib_q->q_full_delta), in bnxt_qplib_queue_full() 341 &qplib_q->hwq) == HWQ_CMP(qplib_q->hwq.cons, in bnxt_qplib_queue_full() 342 &qplib_q->hwq); in bnxt_qplib_queue_full() 377 struct bnxt_qplib_hwq hwq; member 448 struct bnxt_qplib_hwq hwq; member
|
D | qplib_sp.h | 115 struct bnxt_qplib_hwq hwq; member 120 struct bnxt_qplib_hwq hwq; member
|
D | main.c | 867 if (rdev->nq[0].hwq.max_elements) { in bnxt_re_cleanup_res() 948 rdev->nq[i].hwq.max_elements = BNXT_RE_MAX_CQ_COUNT + in bnxt_re_alloc_res() 957 (rdev, rdev->nq[i].hwq.pbl[PBL_LVL_0].pg_map_arr, in bnxt_re_alloc_res() 958 rdev->nq[i].hwq.pbl[rdev->nq[i].hwq.level].pg_count, in bnxt_re_alloc_res()
|
D | ib_verbs.c | 2221 wqe->frmr.pbl_ptr = (__le64 *)qplib_frpl->hwq.pbl_ptr[0]; in bnxt_re_build_reg_wqe() 2222 wqe->frmr.pbl_dma_ptr = qplib_frpl->hwq.pbl_dma_ptr[0]; in bnxt_re_build_reg_wqe() 2225 wqe->frmr.levels = qplib_frpl->hwq.level + 1; in bnxt_re_build_reg_wqe() 2672 resp.tail = cq->qplib_cq.hwq.cons; in bnxt_re_create_cq() 3148 dev_dbg(&lib_qp->sq.hwq.pdev->dev, in send_phantom_wqe() 3150 lib_qp->id, lib_qp->sq.hwq.prod, in send_phantom_wqe() 3151 HWQ_CMP(lib_qp->sq.hwq.prod, &lib_qp->sq.hwq), in send_phantom_wqe() 3342 mr->qplib_mr.hwq.level = PBL_LVL_MAX; in bnxt_re_get_dma_mr()
|
/Linux-v4.19/drivers/scsi/cxlflash/ |
D | main.c | 168 struct hwq *hwq = get_hwq(afu, cmd->hwq_index); in cmd_complete() local 170 spin_lock_irqsave(&hwq->hsq_slock, lock_flags); in cmd_complete() 172 spin_unlock_irqrestore(&hwq->hsq_slock, lock_flags); in cmd_complete() 200 static void flush_pending_cmds(struct hwq *hwq) in flush_pending_cmds() argument 202 struct cxlflash_cfg *cfg = hwq->afu->parent; in flush_pending_cmds() 207 list_for_each_entry_safe(cmd, tmp, &hwq->pending_cmds, list) { in flush_pending_cmds() 244 static int context_reset(struct hwq *hwq, __be64 __iomem *reset_reg) in context_reset() argument 246 struct cxlflash_cfg *cfg = hwq->afu->parent; in context_reset() 253 dev_dbg(dev, "%s: hwq=%p\n", __func__, hwq); in context_reset() 255 spin_lock_irqsave(&hwq->hsq_slock, lock_flags); in context_reset() [all …]
|
D | common.h | 200 struct hwq { struct 235 struct hwq hwqs[CXLFLASH_MAX_HWQS]; argument 237 int (*context_reset)(struct hwq *hwq); 259 static inline struct hwq *get_hwq(struct afu *afu, u32 index) in get_hwq()
|
D | superpipe.c | 271 struct hwq *hwq = get_hwq(afu, PRIMARY_HWQ); in afu_attach() local 295 val = hwq->ctx_hndl; in afu_attach() 302 val = SISL_RHT_CNT_ID((u64)MAX_RHT_PER_CONTEXT, (u64)(hwq->ctx_hndl)); in afu_attach() 1664 struct hwq *hwq = get_hwq(afu, PRIMARY_HWQ); in cxlflash_afu_recover() local 1741 reg = readq_be(&hwq->ctrl_map->mbox_r); in cxlflash_afu_recover()
|
/Linux-v4.19/drivers/net/wireless/mediatek/mt76/ |
D | tx.c | 163 struct mt76_queue *hwq = &dev->q_tx[MT_TXQ_PSD]; in mt76_queue_ps_skb() local 170 dev->queue_ops->tx_queue_skb(dev, hwq, skb, wcid, sta); in mt76_queue_ps_skb() 181 struct mt76_queue *hwq = &dev->q_tx[MT_TXQ_PSD]; in mt76_release_buffered_frames() local 184 spin_lock_bh(&hwq->lock); in mt76_release_buffered_frames() 211 dev->queue_ops->kick(dev, hwq); in mt76_release_buffered_frames() 213 spin_unlock_bh(&hwq->lock); in mt76_release_buffered_frames() 218 mt76_txq_send_burst(struct mt76_dev *dev, struct mt76_queue *hwq, in mt76_txq_send_burst() argument 250 idx = dev->queue_ops->tx_queue_skb(dev, hwq, skb, wcid, txq->sta); in mt76_txq_send_burst() 285 idx = dev->queue_ops->tx_queue_skb(dev, hwq, skb, wcid, in mt76_txq_send_burst() 294 hwq->swq_queued++; in mt76_txq_send_burst() [all …]
|
D | mt76.h | 170 struct mt76_queue *hwq; member 527 void mt76_txq_schedule(struct mt76_dev *dev, struct mt76_queue *hwq);
|
/Linux-v4.19/include/scsi/ |
D | scsi_tcq.h | 31 u16 hwq = blk_mq_unique_tag_to_hwq(tag); in scsi_host_find_tag() local 33 if (hwq < shost->tag_set.nr_hw_queues) { in scsi_host_find_tag() 34 req = blk_mq_tag_to_rq(shost->tag_set.tags[hwq], in scsi_host_find_tag()
|
/Linux-v4.19/drivers/net/wireless/ti/wlcore/ |
D | tx.c | 1214 int hwq = wlcore_tx_get_mac80211_queue(wlvif, queue); in wlcore_stop_queue_locked() local 1215 bool stopped = !!wl->queue_stop_reasons[hwq]; in wlcore_stop_queue_locked() 1218 WARN_ON_ONCE(test_and_set_bit(reason, &wl->queue_stop_reasons[hwq])); in wlcore_stop_queue_locked() 1223 ieee80211_stop_queue(wl->hw, hwq); in wlcore_stop_queue_locked() 1240 int hwq = wlcore_tx_get_mac80211_queue(wlvif, queue); in wlcore_wake_queue() local 1245 WARN_ON_ONCE(!test_and_clear_bit(reason, &wl->queue_stop_reasons[hwq])); in wlcore_wake_queue() 1247 if (wl->queue_stop_reasons[hwq]) in wlcore_wake_queue() 1250 ieee80211_wake_queue(wl->hw, hwq); in wlcore_wake_queue() 1317 int hwq = wlcore_tx_get_mac80211_queue(wlvif, queue); in wlcore_is_queue_stopped_by_reason_locked() local 1320 return test_bit(reason, &wl->queue_stop_reasons[hwq]); in wlcore_is_queue_stopped_by_reason_locked() [all …]
|
/Linux-v4.19/block/ |
D | blk-mq-tag.c | 472 int hwq = 0; in blk_mq_unique_tag() local 476 hwq = hctx->queue_num; in blk_mq_unique_tag() 479 return (hwq << BLK_MQ_UNIQUE_TAG_BITS) | in blk_mq_unique_tag()
|
/Linux-v4.19/drivers/net/wireless/intel/iwlegacy/ |
D | common.h | 2270 il_set_swq_id(struct il_tx_queue *txq, u8 ac, u8 hwq) in il_set_swq_id() argument 2273 BUG_ON(hwq > 31); /* only use 5 bits */ in il_set_swq_id() 2275 txq->swq_id = (hwq << 2) | ac; in il_set_swq_id() 2296 u8 hwq = (queue >> 2) & 0x1f; in il_wake_queue() local 2298 if (test_and_clear_bit(hwq, il->queue_stopped)) in il_wake_queue() 2307 u8 hwq = (queue >> 2) & 0x1f; in il_stop_queue() local 2309 if (!test_and_set_bit(hwq, il->queue_stopped)) in il_stop_queue()
|
/Linux-v4.19/drivers/scsi/ |
D | virtio_scsi.c | 529 u16 hwq = blk_mq_unique_tag_to_hwq(tag); in virtscsi_pick_vq_mq() local 531 return &vscsi->req_vqs[hwq]; in virtscsi_pick_vq_mq()
|
D | scsi_debug.c | 3786 u16 hwq = blk_mq_unique_tag_to_hwq(tag); in get_queue() local 3788 pr_debug("tag=%#x, hwq=%d\n", tag, hwq); in get_queue() 3789 if (WARN_ON_ONCE(hwq >= submit_queues)) in get_queue() 3790 hwq = 0; in get_queue() 3791 return sdebug_q_arr + hwq; in get_queue()
|
/Linux-v4.19/drivers/block/ |
D | nbd.c | 620 u16 hwq; in nbd_read_stat() local 644 hwq = blk_mq_unique_tag_to_hwq(tag); in nbd_read_stat() 645 if (hwq < nbd->tag_set.nr_hw_queues) in nbd_read_stat() 646 req = blk_mq_tag_to_rq(nbd->tag_set.tags[hwq], in nbd_read_stat()
|
D | skd_main.c | 1458 u16 hwq = 0; in skd_isr_completion_posted() local 1514 WARN_ON_ONCE(blk_mq_tag_to_rq(skdev->tag_set.tags[hwq], in skd_isr_completion_posted() 1520 rq = blk_mq_tag_to_rq(skdev->tag_set.tags[hwq], tag); in skd_isr_completion_posted()
|
/Linux-v4.19/drivers/scsi/lpfc/ |
D | lpfc_scsi.c | 3913 uint16_t hwq; in lpfc_sli4_scmd_to_wqidx_distr() local 3917 hwq = blk_mq_unique_tag_to_hwq(tag); in lpfc_sli4_scmd_to_wqidx_distr() 3919 return hwq; in lpfc_sli4_scmd_to_wqidx_distr()
|
/Linux-v4.19/drivers/scsi/qla2xxx/ |
D | qla_os.c | 848 uint16_t hwq; in qla2xxx_queuecommand() local 858 hwq = blk_mq_unique_tag_to_hwq(tag); in qla2xxx_queuecommand() 859 qpair = ha->queue_pair_map[hwq]; in qla2xxx_queuecommand()
|