Home
last modified time | relevance | path

Searched refs:mq (Results 1 – 25 of 47) sorted by relevance

12

/Linux-v5.4/drivers/mmc/core/
Dqueue.c26 static inline bool mmc_cqe_dcmd_busy(struct mmc_queue *mq) in mmc_cqe_dcmd_busy() argument
29 return mq->in_flight[MMC_ISSUE_DCMD]; in mmc_cqe_dcmd_busy()
32 void mmc_cqe_check_busy(struct mmc_queue *mq) in mmc_cqe_check_busy() argument
34 if ((mq->cqe_busy & MMC_CQE_DCMD_BUSY) && !mmc_cqe_dcmd_busy(mq)) in mmc_cqe_check_busy()
35 mq->cqe_busy &= ~MMC_CQE_DCMD_BUSY; in mmc_cqe_check_busy()
37 mq->cqe_busy &= ~MMC_CQE_QUEUE_FULL; in mmc_cqe_check_busy()
61 enum mmc_issue_type mmc_issue_type(struct mmc_queue *mq, struct request *req) in mmc_issue_type() argument
63 struct mmc_host *host = mq->card->host; in mmc_issue_type()
65 if (mq->use_cqe) in mmc_issue_type()
74 static void __mmc_cqe_recovery_notifier(struct mmc_queue *mq) in __mmc_cqe_recovery_notifier() argument
[all …]
Dblock.c232 struct mmc_queue *mq; in power_ro_lock_store() local
243 mq = &md->queue; in power_ro_lock_store()
246 req = blk_get_request(mq->queue, REQ_OP_DRV_OUT, 0); in power_ro_lock_store()
252 blk_execute_rq(mq->queue, NULL, req, 0); in power_ro_lock_store()
635 struct mmc_queue *mq; in mmc_blk_ioctl_cmd() local
655 mq = &md->queue; in mmc_blk_ioctl_cmd()
656 req = blk_get_request(mq->queue, in mmc_blk_ioctl_cmd()
667 blk_execute_rq(mq->queue, NULL, req, 0); in mmc_blk_ioctl_cmd()
685 struct mmc_queue *mq; in mmc_blk_ioctl_multi_cmd() local
725 mq = &md->queue; in mmc_blk_ioctl_multi_cmd()
[all …]
Dqueue.h106 void mmc_cqe_check_busy(struct mmc_queue *mq);
109 enum mmc_issue_type mmc_issue_type(struct mmc_queue *mq, struct request *req);
111 static inline int mmc_tot_in_flight(struct mmc_queue *mq) in mmc_tot_in_flight() argument
113 return mq->in_flight[MMC_ISSUE_SYNC] + in mmc_tot_in_flight()
114 mq->in_flight[MMC_ISSUE_DCMD] + in mmc_tot_in_flight()
115 mq->in_flight[MMC_ISSUE_ASYNC]; in mmc_tot_in_flight()
118 static inline int mmc_cqe_qcnt(struct mmc_queue *mq) in mmc_cqe_qcnt() argument
120 return mq->in_flight[MMC_ISSUE_DCMD] + in mmc_cqe_qcnt()
121 mq->in_flight[MMC_ISSUE_ASYNC]; in mmc_cqe_qcnt()
Dblock.h8 void mmc_blk_cqe_recovery(struct mmc_queue *mq);
12 enum mmc_issued mmc_blk_mq_issue_rq(struct mmc_queue *mq, struct request *req);
14 void mmc_blk_mq_recovery(struct mmc_queue *mq);
/Linux-v5.4/drivers/md/
Ddm-cache-policy-smq.c867 static struct entry *writeback_sentinel(struct smq_policy *mq, unsigned level) in writeback_sentinel() argument
869 return get_sentinel(&mq->writeback_sentinel_alloc, level, mq->current_writeback_sentinels); in writeback_sentinel()
872 static struct entry *demote_sentinel(struct smq_policy *mq, unsigned level) in demote_sentinel() argument
874 return get_sentinel(&mq->demote_sentinel_alloc, level, mq->current_demote_sentinels); in demote_sentinel()
877 static void __update_writeback_sentinels(struct smq_policy *mq) in __update_writeback_sentinels() argument
880 struct queue *q = &mq->dirty; in __update_writeback_sentinels()
884 sentinel = writeback_sentinel(mq, level); in __update_writeback_sentinels()
890 static void __update_demote_sentinels(struct smq_policy *mq) in __update_demote_sentinels() argument
893 struct queue *q = &mq->clean; in __update_demote_sentinels()
897 sentinel = demote_sentinel(mq, level); in __update_demote_sentinels()
[all …]
/Linux-v5.4/drivers/scsi/arm/
Dmsgqueue.c24 struct msgqueue_entry *mq; in mqe_alloc() local
26 if ((mq = msgq->free) != NULL) in mqe_alloc()
27 msgq->free = mq->next; in mqe_alloc()
29 return mq; in mqe_alloc()
38 static void mqe_free(MsgQueue_t *msgq, struct msgqueue_entry *mq) in mqe_free() argument
40 if (mq) { in mqe_free()
41 mq->next = msgq->free; in mqe_free()
42 msgq->free = mq; in mqe_free()
82 struct msgqueue_entry *mq = msgq->qe; in msgqueue_msglength() local
85 for (mq = msgq->qe; mq; mq = mq->next) in msgqueue_msglength()
[all …]
/Linux-v5.4/drivers/sh/maple/
Dmaple.c122 void (*callback) (struct mapleq *mq), in maple_getcond_callback()
140 struct mapleq *mq; in maple_release_device() local
143 mq = mdev->mq; in maple_release_device()
144 kmem_cache_free(maple_queue_cache, mq->recvbuf); in maple_release_device()
145 kfree(mq); in maple_release_device()
172 mdev->mq->command = command; in maple_add_packet()
173 mdev->mq->length = length; in maple_add_packet()
176 mdev->mq->sendbuf = sendbuf; in maple_add_packet()
179 list_add_tail(&mdev->mq->list, &maple_waitq); in maple_add_packet()
188 struct mapleq *mq; in maple_allocq() local
[all …]
/Linux-v5.4/block/
DMakefile9 blk-lib.o blk-mq.o blk-mq-tag.o blk-stat.o \
10 blk-mq-sysfs.o blk-mq-cpumap.o blk-mq-sched.o ioctl.o \
22 obj-$(CONFIG_MQ_IOSCHED_DEADLINE) += mq-deadline.o
30 obj-$(CONFIG_BLK_MQ_PCI) += blk-mq-pci.o
31 obj-$(CONFIG_BLK_MQ_VIRTIO) += blk-mq-virtio.o
32 obj-$(CONFIG_BLK_MQ_RDMA) += blk-mq-rdma.o
35 obj-$(CONFIG_BLK_DEBUG_FS) += blk-mq-debugfs.o
36 obj-$(CONFIG_BLK_DEBUG_FS_ZONED)+= blk-mq-debugfs-zoned.o
/Linux-v5.4/drivers/misc/sgi-xp/
Dxpc_uv.c112 xpc_get_gru_mq_irq_uv(struct xpc_gru_mq_uv *mq, int cpu, char *irq_name) in xpc_get_gru_mq_irq_uv() argument
114 int mmr_pnode = uv_blade_to_pnode(mq->mmr_blade); in xpc_get_gru_mq_irq_uv()
117 mq->irq = uv_setup_irq(irq_name, cpu, mq->mmr_blade, mq->mmr_offset, in xpc_get_gru_mq_irq_uv()
119 if (mq->irq < 0) in xpc_get_gru_mq_irq_uv()
120 return mq->irq; in xpc_get_gru_mq_irq_uv()
122 mq->mmr_value = uv_read_global_mmr64(mmr_pnode, mq->mmr_offset); in xpc_get_gru_mq_irq_uv()
126 mq->irq = SGI_XPC_ACTIVATE; in xpc_get_gru_mq_irq_uv()
128 mq->irq = SGI_XPC_NOTIFY; in xpc_get_gru_mq_irq_uv()
132 mq->mmr_value = (unsigned long)cpu_physical_id(cpu) << 32 | mq->irq; in xpc_get_gru_mq_irq_uv()
133 uv_write_global_mmr64(mmr_pnode, mq->mmr_offset, mq->mmr_value); in xpc_get_gru_mq_irq_uv()
[all …]
/Linux-v5.4/drivers/isdn/capi/
Dcapilib.c48 struct capilib_msgidqueue *mq; in mq_enqueue() local
49 if ((mq = np->msgidfree) == NULL) in mq_enqueue()
51 np->msgidfree = mq->next; in mq_enqueue()
52 mq->msgid = msgid; in mq_enqueue()
53 mq->next = NULL; in mq_enqueue()
55 np->msgidlast->next = mq; in mq_enqueue()
56 np->msgidlast = mq; in mq_enqueue()
58 np->msgidqueue = mq; in mq_enqueue()
68 struct capilib_msgidqueue *mq = *pp; in mq_dequeue() local
69 *pp = mq->next; in mq_dequeue()
[all …]
/Linux-v5.4/drivers/misc/sgi-gru/
Dgrukservices.c133 #define HSTATUS(mq, h) ((mq) + offsetof(struct message_queue, hstatus[h])) argument
547 struct message_queue *mq = p; in gru_create_message_queue() local
551 memset(mq, 0, bytes); in gru_create_message_queue()
552 mq->start = &mq->data; in gru_create_message_queue()
553 mq->start2 = &mq->data + (qlines / 2 - 1) * GRU_CACHE_LINE_BYTES; in gru_create_message_queue()
554 mq->next = &mq->data; in gru_create_message_queue()
555 mq->limit = &mq->data + (qlines - 2) * GRU_CACHE_LINE_BYTES; in gru_create_message_queue()
556 mq->qlines = qlines; in gru_create_message_queue()
557 mq->hstatus[0] = 0; in gru_create_message_queue()
558 mq->hstatus[1] = 1; in gru_create_message_queue()
[all …]
/Linux-v5.4/drivers/mailbox/
Domap-mailbox.c257 struct omap_mbox_queue *mq = in mbox_rx_work() local
263 while (kfifo_len(&mq->fifo) >= sizeof(msg)) { in mbox_rx_work()
264 len = kfifo_out(&mq->fifo, (unsigned char *)&msg, sizeof(msg)); in mbox_rx_work()
268 mbox_chan_received_data(mq->mbox->chan, (void *)data); in mbox_rx_work()
269 spin_lock_irq(&mq->lock); in mbox_rx_work()
270 if (mq->full) { in mbox_rx_work()
271 mq->full = false; in mbox_rx_work()
272 _omap_mbox_enable_irq(mq->mbox, IRQ_RX); in mbox_rx_work()
274 spin_unlock_irq(&mq->lock); in mbox_rx_work()
290 struct omap_mbox_queue *mq = mbox->rxq; in __mbox_rx_interrupt() local
[all …]
/Linux-v5.4/Documentation/admin-guide/device-mapper/
Dcache-policies.rst29 multiqueue (mq)
48 with the multiqueue (mq) policy.
50 The smq policy (vs mq) offers the promise of less memory utilization,
54 Users may switch from "mq" to "smq" simply by appropriately reloading a
56 mq policy's hints to be dropped. Also, performance of the cache may
63 The mq policy used a lot of memory; 88 bytes per cache block on a 64
78 mq placed entries in different levels of the multiqueue structures
91 The mq policy maintained a hit count for each cache block. For a
105 Testing smq shows substantially better performance than mq.
129 /dev/sdd 512 0 mq 4 sequential_threshold 1024 random_threshold 8"
/Linux-v5.4/include/linux/
Dmaple.h70 struct mapleq *mq; member
71 void (*callback) (struct mapleq * mq);
90 void (*callback) (struct mapleq * mq),
/Linux-v5.4/drivers/mtd/maps/
Dvmu-flash.c89 static void vmu_blockread(struct mapleq *mq) in vmu_blockread() argument
94 mdev = mq->dev; in vmu_blockread()
101 memcpy(card->blockread, mq->recvbuf->buf + 12, in vmu_blockread()
191 list_del_init(&(mdev->mq->list)); in maple_vmu_read_block()
192 kfree(mdev->mq->sendbuf); in maple_vmu_read_block()
193 mdev->mq->sendbuf = NULL; in maple_vmu_read_block()
283 kfree(mdev->mq->sendbuf); in maple_vmu_write_block()
284 mdev->mq->sendbuf = NULL; in maple_vmu_write_block()
285 list_del_init(&(mdev->mq->list)); in maple_vmu_write_block()
499 static void vmu_queryblocks(struct mapleq *mq) in vmu_queryblocks() argument
[all …]
/Linux-v5.4/drivers/net/ethernet/netronome/nfp/abm/
Dqdisc.c304 struct nfp_qdisc *mq = nfp_abm_qdisc_tree_deref_slot(slot); in nfp_abm_qdisc_clear_mq() local
307 if (mq->type != NFP_QDISC_MQ || mq->netdev != netdev) in nfp_abm_qdisc_clear_mq()
309 for (i = 0; i < mq->num_children; i++) in nfp_abm_qdisc_clear_mq()
310 if (mq->children[i] == qdisc) { in nfp_abm_qdisc_clear_mq()
311 mq->children[i] = NULL; in nfp_abm_qdisc_clear_mq()
791 memset(&qdisc->mq.stats, 0, sizeof(qdisc->mq.stats)); in nfp_abm_mq_stats()
792 memset(&qdisc->mq.prev_stats, 0, sizeof(qdisc->mq.prev_stats)); in nfp_abm_mq_stats()
803 nfp_abm_stats_propagate(&qdisc->mq.stats, in nfp_abm_mq_stats()
805 nfp_abm_stats_propagate(&qdisc->mq.prev_stats, in nfp_abm_mq_stats()
810 nfp_abm_stats_calculate(&qdisc->mq.stats, &qdisc->mq.prev_stats, in nfp_abm_mq_stats()
/Linux-v5.4/Documentation/block/
Dswitching-sched.rst17 the fly to select one of mq-deadline, none, bfq, or kyber schedulers -
32 [mq-deadline] kyber bfq none
35 [none] mq-deadline kyber bfq
Dnull_blk.rst92 0 nullb* use default blk-mq io scheduler
100 0 Register as a non-blocking blk-mq driver device.
101 1 Register as a blocking blk-mq driver device, null_blk will set
111 1 Tag set shared between devices for blk-mq. Only makes sense with
/Linux-v5.4/drivers/input/mouse/
Dmaplemouse.c27 static void dc_mouse_callback(struct mapleq *mq) in dc_mouse_callback() argument
30 struct maple_device *mapledev = mq->dev; in dc_mouse_callback()
33 unsigned char *res = mq->recvbuf->buf; in dc_mouse_callback()
/Linux-v5.4/Documentation/devicetree/bindings/powerpc/4xx/
Dppc440spe-adma.txt82 - compatible : "ibm,mq-440spe";
87 MQ0: mq {
88 compatible = "ibm,mq-440spe";
/Linux-v5.4/drivers/net/wireless/intel/iwlwifi/dvm/
Dmain.c2053 int mq = priv->queue_to_mac80211[queue]; in iwl_stop_sw_queue() local
2055 if (WARN_ON_ONCE(mq == IWL_INVALID_MAC80211_QUEUE)) in iwl_stop_sw_queue()
2058 if (atomic_inc_return(&priv->queue_stop_count[mq]) > 1) { in iwl_stop_sw_queue()
2061 queue, mq); in iwl_stop_sw_queue()
2065 set_bit(mq, &priv->transport_queue_stop); in iwl_stop_sw_queue()
2066 ieee80211_stop_queue(priv->hw, mq); in iwl_stop_sw_queue()
2072 int mq = priv->queue_to_mac80211[queue]; in iwl_wake_sw_queue() local
2074 if (WARN_ON_ONCE(mq == IWL_INVALID_MAC80211_QUEUE)) in iwl_wake_sw_queue()
2077 if (atomic_dec_return(&priv->queue_stop_count[mq]) > 0) { in iwl_wake_sw_queue()
2080 queue, mq); in iwl_wake_sw_queue()
[all …]
/Linux-v5.4/arch/ia64/include/asm/sn/
Dsn_sal.h92 sn_mq_watchlist_alloc(int blade, void *mq, unsigned int mq_size, in sn_mq_watchlist_alloc() argument
100 addr = (unsigned long)mq; in sn_mq_watchlist_alloc()
/Linux-v5.4/drivers/input/joystick/
Dmaplecontrol.c27 static void dc_pad_callback(struct mapleq *mq) in dc_pad_callback() argument
30 struct maple_device *mapledev = mq->dev; in dc_pad_callback()
33 unsigned char *res = mq->recvbuf->buf; in dc_pad_callback()
/Linux-v5.4/drivers/infiniband/hw/ocrdma/
Docrdma_hw.c124 (dev->mq.cq.va + (dev->mq.cq.tail * sizeof(struct ocrdma_mcqe))); in ocrdma_get_mcqe()
133 dev->mq.cq.tail = (dev->mq.cq.tail + 1) & (OCRDMA_MQ_CQ_LEN - 1); in ocrdma_mcq_inc_tail()
138 return dev->mq.sq.va + (dev->mq.sq.head * sizeof(struct ocrdma_mqe)); in ocrdma_get_mqe()
143 dev->mq.sq.head = (dev->mq.sq.head + 1) & (OCRDMA_MQ_LEN - 1); in ocrdma_mq_inc_head()
148 return dev->mq.sq.va + (dev->mqe_ctx.tag * sizeof(struct ocrdma_mqe)); in ocrdma_get_mqe_rsp()
325 val |= dev->mq.sq.id & OCRDMA_MQ_ID_MASK; in ocrdma_ring_mq_db()
565 struct ocrdma_queue_info *mq, in ocrdma_mbx_create_mq() argument
574 num_pages = PAGES_4K_SPANNED(mq->va, mq->size); in ocrdma_mbx_create_mq()
589 cmd->async_cqid_ringsize |= (ocrdma_encoded_q_len(mq->len) << in ocrdma_mbx_create_mq()
594 ocrdma_build_q_pages(pa, num_pages, mq->dma, PAGE_SIZE_4K); in ocrdma_mbx_create_mq()
[all …]
/Linux-v5.4/Documentation/ABI/testing/
Dsysfs-block-dm45 Description: Request-based Device-mapper blk-mq I/O path mode.
46 Contains the value 1 if the device is using blk-mq.

12