Lines Matching refs:mq
179 struct mmc_queue *mq);
249 struct mmc_queue *mq; in power_ro_lock_store() local
260 mq = &md->queue; in power_ro_lock_store()
263 req = blk_mq_alloc_request(mq->queue, REQ_OP_DRV_OUT, 0); in power_ro_lock_store()
640 struct mmc_queue *mq; in mmc_blk_ioctl_cmd() local
660 mq = &md->queue; in mmc_blk_ioctl_cmd()
661 req = blk_mq_alloc_request(mq->queue, in mmc_blk_ioctl_cmd()
691 struct mmc_queue *mq; in mmc_blk_ioctl_multi_cmd() local
733 mq = &md->queue; in mmc_blk_ioctl_multi_cmd()
734 req = blk_mq_alloc_request(mq->queue, in mmc_blk_ioctl_multi_cmd()
1040 static void mmc_blk_issue_drv_op(struct mmc_queue *mq, struct request *req) in mmc_blk_issue_drv_op() argument
1043 struct mmc_card *card = mq->card; in mmc_blk_issue_drv_op()
1044 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_issue_drv_op()
1107 static void mmc_blk_issue_erase_rq(struct mmc_queue *mq, struct request *req, in mmc_blk_issue_erase_rq() argument
1110 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_issue_erase_rq()
1145 static void mmc_blk_issue_trim_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_issue_trim_rq() argument
1147 mmc_blk_issue_erase_rq(mq, req, MMC_BLK_TRIM, MMC_TRIM_ARG); in mmc_blk_issue_trim_rq()
1150 static void mmc_blk_issue_discard_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_issue_discard_rq() argument
1152 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_issue_discard_rq()
1159 mmc_blk_issue_erase_rq(mq, req, MMC_BLK_DISCARD, arg); in mmc_blk_issue_discard_rq()
1162 static void mmc_blk_issue_secdiscard_rq(struct mmc_queue *mq, in mmc_blk_issue_secdiscard_rq() argument
1165 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_issue_secdiscard_rq()
1232 static void mmc_blk_issue_flush(struct mmc_queue *mq, struct request *req) in mmc_blk_issue_flush() argument
1234 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_issue_flush()
1319 static void mmc_blk_data_prep(struct mmc_queue *mq, struct mmc_queue_req *mqrq, in mmc_blk_data_prep() argument
1323 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_data_prep()
1391 brq->data.blocks = queue_physical_block_size(mq->queue) >> 9; in mmc_blk_data_prep()
1425 brq->data.sg_len = mmc_queue_map_sg(mq, mqrq); in mmc_blk_data_prep()
1455 static void mmc_blk_cqe_complete_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_cqe_complete_rq() argument
1460 struct mmc_host *host = mq->card->host; in mmc_blk_cqe_complete_rq()
1461 enum mmc_issue_type issue_type = mmc_issue_type(mq, req); in mmc_blk_cqe_complete_rq()
1489 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_cqe_complete_rq()
1491 mq->in_flight[issue_type] -= 1; in mmc_blk_cqe_complete_rq()
1493 put_card = (mmc_tot_in_flight(mq) == 0); in mmc_blk_cqe_complete_rq()
1495 mmc_cqe_check_busy(mq); in mmc_blk_cqe_complete_rq()
1497 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_cqe_complete_rq()
1499 if (!mq->cqe_busy) in mmc_blk_cqe_complete_rq()
1503 mmc_put_card(mq->card, &mq->ctx); in mmc_blk_cqe_complete_rq()
1506 void mmc_blk_cqe_recovery(struct mmc_queue *mq) in mmc_blk_cqe_recovery() argument
1508 struct mmc_card *card = mq->card; in mmc_blk_cqe_recovery()
1516 mmc_blk_reset(mq->blkdata, host, MMC_BLK_CQE_RECOVERY); in mmc_blk_cqe_recovery()
1517 mmc_blk_reset_success(mq->blkdata, MMC_BLK_CQE_RECOVERY); in mmc_blk_cqe_recovery()
1528 struct mmc_queue *mq = q->queuedata; in mmc_blk_cqe_req_done() local
1534 if (mq->in_recovery) in mmc_blk_cqe_req_done()
1535 mmc_blk_cqe_complete_rq(mq, req); in mmc_blk_cqe_req_done()
1561 static int mmc_blk_cqe_issue_flush(struct mmc_queue *mq, struct request *req) in mmc_blk_cqe_issue_flush() argument
1573 return mmc_blk_cqe_start_req(mq->card->host, mrq); in mmc_blk_cqe_issue_flush()
1576 static int mmc_blk_hsq_issue_rw_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_hsq_issue_rw_rq() argument
1579 struct mmc_host *host = mq->card->host; in mmc_blk_hsq_issue_rw_rq()
1582 mmc_blk_rw_rq_prep(mqrq, mq->card, 0, mq); in mmc_blk_hsq_issue_rw_rq()
1593 static int mmc_blk_cqe_issue_rw_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_cqe_issue_rw_rq() argument
1596 struct mmc_host *host = mq->card->host; in mmc_blk_cqe_issue_rw_rq()
1599 return mmc_blk_hsq_issue_rw_rq(mq, req); in mmc_blk_cqe_issue_rw_rq()
1601 mmc_blk_data_prep(mq, mqrq, 0, NULL, NULL); in mmc_blk_cqe_issue_rw_rq()
1603 return mmc_blk_cqe_start_req(mq->card->host, &mqrq->brq.mrq); in mmc_blk_cqe_issue_rw_rq()
1609 struct mmc_queue *mq) in mmc_blk_rw_rq_prep() argument
1614 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_rw_rq_prep()
1617 mmc_blk_data_prep(mq, mqrq, recovery_mode, &do_rel_wr, &do_data_tag); in mmc_blk_rw_rq_prep()
1709 static void mmc_blk_read_single(struct mmc_queue *mq, struct request *req) in mmc_blk_read_single() argument
1713 struct mmc_card *card = mq->card; in mmc_blk_read_single()
1716 size_t bytes_per_read = queue_physical_block_size(mq->queue); in mmc_blk_read_single()
1724 mmc_blk_rw_rq_prep(mqrq, card, 1, mq); in mmc_blk_read_single()
1781 struct mmc_queue *mq = req->q->queuedata; in mmc_blk_status_error() local
1784 if (mmc_host_is_spi(mq->card->host)) in mmc_blk_status_error()
1817 static void mmc_blk_mq_rw_recovery(struct mmc_queue *mq, struct request *req) in mmc_blk_mq_rw_recovery() argument
1822 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_mq_rw_recovery()
1823 struct mmc_card *card = mq->card; in mmc_blk_mq_rw_recovery()
1853 if (!mmc_host_is_spi(mq->card->host) && in mmc_blk_mq_rw_recovery()
1855 err = mmc_blk_fix_state(mq->card, req); in mmc_blk_mq_rw_recovery()
1870 if (!mmc_host_is_spi(mq->card->host) && in mmc_blk_mq_rw_recovery()
1900 queue_physical_block_size(mq->queue) >> 9) { in mmc_blk_mq_rw_recovery()
1902 mmc_blk_read_single(mq, req); in mmc_blk_mq_rw_recovery()
1994 static inline void mmc_blk_rw_reset_success(struct mmc_queue *mq, in mmc_blk_rw_reset_success() argument
1999 mmc_blk_reset_success(mq->blkdata, type); in mmc_blk_rw_reset_success()
2002 static void mmc_blk_mq_complete_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_mq_complete_rq() argument
2017 if (mmc_card_removed(mq->card)) in mmc_blk_mq_complete_rq()
2023 static bool mmc_blk_urgent_bkops_needed(struct mmc_queue *mq, in mmc_blk_urgent_bkops_needed() argument
2026 return mmc_card_mmc(mq->card) && !mmc_host_is_spi(mq->card->host) && in mmc_blk_urgent_bkops_needed()
2031 static void mmc_blk_urgent_bkops(struct mmc_queue *mq, in mmc_blk_urgent_bkops() argument
2034 if (mmc_blk_urgent_bkops_needed(mq, mqrq)) in mmc_blk_urgent_bkops()
2035 mmc_run_bkops(mq->card); in mmc_blk_urgent_bkops()
2044 struct mmc_queue *mq = q->queuedata; in mmc_blk_hsq_req_done() local
2045 struct mmc_host *host = mq->card->host; in mmc_blk_hsq_req_done()
2049 mmc_blk_urgent_bkops_needed(mq, mqrq)) { in mmc_blk_hsq_req_done()
2050 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_hsq_req_done()
2051 mq->recovery_needed = true; in mmc_blk_hsq_req_done()
2052 mq->recovery_req = req; in mmc_blk_hsq_req_done()
2053 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_hsq_req_done()
2057 schedule_work(&mq->recovery_work); in mmc_blk_hsq_req_done()
2061 mmc_blk_rw_reset_success(mq, req); in mmc_blk_hsq_req_done()
2067 if (mq->in_recovery) in mmc_blk_hsq_req_done()
2068 mmc_blk_cqe_complete_rq(mq, req); in mmc_blk_hsq_req_done()
2075 struct mmc_queue *mq = req->q->queuedata; in mmc_blk_mq_complete() local
2076 struct mmc_host *host = mq->card->host; in mmc_blk_mq_complete()
2079 mmc_blk_cqe_complete_rq(mq, req); in mmc_blk_mq_complete()
2081 mmc_blk_mq_complete_rq(mq, req); in mmc_blk_mq_complete()
2084 static void mmc_blk_mq_poll_completion(struct mmc_queue *mq, in mmc_blk_mq_poll_completion() argument
2088 struct mmc_host *host = mq->card->host; in mmc_blk_mq_poll_completion()
2091 mmc_blk_card_busy(mq->card, req)) { in mmc_blk_mq_poll_completion()
2092 mmc_blk_mq_rw_recovery(mq, req); in mmc_blk_mq_poll_completion()
2094 mmc_blk_rw_reset_success(mq, req); in mmc_blk_mq_poll_completion()
2098 mmc_blk_urgent_bkops(mq, mqrq); in mmc_blk_mq_poll_completion()
2101 static void mmc_blk_mq_dec_in_flight(struct mmc_queue *mq, enum mmc_issue_type issue_type) in mmc_blk_mq_dec_in_flight() argument
2106 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_mq_dec_in_flight()
2108 mq->in_flight[issue_type] -= 1; in mmc_blk_mq_dec_in_flight()
2110 put_card = (mmc_tot_in_flight(mq) == 0); in mmc_blk_mq_dec_in_flight()
2112 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_mq_dec_in_flight()
2115 mmc_put_card(mq->card, &mq->ctx); in mmc_blk_mq_dec_in_flight()
2118 static void mmc_blk_mq_post_req(struct mmc_queue *mq, struct request *req, in mmc_blk_mq_post_req() argument
2121 enum mmc_issue_type issue_type = mmc_issue_type(mq, req); in mmc_blk_mq_post_req()
2124 struct mmc_host *host = mq->card->host; in mmc_blk_mq_post_req()
2132 if (mq->in_recovery) { in mmc_blk_mq_post_req()
2133 mmc_blk_mq_complete_rq(mq, req); in mmc_blk_mq_post_req()
2141 mmc_blk_mq_dec_in_flight(mq, issue_type); in mmc_blk_mq_post_req()
2144 void mmc_blk_mq_recovery(struct mmc_queue *mq) in mmc_blk_mq_recovery() argument
2146 struct request *req = mq->recovery_req; in mmc_blk_mq_recovery()
2147 struct mmc_host *host = mq->card->host; in mmc_blk_mq_recovery()
2150 mq->recovery_req = NULL; in mmc_blk_mq_recovery()
2151 mq->rw_wait = false; in mmc_blk_mq_recovery()
2155 mmc_blk_mq_rw_recovery(mq, req); in mmc_blk_mq_recovery()
2158 mmc_blk_urgent_bkops(mq, mqrq); in mmc_blk_mq_recovery()
2160 mmc_blk_mq_post_req(mq, req, true); in mmc_blk_mq_recovery()
2163 static void mmc_blk_mq_complete_prev_req(struct mmc_queue *mq, in mmc_blk_mq_complete_prev_req() argument
2166 if (mmc_host_done_complete(mq->card->host)) in mmc_blk_mq_complete_prev_req()
2169 mutex_lock(&mq->complete_lock); in mmc_blk_mq_complete_prev_req()
2171 if (!mq->complete_req) in mmc_blk_mq_complete_prev_req()
2174 mmc_blk_mq_poll_completion(mq, mq->complete_req); in mmc_blk_mq_complete_prev_req()
2177 *prev_req = mq->complete_req; in mmc_blk_mq_complete_prev_req()
2179 mmc_blk_mq_post_req(mq, mq->complete_req, true); in mmc_blk_mq_complete_prev_req()
2181 mq->complete_req = NULL; in mmc_blk_mq_complete_prev_req()
2184 mutex_unlock(&mq->complete_lock); in mmc_blk_mq_complete_prev_req()
2189 struct mmc_queue *mq = container_of(work, struct mmc_queue, in mmc_blk_mq_complete_work() local
2192 mmc_blk_mq_complete_prev_req(mq, NULL); in mmc_blk_mq_complete_work()
2201 struct mmc_queue *mq = q->queuedata; in mmc_blk_mq_req_done() local
2202 struct mmc_host *host = mq->card->host; in mmc_blk_mq_req_done()
2214 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_mq_req_done()
2215 mq->complete_req = req; in mmc_blk_mq_req_done()
2216 mq->rw_wait = false; in mmc_blk_mq_req_done()
2217 waiting = mq->waiting; in mmc_blk_mq_req_done()
2218 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_mq_req_done()
2227 wake_up(&mq->wait); in mmc_blk_mq_req_done()
2229 queue_work(mq->card->complete_wq, &mq->complete_work); in mmc_blk_mq_req_done()
2236 mmc_blk_urgent_bkops_needed(mq, mqrq)) { in mmc_blk_mq_req_done()
2237 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_mq_req_done()
2238 mq->recovery_needed = true; in mmc_blk_mq_req_done()
2239 mq->recovery_req = req; in mmc_blk_mq_req_done()
2240 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_mq_req_done()
2241 wake_up(&mq->wait); in mmc_blk_mq_req_done()
2242 schedule_work(&mq->recovery_work); in mmc_blk_mq_req_done()
2246 mmc_blk_rw_reset_success(mq, req); in mmc_blk_mq_req_done()
2248 mq->rw_wait = false; in mmc_blk_mq_req_done()
2249 wake_up(&mq->wait); in mmc_blk_mq_req_done()
2252 mmc_blk_mq_post_req(mq, req, false); in mmc_blk_mq_req_done()
2255 static bool mmc_blk_rw_wait_cond(struct mmc_queue *mq, int *err) in mmc_blk_rw_wait_cond() argument
2264 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_rw_wait_cond()
2265 if (mq->recovery_needed) { in mmc_blk_rw_wait_cond()
2269 done = !mq->rw_wait; in mmc_blk_rw_wait_cond()
2271 mq->waiting = !done; in mmc_blk_rw_wait_cond()
2272 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_rw_wait_cond()
2277 static int mmc_blk_rw_wait(struct mmc_queue *mq, struct request **prev_req) in mmc_blk_rw_wait() argument
2281 wait_event(mq->wait, mmc_blk_rw_wait_cond(mq, &err)); in mmc_blk_rw_wait()
2284 mmc_blk_mq_complete_prev_req(mq, prev_req); in mmc_blk_rw_wait()
2289 static int mmc_blk_mq_issue_rw_rq(struct mmc_queue *mq, in mmc_blk_mq_issue_rw_rq() argument
2293 struct mmc_host *host = mq->card->host; in mmc_blk_mq_issue_rw_rq()
2297 mmc_blk_rw_rq_prep(mqrq, mq->card, 0, mq); in mmc_blk_mq_issue_rw_rq()
2303 err = mmc_blk_rw_wait(mq, &prev_req); in mmc_blk_mq_issue_rw_rq()
2307 mq->rw_wait = true; in mmc_blk_mq_issue_rw_rq()
2312 mmc_blk_mq_post_req(mq, prev_req, true); in mmc_blk_mq_issue_rw_rq()
2315 mq->rw_wait = false; in mmc_blk_mq_issue_rw_rq()
2328 static int mmc_blk_wait_for_idle(struct mmc_queue *mq, struct mmc_host *host) in mmc_blk_wait_for_idle() argument
2333 return mmc_blk_rw_wait(mq, NULL); in mmc_blk_wait_for_idle()
2336 enum mmc_issued mmc_blk_mq_issue_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_mq_issue_rq() argument
2338 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_mq_issue_rq()
2347 switch (mmc_issue_type(mq, req)) { in mmc_blk_mq_issue_rq()
2349 ret = mmc_blk_wait_for_idle(mq, host); in mmc_blk_mq_issue_rq()
2355 mmc_blk_issue_drv_op(mq, req); in mmc_blk_mq_issue_rq()
2358 mmc_blk_issue_discard_rq(mq, req); in mmc_blk_mq_issue_rq()
2361 mmc_blk_issue_secdiscard_rq(mq, req); in mmc_blk_mq_issue_rq()
2364 mmc_blk_issue_trim_rq(mq, req); in mmc_blk_mq_issue_rq()
2367 mmc_blk_issue_flush(mq, req); in mmc_blk_mq_issue_rq()
2382 ret = mmc_blk_cqe_issue_flush(mq, req); in mmc_blk_mq_issue_rq()
2387 ret = mmc_blk_cqe_issue_rw_rq(mq, req); in mmc_blk_mq_issue_rq()
2389 ret = mmc_blk_mq_issue_rw_rq(mq, req); in mmc_blk_mq_issue_rq()
2811 struct mmc_queue *mq = &md->queue; in mmc_dbg_card_status_get() local
2816 req = blk_mq_alloc_request(mq->queue, REQ_OP_DRV_IN, 0); in mmc_dbg_card_status_get()
2841 struct mmc_queue *mq = &md->queue; in mmc_ext_csd_open() local
2853 req = blk_mq_alloc_request(mq->queue, REQ_OP_DRV_IN, 0); in mmc_ext_csd_open()