Lines Matching refs:mq
181 struct mmc_queue *mq);
249 struct mmc_queue *mq; in power_ro_lock_store() local
260 mq = &md->queue; in power_ro_lock_store()
263 req = blk_mq_alloc_request(mq->queue, REQ_OP_DRV_OUT, 0); in power_ro_lock_store()
630 struct mmc_queue *mq; in mmc_blk_ioctl_cmd() local
650 mq = &md->queue; in mmc_blk_ioctl_cmd()
651 req = blk_mq_alloc_request(mq->queue, in mmc_blk_ioctl_cmd()
680 struct mmc_queue *mq; in mmc_blk_ioctl_multi_cmd() local
722 mq = &md->queue; in mmc_blk_ioctl_multi_cmd()
723 req = blk_mq_alloc_request(mq->queue, in mmc_blk_ioctl_multi_cmd()
1036 static void mmc_blk_issue_drv_op(struct mmc_queue *mq, struct request *req) in mmc_blk_issue_drv_op() argument
1039 struct mmc_card *card = mq->card; in mmc_blk_issue_drv_op()
1040 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_issue_drv_op()
1103 static void mmc_blk_issue_erase_rq(struct mmc_queue *mq, struct request *req, in mmc_blk_issue_erase_rq() argument
1106 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_issue_erase_rq()
1141 static void mmc_blk_issue_trim_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_issue_trim_rq() argument
1143 mmc_blk_issue_erase_rq(mq, req, MMC_BLK_TRIM, MMC_TRIM_ARG); in mmc_blk_issue_trim_rq()
1146 static void mmc_blk_issue_discard_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_issue_discard_rq() argument
1148 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_issue_discard_rq()
1155 mmc_blk_issue_erase_rq(mq, req, MMC_BLK_DISCARD, arg); in mmc_blk_issue_discard_rq()
1158 static void mmc_blk_issue_secdiscard_rq(struct mmc_queue *mq, in mmc_blk_issue_secdiscard_rq() argument
1161 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_issue_secdiscard_rq()
1228 static void mmc_blk_issue_flush(struct mmc_queue *mq, struct request *req) in mmc_blk_issue_flush() argument
1230 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_issue_flush()
1315 static void mmc_blk_data_prep(struct mmc_queue *mq, struct mmc_queue_req *mqrq, in mmc_blk_data_prep() argument
1319 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_data_prep()
1387 brq->data.blocks = queue_physical_block_size(mq->queue) >> 9; in mmc_blk_data_prep()
1421 brq->data.sg_len = mmc_queue_map_sg(mq, mqrq); in mmc_blk_data_prep()
1451 static void mmc_blk_cqe_complete_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_cqe_complete_rq() argument
1456 struct mmc_host *host = mq->card->host; in mmc_blk_cqe_complete_rq()
1457 enum mmc_issue_type issue_type = mmc_issue_type(mq, req); in mmc_blk_cqe_complete_rq()
1485 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_cqe_complete_rq()
1487 mq->in_flight[issue_type] -= 1; in mmc_blk_cqe_complete_rq()
1489 put_card = (mmc_tot_in_flight(mq) == 0); in mmc_blk_cqe_complete_rq()
1491 mmc_cqe_check_busy(mq); in mmc_blk_cqe_complete_rq()
1493 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_cqe_complete_rq()
1495 if (!mq->cqe_busy) in mmc_blk_cqe_complete_rq()
1499 mmc_put_card(mq->card, &mq->ctx); in mmc_blk_cqe_complete_rq()
1502 void mmc_blk_cqe_recovery(struct mmc_queue *mq) in mmc_blk_cqe_recovery() argument
1504 struct mmc_card *card = mq->card; in mmc_blk_cqe_recovery()
1512 mmc_blk_reset(mq->blkdata, host, MMC_BLK_CQE_RECOVERY); in mmc_blk_cqe_recovery()
1513 mmc_blk_reset_success(mq->blkdata, MMC_BLK_CQE_RECOVERY); in mmc_blk_cqe_recovery()
1524 struct mmc_queue *mq = q->queuedata; in mmc_blk_cqe_req_done() local
1530 if (mq->in_recovery) in mmc_blk_cqe_req_done()
1531 mmc_blk_cqe_complete_rq(mq, req); in mmc_blk_cqe_req_done()
1557 static int mmc_blk_cqe_issue_flush(struct mmc_queue *mq, struct request *req) in mmc_blk_cqe_issue_flush() argument
1569 return mmc_blk_cqe_start_req(mq->card->host, mrq); in mmc_blk_cqe_issue_flush()
1572 static int mmc_blk_hsq_issue_rw_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_hsq_issue_rw_rq() argument
1575 struct mmc_host *host = mq->card->host; in mmc_blk_hsq_issue_rw_rq()
1578 mmc_blk_rw_rq_prep(mqrq, mq->card, 0, mq); in mmc_blk_hsq_issue_rw_rq()
1589 static int mmc_blk_cqe_issue_rw_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_cqe_issue_rw_rq() argument
1592 struct mmc_host *host = mq->card->host; in mmc_blk_cqe_issue_rw_rq()
1595 return mmc_blk_hsq_issue_rw_rq(mq, req); in mmc_blk_cqe_issue_rw_rq()
1597 mmc_blk_data_prep(mq, mqrq, 0, NULL, NULL); in mmc_blk_cqe_issue_rw_rq()
1599 return mmc_blk_cqe_start_req(mq->card->host, &mqrq->brq.mrq); in mmc_blk_cqe_issue_rw_rq()
1605 struct mmc_queue *mq) in mmc_blk_rw_rq_prep() argument
1610 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_rw_rq_prep()
1613 mmc_blk_data_prep(mq, mqrq, recovery_mode, &do_rel_wr, &do_data_tag); in mmc_blk_rw_rq_prep()
1705 static void mmc_blk_read_single(struct mmc_queue *mq, struct request *req) in mmc_blk_read_single() argument
1709 struct mmc_card *card = mq->card; in mmc_blk_read_single()
1712 size_t bytes_per_read = queue_physical_block_size(mq->queue); in mmc_blk_read_single()
1720 mmc_blk_rw_rq_prep(mqrq, card, 1, mq); in mmc_blk_read_single()
1777 struct mmc_queue *mq = req->q->queuedata; in mmc_blk_status_error() local
1780 if (mmc_host_is_spi(mq->card->host)) in mmc_blk_status_error()
1813 static void mmc_blk_mq_rw_recovery(struct mmc_queue *mq, struct request *req) in mmc_blk_mq_rw_recovery() argument
1818 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_mq_rw_recovery()
1819 struct mmc_card *card = mq->card; in mmc_blk_mq_rw_recovery()
1849 if (!mmc_host_is_spi(mq->card->host) && in mmc_blk_mq_rw_recovery()
1851 err = mmc_blk_fix_state(mq->card, req); in mmc_blk_mq_rw_recovery()
1866 if (!mmc_host_is_spi(mq->card->host) && in mmc_blk_mq_rw_recovery()
1896 queue_physical_block_size(mq->queue) >> 9) { in mmc_blk_mq_rw_recovery()
1898 mmc_blk_read_single(mq, req); in mmc_blk_mq_rw_recovery()
1990 static inline void mmc_blk_rw_reset_success(struct mmc_queue *mq, in mmc_blk_rw_reset_success() argument
1995 mmc_blk_reset_success(mq->blkdata, type); in mmc_blk_rw_reset_success()
1998 static void mmc_blk_mq_complete_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_mq_complete_rq() argument
2013 if (mmc_card_removed(mq->card)) in mmc_blk_mq_complete_rq()
2019 static bool mmc_blk_urgent_bkops_needed(struct mmc_queue *mq, in mmc_blk_urgent_bkops_needed() argument
2022 return mmc_card_mmc(mq->card) && !mmc_host_is_spi(mq->card->host) && in mmc_blk_urgent_bkops_needed()
2027 static void mmc_blk_urgent_bkops(struct mmc_queue *mq, in mmc_blk_urgent_bkops() argument
2030 if (mmc_blk_urgent_bkops_needed(mq, mqrq)) in mmc_blk_urgent_bkops()
2031 mmc_run_bkops(mq->card); in mmc_blk_urgent_bkops()
2040 struct mmc_queue *mq = q->queuedata; in mmc_blk_hsq_req_done() local
2041 struct mmc_host *host = mq->card->host; in mmc_blk_hsq_req_done()
2045 mmc_blk_urgent_bkops_needed(mq, mqrq)) { in mmc_blk_hsq_req_done()
2046 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_hsq_req_done()
2047 mq->recovery_needed = true; in mmc_blk_hsq_req_done()
2048 mq->recovery_req = req; in mmc_blk_hsq_req_done()
2049 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_hsq_req_done()
2053 schedule_work(&mq->recovery_work); in mmc_blk_hsq_req_done()
2057 mmc_blk_rw_reset_success(mq, req); in mmc_blk_hsq_req_done()
2063 if (mq->in_recovery) in mmc_blk_hsq_req_done()
2064 mmc_blk_cqe_complete_rq(mq, req); in mmc_blk_hsq_req_done()
2071 struct mmc_queue *mq = req->q->queuedata; in mmc_blk_mq_complete() local
2072 struct mmc_host *host = mq->card->host; in mmc_blk_mq_complete()
2075 mmc_blk_cqe_complete_rq(mq, req); in mmc_blk_mq_complete()
2077 mmc_blk_mq_complete_rq(mq, req); in mmc_blk_mq_complete()
2080 static void mmc_blk_mq_poll_completion(struct mmc_queue *mq, in mmc_blk_mq_poll_completion() argument
2084 struct mmc_host *host = mq->card->host; in mmc_blk_mq_poll_completion()
2087 mmc_blk_card_busy(mq->card, req)) { in mmc_blk_mq_poll_completion()
2088 mmc_blk_mq_rw_recovery(mq, req); in mmc_blk_mq_poll_completion()
2090 mmc_blk_rw_reset_success(mq, req); in mmc_blk_mq_poll_completion()
2094 mmc_blk_urgent_bkops(mq, mqrq); in mmc_blk_mq_poll_completion()
2097 static void mmc_blk_mq_dec_in_flight(struct mmc_queue *mq, struct request *req) in mmc_blk_mq_dec_in_flight() argument
2102 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_mq_dec_in_flight()
2104 mq->in_flight[mmc_issue_type(mq, req)] -= 1; in mmc_blk_mq_dec_in_flight()
2106 put_card = (mmc_tot_in_flight(mq) == 0); in mmc_blk_mq_dec_in_flight()
2108 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_mq_dec_in_flight()
2111 mmc_put_card(mq->card, &mq->ctx); in mmc_blk_mq_dec_in_flight()
2114 static void mmc_blk_mq_post_req(struct mmc_queue *mq, struct request *req, in mmc_blk_mq_post_req() argument
2119 struct mmc_host *host = mq->card->host; in mmc_blk_mq_post_req()
2127 if (mq->in_recovery) { in mmc_blk_mq_post_req()
2128 mmc_blk_mq_complete_rq(mq, req); in mmc_blk_mq_post_req()
2136 mmc_blk_mq_dec_in_flight(mq, req); in mmc_blk_mq_post_req()
2139 void mmc_blk_mq_recovery(struct mmc_queue *mq) in mmc_blk_mq_recovery() argument
2141 struct request *req = mq->recovery_req; in mmc_blk_mq_recovery()
2142 struct mmc_host *host = mq->card->host; in mmc_blk_mq_recovery()
2145 mq->recovery_req = NULL; in mmc_blk_mq_recovery()
2146 mq->rw_wait = false; in mmc_blk_mq_recovery()
2150 mmc_blk_mq_rw_recovery(mq, req); in mmc_blk_mq_recovery()
2153 mmc_blk_urgent_bkops(mq, mqrq); in mmc_blk_mq_recovery()
2155 mmc_blk_mq_post_req(mq, req, true); in mmc_blk_mq_recovery()
2158 static void mmc_blk_mq_complete_prev_req(struct mmc_queue *mq, in mmc_blk_mq_complete_prev_req() argument
2161 if (mmc_host_done_complete(mq->card->host)) in mmc_blk_mq_complete_prev_req()
2164 mutex_lock(&mq->complete_lock); in mmc_blk_mq_complete_prev_req()
2166 if (!mq->complete_req) in mmc_blk_mq_complete_prev_req()
2169 mmc_blk_mq_poll_completion(mq, mq->complete_req); in mmc_blk_mq_complete_prev_req()
2172 *prev_req = mq->complete_req; in mmc_blk_mq_complete_prev_req()
2174 mmc_blk_mq_post_req(mq, mq->complete_req, true); in mmc_blk_mq_complete_prev_req()
2176 mq->complete_req = NULL; in mmc_blk_mq_complete_prev_req()
2179 mutex_unlock(&mq->complete_lock); in mmc_blk_mq_complete_prev_req()
2184 struct mmc_queue *mq = container_of(work, struct mmc_queue, in mmc_blk_mq_complete_work() local
2187 mmc_blk_mq_complete_prev_req(mq, NULL); in mmc_blk_mq_complete_work()
2196 struct mmc_queue *mq = q->queuedata; in mmc_blk_mq_req_done() local
2197 struct mmc_host *host = mq->card->host; in mmc_blk_mq_req_done()
2209 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_mq_req_done()
2210 mq->complete_req = req; in mmc_blk_mq_req_done()
2211 mq->rw_wait = false; in mmc_blk_mq_req_done()
2212 waiting = mq->waiting; in mmc_blk_mq_req_done()
2213 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_mq_req_done()
2222 wake_up(&mq->wait); in mmc_blk_mq_req_done()
2224 queue_work(mq->card->complete_wq, &mq->complete_work); in mmc_blk_mq_req_done()
2231 mmc_blk_urgent_bkops_needed(mq, mqrq)) { in mmc_blk_mq_req_done()
2232 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_mq_req_done()
2233 mq->recovery_needed = true; in mmc_blk_mq_req_done()
2234 mq->recovery_req = req; in mmc_blk_mq_req_done()
2235 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_mq_req_done()
2236 wake_up(&mq->wait); in mmc_blk_mq_req_done()
2237 schedule_work(&mq->recovery_work); in mmc_blk_mq_req_done()
2241 mmc_blk_rw_reset_success(mq, req); in mmc_blk_mq_req_done()
2243 mq->rw_wait = false; in mmc_blk_mq_req_done()
2244 wake_up(&mq->wait); in mmc_blk_mq_req_done()
2247 mmc_blk_mq_post_req(mq, req, false); in mmc_blk_mq_req_done()
2250 static bool mmc_blk_rw_wait_cond(struct mmc_queue *mq, int *err) in mmc_blk_rw_wait_cond() argument
2259 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_rw_wait_cond()
2260 if (mq->recovery_needed) { in mmc_blk_rw_wait_cond()
2264 done = !mq->rw_wait; in mmc_blk_rw_wait_cond()
2266 mq->waiting = !done; in mmc_blk_rw_wait_cond()
2267 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_rw_wait_cond()
2272 static int mmc_blk_rw_wait(struct mmc_queue *mq, struct request **prev_req) in mmc_blk_rw_wait() argument
2276 wait_event(mq->wait, mmc_blk_rw_wait_cond(mq, &err)); in mmc_blk_rw_wait()
2279 mmc_blk_mq_complete_prev_req(mq, prev_req); in mmc_blk_rw_wait()
2284 static int mmc_blk_mq_issue_rw_rq(struct mmc_queue *mq, in mmc_blk_mq_issue_rw_rq() argument
2288 struct mmc_host *host = mq->card->host; in mmc_blk_mq_issue_rw_rq()
2292 mmc_blk_rw_rq_prep(mqrq, mq->card, 0, mq); in mmc_blk_mq_issue_rw_rq()
2298 err = mmc_blk_rw_wait(mq, &prev_req); in mmc_blk_mq_issue_rw_rq()
2302 mq->rw_wait = true; in mmc_blk_mq_issue_rw_rq()
2307 mmc_blk_mq_post_req(mq, prev_req, true); in mmc_blk_mq_issue_rw_rq()
2310 mq->rw_wait = false; in mmc_blk_mq_issue_rw_rq()
2323 static int mmc_blk_wait_for_idle(struct mmc_queue *mq, struct mmc_host *host) in mmc_blk_wait_for_idle() argument
2328 return mmc_blk_rw_wait(mq, NULL); in mmc_blk_wait_for_idle()
2331 enum mmc_issued mmc_blk_mq_issue_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_mq_issue_rq() argument
2333 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_mq_issue_rq()
2342 switch (mmc_issue_type(mq, req)) { in mmc_blk_mq_issue_rq()
2344 ret = mmc_blk_wait_for_idle(mq, host); in mmc_blk_mq_issue_rq()
2350 mmc_blk_issue_drv_op(mq, req); in mmc_blk_mq_issue_rq()
2353 mmc_blk_issue_discard_rq(mq, req); in mmc_blk_mq_issue_rq()
2356 mmc_blk_issue_secdiscard_rq(mq, req); in mmc_blk_mq_issue_rq()
2359 mmc_blk_issue_trim_rq(mq, req); in mmc_blk_mq_issue_rq()
2362 mmc_blk_issue_flush(mq, req); in mmc_blk_mq_issue_rq()
2377 ret = mmc_blk_cqe_issue_flush(mq, req); in mmc_blk_mq_issue_rq()
2382 ret = mmc_blk_cqe_issue_rw_rq(mq, req); in mmc_blk_mq_issue_rq()
2384 ret = mmc_blk_mq_issue_rw_rq(mq, req); in mmc_blk_mq_issue_rq()
2806 struct mmc_queue *mq = &md->queue; in mmc_dbg_card_status_get() local
2811 req = blk_mq_alloc_request(mq->queue, REQ_OP_DRV_IN, 0); in mmc_dbg_card_status_get()
2835 struct mmc_queue *mq = &md->queue; in mmc_ext_csd_open() local
2847 req = blk_mq_alloc_request(mq->queue, REQ_OP_DRV_IN, 0); in mmc_ext_csd_open()