Lines Matching refs:mq

179 			       struct mmc_queue *mq);
247 struct mmc_queue *mq; in power_ro_lock_store() local
258 mq = &md->queue; in power_ro_lock_store()
261 req = blk_get_request(mq->queue, REQ_OP_DRV_OUT, 0); in power_ro_lock_store()
628 struct mmc_queue *mq; in mmc_blk_ioctl_cmd() local
648 mq = &md->queue; in mmc_blk_ioctl_cmd()
649 req = blk_get_request(mq->queue, in mmc_blk_ioctl_cmd()
678 struct mmc_queue *mq; in mmc_blk_ioctl_multi_cmd() local
718 mq = &md->queue; in mmc_blk_ioctl_multi_cmd()
719 req = blk_get_request(mq->queue, in mmc_blk_ioctl_multi_cmd()
1026 static void mmc_blk_issue_drv_op(struct mmc_queue *mq, struct request *req) in mmc_blk_issue_drv_op() argument
1029 struct mmc_card *card = mq->card; in mmc_blk_issue_drv_op()
1030 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_issue_drv_op()
1093 static void mmc_blk_issue_discard_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_issue_discard_rq() argument
1095 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_issue_discard_rq()
1130 static void mmc_blk_issue_secdiscard_rq(struct mmc_queue *mq, in mmc_blk_issue_secdiscard_rq() argument
1133 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_issue_secdiscard_rq()
1200 static void mmc_blk_issue_flush(struct mmc_queue *mq, struct request *req) in mmc_blk_issue_flush() argument
1202 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_issue_flush()
1287 static void mmc_blk_data_prep(struct mmc_queue *mq, struct mmc_queue_req *mqrq, in mmc_blk_data_prep() argument
1291 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_data_prep()
1393 brq->data.sg_len = mmc_queue_map_sg(mq, mqrq); in mmc_blk_data_prep()
1423 static void mmc_blk_cqe_complete_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_cqe_complete_rq() argument
1428 struct mmc_host *host = mq->card->host; in mmc_blk_cqe_complete_rq()
1429 enum mmc_issue_type issue_type = mmc_issue_type(mq, req); in mmc_blk_cqe_complete_rq()
1457 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_cqe_complete_rq()
1459 mq->in_flight[issue_type] -= 1; in mmc_blk_cqe_complete_rq()
1461 put_card = (mmc_tot_in_flight(mq) == 0); in mmc_blk_cqe_complete_rq()
1463 mmc_cqe_check_busy(mq); in mmc_blk_cqe_complete_rq()
1465 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_cqe_complete_rq()
1467 if (!mq->cqe_busy) in mmc_blk_cqe_complete_rq()
1471 mmc_put_card(mq->card, &mq->ctx); in mmc_blk_cqe_complete_rq()
1474 void mmc_blk_cqe_recovery(struct mmc_queue *mq) in mmc_blk_cqe_recovery() argument
1476 struct mmc_card *card = mq->card; in mmc_blk_cqe_recovery()
1484 mmc_blk_reset(mq->blkdata, host, MMC_BLK_CQE_RECOVERY); in mmc_blk_cqe_recovery()
1486 mmc_blk_reset_success(mq->blkdata, MMC_BLK_CQE_RECOVERY); in mmc_blk_cqe_recovery()
1497 struct mmc_queue *mq = q->queuedata; in mmc_blk_cqe_req_done() local
1503 if (mq->in_recovery) in mmc_blk_cqe_req_done()
1504 mmc_blk_cqe_complete_rq(mq, req); in mmc_blk_cqe_req_done()
1530 static int mmc_blk_cqe_issue_flush(struct mmc_queue *mq, struct request *req) in mmc_blk_cqe_issue_flush() argument
1542 return mmc_blk_cqe_start_req(mq->card->host, mrq); in mmc_blk_cqe_issue_flush()
1545 static int mmc_blk_hsq_issue_rw_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_hsq_issue_rw_rq() argument
1548 struct mmc_host *host = mq->card->host; in mmc_blk_hsq_issue_rw_rq()
1551 mmc_blk_rw_rq_prep(mqrq, mq->card, 0, mq); in mmc_blk_hsq_issue_rw_rq()
1562 static int mmc_blk_cqe_issue_rw_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_cqe_issue_rw_rq() argument
1565 struct mmc_host *host = mq->card->host; in mmc_blk_cqe_issue_rw_rq()
1568 return mmc_blk_hsq_issue_rw_rq(mq, req); in mmc_blk_cqe_issue_rw_rq()
1570 mmc_blk_data_prep(mq, mqrq, 0, NULL, NULL); in mmc_blk_cqe_issue_rw_rq()
1572 return mmc_blk_cqe_start_req(mq->card->host, &mqrq->brq.mrq); in mmc_blk_cqe_issue_rw_rq()
1578 struct mmc_queue *mq) in mmc_blk_rw_rq_prep() argument
1583 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_rw_rq_prep()
1586 mmc_blk_data_prep(mq, mqrq, disable_multi, &do_rel_wr, &do_data_tag); in mmc_blk_rw_rq_prep()
1678 static void mmc_blk_read_single(struct mmc_queue *mq, struct request *req) in mmc_blk_read_single() argument
1682 struct mmc_card *card = mq->card; in mmc_blk_read_single()
1691 mmc_blk_rw_rq_prep(mqrq, card, 1, mq); in mmc_blk_read_single()
1749 struct mmc_queue *mq = req->q->queuedata; in mmc_blk_status_error() local
1752 if (mmc_host_is_spi(mq->card->host)) in mmc_blk_status_error()
1785 static void mmc_blk_mq_rw_recovery(struct mmc_queue *mq, struct request *req) in mmc_blk_mq_rw_recovery() argument
1790 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_mq_rw_recovery()
1791 struct mmc_card *card = mq->card; in mmc_blk_mq_rw_recovery()
1821 if (!mmc_host_is_spi(mq->card->host) && in mmc_blk_mq_rw_recovery()
1823 err = mmc_blk_fix_state(mq->card, req); in mmc_blk_mq_rw_recovery()
1838 if (!mmc_host_is_spi(mq->card->host) && in mmc_blk_mq_rw_recovery()
1870 mmc_blk_read_single(mq, req); in mmc_blk_mq_rw_recovery()
1930 static inline void mmc_blk_rw_reset_success(struct mmc_queue *mq, in mmc_blk_rw_reset_success() argument
1935 mmc_blk_reset_success(mq->blkdata, type); in mmc_blk_rw_reset_success()
1938 static void mmc_blk_mq_complete_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_mq_complete_rq() argument
1953 if (mmc_card_removed(mq->card)) in mmc_blk_mq_complete_rq()
1959 static bool mmc_blk_urgent_bkops_needed(struct mmc_queue *mq, in mmc_blk_urgent_bkops_needed() argument
1962 return mmc_card_mmc(mq->card) && !mmc_host_is_spi(mq->card->host) && in mmc_blk_urgent_bkops_needed()
1967 static void mmc_blk_urgent_bkops(struct mmc_queue *mq, in mmc_blk_urgent_bkops() argument
1970 if (mmc_blk_urgent_bkops_needed(mq, mqrq)) in mmc_blk_urgent_bkops()
1971 mmc_run_bkops(mq->card); in mmc_blk_urgent_bkops()
1980 struct mmc_queue *mq = q->queuedata; in mmc_blk_hsq_req_done() local
1981 struct mmc_host *host = mq->card->host; in mmc_blk_hsq_req_done()
1985 mmc_blk_urgent_bkops_needed(mq, mqrq)) { in mmc_blk_hsq_req_done()
1986 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_hsq_req_done()
1987 mq->recovery_needed = true; in mmc_blk_hsq_req_done()
1988 mq->recovery_req = req; in mmc_blk_hsq_req_done()
1989 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_hsq_req_done()
1993 schedule_work(&mq->recovery_work); in mmc_blk_hsq_req_done()
1997 mmc_blk_rw_reset_success(mq, req); in mmc_blk_hsq_req_done()
2003 if (mq->in_recovery) in mmc_blk_hsq_req_done()
2004 mmc_blk_cqe_complete_rq(mq, req); in mmc_blk_hsq_req_done()
2011 struct mmc_queue *mq = req->q->queuedata; in mmc_blk_mq_complete() local
2012 struct mmc_host *host = mq->card->host; in mmc_blk_mq_complete()
2015 mmc_blk_cqe_complete_rq(mq, req); in mmc_blk_mq_complete()
2017 mmc_blk_mq_complete_rq(mq, req); in mmc_blk_mq_complete()
2020 static void mmc_blk_mq_poll_completion(struct mmc_queue *mq, in mmc_blk_mq_poll_completion() argument
2024 struct mmc_host *host = mq->card->host; in mmc_blk_mq_poll_completion()
2027 mmc_blk_card_busy(mq->card, req)) { in mmc_blk_mq_poll_completion()
2028 mmc_blk_mq_rw_recovery(mq, req); in mmc_blk_mq_poll_completion()
2030 mmc_blk_rw_reset_success(mq, req); in mmc_blk_mq_poll_completion()
2034 mmc_blk_urgent_bkops(mq, mqrq); in mmc_blk_mq_poll_completion()
2037 static void mmc_blk_mq_dec_in_flight(struct mmc_queue *mq, struct request *req) in mmc_blk_mq_dec_in_flight() argument
2042 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_mq_dec_in_flight()
2044 mq->in_flight[mmc_issue_type(mq, req)] -= 1; in mmc_blk_mq_dec_in_flight()
2046 put_card = (mmc_tot_in_flight(mq) == 0); in mmc_blk_mq_dec_in_flight()
2048 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_mq_dec_in_flight()
2051 mmc_put_card(mq->card, &mq->ctx); in mmc_blk_mq_dec_in_flight()
2054 static void mmc_blk_mq_post_req(struct mmc_queue *mq, struct request *req) in mmc_blk_mq_post_req() argument
2058 struct mmc_host *host = mq->card->host; in mmc_blk_mq_post_req()
2066 if (mq->in_recovery) in mmc_blk_mq_post_req()
2067 mmc_blk_mq_complete_rq(mq, req); in mmc_blk_mq_post_req()
2071 mmc_blk_mq_dec_in_flight(mq, req); in mmc_blk_mq_post_req()
2074 void mmc_blk_mq_recovery(struct mmc_queue *mq) in mmc_blk_mq_recovery() argument
2076 struct request *req = mq->recovery_req; in mmc_blk_mq_recovery()
2077 struct mmc_host *host = mq->card->host; in mmc_blk_mq_recovery()
2080 mq->recovery_req = NULL; in mmc_blk_mq_recovery()
2081 mq->rw_wait = false; in mmc_blk_mq_recovery()
2085 mmc_blk_mq_rw_recovery(mq, req); in mmc_blk_mq_recovery()
2088 mmc_blk_urgent_bkops(mq, mqrq); in mmc_blk_mq_recovery()
2090 mmc_blk_mq_post_req(mq, req); in mmc_blk_mq_recovery()
2093 static void mmc_blk_mq_complete_prev_req(struct mmc_queue *mq, in mmc_blk_mq_complete_prev_req() argument
2096 if (mmc_host_done_complete(mq->card->host)) in mmc_blk_mq_complete_prev_req()
2099 mutex_lock(&mq->complete_lock); in mmc_blk_mq_complete_prev_req()
2101 if (!mq->complete_req) in mmc_blk_mq_complete_prev_req()
2104 mmc_blk_mq_poll_completion(mq, mq->complete_req); in mmc_blk_mq_complete_prev_req()
2107 *prev_req = mq->complete_req; in mmc_blk_mq_complete_prev_req()
2109 mmc_blk_mq_post_req(mq, mq->complete_req); in mmc_blk_mq_complete_prev_req()
2111 mq->complete_req = NULL; in mmc_blk_mq_complete_prev_req()
2114 mutex_unlock(&mq->complete_lock); in mmc_blk_mq_complete_prev_req()
2119 struct mmc_queue *mq = container_of(work, struct mmc_queue, in mmc_blk_mq_complete_work() local
2122 mmc_blk_mq_complete_prev_req(mq, NULL); in mmc_blk_mq_complete_work()
2131 struct mmc_queue *mq = q->queuedata; in mmc_blk_mq_req_done() local
2132 struct mmc_host *host = mq->card->host; in mmc_blk_mq_req_done()
2144 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_mq_req_done()
2145 mq->complete_req = req; in mmc_blk_mq_req_done()
2146 mq->rw_wait = false; in mmc_blk_mq_req_done()
2147 waiting = mq->waiting; in mmc_blk_mq_req_done()
2148 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_mq_req_done()
2157 wake_up(&mq->wait); in mmc_blk_mq_req_done()
2159 queue_work(mq->card->complete_wq, &mq->complete_work); in mmc_blk_mq_req_done()
2166 mmc_blk_urgent_bkops_needed(mq, mqrq)) { in mmc_blk_mq_req_done()
2167 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_mq_req_done()
2168 mq->recovery_needed = true; in mmc_blk_mq_req_done()
2169 mq->recovery_req = req; in mmc_blk_mq_req_done()
2170 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_mq_req_done()
2171 wake_up(&mq->wait); in mmc_blk_mq_req_done()
2172 schedule_work(&mq->recovery_work); in mmc_blk_mq_req_done()
2176 mmc_blk_rw_reset_success(mq, req); in mmc_blk_mq_req_done()
2178 mq->rw_wait = false; in mmc_blk_mq_req_done()
2179 wake_up(&mq->wait); in mmc_blk_mq_req_done()
2181 mmc_blk_mq_post_req(mq, req); in mmc_blk_mq_req_done()
2184 static bool mmc_blk_rw_wait_cond(struct mmc_queue *mq, int *err) in mmc_blk_rw_wait_cond() argument
2193 spin_lock_irqsave(&mq->lock, flags); in mmc_blk_rw_wait_cond()
2194 if (mq->recovery_needed) { in mmc_blk_rw_wait_cond()
2198 done = !mq->rw_wait; in mmc_blk_rw_wait_cond()
2200 mq->waiting = !done; in mmc_blk_rw_wait_cond()
2201 spin_unlock_irqrestore(&mq->lock, flags); in mmc_blk_rw_wait_cond()
2206 static int mmc_blk_rw_wait(struct mmc_queue *mq, struct request **prev_req) in mmc_blk_rw_wait() argument
2210 wait_event(mq->wait, mmc_blk_rw_wait_cond(mq, &err)); in mmc_blk_rw_wait()
2213 mmc_blk_mq_complete_prev_req(mq, prev_req); in mmc_blk_rw_wait()
2218 static int mmc_blk_mq_issue_rw_rq(struct mmc_queue *mq, in mmc_blk_mq_issue_rw_rq() argument
2222 struct mmc_host *host = mq->card->host; in mmc_blk_mq_issue_rw_rq()
2226 mmc_blk_rw_rq_prep(mqrq, mq->card, 0, mq); in mmc_blk_mq_issue_rw_rq()
2232 err = mmc_blk_rw_wait(mq, &prev_req); in mmc_blk_mq_issue_rw_rq()
2236 mq->rw_wait = true; in mmc_blk_mq_issue_rw_rq()
2241 mmc_blk_mq_post_req(mq, prev_req); in mmc_blk_mq_issue_rw_rq()
2244 mq->rw_wait = false; in mmc_blk_mq_issue_rw_rq()
2257 static int mmc_blk_wait_for_idle(struct mmc_queue *mq, struct mmc_host *host) in mmc_blk_wait_for_idle() argument
2262 return mmc_blk_rw_wait(mq, NULL); in mmc_blk_wait_for_idle()
2265 enum mmc_issued mmc_blk_mq_issue_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_mq_issue_rq() argument
2267 struct mmc_blk_data *md = mq->blkdata; in mmc_blk_mq_issue_rq()
2276 switch (mmc_issue_type(mq, req)) { in mmc_blk_mq_issue_rq()
2278 ret = mmc_blk_wait_for_idle(mq, host); in mmc_blk_mq_issue_rq()
2284 mmc_blk_issue_drv_op(mq, req); in mmc_blk_mq_issue_rq()
2287 mmc_blk_issue_discard_rq(mq, req); in mmc_blk_mq_issue_rq()
2290 mmc_blk_issue_secdiscard_rq(mq, req); in mmc_blk_mq_issue_rq()
2293 mmc_blk_issue_flush(mq, req); in mmc_blk_mq_issue_rq()
2308 ret = mmc_blk_cqe_issue_flush(mq, req); in mmc_blk_mq_issue_rq()
2313 ret = mmc_blk_cqe_issue_rw_rq(mq, req); in mmc_blk_mq_issue_rq()
2315 ret = mmc_blk_mq_issue_rw_rq(mq, req); in mmc_blk_mq_issue_rq()
2728 struct mmc_queue *mq = &md->queue; in mmc_dbg_card_status_get() local
2733 req = blk_get_request(mq->queue, REQ_OP_DRV_IN, 0); in mmc_dbg_card_status_get()
2757 struct mmc_queue *mq = &md->queue; in mmc_ext_csd_open() local
2769 req = blk_get_request(mq->queue, REQ_OP_DRV_IN, 0); in mmc_ext_csd_open()