/Linux-v4.19/drivers/lightnvm/ |
D | pblk-read.c | 41 static void pblk_read_ppalist_rq(struct pblk *pblk, struct nvm_rq *rqd, in pblk_read_ppalist_rq() argument 45 struct pblk_sec_meta *meta_list = rqd->meta_list; in pblk_read_ppalist_rq() 47 int nr_secs = rqd->nr_ppas; in pblk_read_ppalist_rq() 87 rqd->ppa_list[j++] = p; in pblk_read_ppalist_rq() 96 rqd->flags = pblk_set_read_mode(pblk, PBLK_READ_SEQUENTIAL); in pblk_read_ppalist_rq() 98 rqd->flags = pblk_set_read_mode(pblk, PBLK_READ_RANDOM); in pblk_read_ppalist_rq() 106 static void pblk_read_check_seq(struct pblk *pblk, struct nvm_rq *rqd, in pblk_read_check_seq() argument 109 struct pblk_sec_meta *meta_lba_list = rqd->meta_list; in pblk_read_check_seq() 110 int nr_lbas = rqd->nr_ppas; in pblk_read_check_seq() 123 p = (nr_lbas == 1) ? &rqd->ppa_list[i] : &rqd->ppa_addr; in pblk_read_check_seq() [all …]
|
D | pblk-write.c | 20 static unsigned long pblk_end_w_bio(struct pblk *pblk, struct nvm_rq *rqd, in pblk_end_w_bio() argument 51 pblk_bio_free_pages(pblk, rqd->bio, c_ctx->nr_valid, in pblk_end_w_bio() 55 atomic_long_add(rqd->nr_ppas, &pblk->sync_writes); in pblk_end_w_bio() 60 bio_put(rqd->bio); in pblk_end_w_bio() 61 pblk_free_rqd(pblk, rqd, PBLK_WRITE); in pblk_end_w_bio() 67 struct nvm_rq *rqd, in pblk_end_queued_w_bio() argument 71 return pblk_end_w_bio(pblk, rqd, c_ctx); in pblk_end_queued_w_bio() 74 static void pblk_complete_write(struct pblk *pblk, struct nvm_rq *rqd, in pblk_complete_write() argument 85 pblk_up_rq(pblk, rqd->ppa_list, rqd->nr_ppas, c_ctx->lun_bitmap); in pblk_complete_write() 89 pos = pblk_end_w_bio(pblk, rqd, c_ctx); in pblk_complete_write() [all …]
|
D | pblk-recovery.c | 102 struct nvm_rq *rqd; member 115 struct nvm_rq *rqd; in pblk_recov_read_oob() local 127 rqd = p.rqd; in pblk_recov_read_oob() 139 memset(rqd, 0, pblk_g_rq_size); in pblk_recov_read_oob() 153 rqd->bio = bio; in pblk_recov_read_oob() 154 rqd->opcode = NVM_OP_PREAD; in pblk_recov_read_oob() 155 rqd->meta_list = meta_list; in pblk_recov_read_oob() 156 rqd->nr_ppas = rq_ppas; in pblk_recov_read_oob() 157 rqd->ppa_list = ppa_list; in pblk_recov_read_oob() 158 rqd->dma_ppa_list = dma_ppa_list; in pblk_recov_read_oob() [all …]
|
D | pblk-map.c | 86 void pblk_map_rq(struct pblk *pblk, struct nvm_rq *rqd, unsigned int sentry, in pblk_map_rq() argument 90 struct pblk_sec_meta *meta_list = rqd->meta_list; in pblk_map_rq() 95 for (i = off; i < rqd->nr_ppas; i += min) { in pblk_map_rq() 97 if (pblk_map_page_data(pblk, sentry + i, &rqd->ppa_list[i], in pblk_map_rq() 99 bio_put(rqd->bio); in pblk_map_rq() 100 pblk_free_rqd(pblk, rqd, PBLK_WRITE); in pblk_map_rq() 107 void pblk_map_erase_rq(struct pblk *pblk, struct nvm_rq *rqd, in pblk_map_erase_rq() argument 114 struct pblk_sec_meta *meta_list = rqd->meta_list; in pblk_map_erase_rq() 120 for (i = 0; i < rqd->nr_ppas; i += min) { in pblk_map_erase_rq() 122 if (pblk_map_page_data(pblk, sentry + i, &rqd->ppa_list[i], in pblk_map_erase_rq() [all …]
|
D | pblk-core.c | 75 static void __pblk_end_io_erase(struct pblk *pblk, struct nvm_rq *rqd) in __pblk_end_io_erase() argument 83 line = &pblk->lines[pblk_ppa_to_line(rqd->ppa_addr)]; in __pblk_end_io_erase() 84 pos = pblk_ppa_to_pos(geo, rqd->ppa_addr); in __pblk_end_io_erase() 89 if (rqd->error) { in __pblk_end_io_erase() 91 pblk_mark_bb(pblk, line, rqd->ppa_addr); in __pblk_end_io_erase() 100 static void pblk_end_io_erase(struct nvm_rq *rqd) in pblk_end_io_erase() argument 102 struct pblk *pblk = rqd->private; in pblk_end_io_erase() 104 __pblk_end_io_erase(pblk, rqd); in pblk_end_io_erase() 105 mempool_free(rqd, &pblk->e_rq_pool); in pblk_end_io_erase() 234 struct nvm_rq *rqd; in pblk_alloc_rqd() local [all …]
|
D | core.c | 599 static void nvm_rq_tgt_to_dev(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd) in nvm_rq_tgt_to_dev() argument 601 if (rqd->nr_ppas == 1) { in nvm_rq_tgt_to_dev() 602 nvm_ppa_tgt_to_dev(tgt_dev, &rqd->ppa_addr, 1); in nvm_rq_tgt_to_dev() 606 nvm_ppa_tgt_to_dev(tgt_dev, rqd->ppa_list, rqd->nr_ppas); in nvm_rq_tgt_to_dev() 609 static void nvm_rq_dev_to_tgt(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd) in nvm_rq_dev_to_tgt() argument 611 if (rqd->nr_ppas == 1) { in nvm_rq_dev_to_tgt() 612 nvm_ppa_dev_to_tgt(tgt_dev, &rqd->ppa_addr, 1); in nvm_rq_dev_to_tgt() 616 nvm_ppa_dev_to_tgt(tgt_dev, rqd->ppa_list, rqd->nr_ppas); in nvm_rq_dev_to_tgt() 670 static int nvm_set_rqd_ppalist(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd, in nvm_set_rqd_ppalist() argument 679 rqd->nr_ppas = nr_ppas; in nvm_set_rqd_ppalist() [all …]
|
D | pblk.h | 142 struct nvm_rq *rqd; member 745 unsigned int pblk_rb_read_to_bio(struct pblk_rb *rb, struct nvm_rq *rqd, 772 void pblk_free_rqd(struct pblk *pblk, struct nvm_rq *rqd, int type); 774 int pblk_setup_w_rec_rq(struct pblk *pblk, struct nvm_rq *rqd, 781 void pblk_log_write_err(struct pblk *pblk, struct nvm_rq *rqd); 782 void pblk_log_read_err(struct pblk *pblk, struct nvm_rq *rqd); 783 int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd); 784 int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd); 857 void pblk_map_erase_rq(struct pblk *pblk, struct nvm_rq *rqd, 860 void pblk_map_rq(struct pblk *pblk, struct nvm_rq *rqd, unsigned int sentry, [all …]
|
D | pblk-rb.c | 512 unsigned int pblk_rb_read_to_bio(struct pblk_rb *rb, struct nvm_rq *rqd, in pblk_rb_read_to_bio() argument 518 struct pblk_c_ctx *c_ctx = nvm_rq_to_pdu(rqd); in pblk_rb_read_to_bio() 519 struct bio *bio = rqd->bio; in pblk_rb_read_to_bio()
|
/Linux-v4.19/block/ |
D | blk-rq-qos.c | 104 bool rq_depth_calc_max_depth(struct rq_depth *rqd) in rq_depth_calc_max_depth() argument 116 if (rqd->queue_depth == 1) { in rq_depth_calc_max_depth() 117 if (rqd->scale_step > 0) in rq_depth_calc_max_depth() 118 rqd->max_depth = 1; in rq_depth_calc_max_depth() 120 rqd->max_depth = 2; in rq_depth_calc_max_depth() 131 depth = min_t(unsigned int, rqd->default_depth, in rq_depth_calc_max_depth() 132 rqd->queue_depth); in rq_depth_calc_max_depth() 133 if (rqd->scale_step > 0) in rq_depth_calc_max_depth() 134 depth = 1 + ((depth - 1) >> min(31, rqd->scale_step)); in rq_depth_calc_max_depth() 135 else if (rqd->scale_step < 0) { in rq_depth_calc_max_depth() [all …]
|
D | blk-wbt.c | 236 struct rq_depth *rqd = &rwb->rq_depth; in latency_exceeded() local 280 if (rqd->scale_step) in latency_exceeded() 289 struct rq_depth *rqd = &rwb->rq_depth; in rwb_trace_step() local 291 trace_wbt_step(bdi, msg, rqd->scale_step, rwb->cur_win_nsec, in rwb_trace_step() 292 rwb->wb_background, rwb->wb_normal, rqd->max_depth); in rwb_trace_step() 327 struct rq_depth *rqd = &rwb->rq_depth; in rwb_arm_timer() local 329 if (rqd->scale_step > 0) { in rwb_arm_timer() 337 int_sqrt((rqd->scale_step + 1) << 8)); in rwb_arm_timer() 352 struct rq_depth *rqd = &rwb->rq_depth; in wb_timer_fn() local 358 trace_wbt_timer(rwb->rqos.q->backing_dev_info, status, rqd->scale_step, in wb_timer_fn() [all …]
|
D | blk-rq-qos.h | 97 void rq_depth_scale_up(struct rq_depth *rqd); 98 void rq_depth_scale_down(struct rq_depth *rqd, bool hard_throttle); 99 bool rq_depth_calc_max_depth(struct rq_depth *rqd);
|
/Linux-v4.19/drivers/nvme/host/ |
D | lightnvm.c | 628 static inline void nvme_nvm_rqtocmd(struct nvm_rq *rqd, struct nvme_ns *ns, in nvme_nvm_rqtocmd() argument 631 c->ph_rw.opcode = rqd->opcode; in nvme_nvm_rqtocmd() 633 c->ph_rw.spba = cpu_to_le64(rqd->ppa_addr.ppa); in nvme_nvm_rqtocmd() 634 c->ph_rw.metadata = cpu_to_le64(rqd->dma_meta_list); in nvme_nvm_rqtocmd() 635 c->ph_rw.control = cpu_to_le16(rqd->flags); in nvme_nvm_rqtocmd() 636 c->ph_rw.length = cpu_to_le16(rqd->nr_ppas - 1); in nvme_nvm_rqtocmd() 641 struct nvm_rq *rqd = rq->end_io_data; in nvme_nvm_end_io() local 643 rqd->ppa_status = le64_to_cpu(nvme_req(rq)->result.u64); in nvme_nvm_end_io() 644 rqd->error = nvme_req(rq)->status; in nvme_nvm_end_io() 645 nvm_end_io(rqd); in nvme_nvm_end_io() [all …]
|
/Linux-v4.19/drivers/dma/ |
D | pl330.c | 537 struct list_head rqd; member 1692 list_add_tail(&descdone->rqd, &pl330->req_done); in pl330_update() 1699 struct dma_pl330_desc, rqd); in pl330_update() 1700 list_del(&descdone->rqd); in pl330_update()
|
/Linux-v4.19/arch/m68k/ifpsp060/src/ |
D | ilsp.S | 299 swap %d5 # same as r*b if previous step rqd
|