/Linux-v5.4/drivers/lightnvm/ |
D | pblk-read.c | 40 static int pblk_read_ppalist_rq(struct pblk *pblk, struct nvm_rq *rqd, in pblk_read_ppalist_rq() argument 44 void *meta_list = rqd->meta_list; in pblk_read_ppalist_rq() 48 nr_secs = pblk_lookup_l2p_seq(pblk, rqd->ppa_list, blba, rqd->nr_ppas, in pblk_read_ppalist_rq() 58 if (pblk_ppa_empty(rqd->ppa_list[i])) { in pblk_read_ppalist_rq() 62 } else if (pblk_addr_in_cache(rqd->ppa_list[i])) { in pblk_read_ppalist_rq() 69 rqd->ppa_list[i])) { in pblk_read_ppalist_rq() 99 rqd->is_seq = 1; in pblk_read_ppalist_rq() 109 static void pblk_read_check_seq(struct pblk *pblk, struct nvm_rq *rqd, in pblk_read_check_seq() argument 112 void *meta_list = rqd->meta_list; in pblk_read_check_seq() 113 int nr_lbas = rqd->nr_ppas; in pblk_read_check_seq() [all …]
|
D | pblk-write.c | 22 static unsigned long pblk_end_w_bio(struct pblk *pblk, struct nvm_rq *rqd, in pblk_end_w_bio() argument 53 pblk_bio_free_pages(pblk, rqd->bio, c_ctx->nr_valid, in pblk_end_w_bio() 57 atomic_long_add(rqd->nr_ppas, &pblk->sync_writes); in pblk_end_w_bio() 62 bio_put(rqd->bio); in pblk_end_w_bio() 63 pblk_free_rqd(pblk, rqd, PBLK_WRITE); in pblk_end_w_bio() 69 struct nvm_rq *rqd, in pblk_end_queued_w_bio() argument 73 return pblk_end_w_bio(pblk, rqd, c_ctx); in pblk_end_queued_w_bio() 76 static void pblk_complete_write(struct pblk *pblk, struct nvm_rq *rqd, in pblk_complete_write() argument 90 pos = pblk_end_w_bio(pblk, rqd, c_ctx); in pblk_complete_write() 94 rqd = nvm_rq_from_c_ctx(c); in pblk_complete_write() [all …]
|
D | pblk-core.c | 79 static void __pblk_end_io_erase(struct pblk *pblk, struct nvm_rq *rqd) in __pblk_end_io_erase() argument 87 line = pblk_ppa_to_line(pblk, rqd->ppa_addr); in __pblk_end_io_erase() 88 pos = pblk_ppa_to_pos(geo, rqd->ppa_addr); in __pblk_end_io_erase() 93 if (rqd->error) { in __pblk_end_io_erase() 95 &rqd->ppa_addr, PBLK_CHUNK_RESET_FAILED); in __pblk_end_io_erase() 98 pblk_mark_bb(pblk, line, rqd->ppa_addr); in __pblk_end_io_erase() 101 &rqd->ppa_addr, PBLK_CHUNK_RESET_DONE); in __pblk_end_io_erase() 106 trace_pblk_chunk_state(pblk_disk_name(pblk), &rqd->ppa_addr, in __pblk_end_io_erase() 113 static void pblk_end_io_erase(struct nvm_rq *rqd) in pblk_end_io_erase() argument 115 struct pblk *pblk = rqd->private; in pblk_end_io_erase() [all …]
|
D | core.c | 611 static void nvm_rq_tgt_to_dev(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd) in nvm_rq_tgt_to_dev() argument 613 struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); in nvm_rq_tgt_to_dev() 615 nvm_ppa_tgt_to_dev(tgt_dev, ppa_list, rqd->nr_ppas); in nvm_rq_tgt_to_dev() 618 static void nvm_rq_dev_to_tgt(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd) in nvm_rq_dev_to_tgt() argument 620 struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); in nvm_rq_dev_to_tgt() 622 nvm_ppa_dev_to_tgt(tgt_dev, ppa_list, rqd->nr_ppas); in nvm_rq_dev_to_tgt() 676 static int nvm_set_rqd_ppalist(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd, in nvm_set_rqd_ppalist() argument 685 rqd->nr_ppas = nr_ppas; in nvm_set_rqd_ppalist() 686 rqd->ppa_addr = ppas[0]; in nvm_set_rqd_ppalist() 691 rqd->nr_ppas = nr_ppas; in nvm_set_rqd_ppalist() [all …]
|
D | pblk-recovery.c | 145 struct nvm_rq *rqd; member 158 static void pblk_end_io_recov(struct nvm_rq *rqd) in pblk_end_io_recov() argument 160 struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); in pblk_end_io_recov() 161 struct pblk_pad_rq *pad_rq = rqd->private; in pblk_end_io_recov() 166 pblk_free_rqd(pblk, rqd, PBLK_WRITE_INT); in pblk_end_io_recov() 180 struct nvm_rq *rqd; in pblk_recov_pad_line() local 214 rqd = pblk_alloc_rqd(pblk, PBLK_WRITE_INT); in pblk_recov_pad_line() 216 ret = pblk_alloc_rqd_meta(pblk, rqd); in pblk_recov_pad_line() 218 pblk_free_rqd(pblk, rqd, PBLK_WRITE_INT); in pblk_recov_pad_line() 222 rqd->bio = NULL; in pblk_recov_pad_line() [all …]
|
D | pblk-map.c | 96 int pblk_map_rq(struct pblk *pblk, struct nvm_rq *rqd, unsigned int sentry, in pblk_map_rq() argument 100 void *meta_list = pblk_get_meta_for_writes(pblk, rqd); in pblk_map_rq() 102 struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); in pblk_map_rq() 108 for (i = off; i < rqd->nr_ppas; i += min) { in pblk_map_rq() 122 int pblk_map_erase_rq(struct pblk *pblk, struct nvm_rq *rqd, in pblk_map_erase_rq() argument 129 void *meta_list = pblk_get_meta_for_writes(pblk, rqd); in pblk_map_erase_rq() 131 struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); in pblk_map_erase_rq() 139 for (i = 0; i < rqd->nr_ppas; i += min) { in pblk_map_erase_rq() 155 return pblk_map_rq(pblk, rqd, sentry, lun_bitmap, in pblk_map_erase_rq() 170 return pblk_map_rq(pblk, rqd, sentry, lun_bitmap, in pblk_map_erase_rq()
|
D | pblk.h | 134 struct nvm_rq *rqd; member 738 unsigned int pblk_rb_read_to_bio(struct pblk_rb *rb, struct nvm_rq *rqd, 765 void pblk_free_rqd(struct pblk *pblk, struct nvm_rq *rqd, int type); 766 int pblk_alloc_rqd_meta(struct pblk *pblk, struct nvm_rq *rqd); 767 void pblk_free_rqd_meta(struct pblk *pblk, struct nvm_rq *rqd); 769 int pblk_setup_w_rec_rq(struct pblk *pblk, struct nvm_rq *rqd, 776 void pblk_log_write_err(struct pblk *pblk, struct nvm_rq *rqd); 777 void pblk_log_read_err(struct pblk *pblk, struct nvm_rq *rqd); 778 int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd, void *buf); 779 int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd, void *buf); [all …]
|
D | pblk-rb.c | 550 unsigned int pblk_rb_read_to_bio(struct pblk_rb *rb, struct nvm_rq *rqd, in pblk_rb_read_to_bio() argument 556 struct pblk_c_ctx *c_ctx = nvm_rq_to_pdu(rqd); in pblk_rb_read_to_bio() 557 struct bio *bio = rqd->bio; in pblk_rb_read_to_bio()
|
/Linux-v5.4/block/ |
D | blk-rq-qos.c | 116 bool rq_depth_calc_max_depth(struct rq_depth *rqd) in rq_depth_calc_max_depth() argument 128 if (rqd->queue_depth == 1) { in rq_depth_calc_max_depth() 129 if (rqd->scale_step > 0) in rq_depth_calc_max_depth() 130 rqd->max_depth = 1; in rq_depth_calc_max_depth() 132 rqd->max_depth = 2; in rq_depth_calc_max_depth() 143 depth = min_t(unsigned int, rqd->default_depth, in rq_depth_calc_max_depth() 144 rqd->queue_depth); in rq_depth_calc_max_depth() 145 if (rqd->scale_step > 0) in rq_depth_calc_max_depth() 146 depth = 1 + ((depth - 1) >> min(31, rqd->scale_step)); in rq_depth_calc_max_depth() 147 else if (rqd->scale_step < 0) { in rq_depth_calc_max_depth() [all …]
|
D | blk-wbt.c | 237 struct rq_depth *rqd = &rwb->rq_depth; in latency_exceeded() local 281 if (rqd->scale_step) in latency_exceeded() 290 struct rq_depth *rqd = &rwb->rq_depth; in rwb_trace_step() local 292 trace_wbt_step(bdi, msg, rqd->scale_step, rwb->cur_win_nsec, in rwb_trace_step() 293 rwb->wb_background, rwb->wb_normal, rqd->max_depth); in rwb_trace_step() 330 struct rq_depth *rqd = &rwb->rq_depth; in rwb_arm_timer() local 332 if (rqd->scale_step > 0) { in rwb_arm_timer() 340 int_sqrt((rqd->scale_step + 1) << 8)); in rwb_arm_timer() 355 struct rq_depth *rqd = &rwb->rq_depth; in wb_timer_fn() local 361 trace_wbt_timer(rwb->rqos.q->backing_dev_info, status, rqd->scale_step, in wb_timer_fn() [all …]
|
D | blk-rq-qos.h | 130 bool rq_depth_scale_up(struct rq_depth *rqd); 131 bool rq_depth_scale_down(struct rq_depth *rqd, bool hard_throttle); 132 bool rq_depth_calc_max_depth(struct rq_depth *rqd);
|
/Linux-v5.4/drivers/nvme/host/ |
D | lightnvm.c | 624 static inline void nvme_nvm_rqtocmd(struct nvm_rq *rqd, struct nvme_ns *ns, in nvme_nvm_rqtocmd() argument 627 c->ph_rw.opcode = rqd->opcode; in nvme_nvm_rqtocmd() 629 c->ph_rw.spba = cpu_to_le64(rqd->ppa_addr.ppa); in nvme_nvm_rqtocmd() 630 c->ph_rw.metadata = cpu_to_le64(rqd->dma_meta_list); in nvme_nvm_rqtocmd() 631 c->ph_rw.control = cpu_to_le16(rqd->flags); in nvme_nvm_rqtocmd() 632 c->ph_rw.length = cpu_to_le16(rqd->nr_ppas - 1); in nvme_nvm_rqtocmd() 637 struct nvm_rq *rqd = rq->end_io_data; in nvme_nvm_end_io() local 639 rqd->ppa_status = le64_to_cpu(nvme_req(rq)->result.u64); in nvme_nvm_end_io() 640 rqd->error = nvme_req(rq)->status; in nvme_nvm_end_io() 641 nvm_end_io(rqd); in nvme_nvm_end_io() [all …]
|
/Linux-v5.4/include/linux/ |
D | lightnvm.h | 321 static inline struct ppa_addr *nvm_rq_to_ppa_list(struct nvm_rq *rqd) in nvm_rq_to_ppa_list() argument 323 return (rqd->nr_ppas > 1) ? rqd->ppa_list : &rqd->ppa_addr; in nvm_rq_to_ppa_list()
|
/Linux-v5.4/drivers/dma/ |
D | pl330.c | 539 struct list_head rqd; member 1702 list_add_tail(&descdone->rqd, &pl330->req_done); in pl330_update() 1709 struct dma_pl330_desc, rqd); in pl330_update() 1710 list_del(&descdone->rqd); in pl330_update()
|
/Linux-v5.4/arch/m68k/ifpsp060/src/ |
D | ilsp.S | 299 swap %d5 # same as r*b if previous step rqd
|