Home
last modified time | relevance | path

Searched refs:rqd (Results 1 – 15 of 15) sorted by relevance

/Linux-v5.4/drivers/lightnvm/
Dpblk-read.c40 static int pblk_read_ppalist_rq(struct pblk *pblk, struct nvm_rq *rqd, in pblk_read_ppalist_rq() argument
44 void *meta_list = rqd->meta_list; in pblk_read_ppalist_rq()
48 nr_secs = pblk_lookup_l2p_seq(pblk, rqd->ppa_list, blba, rqd->nr_ppas, in pblk_read_ppalist_rq()
58 if (pblk_ppa_empty(rqd->ppa_list[i])) { in pblk_read_ppalist_rq()
62 } else if (pblk_addr_in_cache(rqd->ppa_list[i])) { in pblk_read_ppalist_rq()
69 rqd->ppa_list[i])) { in pblk_read_ppalist_rq()
99 rqd->is_seq = 1; in pblk_read_ppalist_rq()
109 static void pblk_read_check_seq(struct pblk *pblk, struct nvm_rq *rqd, in pblk_read_check_seq() argument
112 void *meta_list = rqd->meta_list; in pblk_read_check_seq()
113 int nr_lbas = rqd->nr_ppas; in pblk_read_check_seq()
[all …]
Dpblk-write.c22 static unsigned long pblk_end_w_bio(struct pblk *pblk, struct nvm_rq *rqd, in pblk_end_w_bio() argument
53 pblk_bio_free_pages(pblk, rqd->bio, c_ctx->nr_valid, in pblk_end_w_bio()
57 atomic_long_add(rqd->nr_ppas, &pblk->sync_writes); in pblk_end_w_bio()
62 bio_put(rqd->bio); in pblk_end_w_bio()
63 pblk_free_rqd(pblk, rqd, PBLK_WRITE); in pblk_end_w_bio()
69 struct nvm_rq *rqd, in pblk_end_queued_w_bio() argument
73 return pblk_end_w_bio(pblk, rqd, c_ctx); in pblk_end_queued_w_bio()
76 static void pblk_complete_write(struct pblk *pblk, struct nvm_rq *rqd, in pblk_complete_write() argument
90 pos = pblk_end_w_bio(pblk, rqd, c_ctx); in pblk_complete_write()
94 rqd = nvm_rq_from_c_ctx(c); in pblk_complete_write()
[all …]
Dpblk-core.c79 static void __pblk_end_io_erase(struct pblk *pblk, struct nvm_rq *rqd) in __pblk_end_io_erase() argument
87 line = pblk_ppa_to_line(pblk, rqd->ppa_addr); in __pblk_end_io_erase()
88 pos = pblk_ppa_to_pos(geo, rqd->ppa_addr); in __pblk_end_io_erase()
93 if (rqd->error) { in __pblk_end_io_erase()
95 &rqd->ppa_addr, PBLK_CHUNK_RESET_FAILED); in __pblk_end_io_erase()
98 pblk_mark_bb(pblk, line, rqd->ppa_addr); in __pblk_end_io_erase()
101 &rqd->ppa_addr, PBLK_CHUNK_RESET_DONE); in __pblk_end_io_erase()
106 trace_pblk_chunk_state(pblk_disk_name(pblk), &rqd->ppa_addr, in __pblk_end_io_erase()
113 static void pblk_end_io_erase(struct nvm_rq *rqd) in pblk_end_io_erase() argument
115 struct pblk *pblk = rqd->private; in pblk_end_io_erase()
[all …]
Dcore.c611 static void nvm_rq_tgt_to_dev(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd) in nvm_rq_tgt_to_dev() argument
613 struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); in nvm_rq_tgt_to_dev()
615 nvm_ppa_tgt_to_dev(tgt_dev, ppa_list, rqd->nr_ppas); in nvm_rq_tgt_to_dev()
618 static void nvm_rq_dev_to_tgt(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd) in nvm_rq_dev_to_tgt() argument
620 struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); in nvm_rq_dev_to_tgt()
622 nvm_ppa_dev_to_tgt(tgt_dev, ppa_list, rqd->nr_ppas); in nvm_rq_dev_to_tgt()
676 static int nvm_set_rqd_ppalist(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd, in nvm_set_rqd_ppalist() argument
685 rqd->nr_ppas = nr_ppas; in nvm_set_rqd_ppalist()
686 rqd->ppa_addr = ppas[0]; in nvm_set_rqd_ppalist()
691 rqd->nr_ppas = nr_ppas; in nvm_set_rqd_ppalist()
[all …]
Dpblk-recovery.c145 struct nvm_rq *rqd; member
158 static void pblk_end_io_recov(struct nvm_rq *rqd) in pblk_end_io_recov() argument
160 struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); in pblk_end_io_recov()
161 struct pblk_pad_rq *pad_rq = rqd->private; in pblk_end_io_recov()
166 pblk_free_rqd(pblk, rqd, PBLK_WRITE_INT); in pblk_end_io_recov()
180 struct nvm_rq *rqd; in pblk_recov_pad_line() local
214 rqd = pblk_alloc_rqd(pblk, PBLK_WRITE_INT); in pblk_recov_pad_line()
216 ret = pblk_alloc_rqd_meta(pblk, rqd); in pblk_recov_pad_line()
218 pblk_free_rqd(pblk, rqd, PBLK_WRITE_INT); in pblk_recov_pad_line()
222 rqd->bio = NULL; in pblk_recov_pad_line()
[all …]
Dpblk-map.c96 int pblk_map_rq(struct pblk *pblk, struct nvm_rq *rqd, unsigned int sentry, in pblk_map_rq() argument
100 void *meta_list = pblk_get_meta_for_writes(pblk, rqd); in pblk_map_rq()
102 struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); in pblk_map_rq()
108 for (i = off; i < rqd->nr_ppas; i += min) { in pblk_map_rq()
122 int pblk_map_erase_rq(struct pblk *pblk, struct nvm_rq *rqd, in pblk_map_erase_rq() argument
129 void *meta_list = pblk_get_meta_for_writes(pblk, rqd); in pblk_map_erase_rq()
131 struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); in pblk_map_erase_rq()
139 for (i = 0; i < rqd->nr_ppas; i += min) { in pblk_map_erase_rq()
155 return pblk_map_rq(pblk, rqd, sentry, lun_bitmap, in pblk_map_erase_rq()
170 return pblk_map_rq(pblk, rqd, sentry, lun_bitmap, in pblk_map_erase_rq()
Dpblk.h134 struct nvm_rq *rqd; member
738 unsigned int pblk_rb_read_to_bio(struct pblk_rb *rb, struct nvm_rq *rqd,
765 void pblk_free_rqd(struct pblk *pblk, struct nvm_rq *rqd, int type);
766 int pblk_alloc_rqd_meta(struct pblk *pblk, struct nvm_rq *rqd);
767 void pblk_free_rqd_meta(struct pblk *pblk, struct nvm_rq *rqd);
769 int pblk_setup_w_rec_rq(struct pblk *pblk, struct nvm_rq *rqd,
776 void pblk_log_write_err(struct pblk *pblk, struct nvm_rq *rqd);
777 void pblk_log_read_err(struct pblk *pblk, struct nvm_rq *rqd);
778 int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd, void *buf);
779 int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd, void *buf);
[all …]
Dpblk-rb.c550 unsigned int pblk_rb_read_to_bio(struct pblk_rb *rb, struct nvm_rq *rqd, in pblk_rb_read_to_bio() argument
556 struct pblk_c_ctx *c_ctx = nvm_rq_to_pdu(rqd); in pblk_rb_read_to_bio()
557 struct bio *bio = rqd->bio; in pblk_rb_read_to_bio()
/Linux-v5.4/block/
Dblk-rq-qos.c116 bool rq_depth_calc_max_depth(struct rq_depth *rqd) in rq_depth_calc_max_depth() argument
128 if (rqd->queue_depth == 1) { in rq_depth_calc_max_depth()
129 if (rqd->scale_step > 0) in rq_depth_calc_max_depth()
130 rqd->max_depth = 1; in rq_depth_calc_max_depth()
132 rqd->max_depth = 2; in rq_depth_calc_max_depth()
143 depth = min_t(unsigned int, rqd->default_depth, in rq_depth_calc_max_depth()
144 rqd->queue_depth); in rq_depth_calc_max_depth()
145 if (rqd->scale_step > 0) in rq_depth_calc_max_depth()
146 depth = 1 + ((depth - 1) >> min(31, rqd->scale_step)); in rq_depth_calc_max_depth()
147 else if (rqd->scale_step < 0) { in rq_depth_calc_max_depth()
[all …]
Dblk-wbt.c237 struct rq_depth *rqd = &rwb->rq_depth; in latency_exceeded() local
281 if (rqd->scale_step) in latency_exceeded()
290 struct rq_depth *rqd = &rwb->rq_depth; in rwb_trace_step() local
292 trace_wbt_step(bdi, msg, rqd->scale_step, rwb->cur_win_nsec, in rwb_trace_step()
293 rwb->wb_background, rwb->wb_normal, rqd->max_depth); in rwb_trace_step()
330 struct rq_depth *rqd = &rwb->rq_depth; in rwb_arm_timer() local
332 if (rqd->scale_step > 0) { in rwb_arm_timer()
340 int_sqrt((rqd->scale_step + 1) << 8)); in rwb_arm_timer()
355 struct rq_depth *rqd = &rwb->rq_depth; in wb_timer_fn() local
361 trace_wbt_timer(rwb->rqos.q->backing_dev_info, status, rqd->scale_step, in wb_timer_fn()
[all …]
Dblk-rq-qos.h130 bool rq_depth_scale_up(struct rq_depth *rqd);
131 bool rq_depth_scale_down(struct rq_depth *rqd, bool hard_throttle);
132 bool rq_depth_calc_max_depth(struct rq_depth *rqd);
/Linux-v5.4/drivers/nvme/host/
Dlightnvm.c624 static inline void nvme_nvm_rqtocmd(struct nvm_rq *rqd, struct nvme_ns *ns, in nvme_nvm_rqtocmd() argument
627 c->ph_rw.opcode = rqd->opcode; in nvme_nvm_rqtocmd()
629 c->ph_rw.spba = cpu_to_le64(rqd->ppa_addr.ppa); in nvme_nvm_rqtocmd()
630 c->ph_rw.metadata = cpu_to_le64(rqd->dma_meta_list); in nvme_nvm_rqtocmd()
631 c->ph_rw.control = cpu_to_le16(rqd->flags); in nvme_nvm_rqtocmd()
632 c->ph_rw.length = cpu_to_le16(rqd->nr_ppas - 1); in nvme_nvm_rqtocmd()
637 struct nvm_rq *rqd = rq->end_io_data; in nvme_nvm_end_io() local
639 rqd->ppa_status = le64_to_cpu(nvme_req(rq)->result.u64); in nvme_nvm_end_io()
640 rqd->error = nvme_req(rq)->status; in nvme_nvm_end_io()
641 nvm_end_io(rqd); in nvme_nvm_end_io()
[all …]
/Linux-v5.4/include/linux/
Dlightnvm.h321 static inline struct ppa_addr *nvm_rq_to_ppa_list(struct nvm_rq *rqd) in nvm_rq_to_ppa_list() argument
323 return (rqd->nr_ppas > 1) ? rqd->ppa_list : &rqd->ppa_addr; in nvm_rq_to_ppa_list()
/Linux-v5.4/drivers/dma/
Dpl330.c539 struct list_head rqd; member
1702 list_add_tail(&descdone->rqd, &pl330->req_done); in pl330_update()
1709 struct dma_pl330_desc, rqd); in pl330_update()
1710 list_del(&descdone->rqd); in pl330_update()
/Linux-v5.4/arch/m68k/ifpsp060/src/
Dilsp.S299 swap %d5 # same as r*b if previous step rqd