Home
last modified time | relevance | path

Searched refs:rqd (Results 1 – 14 of 14) sorted by relevance

/Linux-v4.19/drivers/lightnvm/
Dpblk-read.c41 static void pblk_read_ppalist_rq(struct pblk *pblk, struct nvm_rq *rqd, in pblk_read_ppalist_rq() argument
45 struct pblk_sec_meta *meta_list = rqd->meta_list; in pblk_read_ppalist_rq()
47 int nr_secs = rqd->nr_ppas; in pblk_read_ppalist_rq()
87 rqd->ppa_list[j++] = p; in pblk_read_ppalist_rq()
96 rqd->flags = pblk_set_read_mode(pblk, PBLK_READ_SEQUENTIAL); in pblk_read_ppalist_rq()
98 rqd->flags = pblk_set_read_mode(pblk, PBLK_READ_RANDOM); in pblk_read_ppalist_rq()
106 static void pblk_read_check_seq(struct pblk *pblk, struct nvm_rq *rqd, in pblk_read_check_seq() argument
109 struct pblk_sec_meta *meta_lba_list = rqd->meta_list; in pblk_read_check_seq()
110 int nr_lbas = rqd->nr_ppas; in pblk_read_check_seq()
123 p = (nr_lbas == 1) ? &rqd->ppa_list[i] : &rqd->ppa_addr; in pblk_read_check_seq()
[all …]
Dpblk-write.c20 static unsigned long pblk_end_w_bio(struct pblk *pblk, struct nvm_rq *rqd, in pblk_end_w_bio() argument
51 pblk_bio_free_pages(pblk, rqd->bio, c_ctx->nr_valid, in pblk_end_w_bio()
55 atomic_long_add(rqd->nr_ppas, &pblk->sync_writes); in pblk_end_w_bio()
60 bio_put(rqd->bio); in pblk_end_w_bio()
61 pblk_free_rqd(pblk, rqd, PBLK_WRITE); in pblk_end_w_bio()
67 struct nvm_rq *rqd, in pblk_end_queued_w_bio() argument
71 return pblk_end_w_bio(pblk, rqd, c_ctx); in pblk_end_queued_w_bio()
74 static void pblk_complete_write(struct pblk *pblk, struct nvm_rq *rqd, in pblk_complete_write() argument
85 pblk_up_rq(pblk, rqd->ppa_list, rqd->nr_ppas, c_ctx->lun_bitmap); in pblk_complete_write()
89 pos = pblk_end_w_bio(pblk, rqd, c_ctx); in pblk_complete_write()
[all …]
Dpblk-recovery.c102 struct nvm_rq *rqd; member
115 struct nvm_rq *rqd; in pblk_recov_read_oob() local
127 rqd = p.rqd; in pblk_recov_read_oob()
139 memset(rqd, 0, pblk_g_rq_size); in pblk_recov_read_oob()
153 rqd->bio = bio; in pblk_recov_read_oob()
154 rqd->opcode = NVM_OP_PREAD; in pblk_recov_read_oob()
155 rqd->meta_list = meta_list; in pblk_recov_read_oob()
156 rqd->nr_ppas = rq_ppas; in pblk_recov_read_oob()
157 rqd->ppa_list = ppa_list; in pblk_recov_read_oob()
158 rqd->dma_ppa_list = dma_ppa_list; in pblk_recov_read_oob()
[all …]
Dpblk-map.c86 void pblk_map_rq(struct pblk *pblk, struct nvm_rq *rqd, unsigned int sentry, in pblk_map_rq() argument
90 struct pblk_sec_meta *meta_list = rqd->meta_list; in pblk_map_rq()
95 for (i = off; i < rqd->nr_ppas; i += min) { in pblk_map_rq()
97 if (pblk_map_page_data(pblk, sentry + i, &rqd->ppa_list[i], in pblk_map_rq()
99 bio_put(rqd->bio); in pblk_map_rq()
100 pblk_free_rqd(pblk, rqd, PBLK_WRITE); in pblk_map_rq()
107 void pblk_map_erase_rq(struct pblk *pblk, struct nvm_rq *rqd, in pblk_map_erase_rq() argument
114 struct pblk_sec_meta *meta_list = rqd->meta_list; in pblk_map_erase_rq()
120 for (i = 0; i < rqd->nr_ppas; i += min) { in pblk_map_erase_rq()
122 if (pblk_map_page_data(pblk, sentry + i, &rqd->ppa_list[i], in pblk_map_erase_rq()
[all …]
Dpblk-core.c75 static void __pblk_end_io_erase(struct pblk *pblk, struct nvm_rq *rqd) in __pblk_end_io_erase() argument
83 line = &pblk->lines[pblk_ppa_to_line(rqd->ppa_addr)]; in __pblk_end_io_erase()
84 pos = pblk_ppa_to_pos(geo, rqd->ppa_addr); in __pblk_end_io_erase()
89 if (rqd->error) { in __pblk_end_io_erase()
91 pblk_mark_bb(pblk, line, rqd->ppa_addr); in __pblk_end_io_erase()
100 static void pblk_end_io_erase(struct nvm_rq *rqd) in pblk_end_io_erase() argument
102 struct pblk *pblk = rqd->private; in pblk_end_io_erase()
104 __pblk_end_io_erase(pblk, rqd); in pblk_end_io_erase()
105 mempool_free(rqd, &pblk->e_rq_pool); in pblk_end_io_erase()
234 struct nvm_rq *rqd; in pblk_alloc_rqd() local
[all …]
Dcore.c599 static void nvm_rq_tgt_to_dev(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd) in nvm_rq_tgt_to_dev() argument
601 if (rqd->nr_ppas == 1) { in nvm_rq_tgt_to_dev()
602 nvm_ppa_tgt_to_dev(tgt_dev, &rqd->ppa_addr, 1); in nvm_rq_tgt_to_dev()
606 nvm_ppa_tgt_to_dev(tgt_dev, rqd->ppa_list, rqd->nr_ppas); in nvm_rq_tgt_to_dev()
609 static void nvm_rq_dev_to_tgt(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd) in nvm_rq_dev_to_tgt() argument
611 if (rqd->nr_ppas == 1) { in nvm_rq_dev_to_tgt()
612 nvm_ppa_dev_to_tgt(tgt_dev, &rqd->ppa_addr, 1); in nvm_rq_dev_to_tgt()
616 nvm_ppa_dev_to_tgt(tgt_dev, rqd->ppa_list, rqd->nr_ppas); in nvm_rq_dev_to_tgt()
670 static int nvm_set_rqd_ppalist(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd, in nvm_set_rqd_ppalist() argument
679 rqd->nr_ppas = nr_ppas; in nvm_set_rqd_ppalist()
[all …]
Dpblk.h142 struct nvm_rq *rqd; member
745 unsigned int pblk_rb_read_to_bio(struct pblk_rb *rb, struct nvm_rq *rqd,
772 void pblk_free_rqd(struct pblk *pblk, struct nvm_rq *rqd, int type);
774 int pblk_setup_w_rec_rq(struct pblk *pblk, struct nvm_rq *rqd,
781 void pblk_log_write_err(struct pblk *pblk, struct nvm_rq *rqd);
782 void pblk_log_read_err(struct pblk *pblk, struct nvm_rq *rqd);
783 int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd);
784 int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd);
857 void pblk_map_erase_rq(struct pblk *pblk, struct nvm_rq *rqd,
860 void pblk_map_rq(struct pblk *pblk, struct nvm_rq *rqd, unsigned int sentry,
[all …]
Dpblk-rb.c512 unsigned int pblk_rb_read_to_bio(struct pblk_rb *rb, struct nvm_rq *rqd, in pblk_rb_read_to_bio() argument
518 struct pblk_c_ctx *c_ctx = nvm_rq_to_pdu(rqd); in pblk_rb_read_to_bio()
519 struct bio *bio = rqd->bio; in pblk_rb_read_to_bio()
/Linux-v4.19/block/
Dblk-rq-qos.c104 bool rq_depth_calc_max_depth(struct rq_depth *rqd) in rq_depth_calc_max_depth() argument
116 if (rqd->queue_depth == 1) { in rq_depth_calc_max_depth()
117 if (rqd->scale_step > 0) in rq_depth_calc_max_depth()
118 rqd->max_depth = 1; in rq_depth_calc_max_depth()
120 rqd->max_depth = 2; in rq_depth_calc_max_depth()
131 depth = min_t(unsigned int, rqd->default_depth, in rq_depth_calc_max_depth()
132 rqd->queue_depth); in rq_depth_calc_max_depth()
133 if (rqd->scale_step > 0) in rq_depth_calc_max_depth()
134 depth = 1 + ((depth - 1) >> min(31, rqd->scale_step)); in rq_depth_calc_max_depth()
135 else if (rqd->scale_step < 0) { in rq_depth_calc_max_depth()
[all …]
Dblk-wbt.c236 struct rq_depth *rqd = &rwb->rq_depth; in latency_exceeded() local
280 if (rqd->scale_step) in latency_exceeded()
289 struct rq_depth *rqd = &rwb->rq_depth; in rwb_trace_step() local
291 trace_wbt_step(bdi, msg, rqd->scale_step, rwb->cur_win_nsec, in rwb_trace_step()
292 rwb->wb_background, rwb->wb_normal, rqd->max_depth); in rwb_trace_step()
327 struct rq_depth *rqd = &rwb->rq_depth; in rwb_arm_timer() local
329 if (rqd->scale_step > 0) { in rwb_arm_timer()
337 int_sqrt((rqd->scale_step + 1) << 8)); in rwb_arm_timer()
352 struct rq_depth *rqd = &rwb->rq_depth; in wb_timer_fn() local
358 trace_wbt_timer(rwb->rqos.q->backing_dev_info, status, rqd->scale_step, in wb_timer_fn()
[all …]
Dblk-rq-qos.h97 void rq_depth_scale_up(struct rq_depth *rqd);
98 void rq_depth_scale_down(struct rq_depth *rqd, bool hard_throttle);
99 bool rq_depth_calc_max_depth(struct rq_depth *rqd);
/Linux-v4.19/drivers/nvme/host/
Dlightnvm.c628 static inline void nvme_nvm_rqtocmd(struct nvm_rq *rqd, struct nvme_ns *ns, in nvme_nvm_rqtocmd() argument
631 c->ph_rw.opcode = rqd->opcode; in nvme_nvm_rqtocmd()
633 c->ph_rw.spba = cpu_to_le64(rqd->ppa_addr.ppa); in nvme_nvm_rqtocmd()
634 c->ph_rw.metadata = cpu_to_le64(rqd->dma_meta_list); in nvme_nvm_rqtocmd()
635 c->ph_rw.control = cpu_to_le16(rqd->flags); in nvme_nvm_rqtocmd()
636 c->ph_rw.length = cpu_to_le16(rqd->nr_ppas - 1); in nvme_nvm_rqtocmd()
641 struct nvm_rq *rqd = rq->end_io_data; in nvme_nvm_end_io() local
643 rqd->ppa_status = le64_to_cpu(nvme_req(rq)->result.u64); in nvme_nvm_end_io()
644 rqd->error = nvme_req(rq)->status; in nvme_nvm_end_io()
645 nvm_end_io(rqd); in nvme_nvm_end_io()
[all …]
/Linux-v4.19/drivers/dma/
Dpl330.c537 struct list_head rqd; member
1692 list_add_tail(&descdone->rqd, &pl330->req_done); in pl330_update()
1699 struct dma_pl330_desc, rqd); in pl330_update()
1700 list_del(&descdone->rqd); in pl330_update()
/Linux-v4.19/arch/m68k/ifpsp060/src/
Dilsp.S299 swap %d5 # same as r*b if previous step rqd