Home
last modified time | relevance | path

Searched refs:req_op (Results 1 – 25 of 89) sorted by relevance

1234

/Linux-v6.1/include/linux/
Dblk_types.h358 enum req_op { enum
470 static inline enum req_op bio_op(const struct bio *bio) in bio_op()
476 static inline void bio_set_op_attrs(struct bio *bio, enum req_op op, in bio_set_op_attrs()
518 static inline bool op_is_zone_mgmt(enum req_op op) in op_is_zone_mgmt()
531 static inline int op_stat_group(enum req_op op) in op_stat_group()
Dblk-mq.h206 static inline enum req_op req_op(const struct request *req) in req_op() function
213 return blk_op_is_passthrough(req_op(rq)); in blk_rq_is_passthrough()
221 #define rq_data_dir(rq) (op_is_write(req_op(rq)) ? WRITE : READ)
224 (op_is_write(req_op(rq)) ? DMA_TO_DEVICE : DMA_FROM_DEVICE)
Dblkdev.h334 extern int blkdev_zone_mgmt(struct block_device *bdev, enum req_op op,
876 extern const char *blk_op_str(enum req_op op);
1403 int (*rw_page)(struct block_device *, sector_t, struct page *, enum req_op);
1456 unsigned int sectors, enum req_op op,
1458 void bdev_end_io_acct(struct block_device *bdev, enum req_op op,
/Linux-v6.1/drivers/block/null_blk/
Dnull_blk.h141 blk_status_t null_process_cmd(struct nullb_cmd *cmd, enum req_op op,
150 blk_status_t null_process_zoned_cmd(struct nullb_cmd *cmd, enum req_op op,
167 enum req_op op, sector_t sector, sector_t nr_sectors) in null_process_zoned_cmd()
Dtrace.h39 __field(enum req_op, op)
44 __entry->op = req_op(cmd->rq);
/Linux-v6.1/block/
Dblk-merge.c569 if (req_op(rq) == REQ_OP_DISCARD) in blk_rq_get_max_segments()
583 max_sectors = blk_queue_get_max_sectors(q, req_op(rq)); in blk_rq_get_max_sectors()
585 req_op(rq) == REQ_OP_DISCARD || in blk_rq_get_max_sectors()
586 req_op(rq) == REQ_OP_SECURE_ERASE) in blk_rq_get_max_sectors()
602 if (req_op(req) == REQ_OP_DISCARD) in ll_new_hw_segment()
742 part_stat_inc(req->part, merges[op_stat_group(req_op(req))]); in blk_account_io_merge_request()
768 if (req_op(req) != req_op(next)) in attempt_merge()
880 if (req_op(rq) != bio_op(bio)) in blk_rq_merge_ok()
922 part_stat_inc(req->part, merges[op_stat_group(req_op(req))]); in blk_account_io_merge_bio()
Dblk.h131 if (req_op(rq) == REQ_OP_FLUSH) in rq_mergeable()
134 if (req_op(rq) == REQ_OP_WRITE_ZEROES) in rq_mergeable()
137 if (req_op(rq) == REQ_OP_ZONE_APPEND) in rq_mergeable()
158 if (req_op(req) == REQ_OP_DISCARD && in blk_discard_mergable()
165 enum req_op op) in blk_queue_get_max_sectors()
Dblk-zoned.c66 if (bdev_op_is_zoned_write(rq->q->disk->part0, req_op(rq))) in blk_req_needs_zone_write_lock()
256 int blkdev_zone_mgmt(struct block_device *bdev, enum req_op op, in blkdev_zone_mgmt()
396 enum req_op op; in blkdev_zone_mgmt_ioctl()
Dblk-map.c158 bio_init(bio, NULL, bio->bi_inline_vecs, nr_pages, req_op(rq)); in bio_copy_user_iov()
260 bio_init(bio, NULL, bio->bi_inline_vecs, nr_vecs, req_op(rq)); in blk_rq_map_bio_alloc()
796 bio->bi_opf |= req_op(rq); in blk_rq_map_kern()
/Linux-v6.1/fs/zonefs/
Dtrace.h23 TP_PROTO(struct inode *inode, enum req_op op),
28 __field(enum req_op, op)
/Linux-v6.1/drivers/block/rnbd/
Drnbd-proto.h271 switch (req_op(rq)) { in rq_to_rnbd_flags()
289 (__force u32)req_op(rq), in rq_to_rnbd_flags()
/Linux-v6.1/fs/
Ddirect-io.c169 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in dio_refill_pages()
236 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in dio_complete()
332 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in dio_bio_end_aio()
419 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in dio_bio_submit()
495 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in dio_bio_complete()
623 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in get_more_blocks()
806 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in submit_page_section()
923 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in do_direct_IO()
/Linux-v6.1/include/trace/events/
Dnilfs2.h195 enum req_op mode),
203 __field(enum req_op, mode)
/Linux-v6.1/arch/um/drivers/
Dubd_kern.c483 if ((io_req->error == BLK_STS_NOTSUPP) && (req_op(io_req->req) == REQ_OP_DISCARD)) { in ubd_handler()
1245 if (req_op(req->req) == REQ_OP_READ) { in cowify_req()
1265 enum req_op op = req_op(req); in ubd_map_req()
1328 enum req_op op = req_op(req); in ubd_submit_request()
1365 switch (req_op(req)) { in ubd_queue_rq()
1483 if (req_op(req->req) == REQ_OP_FLUSH) { in do_io()
1504 switch (req_op(req->req)) { in do_io()
/Linux-v6.1/drivers/scsi/
Dsd_zbc.c503 switch (req_op(rq)) { in sd_zbc_need_zone_wp_update()
532 enum req_op op = req_op(rq); in sd_zbc_zone_wp_update()
604 if (op_is_zone_mgmt(req_op(rq)) && in sd_zbc_complete()
617 if (req_op(rq) == REQ_OP_ZONE_APPEND) in sd_zbc_complete()
/Linux-v6.1/fs/xfs/
Dxfs_bio_io.c18 enum req_op op) in xfs_rw_bdev()
Dxfs_linux.h199 char *data, enum req_op op);
/Linux-v6.1/drivers/mmc/core/
Dqueue.c46 switch (req_op(req)) { in mmc_cqe_issue_type()
67 if (req_op(req) == REQ_OP_READ || req_op(req) == REQ_OP_WRITE) in mmc_issue_type()
/Linux-v6.1/drivers/crypto/hisilicon/sec2/
Dsec_crypto.c269 ctx->req_op->buf_unmap(ctx, req); in sec_req_cb()
271 ctx->req_op->callback(ctx, req, err); in sec_req_cb()
1245 ret = ctx->req_op->buf_map(ctx, req); in sec_request_transfer()
1249 ctx->req_op->do_transfer(ctx, req); in sec_request_transfer()
1251 ret = ctx->req_op->bd_fill(ctx, req); in sec_request_transfer()
1258 ctx->req_op->buf_unmap(ctx, req); in sec_request_transfer()
1264 ctx->req_op->buf_unmap(ctx, req); in sec_request_untransfer()
1790 ret = ctx->req_op->bd_send(ctx, req); in sec_process()
1867 ctx->req_op = &sec_skcipher_req_ops; in sec_skcipher_ctx_init()
1870 ctx->req_op = &sec_skcipher_req_ops_v3; in sec_skcipher_ctx_init()
[all …]
Dsec.h138 const struct sec_req_op *req_op; member
/Linux-v6.1/drivers/block/
Dublk_drv.c478 if (req_op(req) != REQ_OP_WRITE && req_op(req) != REQ_OP_FLUSH) in ublk_map_io()
502 if (req_op(req) == REQ_OP_READ && ublk_rq_has_data(req)) { in ublk_unmap_io()
553 switch (req_op(req)) { in ublk_setup_iod()
601 if (!io->res && req_op(req) == REQ_OP_READ) in ublk_complete_rq()
615 if (req_op(req) != REQ_OP_READ && req_op(req) != REQ_OP_WRITE) { in ublk_complete_rq()
714 (req_op(req) == REQ_OP_WRITE || in __ublk_rq_task_work()
715 req_op(req) == REQ_OP_FLUSH)) { in __ublk_rq_task_work()
Dbrd.c259 unsigned int len, unsigned int off, enum req_op op, in brd_do_bvec()
313 struct page *page, enum req_op op) in brd_rw_page()
Dxen-blkfront.c566 if (req_op(req) == REQ_OP_SECURE_ERASE && info->feature_secdiscard) in blkif_queue_discard_req()
771 BUG_ON(req_op(req) == REQ_OP_FLUSH || req->cmd_flags & REQ_FUA); in blkif_queue_rw_req()
783 if (req_op(req) == REQ_OP_FLUSH || req->cmd_flags & REQ_FUA) { in blkif_queue_rw_req()
872 if (unlikely(req_op(req) == REQ_OP_DISCARD || in blkif_queue_request()
873 req_op(req) == REQ_OP_SECURE_ERASE)) in blkif_queue_request()
893 ((req_op(req) == REQ_OP_FLUSH) && in blkif_request_flush_invalid()
2082 if (req_op(shadow[j].request) == REQ_OP_FLUSH || in blkfront_resume()
2083 req_op(shadow[j].request) == REQ_OP_DISCARD || in blkfront_resume()
2084 req_op(shadow[j].request) == REQ_OP_SECURE_ERASE || in blkfront_resume()
/Linux-v6.1/drivers/md/
Ddm-rq.c216 if (req_op(clone) == REQ_OP_DISCARD && in dm_done()
219 else if (req_op(clone) == REQ_OP_WRITE_ZEROES && in dm_done()
Ddm-ebs-target.c64 static int __ebs_rw_bvec(struct ebs_c *ec, enum req_op op, struct bio_vec *bv, in __ebs_rw_bvec()
121 static int __ebs_rw_bio(struct ebs_c *ec, enum req_op op, struct bio *bio) in __ebs_rw_bio()

1234