/Linux-v4.19/block/ |
D | blk-core.c | 270 if (unlikely(rq->rq_flags & RQF_QUIET)) in req_bio_endio() 276 if (bio->bi_iter.bi_size == 0 && !(rq->rq_flags & RQF_FLUSH_SEQ)) in req_bio_endio() 1210 if (rq->rq_flags & RQF_ELVPRIV) { in blk_free_request() 1273 req_flags_t rq_flags) in freed_request() argument 1279 if (rq_flags & RQF_ELVPRIV) in freed_request() 1356 req_flags_t rq_flags = RQF_ALLOCED; in __get_request() local 1419 rq_flags |= RQF_ELVPRIV; in __get_request() 1426 rq_flags |= RQF_IO_STAT; in __get_request() 1437 rq->rq_flags = rq_flags; in __get_request() 1439 rq->rq_flags |= RQF_PREEMPT; in __get_request() [all …]
|
D | elevator.c | 258 rq->rq_flags &= ~RQF_HASHED; in __elv_rqhash_del() 274 rq->rq_flags |= RQF_HASHED; in elv_rqhash_add() 383 if (pos->rq_flags & (RQF_STARTED | RQF_SOFTBARRIER)) in elv_dispatch_sort() 533 next_sorted = (__force bool)(next->rq_flags & RQF_SORTED); in elv_merge_requests() 563 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_requeue_request() 569 if (q->dev && !(rq->rq_flags & RQF_PM) && q->nr_pending++ == 0 && in blk_pm_add_request() 589 if (rq->rq_flags & RQF_SORTED) in elv_requeue_request() 593 rq->rq_flags &= ~RQF_STARTED; in elv_requeue_request() 627 if (rq->rq_flags & RQF_SOFTBARRIER) { in __elv_add_request() 633 } else if (!(rq->rq_flags & RQF_ELVPRIV) && in __elv_add_request() [all …]
|
D | blk-tag.c | 272 rq->rq_flags &= ~RQF_QUEUED; in blk_queue_end_tag() 317 if (unlikely((rq->rq_flags & RQF_QUEUED))) { in blk_queue_start_tag() 372 rq->rq_flags |= RQF_QUEUED; in blk_queue_start_tag()
|
D | blk-mq.c | 282 req_flags_t rq_flags = 0; in blk_mq_rq_ctx_init() local 289 rq_flags = RQF_MQ_INFLIGHT; in blk_mq_rq_ctx_init() 300 rq->rq_flags = rq_flags; in blk_mq_rq_ctx_init() 304 rq->rq_flags |= RQF_PREEMPT; in blk_mq_rq_ctx_init() 306 rq->rq_flags |= RQF_IO_STAT; in blk_mq_rq_ctx_init() 392 rq->rq_flags |= RQF_ELVPRIV; in blk_mq_get_request() 493 if (rq->rq_flags & RQF_ELVPRIV) { in blk_mq_free_request() 503 if (rq->rq_flags & RQF_MQ_INFLIGHT) in blk_mq_free_request() 524 if (rq->rq_flags & RQF_STATS) { in __blk_mq_end_request() 643 rq->rq_flags |= RQF_STATS; in blk_mq_start_request() [all …]
|
D | blk-flush.c | 131 rq->rq_flags &= ~RQF_FLUSH_SEQ; in blk_flush_restore_request() 354 flush_rq->rq_flags |= RQF_FLUSH_SEQ; in blk_kick_flush() 394 rq->rq_flags &= ~RQF_STARTED; in flush_data_end_io() 500 rq->rq_flags |= RQF_FLUSH_SEQ; in blk_insert_flush()
|
D | blk-zoned.c | 52 WARN_ON_ONCE(rq->rq_flags & RQF_ZONE_WRITE_LOCKED); in __blk_req_zone_write_lock() 53 rq->rq_flags |= RQF_ZONE_WRITE_LOCKED; in __blk_req_zone_write_lock() 59 rq->rq_flags &= ~RQF_ZONE_WRITE_LOCKED; in __blk_req_zone_write_unlock()
|
D | blk-mq.h | 181 if (rq->rq_flags & RQF_MQ_INFLIGHT) { in __blk_mq_put_driver_tag() 182 rq->rq_flags &= ~RQF_MQ_INFLIGHT; in __blk_mq_put_driver_tag()
|
D | blk-merge.c | 442 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) in blk_rq_map_sg() 449 if (unlikely(rq->rq_flags & RQF_COPY_USER) && in blk_rq_map_sg() 639 if (rq->rq_flags & RQF_MIXED_MERGE) in blk_rq_set_mixed_merge() 652 rq->rq_flags |= RQF_MIXED_MERGE; in blk_rq_set_mixed_merge() 729 if (((req->rq_flags | next->rq_flags) & RQF_MIXED_MERGE) || in attempt_merge()
|
D | blk-map.c | 144 rq->rq_flags |= RQF_COPY_USER; in blk_rq_map_user_iov() 242 rq->rq_flags |= RQF_COPY_USER; in blk_rq_map_kern()
|
D | blk-exec.c | 71 rq->rq_flags |= RQF_QUIET; in blk_execute_rq_nowait()
|
D | blk-mq-sched.c | 351 if (rq->rq_flags & RQF_FLUSH_SEQ) { in blk_mq_sched_bypass_insert() 359 rq->rq_flags |= RQF_SORTED; in blk_mq_sched_bypass_insert() 373 if (!(rq->rq_flags & RQF_FLUSH_SEQ) && op_is_flush(rq->cmd_flags)) { in blk_mq_sched_insert_request()
|
D | blk.h | 215 #define ELV_ON_HASH(rq) ((rq)->rq_flags & RQF_HASHED) 305 (rq->rq_flags & RQF_IO_STAT) && in blk_do_io_stat()
|
/Linux-v4.19/kernel/sched/ |
D | sched.h | 1029 struct rq_flags { struct 1042 static inline void rq_pin_lock(struct rq *rq, struct rq_flags *rf) in rq_pin_lock() argument 1052 static inline void rq_unpin_lock(struct rq *rq, struct rq_flags *rf) in rq_unpin_lock() 1062 static inline void rq_repin_lock(struct rq *rq, struct rq_flags *rf) in rq_repin_lock() 1524 struct rq_flags *rf); 1780 struct rq *__task_rq_lock(struct task_struct *p, struct rq_flags *rf) 1783 struct rq *task_rq_lock(struct task_struct *p, struct rq_flags *rf) 1787 static inline void __task_rq_unlock(struct rq *rq, struct rq_flags *rf) in __task_rq_unlock() 1795 task_rq_unlock(struct rq *rq, struct task_struct *p, struct rq_flags *rf) in task_rq_unlock() 1805 rq_lock_irqsave(struct rq *rq, struct rq_flags *rf) in rq_lock_irqsave() [all …]
|
D | core.c | 66 struct rq *__task_rq_lock(struct task_struct *p, struct rq_flags *rf) in __task_rq_lock() 90 struct rq *task_rq_lock(struct task_struct *p, struct rq_flags *rf) in task_rq_lock() 227 struct rq_flags rf; in hrtick() 254 struct rq_flags rf; in __hrtick_start() 908 static struct rq *move_queued_task(struct rq *rq, struct rq_flags *rf, in move_queued_task() 943 static struct rq *__migrate_task(struct rq *rq, struct rq_flags *rf, in __migrate_task() 966 struct rq_flags rf; in migration_cpu_stop() 1053 struct rq_flags rf; in __set_cpus_allowed_ptr() 1184 struct rq_flags srf, drf; in __migrate_swap_task() 1317 struct rq_flags rf; in wait_task_inactive() [all …]
|
D | stop_task.c | 27 pick_next_task_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in pick_next_task_stop()
|
/Linux-v4.19/drivers/ide/ |
D | ide-cd.c | 102 if (!sense || !rq || (rq->rq_flags & RQF_QUIET)) in cdrom_log_sense() 292 rq->rq_flags |= RQF_FAILED; in cdrom_decode_status() 312 !(rq->rq_flags & RQF_QUIET)) in cdrom_decode_status() 347 if (!(rq->rq_flags & RQF_QUIET)) in cdrom_decode_status() 356 if (!(rq->rq_flags & RQF_QUIET)) in cdrom_decode_status() 363 if (!(rq->rq_flags & RQF_QUIET)) in cdrom_decode_status() 381 rq->rq_flags |= RQF_FAILED; in cdrom_decode_status() 423 req_flags_t rq_flags) in ide_cd_queue_pc() argument 432 cmd[0], write, timeout, rq_flags); in ide_cd_queue_pc() 447 rq->rq_flags |= rq_flags; in ide_cd_queue_pc() [all …]
|
D | ide-io.c | 310 BUG_ON(!(rq->rq_flags & RQF_STARTED)); in start_request() 319 rq->rq_flags |= RQF_FAILED; in start_request() 538 (rq->rq_flags & RQF_PREEMPT) == 0) { in do_ide_request()
|
D | ide-atapi.c | 216 sense_rq->rq_flags |= RQF_PREEMPT; in ide_prep_sense() 302 if (!(rq->rq_flags & RQF_QUIET)) in ide_cd_expiry() 388 rq->rq_flags |= RQF_FAILED; in ide_check_ireason()
|
/Linux-v4.19/net/sunrpc/ |
D | svc_xprt.c | 347 if (!test_bit(RQ_DATA, &rqstp->rq_flags)) { in svc_xprt_reserve_slot() 351 set_bit(RQ_DATA, &rqstp->rq_flags); in svc_xprt_reserve_slot() 359 if (test_and_clear_bit(RQ_DATA, &rqstp->rq_flags)) { in svc_xprt_release_slot() 409 if (test_and_set_bit(RQ_BUSY, &rqstp->rq_flags)) in svc_xprt_do_enqueue() 531 if (test_bit(RQ_BUSY, &rqstp->rq_flags)) in svc_wake_up() 699 clear_bit(RQ_BUSY, &rqstp->rq_flags); in svc_get_next_xprt() 709 set_bit(RQ_BUSY, &rqstp->rq_flags); in svc_get_next_xprt() 1159 if (rqstp->rq_arg.page_len || !test_bit(RQ_USEDEFERRAL, &rqstp->rq_flags)) in svc_defer() 1188 set_bit(RQ_DROPME, &rqstp->rq_flags); in svc_defer()
|
D | svc.c | 611 __set_bit(RQ_BUSY, &rqstp->rq_flags); in svc_rqst_alloc() 696 set_bit(RQ_VICTIM, &rqstp->rq_flags); in choose_victim() 859 if (!test_and_set_bit(RQ_VICTIM, &rqstp->rq_flags)) in svc_exit_thread() 1169 set_bit(RQ_SPLICE_OK, &rqstp->rq_flags); in svc_process_common() 1171 set_bit(RQ_USEDEFERRAL, &rqstp->rq_flags); in svc_process_common() 1172 clear_bit(RQ_DROPME, &rqstp->rq_flags); in svc_process_common() 1291 test_bit(RQ_DROPME, &rqstp->rq_flags)) { in svc_process_common()
|
/Linux-v4.19/drivers/scsi/ |
D | scsi_lib.c | 146 if (cmd->request->rq_flags & RQF_DONTPREP) { in scsi_mq_requeue_cmd() 147 cmd->request->rq_flags &= ~RQF_DONTPREP; in scsi_mq_requeue_cmd() 261 int timeout, int retries, u64 flags, req_flags_t rq_flags, in __scsi_execute() argument 284 req->rq_flags |= rq_flags | RQF_QUIET; in __scsi_execute() 898 if (!(req->rq_flags & RQF_QUIET)) { in scsi_io_completion_action() 987 else if (req->rq_flags & RQF_QUIET) in scsi_io_completion_nz_result() 1409 if (req && !(req->rq_flags & RQF_PREEMPT)) in scsi_prep_state_check() 1418 if (req && !(req->rq_flags & RQF_PREEMPT)) in scsi_prep_state_check() 1454 req->rq_flags |= RQF_DONTPREP; in scsi_prep_return() 1923 if (blk_queue_tagged(q) && !(req->rq_flags & RQF_QUEUED)) { in scsi_request_fn() [all …]
|
/Linux-v4.19/drivers/mmc/core/ |
D | queue.c | 255 req->rq_flags |= RQF_QUIET; in mmc_mq_queue_rq() 299 if (!(req->rq_flags & RQF_DONTPREP)) { in mmc_mq_queue_rq() 301 req->rq_flags |= RQF_DONTPREP; in mmc_mq_queue_rq()
|
/Linux-v4.19/include/scsi/ |
D | scsi_device.h | 435 req_flags_t rq_flags, int *resid); 438 sshdr, timeout, retries, flags, rq_flags, resid) \ argument 443 sense, sshdr, timeout, retries, flags, rq_flags, \
|
/Linux-v4.19/drivers/md/ |
D | dm-rq.c | 351 if (rq->rq_flags & RQF_FAILED) in dm_softirq_done() 380 rq->rq_flags |= RQF_FAILED; in dm_kill_unmapped_request() 407 clone->rq_flags |= RQF_IO_STAT; in dm_dispatch_clone_request()
|
/Linux-v4.19/include/linux/ |
D | blkdev.h | 156 req_flags_t rq_flags; member 753 return (rq->rq_flags & RQF_STARTED) && !blk_rq_is_passthrough(rq); in blk_account_rq() 858 if (rq->rq_flags & RQF_NOMERGE_FLAGS) in rq_mergeable() 1088 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) in blk_rq_payload_bytes() 1258 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) in blk_rq_nr_phys_segments() 2013 if (rq->rq_flags & RQF_ZONE_WRITE_LOCKED) in blk_req_zone_write_unlock()
|