/Linux-v5.10/block/ |
D | blk-pm.h | 18 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_mark_last_busy() 26 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_requeue_request() 35 if (q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_add_request() 43 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_put_request()
|
D | blk-zoned.c | 92 WARN_ON_ONCE(rq->rq_flags & RQF_ZONE_WRITE_LOCKED); in blk_req_zone_write_trylock() 93 rq->rq_flags |= RQF_ZONE_WRITE_LOCKED; in blk_req_zone_write_trylock() 105 WARN_ON_ONCE(rq->rq_flags & RQF_ZONE_WRITE_LOCKED); in __blk_req_zone_write_lock() 106 rq->rq_flags |= RQF_ZONE_WRITE_LOCKED; in __blk_req_zone_write_lock() 112 rq->rq_flags &= ~RQF_ZONE_WRITE_LOCKED; in __blk_req_zone_write_unlock()
|
D | blk-flush.c | 131 rq->rq_flags &= ~RQF_FLUSH_SEQ; in blk_flush_restore_request() 325 flush_rq->rq_flags |= RQF_MQ_INFLIGHT; in blk_kick_flush() 331 flush_rq->rq_flags |= RQF_FLUSH_SEQ; in blk_kick_flush() 423 rq->rq_flags |= RQF_FLUSH_SEQ; in blk_insert_flush()
|
D | blk-core.c | 245 if (unlikely(rq->rq_flags & RQF_QUIET)) in req_bio_endio() 262 if (bio->bi_iter.bi_size == 0 && !(rq->rq_flags & RQF_FLUSH_SEQ)) in req_bio_endio() 1241 if (!(rq->rq_flags & RQF_MIXED_MERGE)) in blk_rq_err_bytes() 1299 !(req->rq_flags & RQF_FLUSH_SEQ)) { in blk_account_io_done() 1452 !(req->rq_flags & RQF_QUIET))) in blk_update_request() 1496 if (req->rq_flags & RQF_MIXED_MERGE) { in blk_update_request() 1501 if (!(req->rq_flags & RQF_SPECIAL_PAYLOAD)) { in blk_update_request() 1632 if (rq_src->rq_flags & RQF_SPECIAL_PAYLOAD) { in blk_rq_prep_clone() 1633 rq->rq_flags |= RQF_SPECIAL_PAYLOAD; in blk_rq_prep_clone()
|
D | blk-mq.c | 272 return (rq->rq_flags & (RQF_IO_STAT | RQF_STATS)) || rq->q->elevator; in blk_mq_need_time_stamp() 293 rq->rq_flags = 0; in blk_mq_rq_ctx_init() 296 rq->rq_flags |= RQF_PREEMPT; in blk_mq_rq_ctx_init() 298 rq->rq_flags |= RQF_IO_STAT; in blk_mq_rq_ctx_init() 338 rq->rq_flags |= RQF_ELVPRIV; in blk_mq_rq_ctx_init() 511 if (rq->rq_flags & RQF_ELVPRIV) { in blk_mq_free_request() 521 if (rq->rq_flags & RQF_MQ_INFLIGHT) in blk_mq_free_request() 542 if (rq->rq_flags & RQF_STATS) { in __blk_mq_end_request() 739 rq->rq_flags |= RQF_STATS; in blk_mq_start_request() 766 rq->rq_flags &= ~RQF_TIMED_OUT; in __blk_mq_requeue_request() [all …]
|
D | blk-mq.h | 230 if (rq->rq_flags & RQF_MQ_INFLIGHT) { in __blk_mq_put_driver_tag() 231 rq->rq_flags &= ~RQF_MQ_INFLIGHT; in __blk_mq_put_driver_tag()
|
D | blk-merge.c | 517 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) in __blk_rq_map_sg() 664 if (rq->rq_flags & RQF_MIXED_MERGE) in blk_rq_set_mixed_merge() 677 rq->rq_flags |= RQF_MIXED_MERGE; in blk_rq_set_mixed_merge() 776 if (((req->rq_flags | next->rq_flags) & RQF_MIXED_MERGE) || in attempt_merge()
|
D | blk-mq-sched.c | 408 if ((rq->rq_flags & RQF_FLUSH_SEQ) || blk_rq_is_passthrough(rq)) in blk_mq_sched_bypass_insert() 412 rq->rq_flags |= RQF_SORTED; in blk_mq_sched_bypass_insert() 449 at_head = (rq->rq_flags & RQF_FLUSH_SEQ) ? true : at_head; in blk_mq_sched_insert_request()
|
D | blk-mq-sched.h | 66 if ((rq->rq_flags & RQF_ELVPRIV) && e && e->type->ops.requeue_request) in blk_mq_sched_requeue_request()
|
D | blk-timeout.c | 140 req->rq_flags &= ~RQF_TIMED_OUT; in blk_add_timer()
|
D | blk.h | 198 #define ELV_ON_HASH(rq) ((rq)->rq_flags & RQF_HASHED) 254 return rq->rq_disk && (rq->rq_flags & RQF_IO_STAT); in blk_do_io_stat()
|
/Linux-v5.10/kernel/sched/ |
D | sched.h | 1193 struct rq_flags { struct 1216 static inline void rq_pin_lock(struct rq *rq, struct rq_flags *rf) in rq_pin_lock() 1226 static inline void rq_unpin_lock(struct rq *rq, struct rq_flags *rf) in rq_unpin_lock() 1236 static inline void rq_repin_lock(struct rq *rq, struct rq_flags *rf) in rq_repin_lock() 1248 struct rq *__task_rq_lock(struct task_struct *p, struct rq_flags *rf) 1251 struct rq *task_rq_lock(struct task_struct *p, struct rq_flags *rf) 1255 static inline void __task_rq_unlock(struct rq *rq, struct rq_flags *rf) in __task_rq_unlock() 1263 task_rq_unlock(struct rq *rq, struct task_struct *p, struct rq_flags *rf) in task_rq_unlock() 1273 rq_lock_irqsave(struct rq *rq, struct rq_flags *rf) in rq_lock_irqsave() 1281 rq_lock_irq(struct rq *rq, struct rq_flags *rf) in rq_lock_irq() [all …]
|
D | core.c | 180 struct rq *__task_rq_lock(struct task_struct *p, struct rq_flags *rf) in __task_rq_lock() 204 struct rq *task_rq_lock(struct task_struct *p, struct rq_flags *rf) in task_rq_lock() 349 struct rq_flags rf; in hrtick() 376 struct rq_flags rf; in __hrtick_start() 1026 struct rq_flags rf; in uclamp_update_util_min_rt_default() 1284 struct rq_flags rf; in uclamp_update_active() 1733 static struct rq *move_queued_task(struct rq *rq, struct rq_flags *rf, in move_queued_task() 1766 static struct rq *__migrate_task(struct rq *rq, struct rq_flags *rf, in __migrate_task() 1789 struct rq_flags rf; in migration_cpu_stop() 1876 struct rq_flags rf; in __set_cpus_allowed_ptr() [all …]
|
D | stop_task.c | 20 balance_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in balance_stop()
|
/Linux-v5.10/drivers/ide/ |
D | ide-cd.c | 103 if (!sense || !rq || (rq->rq_flags & RQF_QUIET)) in cdrom_log_sense() 304 rq->rq_flags |= RQF_FAILED; in cdrom_decode_status() 324 !(rq->rq_flags & RQF_QUIET)) in cdrom_decode_status() 359 if (!(rq->rq_flags & RQF_QUIET)) in cdrom_decode_status() 368 if (!(rq->rq_flags & RQF_QUIET)) in cdrom_decode_status() 375 if (!(rq->rq_flags & RQF_QUIET)) in cdrom_decode_status() 393 rq->rq_flags |= RQF_FAILED; in cdrom_decode_status() 435 req_flags_t rq_flags) in ide_cd_queue_pc() argument 444 cmd[0], write, timeout, rq_flags); in ide_cd_queue_pc() 459 rq->rq_flags |= rq_flags; in ide_cd_queue_pc() [all …]
|
D | ide-io.c | 330 rq->rq_flags |= RQF_FAILED; in start_request() 466 if (!blk_rq_is_passthrough(rq) && !(rq->rq_flags & RQF_DONTPREP)) { in ide_issue_rq() 467 rq->rq_flags |= RQF_DONTPREP; in ide_issue_rq() 526 (rq->rq_flags & RQF_PREEMPT) == 0) { in ide_issue_rq()
|
D | ide-atapi.c | 226 sense_rq->rq_flags |= RQF_PREEMPT; in ide_prep_sense() 319 if (!(rq->rq_flags & RQF_QUIET)) in ide_cd_expiry() 405 rq->rq_flags |= RQF_FAILED; in ide_check_ireason()
|
/Linux-v5.10/net/sunrpc/ |
D | svc_xprt.c | 362 if (!test_bit(RQ_DATA, &rqstp->rq_flags)) { in svc_xprt_reserve_slot() 366 set_bit(RQ_DATA, &rqstp->rq_flags); in svc_xprt_reserve_slot() 374 if (test_and_clear_bit(RQ_DATA, &rqstp->rq_flags)) { in svc_xprt_release_slot() 438 if (test_and_set_bit(RQ_BUSY, &rqstp->rq_flags)) in svc_xprt_do_enqueue() 561 if (test_bit(RQ_BUSY, &rqstp->rq_flags)) in svc_wake_up() 729 clear_bit(RQ_BUSY, &rqstp->rq_flags); in svc_get_next_xprt() 739 set_bit(RQ_BUSY, &rqstp->rq_flags); in svc_get_next_xprt() 1175 if (rqstp->rq_arg.page_len || !test_bit(RQ_USEDEFERRAL, &rqstp->rq_flags)) in svc_defer() 1205 set_bit(RQ_DROPME, &rqstp->rq_flags); in svc_defer()
|
D | svc.c | 612 __set_bit(RQ_BUSY, &rqstp->rq_flags); in svc_rqst_alloc() 697 set_bit(RQ_VICTIM, &rqstp->rq_flags); in choose_victim() 860 if (!test_and_set_bit(RQ_VICTIM, &rqstp->rq_flags)) in svc_exit_thread() 1163 set_bit(RQ_AUTHERR, &rqstp->rq_flags); in svc_return_autherr() 1171 if (test_and_clear_bit(RQ_AUTHERR, &rqstp->rq_flags)) in svc_get_autherr() 1196 test_bit(RQ_DROPME, &rqstp->rq_flags)) in svc_generic_dispatch() 1199 if (test_bit(RQ_AUTHERR, &rqstp->rq_flags)) in svc_generic_dispatch() 1290 set_bit(RQ_SPLICE_OK, &rqstp->rq_flags); in svc_process_common() 1292 set_bit(RQ_USEDEFERRAL, &rqstp->rq_flags); in svc_process_common() 1293 clear_bit(RQ_DROPME, &rqstp->rq_flags); in svc_process_common()
|
/Linux-v5.10/drivers/scsi/ |
D | scsi_lib.c | 156 if (cmd->request->rq_flags & RQF_DONTPREP) { in scsi_mq_requeue_cmd() 157 cmd->request->rq_flags &= ~RQF_DONTPREP; in scsi_mq_requeue_cmd() 243 int timeout, int retries, u64 flags, req_flags_t rq_flags, in __scsi_execute() argument 266 req->rq_flags |= rq_flags | RQF_QUIET; in __scsi_execute() 802 if (!(req->rq_flags & RQF_QUIET)) { in scsi_io_completion_action() 891 else if (req->rq_flags & RQF_QUIET) in scsi_io_completion_nz_result() 1125 if (rq->rq_flags & RQF_DONTPREP) { in scsi_cleanup_rq() 1127 rq->rq_flags &= ~RQF_DONTPREP; in scsi_cleanup_rq() 1234 if (req && !(req->rq_flags & RQF_PREEMPT)) in scsi_device_state_check() 1243 if (req && !(req->rq_flags & RQF_PREEMPT)) in scsi_device_state_check() [all …]
|
/Linux-v5.10/include/scsi/ |
D | scsi_device.h | 445 req_flags_t rq_flags, int *resid); 448 sshdr, timeout, retries, flags, rq_flags, resid) \ argument 453 sense, sshdr, timeout, retries, flags, rq_flags, \
|
/Linux-v5.10/drivers/mmc/core/ |
D | queue.c | 260 req->rq_flags |= RQF_QUIET; in mmc_mq_queue_rq() 312 if (!(req->rq_flags & RQF_DONTPREP)) { in mmc_mq_queue_rq() 314 req->rq_flags |= RQF_DONTPREP; in mmc_mq_queue_rq()
|
/Linux-v5.10/drivers/md/ |
D | dm-rq.c | 269 if (rq->rq_flags & RQF_FAILED) in dm_softirq_done() 296 rq->rq_flags |= RQF_FAILED; in dm_kill_unmapped_request() 312 clone->rq_flags |= RQF_IO_STAT; in dm_dispatch_clone_request()
|
/Linux-v5.10/include/linux/ |
D | blkdev.h | 136 req_flags_t rq_flags; member 680 return (rq->rq_flags & RQF_STARTED) && !blk_rq_is_passthrough(rq); in blk_account_rq() 814 if (rq->rq_flags & RQF_NOMERGE_FLAGS) in rq_mergeable() 1039 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) in blk_rq_payload_bytes() 1050 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) in req_bvec() 1193 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) in blk_rq_nr_phys_segments() 1892 if (rq->rq_flags & RQF_ZONE_WRITE_LOCKED) in blk_req_zone_write_unlock()
|
/Linux-v5.10/drivers/scsi/device_handler/ |
D | scsi_dh_hp_sw.c | 167 req->rq_flags |= RQF_QUIET; in hp_sw_prep_fn()
|