Home
last modified time | relevance | path

Searched refs:rq_flags (Results 1 – 25 of 67) sorted by relevance

123

/Linux-v5.10/block/
Dblk-pm.h18 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_mark_last_busy()
26 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_requeue_request()
35 if (q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_add_request()
43 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_put_request()
Dblk-zoned.c92 WARN_ON_ONCE(rq->rq_flags & RQF_ZONE_WRITE_LOCKED); in blk_req_zone_write_trylock()
93 rq->rq_flags |= RQF_ZONE_WRITE_LOCKED; in blk_req_zone_write_trylock()
105 WARN_ON_ONCE(rq->rq_flags & RQF_ZONE_WRITE_LOCKED); in __blk_req_zone_write_lock()
106 rq->rq_flags |= RQF_ZONE_WRITE_LOCKED; in __blk_req_zone_write_lock()
112 rq->rq_flags &= ~RQF_ZONE_WRITE_LOCKED; in __blk_req_zone_write_unlock()
Dblk-flush.c131 rq->rq_flags &= ~RQF_FLUSH_SEQ; in blk_flush_restore_request()
325 flush_rq->rq_flags |= RQF_MQ_INFLIGHT; in blk_kick_flush()
331 flush_rq->rq_flags |= RQF_FLUSH_SEQ; in blk_kick_flush()
423 rq->rq_flags |= RQF_FLUSH_SEQ; in blk_insert_flush()
Dblk-core.c245 if (unlikely(rq->rq_flags & RQF_QUIET)) in req_bio_endio()
262 if (bio->bi_iter.bi_size == 0 && !(rq->rq_flags & RQF_FLUSH_SEQ)) in req_bio_endio()
1241 if (!(rq->rq_flags & RQF_MIXED_MERGE)) in blk_rq_err_bytes()
1299 !(req->rq_flags & RQF_FLUSH_SEQ)) { in blk_account_io_done()
1452 !(req->rq_flags & RQF_QUIET))) in blk_update_request()
1496 if (req->rq_flags & RQF_MIXED_MERGE) { in blk_update_request()
1501 if (!(req->rq_flags & RQF_SPECIAL_PAYLOAD)) { in blk_update_request()
1632 if (rq_src->rq_flags & RQF_SPECIAL_PAYLOAD) { in blk_rq_prep_clone()
1633 rq->rq_flags |= RQF_SPECIAL_PAYLOAD; in blk_rq_prep_clone()
Dblk-mq.c272 return (rq->rq_flags & (RQF_IO_STAT | RQF_STATS)) || rq->q->elevator; in blk_mq_need_time_stamp()
293 rq->rq_flags = 0; in blk_mq_rq_ctx_init()
296 rq->rq_flags |= RQF_PREEMPT; in blk_mq_rq_ctx_init()
298 rq->rq_flags |= RQF_IO_STAT; in blk_mq_rq_ctx_init()
338 rq->rq_flags |= RQF_ELVPRIV; in blk_mq_rq_ctx_init()
511 if (rq->rq_flags & RQF_ELVPRIV) { in blk_mq_free_request()
521 if (rq->rq_flags & RQF_MQ_INFLIGHT) in blk_mq_free_request()
542 if (rq->rq_flags & RQF_STATS) { in __blk_mq_end_request()
739 rq->rq_flags |= RQF_STATS; in blk_mq_start_request()
766 rq->rq_flags &= ~RQF_TIMED_OUT; in __blk_mq_requeue_request()
[all …]
Dblk-mq.h230 if (rq->rq_flags & RQF_MQ_INFLIGHT) { in __blk_mq_put_driver_tag()
231 rq->rq_flags &= ~RQF_MQ_INFLIGHT; in __blk_mq_put_driver_tag()
Dblk-merge.c517 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) in __blk_rq_map_sg()
664 if (rq->rq_flags & RQF_MIXED_MERGE) in blk_rq_set_mixed_merge()
677 rq->rq_flags |= RQF_MIXED_MERGE; in blk_rq_set_mixed_merge()
776 if (((req->rq_flags | next->rq_flags) & RQF_MIXED_MERGE) || in attempt_merge()
Dblk-mq-sched.c408 if ((rq->rq_flags & RQF_FLUSH_SEQ) || blk_rq_is_passthrough(rq)) in blk_mq_sched_bypass_insert()
412 rq->rq_flags |= RQF_SORTED; in blk_mq_sched_bypass_insert()
449 at_head = (rq->rq_flags & RQF_FLUSH_SEQ) ? true : at_head; in blk_mq_sched_insert_request()
Dblk-mq-sched.h66 if ((rq->rq_flags & RQF_ELVPRIV) && e && e->type->ops.requeue_request) in blk_mq_sched_requeue_request()
Dblk-timeout.c140 req->rq_flags &= ~RQF_TIMED_OUT; in blk_add_timer()
Dblk.h198 #define ELV_ON_HASH(rq) ((rq)->rq_flags & RQF_HASHED)
254 return rq->rq_disk && (rq->rq_flags & RQF_IO_STAT); in blk_do_io_stat()
/Linux-v5.10/kernel/sched/
Dsched.h1193 struct rq_flags { struct
1216 static inline void rq_pin_lock(struct rq *rq, struct rq_flags *rf) in rq_pin_lock()
1226 static inline void rq_unpin_lock(struct rq *rq, struct rq_flags *rf) in rq_unpin_lock()
1236 static inline void rq_repin_lock(struct rq *rq, struct rq_flags *rf) in rq_repin_lock()
1248 struct rq *__task_rq_lock(struct task_struct *p, struct rq_flags *rf)
1251 struct rq *task_rq_lock(struct task_struct *p, struct rq_flags *rf)
1255 static inline void __task_rq_unlock(struct rq *rq, struct rq_flags *rf) in __task_rq_unlock()
1263 task_rq_unlock(struct rq *rq, struct task_struct *p, struct rq_flags *rf) in task_rq_unlock()
1273 rq_lock_irqsave(struct rq *rq, struct rq_flags *rf) in rq_lock_irqsave()
1281 rq_lock_irq(struct rq *rq, struct rq_flags *rf) in rq_lock_irq()
[all …]
Dcore.c180 struct rq *__task_rq_lock(struct task_struct *p, struct rq_flags *rf) in __task_rq_lock()
204 struct rq *task_rq_lock(struct task_struct *p, struct rq_flags *rf) in task_rq_lock()
349 struct rq_flags rf; in hrtick()
376 struct rq_flags rf; in __hrtick_start()
1026 struct rq_flags rf; in uclamp_update_util_min_rt_default()
1284 struct rq_flags rf; in uclamp_update_active()
1733 static struct rq *move_queued_task(struct rq *rq, struct rq_flags *rf, in move_queued_task()
1766 static struct rq *__migrate_task(struct rq *rq, struct rq_flags *rf, in __migrate_task()
1789 struct rq_flags rf; in migration_cpu_stop()
1876 struct rq_flags rf; in __set_cpus_allowed_ptr()
[all …]
Dstop_task.c20 balance_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in balance_stop()
/Linux-v5.10/drivers/ide/
Dide-cd.c103 if (!sense || !rq || (rq->rq_flags & RQF_QUIET)) in cdrom_log_sense()
304 rq->rq_flags |= RQF_FAILED; in cdrom_decode_status()
324 !(rq->rq_flags & RQF_QUIET)) in cdrom_decode_status()
359 if (!(rq->rq_flags & RQF_QUIET)) in cdrom_decode_status()
368 if (!(rq->rq_flags & RQF_QUIET)) in cdrom_decode_status()
375 if (!(rq->rq_flags & RQF_QUIET)) in cdrom_decode_status()
393 rq->rq_flags |= RQF_FAILED; in cdrom_decode_status()
435 req_flags_t rq_flags) in ide_cd_queue_pc() argument
444 cmd[0], write, timeout, rq_flags); in ide_cd_queue_pc()
459 rq->rq_flags |= rq_flags; in ide_cd_queue_pc()
[all …]
Dide-io.c330 rq->rq_flags |= RQF_FAILED; in start_request()
466 if (!blk_rq_is_passthrough(rq) && !(rq->rq_flags & RQF_DONTPREP)) { in ide_issue_rq()
467 rq->rq_flags |= RQF_DONTPREP; in ide_issue_rq()
526 (rq->rq_flags & RQF_PREEMPT) == 0) { in ide_issue_rq()
Dide-atapi.c226 sense_rq->rq_flags |= RQF_PREEMPT; in ide_prep_sense()
319 if (!(rq->rq_flags & RQF_QUIET)) in ide_cd_expiry()
405 rq->rq_flags |= RQF_FAILED; in ide_check_ireason()
/Linux-v5.10/net/sunrpc/
Dsvc_xprt.c362 if (!test_bit(RQ_DATA, &rqstp->rq_flags)) { in svc_xprt_reserve_slot()
366 set_bit(RQ_DATA, &rqstp->rq_flags); in svc_xprt_reserve_slot()
374 if (test_and_clear_bit(RQ_DATA, &rqstp->rq_flags)) { in svc_xprt_release_slot()
438 if (test_and_set_bit(RQ_BUSY, &rqstp->rq_flags)) in svc_xprt_do_enqueue()
561 if (test_bit(RQ_BUSY, &rqstp->rq_flags)) in svc_wake_up()
729 clear_bit(RQ_BUSY, &rqstp->rq_flags); in svc_get_next_xprt()
739 set_bit(RQ_BUSY, &rqstp->rq_flags); in svc_get_next_xprt()
1175 if (rqstp->rq_arg.page_len || !test_bit(RQ_USEDEFERRAL, &rqstp->rq_flags)) in svc_defer()
1205 set_bit(RQ_DROPME, &rqstp->rq_flags); in svc_defer()
Dsvc.c612 __set_bit(RQ_BUSY, &rqstp->rq_flags); in svc_rqst_alloc()
697 set_bit(RQ_VICTIM, &rqstp->rq_flags); in choose_victim()
860 if (!test_and_set_bit(RQ_VICTIM, &rqstp->rq_flags)) in svc_exit_thread()
1163 set_bit(RQ_AUTHERR, &rqstp->rq_flags); in svc_return_autherr()
1171 if (test_and_clear_bit(RQ_AUTHERR, &rqstp->rq_flags)) in svc_get_autherr()
1196 test_bit(RQ_DROPME, &rqstp->rq_flags)) in svc_generic_dispatch()
1199 if (test_bit(RQ_AUTHERR, &rqstp->rq_flags)) in svc_generic_dispatch()
1290 set_bit(RQ_SPLICE_OK, &rqstp->rq_flags); in svc_process_common()
1292 set_bit(RQ_USEDEFERRAL, &rqstp->rq_flags); in svc_process_common()
1293 clear_bit(RQ_DROPME, &rqstp->rq_flags); in svc_process_common()
/Linux-v5.10/drivers/scsi/
Dscsi_lib.c156 if (cmd->request->rq_flags & RQF_DONTPREP) { in scsi_mq_requeue_cmd()
157 cmd->request->rq_flags &= ~RQF_DONTPREP; in scsi_mq_requeue_cmd()
243 int timeout, int retries, u64 flags, req_flags_t rq_flags, in __scsi_execute() argument
266 req->rq_flags |= rq_flags | RQF_QUIET; in __scsi_execute()
802 if (!(req->rq_flags & RQF_QUIET)) { in scsi_io_completion_action()
891 else if (req->rq_flags & RQF_QUIET) in scsi_io_completion_nz_result()
1125 if (rq->rq_flags & RQF_DONTPREP) { in scsi_cleanup_rq()
1127 rq->rq_flags &= ~RQF_DONTPREP; in scsi_cleanup_rq()
1234 if (req && !(req->rq_flags & RQF_PREEMPT)) in scsi_device_state_check()
1243 if (req && !(req->rq_flags & RQF_PREEMPT)) in scsi_device_state_check()
[all …]
/Linux-v5.10/include/scsi/
Dscsi_device.h445 req_flags_t rq_flags, int *resid);
448 sshdr, timeout, retries, flags, rq_flags, resid) \ argument
453 sense, sshdr, timeout, retries, flags, rq_flags, \
/Linux-v5.10/drivers/mmc/core/
Dqueue.c260 req->rq_flags |= RQF_QUIET; in mmc_mq_queue_rq()
312 if (!(req->rq_flags & RQF_DONTPREP)) { in mmc_mq_queue_rq()
314 req->rq_flags |= RQF_DONTPREP; in mmc_mq_queue_rq()
/Linux-v5.10/drivers/md/
Ddm-rq.c269 if (rq->rq_flags & RQF_FAILED) in dm_softirq_done()
296 rq->rq_flags |= RQF_FAILED; in dm_kill_unmapped_request()
312 clone->rq_flags |= RQF_IO_STAT; in dm_dispatch_clone_request()
/Linux-v5.10/include/linux/
Dblkdev.h136 req_flags_t rq_flags; member
680 return (rq->rq_flags & RQF_STARTED) && !blk_rq_is_passthrough(rq); in blk_account_rq()
814 if (rq->rq_flags & RQF_NOMERGE_FLAGS) in rq_mergeable()
1039 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) in blk_rq_payload_bytes()
1050 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) in req_bvec()
1193 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) in blk_rq_nr_phys_segments()
1892 if (rq->rq_flags & RQF_ZONE_WRITE_LOCKED) in blk_req_zone_write_unlock()
/Linux-v5.10/drivers/scsi/device_handler/
Dscsi_dh_hp_sw.c167 req->rq_flags |= RQF_QUIET; in hp_sw_prep_fn()

123