Lines Matching defs:request_queue
433 struct request_queue { struct
437 struct list_head queue_head;
438 struct request *last_merge;
439 struct elevator_queue *elevator;
440 int nr_rqs[2]; /* # allocated [a]sync rqs */
441 int nr_rqs_elvpriv; /* # allocated rqs w/ elvpriv */
443 struct blk_queue_stats *stats;
444 struct rq_qos *rq_qos;
452 struct request_list root_rl;
454 request_fn_proc *request_fn;
455 make_request_fn *make_request_fn;
456 poll_q_fn *poll_fn;
457 prep_rq_fn *prep_rq_fn;
458 unprep_rq_fn *unprep_rq_fn;
459 softirq_done_fn *softirq_done_fn;
460 rq_timed_out_fn *rq_timed_out_fn;
461 dma_drain_needed_fn *dma_drain_needed;
462 lld_busy_fn *lld_busy_fn;
464 init_rq_fn *init_rq_fn;
466 exit_rq_fn *exit_rq_fn;
468 void (*initialize_rq_fn)(struct request *rq);
470 const struct blk_mq_ops *mq_ops;
472 unsigned int *mq_map;
475 struct blk_mq_ctx __percpu *queue_ctx;
476 unsigned int nr_queues;
478 unsigned int queue_depth;
481 struct blk_mq_hw_ctx **queue_hw_ctx;
482 unsigned int nr_hw_queues;
487 sector_t end_sector;
488 struct request *boundary_rq;
493 struct delayed_work delay_work;
495 struct backing_dev_info *backing_dev_info;
501 void *queuedata;
506 unsigned long queue_flags;
512 int id;
517 gfp_t bounce_gfp;
524 spinlock_t __queue_lock;
525 spinlock_t *queue_lock;
530 struct kobject kobj;
535 struct kobject mq_kobj;
538 struct blk_integrity integrity;
542 struct device *dev;
543 int rpm_status;
544 unsigned int nr_pending;
550 unsigned long nr_requests; /* Max # of requests */
551 unsigned int nr_congestion_on;
552 unsigned int nr_congestion_off;
553 unsigned int nr_batching;
555 unsigned int dma_drain_size;
556 void *dma_drain_buffer;
557 unsigned int dma_pad_mask;
558 unsigned int dma_alignment;
560 struct blk_queue_tag *queue_tags;
562 unsigned int nr_sorted;
563 unsigned int in_flight[2];
570 unsigned int request_fn_active;
572 unsigned int rq_timeout;
573 int poll_nsec;
575 struct blk_stat_callback *poll_cb;
576 struct blk_rq_stat poll_stat[BLK_MQ_POLL_STATS_BKTS];
578 struct timer_list timeout;
579 struct work_struct timeout_work;
580 struct list_head timeout_list;
582 struct list_head icq_list;
585 struct blkcg_gq *root_blkg;
586 struct list_head blkg_list;
589 struct queue_limits limits;
609 unsigned int nr_zones;
610 unsigned long *seq_zones_bitmap;
611 unsigned long *seq_zones_wlock;
617 unsigned int sg_timeout;
618 unsigned int sg_reserved_size;
619 int node;
621 struct blk_trace *blk_trace;
622 struct mutex blk_trace_mutex;
627 struct blk_flush_queue *fq;
629 struct list_head requeue_list;
630 spinlock_t requeue_lock;
631 struct delayed_work requeue_work;
633 struct mutex sysfs_lock;
635 int bypass_depth;
636 atomic_t mq_freeze_depth;
639 bsg_job_fn *bsg_job_fn;
640 struct bsg_class_device bsg_dev;
645 struct throtl_data *td;
647 struct rcu_head rcu_head;
648 wait_queue_head_t mq_freeze_wq;
649 struct percpu_ref q_usage_counter;
650 struct list_head all_q_node;
652 struct blk_mq_tag_set *tag_set;
653 struct list_head tag_set_list;
654 struct bio_set bio_split;
657 struct dentry *debugfs_dir;
658 struct dentry *sched_debugfs_dir;
661 bool mq_sysfs_init_done;
663 size_t cmd_size;
664 void *rq_alloc_data;
666 struct work_struct release_work;
669 u64 write_hints[BLK_MAX_WRITE_HINTS];