Lines Matching refs:request_queue

25 void blk_queue_rq_timeout(struct request_queue *q, unsigned int timeout)  in blk_queue_rq_timeout()
111 void blk_queue_make_request(struct request_queue *q, make_request_fn *mfn) in blk_queue_make_request()
136 void blk_queue_bounce_limit(struct request_queue *q, u64 max_addr) in blk_queue_bounce_limit()
183 void blk_queue_max_hw_sectors(struct request_queue *q, unsigned int max_hw_sectors) in blk_queue_max_hw_sectors()
215 void blk_queue_chunk_sectors(struct request_queue *q, unsigned int chunk_sectors) in blk_queue_chunk_sectors()
227 void blk_queue_max_discard_sectors(struct request_queue *q, in blk_queue_max_discard_sectors()
240 void blk_queue_max_write_same_sectors(struct request_queue *q, in blk_queue_max_write_same_sectors()
253 void blk_queue_max_write_zeroes_sectors(struct request_queue *q, in blk_queue_max_write_zeroes_sectors()
269 void blk_queue_max_segments(struct request_queue *q, unsigned short max_segments) in blk_queue_max_segments()
290 void blk_queue_max_discard_segments(struct request_queue *q, in blk_queue_max_discard_segments()
306 void blk_queue_max_segment_size(struct request_queue *q, unsigned int max_size) in blk_queue_max_segment_size()
331 void blk_queue_logical_block_size(struct request_queue *q, unsigned short size) in blk_queue_logical_block_size()
353 void blk_queue_physical_block_size(struct request_queue *q, unsigned int size) in blk_queue_physical_block_size()
376 void blk_queue_alignment_offset(struct request_queue *q, unsigned int offset) in blk_queue_alignment_offset()
421 void blk_queue_io_min(struct request_queue *q, unsigned int min) in blk_queue_io_min()
459 void blk_queue_io_opt(struct request_queue *q, unsigned int opt) in blk_queue_io_opt()
470 void blk_queue_stack_limits(struct request_queue *t, struct request_queue *b) in blk_queue_stack_limits()
635 struct request_queue *bq = bdev_get_queue(bdev); in bdev_stack_limits()
656 struct request_queue *t = disk->queue; in disk_stack_limits()
680 void blk_queue_update_dma_pad(struct request_queue *q, unsigned int mask) in blk_queue_update_dma_pad()
708 int blk_queue_dma_drain(struct request_queue *q, in blk_queue_dma_drain()
729 void blk_queue_segment_boundary(struct request_queue *q, unsigned long mask) in blk_queue_segment_boundary()
746 void blk_queue_virt_boundary(struct request_queue *q, unsigned long mask) in blk_queue_virt_boundary()
771 void blk_queue_dma_alignment(struct request_queue *q, int mask) in blk_queue_dma_alignment()
791 void blk_queue_update_dma_alignment(struct request_queue *q, int mask) in blk_queue_update_dma_alignment()
806 void blk_set_queue_depth(struct request_queue *q, unsigned int depth) in blk_set_queue_depth()
821 void blk_queue_write_cache(struct request_queue *q, bool wc, bool fua) in blk_queue_write_cache()
845 void blk_queue_required_elevator_features(struct request_queue *q, in blk_queue_required_elevator_features()
859 bool blk_queue_can_use_dma_map_merging(struct request_queue *q, in blk_queue_can_use_dma_map_merging()