Lines Matching refs:t
521 void blk_queue_stack_limits(struct request_queue *t, struct request_queue *b) in blk_queue_stack_limits() argument
523 blk_stack_limits(&t->limits, &b->limits, 0); in blk_queue_stack_limits()
548 int blk_stack_limits(struct queue_limits *t, struct queue_limits *b, in blk_stack_limits() argument
553 t->max_sectors = min_not_zero(t->max_sectors, b->max_sectors); in blk_stack_limits()
554 t->max_hw_sectors = min_not_zero(t->max_hw_sectors, b->max_hw_sectors); in blk_stack_limits()
555 t->max_dev_sectors = min_not_zero(t->max_dev_sectors, b->max_dev_sectors); in blk_stack_limits()
556 t->max_write_same_sectors = min(t->max_write_same_sectors, in blk_stack_limits()
558 t->max_write_zeroes_sectors = min(t->max_write_zeroes_sectors, in blk_stack_limits()
560 t->bounce_pfn = min_not_zero(t->bounce_pfn, b->bounce_pfn); in blk_stack_limits()
562 t->seg_boundary_mask = min_not_zero(t->seg_boundary_mask, in blk_stack_limits()
564 t->virt_boundary_mask = min_not_zero(t->virt_boundary_mask, in blk_stack_limits()
567 t->max_segments = min_not_zero(t->max_segments, b->max_segments); in blk_stack_limits()
568 t->max_discard_segments = min_not_zero(t->max_discard_segments, in blk_stack_limits()
570 t->max_integrity_segments = min_not_zero(t->max_integrity_segments, in blk_stack_limits()
573 t->max_segment_size = min_not_zero(t->max_segment_size, in blk_stack_limits()
576 t->misaligned |= b->misaligned; in blk_stack_limits()
583 if (t->alignment_offset != alignment) { in blk_stack_limits()
585 top = max(t->physical_block_size, t->io_min) in blk_stack_limits()
586 + t->alignment_offset; in blk_stack_limits()
591 t->misaligned = 1; in blk_stack_limits()
596 t->logical_block_size = max(t->logical_block_size, in blk_stack_limits()
599 t->physical_block_size = max(t->physical_block_size, in blk_stack_limits()
602 t->io_min = max(t->io_min, b->io_min); in blk_stack_limits()
603 t->io_opt = lcm_not_zero(t->io_opt, b->io_opt); in blk_stack_limits()
605 t->cluster &= b->cluster; in blk_stack_limits()
608 if (t->physical_block_size & (t->logical_block_size - 1)) { in blk_stack_limits()
609 t->physical_block_size = t->logical_block_size; in blk_stack_limits()
610 t->misaligned = 1; in blk_stack_limits()
615 if (t->io_min & (t->physical_block_size - 1)) { in blk_stack_limits()
616 t->io_min = t->physical_block_size; in blk_stack_limits()
617 t->misaligned = 1; in blk_stack_limits()
622 if (t->io_opt & (t->physical_block_size - 1)) { in blk_stack_limits()
623 t->io_opt = 0; in blk_stack_limits()
624 t->misaligned = 1; in blk_stack_limits()
628 t->raid_partial_stripes_expensive = in blk_stack_limits()
629 max(t->raid_partial_stripes_expensive, in blk_stack_limits()
633 t->alignment_offset = lcm_not_zero(t->alignment_offset, alignment) in blk_stack_limits()
634 % max(t->physical_block_size, t->io_min); in blk_stack_limits()
637 if (t->alignment_offset & (t->logical_block_size - 1)) { in blk_stack_limits()
638 t->misaligned = 1; in blk_stack_limits()
646 if (t->discard_granularity != 0 && in blk_stack_limits()
647 t->discard_alignment != alignment) { in blk_stack_limits()
648 top = t->discard_granularity + t->discard_alignment; in blk_stack_limits()
653 t->discard_misaligned = 1; in blk_stack_limits()
656 t->max_discard_sectors = min_not_zero(t->max_discard_sectors, in blk_stack_limits()
658 t->max_hw_discard_sectors = min_not_zero(t->max_hw_discard_sectors, in blk_stack_limits()
660 t->discard_granularity = max(t->discard_granularity, in blk_stack_limits()
662 t->discard_alignment = lcm_not_zero(t->discard_alignment, alignment) % in blk_stack_limits()
663 t->discard_granularity; in blk_stack_limits()
667 t->chunk_sectors = min_not_zero(t->chunk_sectors, in blk_stack_limits()
685 int bdev_stack_limits(struct queue_limits *t, struct block_device *bdev, in bdev_stack_limits() argument
692 return blk_stack_limits(t, &bq->limits, start); in bdev_stack_limits()
709 struct request_queue *t = disk->queue; in disk_stack_limits() local
711 if (bdev_stack_limits(&t->limits, bdev, offset >> 9) < 0) { in disk_stack_limits()