Lines Matching refs:t

470 void blk_queue_stack_limits(struct request_queue *t, struct request_queue *b)  in blk_queue_stack_limits()  argument
472 blk_stack_limits(&t->limits, &b->limits, 0); in blk_queue_stack_limits()
497 int blk_stack_limits(struct queue_limits *t, struct queue_limits *b, in blk_stack_limits() argument
502 t->max_sectors = min_not_zero(t->max_sectors, b->max_sectors); in blk_stack_limits()
503 t->max_hw_sectors = min_not_zero(t->max_hw_sectors, b->max_hw_sectors); in blk_stack_limits()
504 t->max_dev_sectors = min_not_zero(t->max_dev_sectors, b->max_dev_sectors); in blk_stack_limits()
505 t->max_write_same_sectors = min(t->max_write_same_sectors, in blk_stack_limits()
507 t->max_write_zeroes_sectors = min(t->max_write_zeroes_sectors, in blk_stack_limits()
509 t->bounce_pfn = min_not_zero(t->bounce_pfn, b->bounce_pfn); in blk_stack_limits()
511 t->seg_boundary_mask = min_not_zero(t->seg_boundary_mask, in blk_stack_limits()
513 t->virt_boundary_mask = min_not_zero(t->virt_boundary_mask, in blk_stack_limits()
516 t->max_segments = min_not_zero(t->max_segments, b->max_segments); in blk_stack_limits()
517 t->max_discard_segments = min_not_zero(t->max_discard_segments, in blk_stack_limits()
519 t->max_integrity_segments = min_not_zero(t->max_integrity_segments, in blk_stack_limits()
522 t->max_segment_size = min_not_zero(t->max_segment_size, in blk_stack_limits()
525 t->misaligned |= b->misaligned; in blk_stack_limits()
532 if (t->alignment_offset != alignment) { in blk_stack_limits()
534 top = max(t->physical_block_size, t->io_min) in blk_stack_limits()
535 + t->alignment_offset; in blk_stack_limits()
540 t->misaligned = 1; in blk_stack_limits()
545 t->logical_block_size = max(t->logical_block_size, in blk_stack_limits()
548 t->physical_block_size = max(t->physical_block_size, in blk_stack_limits()
551 t->io_min = max(t->io_min, b->io_min); in blk_stack_limits()
552 t->io_opt = lcm_not_zero(t->io_opt, b->io_opt); in blk_stack_limits()
555 if (t->physical_block_size & (t->logical_block_size - 1)) { in blk_stack_limits()
556 t->physical_block_size = t->logical_block_size; in blk_stack_limits()
557 t->misaligned = 1; in blk_stack_limits()
562 if (t->io_min & (t->physical_block_size - 1)) { in blk_stack_limits()
563 t->io_min = t->physical_block_size; in blk_stack_limits()
564 t->misaligned = 1; in blk_stack_limits()
569 if (t->io_opt & (t->physical_block_size - 1)) { in blk_stack_limits()
570 t->io_opt = 0; in blk_stack_limits()
571 t->misaligned = 1; in blk_stack_limits()
575 t->raid_partial_stripes_expensive = in blk_stack_limits()
576 max(t->raid_partial_stripes_expensive, in blk_stack_limits()
580 t->alignment_offset = lcm_not_zero(t->alignment_offset, alignment) in blk_stack_limits()
581 % max(t->physical_block_size, t->io_min); in blk_stack_limits()
584 if (t->alignment_offset & (t->logical_block_size - 1)) { in blk_stack_limits()
585 t->misaligned = 1; in blk_stack_limits()
593 if (t->discard_granularity != 0 && in blk_stack_limits()
594 t->discard_alignment != alignment) { in blk_stack_limits()
595 top = t->discard_granularity + t->discard_alignment; in blk_stack_limits()
600 t->discard_misaligned = 1; in blk_stack_limits()
603 t->max_discard_sectors = min_not_zero(t->max_discard_sectors, in blk_stack_limits()
605 t->max_hw_discard_sectors = min_not_zero(t->max_hw_discard_sectors, in blk_stack_limits()
607 t->discard_granularity = max(t->discard_granularity, in blk_stack_limits()
609 t->discard_alignment = lcm_not_zero(t->discard_alignment, alignment) % in blk_stack_limits()
610 t->discard_granularity; in blk_stack_limits()
614 t->chunk_sectors = min_not_zero(t->chunk_sectors, in blk_stack_limits()
632 int bdev_stack_limits(struct queue_limits *t, struct block_device *bdev, in bdev_stack_limits() argument
639 return blk_stack_limits(t, &bq->limits, start); in bdev_stack_limits()
656 struct request_queue *t = disk->queue; in disk_stack_limits() local
658 if (bdev_stack_limits(&t->limits, bdev, offset >> 9) < 0) { in disk_stack_limits()