Lines Matching refs:disk
108 struct gendisk *disk; member
131 #define nbd_name(nbd) ((nbd)->disk->disk_name)
150 return disk_to_dev(nbd->disk); in nbd_to_dev()
197 struct gendisk *disk = dev_to_disk(dev); in pid_show() local
198 struct nbd_device *nbd = (struct nbd_device *)disk->private_data; in pid_show()
210 struct gendisk *disk = nbd->disk; in nbd_dev_remove() local
213 if (disk) { in nbd_dev_remove()
214 q = disk->queue; in nbd_dev_remove()
215 del_gendisk(disk); in nbd_dev_remove()
218 disk->private_data = NULL; in nbd_dev_remove()
219 put_disk(disk); in nbd_dev_remove()
272 set_capacity(nbd->disk, 0); in nbd_size_clear()
280 struct block_device *bdev = bdget_disk(nbd->disk, 0); in nbd_size_update()
283 nbd->disk->queue->limits.discard_granularity = config->blksize; in nbd_size_update()
284 nbd->disk->queue->limits.discard_alignment = config->blksize; in nbd_size_update()
285 blk_queue_max_discard_sectors(nbd->disk->queue, UINT_MAX); in nbd_size_update()
287 blk_queue_logical_block_size(nbd->disk->queue, config->blksize); in nbd_size_update()
288 blk_queue_physical_block_size(nbd->disk->queue, config->blksize); in nbd_size_update()
289 set_capacity(nbd->disk, config->bytesize >> 9); in nbd_size_update()
339 dev_warn(disk_to_dev(nbd->disk), "shutting down sockets\n"); in sock_shutdown()
414 dev_err_ratelimited(disk_to_dev(nbd->disk), in sock_xmit()
497 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_send_cmd()
547 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_send_cmd()
589 dev_err(disk_to_dev(nbd->disk), in nbd_send_cmd()
631 dev_err(disk_to_dev(nbd->disk), in nbd_read_stat()
637 dev_err(disk_to_dev(nbd->disk), "Wrong magic (0x%lx)\n", in nbd_read_stat()
649 dev_err(disk_to_dev(nbd->disk), "Unexpected reply (%d) %p\n", in nbd_read_stat()
657 dev_err(disk_to_dev(nbd->disk), "Double reply on req %p, cmd_cookie %u, handle cookie %u\n", in nbd_read_stat()
663 dev_err(disk_to_dev(nbd->disk), "Raced with timeout on req %p\n", in nbd_read_stat()
669 dev_err(disk_to_dev(nbd->disk), "Other side returned error (%d)\n", in nbd_read_stat()
685 dev_err(disk_to_dev(nbd->disk), "Receive data failed (result %d)\n", in nbd_read_stat()
749 blk_mq_quiesce_queue(nbd->disk->queue); in nbd_clear_que()
751 blk_mq_unquiesce_queue(nbd->disk->queue); in nbd_clear_que()
752 dev_dbg(disk_to_dev(nbd->disk), "queue cleared\n"); in nbd_clear_que()
766 dev_err_ratelimited(disk_to_dev(nbd->disk), in find_fallback()
789 dev_err_ratelimited(disk_to_dev(nbd->disk), in find_fallback()
819 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_handle_cmd()
827 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_handle_cmd()
877 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_handle_cmd()
942 dev_err(disk_to_dev(nbd->disk), in nbd_add_socket()
1047 set_disk_ro(nbd->disk, true); in nbd_parse_flags()
1049 set_disk_ro(nbd->disk, false); in nbd_parse_flags()
1051 blk_queue_flag_set(QUEUE_FLAG_DISCARD, nbd->disk->queue); in nbd_parse_flags()
1054 blk_queue_write_cache(nbd->disk->queue, true, true); in nbd_parse_flags()
1056 blk_queue_write_cache(nbd->disk->queue, true, false); in nbd_parse_flags()
1059 blk_queue_write_cache(nbd->disk->queue, false, false); in nbd_parse_flags()
1080 dev_err(disk_to_dev(nbd->disk), in send_disconnects()
1090 dev_info(disk_to_dev(nbd->disk), "NBD_DISCONNECT\n"); in nbd_disconnect()
1112 device_remove_file(disk_to_dev(nbd->disk), &pid_attr); in nbd_config_put()
1127 nbd->disk->queue->limits.discard_granularity = 0; in nbd_config_put()
1128 nbd->disk->queue->limits.discard_alignment = 0; in nbd_config_put()
1129 blk_queue_max_discard_sectors(nbd->disk->queue, UINT_MAX); in nbd_config_put()
1130 blk_queue_flag_clear(QUEUE_FLAG_DISCARD, nbd->disk->queue); in nbd_config_put()
1150 dev_err(disk_to_dev(nbd->disk), "server does not support multiple connections per device.\n"); in nbd_start_device()
1159 error = device_create_file(disk_to_dev(nbd->disk), &pid_attr); in nbd_start_device()
1161 dev_err(disk_to_dev(nbd->disk), "device_create_file failed!\n"); in nbd_start_device()
1258 blk_queue_rq_timeout(nbd->disk->queue, arg * HZ); in __nbd_ioctl()
1370 static void nbd_release(struct gendisk *disk, fmode_t mode) in nbd_release() argument
1372 struct nbd_device *nbd = disk->private_data; in nbd_release()
1373 struct block_device *bdev = bdget_disk(disk, 0); in nbd_release()
1541 struct gendisk *disk; in nbd_dev_add() local
1549 disk = alloc_disk(1 << part_shift); in nbd_dev_add()
1550 if (!disk) in nbd_dev_add()
1567 nbd->disk = disk; in nbd_dev_add()
1586 disk->queue = q; in nbd_dev_add()
1591 blk_queue_flag_set(QUEUE_FLAG_NONROT, disk->queue); in nbd_dev_add()
1592 blk_queue_flag_clear(QUEUE_FLAG_ADD_RANDOM, disk->queue); in nbd_dev_add()
1593 disk->queue->limits.discard_granularity = 0; in nbd_dev_add()
1594 disk->queue->limits.discard_alignment = 0; in nbd_dev_add()
1595 blk_queue_max_discard_sectors(disk->queue, 0); in nbd_dev_add()
1596 blk_queue_max_segment_size(disk->queue, UINT_MAX); in nbd_dev_add()
1597 blk_queue_max_segments(disk->queue, USHRT_MAX); in nbd_dev_add()
1598 blk_queue_max_hw_sectors(disk->queue, 65536); in nbd_dev_add()
1599 disk->queue->limits.max_sectors = 256; in nbd_dev_add()
1605 disk->major = NBD_MAJOR; in nbd_dev_add()
1606 disk->first_minor = index << part_shift; in nbd_dev_add()
1607 disk->fops = &nbd_fops; in nbd_dev_add()
1608 disk->private_data = nbd; in nbd_dev_add()
1609 sprintf(disk->disk_name, "nbd%d", index); in nbd_dev_add()
1610 add_disk(disk); in nbd_dev_add()
1619 put_disk(disk); in nbd_dev_add()
1764 blk_queue_rq_timeout(nbd->disk->queue, timeout * HZ); in nbd_genl_connect()
1930 blk_queue_rq_timeout(nbd->disk->queue, timeout * HZ); in nbd_genl_reconfigure()