Lines Matching refs:disk

114 	struct gendisk *disk;  member
142 #define nbd_name(nbd) ((nbd)->disk->disk_name)
162 return disk_to_dev(nbd->disk); in nbd_to_dev()
209 struct gendisk *disk = dev_to_disk(dev); in pid_show() local
210 struct nbd_device *nbd = (struct nbd_device *)disk->private_data; in pid_show()
222 struct gendisk *disk = nbd->disk; in nbd_dev_remove() local
225 if (disk) { in nbd_dev_remove()
226 q = disk->queue; in nbd_dev_remove()
227 del_gendisk(disk); in nbd_dev_remove()
230 disk->private_data = NULL; in nbd_dev_remove()
231 put_disk(disk); in nbd_dev_remove()
294 set_capacity(nbd->disk, 0); in nbd_size_clear()
302 struct block_device *bdev = bdget_disk(nbd->disk, 0); in nbd_size_update()
306 nbd->disk->queue->limits.discard_granularity = config->blksize; in nbd_size_update()
307 nbd->disk->queue->limits.discard_alignment = config->blksize; in nbd_size_update()
308 blk_queue_max_discard_sectors(nbd->disk->queue, UINT_MAX); in nbd_size_update()
310 blk_queue_logical_block_size(nbd->disk->queue, config->blksize); in nbd_size_update()
311 blk_queue_physical_block_size(nbd->disk->queue, config->blksize); in nbd_size_update()
312 set_capacity(nbd->disk, nr_sectors); in nbd_size_update()
319 set_bit(GD_NEED_PART_SCAN, &nbd->disk->state); in nbd_size_update()
364 dev_warn(disk_to_dev(nbd->disk), "shutting down sockets\n"); in sock_shutdown()
484 dev_err_ratelimited(disk_to_dev(nbd->disk), in sock_xmit()
554 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_send_cmd()
612 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_send_cmd()
653 dev_err(disk_to_dev(nbd->disk), in nbd_send_cmd()
696 dev_err(disk_to_dev(nbd->disk), in nbd_read_stat()
702 dev_err(disk_to_dev(nbd->disk), "Wrong magic (0x%lx)\n", in nbd_read_stat()
714 dev_err(disk_to_dev(nbd->disk), "Unexpected reply (%d) %p\n", in nbd_read_stat()
723 dev_err(disk_to_dev(nbd->disk), "Double reply on req %p, cmd_cookie %u, handle cookie %u\n", in nbd_read_stat()
729 dev_err(disk_to_dev(nbd->disk), "Command already handled %p\n", in nbd_read_stat()
735 dev_err(disk_to_dev(nbd->disk), "Raced with timeout on req %p\n", in nbd_read_stat()
741 dev_err(disk_to_dev(nbd->disk), "Other side returned error (%d)\n", in nbd_read_stat()
756 dev_err(disk_to_dev(nbd->disk), "Receive data failed (result %d)\n", in nbd_read_stat()
826 blk_mq_quiesce_queue(nbd->disk->queue); in nbd_clear_que()
828 blk_mq_unquiesce_queue(nbd->disk->queue); in nbd_clear_que()
829 dev_dbg(disk_to_dev(nbd->disk), "queue cleared\n"); in nbd_clear_que()
843 dev_err_ratelimited(disk_to_dev(nbd->disk), in find_fallback()
866 dev_err_ratelimited(disk_to_dev(nbd->disk), in find_fallback()
896 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_handle_cmd()
904 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_handle_cmd()
954 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_handle_cmd()
1010 dev_err(disk_to_dev(nbd->disk), "Unsupported socket: shutdown callout must be supported.\n"); in nbd_get_socket()
1039 dev_err(disk_to_dev(nbd->disk), in nbd_add_socket()
1150 set_disk_ro(nbd->disk, true); in nbd_parse_flags()
1152 set_disk_ro(nbd->disk, false); in nbd_parse_flags()
1154 blk_queue_flag_set(QUEUE_FLAG_DISCARD, nbd->disk->queue); in nbd_parse_flags()
1157 blk_queue_write_cache(nbd->disk->queue, true, true); in nbd_parse_flags()
1159 blk_queue_write_cache(nbd->disk->queue, true, false); in nbd_parse_flags()
1162 blk_queue_write_cache(nbd->disk->queue, false, false); in nbd_parse_flags()
1183 dev_err(disk_to_dev(nbd->disk), in send_disconnects()
1193 dev_info(disk_to_dev(nbd->disk), "NBD_DISCONNECT\n"); in nbd_disconnect()
1216 device_remove_file(disk_to_dev(nbd->disk), &pid_attr); in nbd_config_put()
1235 nbd->disk->queue->limits.discard_granularity = 0; in nbd_config_put()
1236 nbd->disk->queue->limits.discard_alignment = 0; in nbd_config_put()
1237 blk_queue_max_discard_sectors(nbd->disk->queue, UINT_MAX); in nbd_config_put()
1238 blk_queue_flag_clear(QUEUE_FLAG_DISCARD, nbd->disk->queue); in nbd_config_put()
1258 dev_err(disk_to_dev(nbd->disk), "server does not support multiple connections per device.\n"); in nbd_start_device()
1266 dev_err(disk_to_dev(nbd->disk), "Could not allocate knbd recv work queue.\n"); in nbd_start_device()
1275 error = device_create_file(disk_to_dev(nbd->disk), &pid_attr); in nbd_start_device()
1277 dev_err(disk_to_dev(nbd->disk), "device_create_file failed!\n"); in nbd_start_device()
1326 set_bit(GD_NEED_PART_SCAN, &nbd->disk->state); in nbd_start_device_ioctl()
1367 blk_queue_rq_timeout(nbd->disk->queue, timeout * HZ); in nbd_set_cmd_timeout()
1369 blk_queue_rq_timeout(nbd->disk->queue, 30 * HZ); in nbd_set_cmd_timeout()
1513 static void nbd_release(struct gendisk *disk, fmode_t mode) in nbd_release() argument
1515 struct nbd_device *nbd = disk->private_data; in nbd_release()
1516 struct block_device *bdev = bdget_disk(disk, 0); in nbd_release()
1685 struct gendisk *disk; in nbd_dev_add() local
1693 disk = alloc_disk(1 << part_shift); in nbd_dev_add()
1694 if (!disk) in nbd_dev_add()
1711 nbd->disk = disk; in nbd_dev_add()
1731 disk->queue = q; in nbd_dev_add()
1736 blk_queue_flag_set(QUEUE_FLAG_NONROT, disk->queue); in nbd_dev_add()
1737 blk_queue_flag_clear(QUEUE_FLAG_ADD_RANDOM, disk->queue); in nbd_dev_add()
1738 disk->queue->limits.discard_granularity = 0; in nbd_dev_add()
1739 disk->queue->limits.discard_alignment = 0; in nbd_dev_add()
1740 blk_queue_max_discard_sectors(disk->queue, 0); in nbd_dev_add()
1741 blk_queue_max_segment_size(disk->queue, UINT_MAX); in nbd_dev_add()
1742 blk_queue_max_segments(disk->queue, USHRT_MAX); in nbd_dev_add()
1743 blk_queue_max_hw_sectors(disk->queue, 65536); in nbd_dev_add()
1744 disk->queue->limits.max_sectors = 256; in nbd_dev_add()
1750 disk->major = NBD_MAJOR; in nbd_dev_add()
1751 disk->first_minor = index << part_shift; in nbd_dev_add()
1752 disk->fops = &nbd_fops; in nbd_dev_add()
1753 disk->private_data = nbd; in nbd_dev_add()
1754 sprintf(disk->disk_name, "nbd%d", index); in nbd_dev_add()
1755 add_disk(disk); in nbd_dev_add()
1764 put_disk(disk); in nbd_dev_add()