Lines Matching refs:ubd_dev
266 struct ubd *ubd_dev; in ubd_setup_common() local
319 ubd_dev = &ubd_devs[n]; in ubd_setup_common()
320 if(ubd_dev->file != NULL){ in ubd_setup_common()
338 ubd_dev->no_cow = 1; in ubd_setup_common()
341 ubd_dev->shared = 1; in ubd_setup_common()
344 ubd_dev->no_trim = 1; in ubd_setup_common()
370 if(ubd_dev->no_cow){ in ubd_setup_common()
380 ubd_dev->file = str; in ubd_setup_common()
381 ubd_dev->cow.file = backing_file; in ubd_setup_common()
382 ubd_dev->boot_openflags = flags; in ubd_setup_common()
556 static inline int ubd_file_size(struct ubd *ubd_dev, __u64 *size_out) in ubd_file_size() argument
570 if (ubd_dev->file && ubd_dev->cow.file) { in ubd_file_size()
571 file = ubd_dev->cow.file; in ubd_file_size()
576 fd = os_open_file(ubd_dev->file, of_read(OPENFLAGS()), 0); in ubd_file_size()
585 file = ubd_dev->file; in ubd_file_size()
775 static void ubd_close_dev(struct ubd *ubd_dev) in ubd_close_dev() argument
777 os_close_file(ubd_dev->fd); in ubd_close_dev()
778 if(ubd_dev->cow.file == NULL) in ubd_close_dev()
781 os_close_file(ubd_dev->cow.fd); in ubd_close_dev()
782 vfree(ubd_dev->cow.bitmap); in ubd_close_dev()
783 ubd_dev->cow.bitmap = NULL; in ubd_close_dev()
786 static int ubd_open_dev(struct ubd *ubd_dev) in ubd_open_dev() argument
793 ubd_dev->openflags = ubd_dev->boot_openflags; in ubd_open_dev()
795 create_ptr = (ubd_dev->cow.file != NULL) ? &create_cow : NULL; in ubd_open_dev()
796 back_ptr = ubd_dev->no_cow ? NULL : &ubd_dev->cow.file; in ubd_open_dev()
798 fd = open_ubd_file(ubd_dev->file, &ubd_dev->openflags, ubd_dev->shared, in ubd_open_dev()
799 back_ptr, &ubd_dev->cow.bitmap_offset, in ubd_open_dev()
800 &ubd_dev->cow.bitmap_len, &ubd_dev->cow.data_offset, in ubd_open_dev()
804 fd = create_cow_file(ubd_dev->file, ubd_dev->cow.file, in ubd_open_dev()
805 ubd_dev->openflags, SECTOR_SIZE, PAGE_SIZE, in ubd_open_dev()
806 &ubd_dev->cow.bitmap_offset, in ubd_open_dev()
807 &ubd_dev->cow.bitmap_len, in ubd_open_dev()
808 &ubd_dev->cow.data_offset); in ubd_open_dev()
811 "\"%s\"\n", ubd_dev->file, ubd_dev->cow.file); in ubd_open_dev()
816 printk("Failed to open '%s', errno = %d\n", ubd_dev->file, in ubd_open_dev()
820 ubd_dev->fd = fd; in ubd_open_dev()
822 if(ubd_dev->cow.file != NULL){ in ubd_open_dev()
823 blk_queue_max_hw_sectors(ubd_dev->queue, 8 * sizeof(long)); in ubd_open_dev()
826 ubd_dev->cow.bitmap = vmalloc(ubd_dev->cow.bitmap_len); in ubd_open_dev()
827 if(ubd_dev->cow.bitmap == NULL){ in ubd_open_dev()
833 err = read_cow_bitmap(ubd_dev->fd, ubd_dev->cow.bitmap, in ubd_open_dev()
834 ubd_dev->cow.bitmap_offset, in ubd_open_dev()
835 ubd_dev->cow.bitmap_len); in ubd_open_dev()
839 flags = ubd_dev->openflags; in ubd_open_dev()
841 err = open_ubd_file(ubd_dev->cow.file, &flags, ubd_dev->shared, NULL, in ubd_open_dev()
844 ubd_dev->cow.fd = err; in ubd_open_dev()
846 if (ubd_dev->no_trim == 0) { in ubd_open_dev()
847 ubd_dev->queue->limits.discard_granularity = SECTOR_SIZE; in ubd_open_dev()
848 ubd_dev->queue->limits.discard_alignment = SECTOR_SIZE; in ubd_open_dev()
849 blk_queue_max_discard_sectors(ubd_dev->queue, UBD_MAX_REQUEST); in ubd_open_dev()
850 blk_queue_max_write_zeroes_sectors(ubd_dev->queue, UBD_MAX_REQUEST); in ubd_open_dev()
851 blk_queue_flag_set(QUEUE_FLAG_DISCARD, ubd_dev->queue); in ubd_open_dev()
853 blk_queue_flag_set(QUEUE_FLAG_NONROT, ubd_dev->queue); in ubd_open_dev()
856 os_close_file(ubd_dev->fd); in ubd_open_dev()
862 struct ubd *ubd_dev = dev_get_drvdata(dev); in ubd_device_release() local
864 blk_cleanup_queue(ubd_dev->queue); in ubd_device_release()
865 blk_mq_free_tag_set(&ubd_dev->tag_set); in ubd_device_release()
866 *ubd_dev = ((struct ubd) DEFAULT_UBD); in ubd_device_release()
914 struct ubd *ubd_dev = &ubd_devs[n]; in ubd_add() local
917 if(ubd_dev->file == NULL) in ubd_add()
920 err = ubd_file_size(ubd_dev, &ubd_dev->size); in ubd_add()
926 ubd_dev->size = ROUND_BLOCK(ubd_dev->size); in ubd_add()
928 ubd_dev->tag_set.ops = &ubd_mq_ops; in ubd_add()
929 ubd_dev->tag_set.queue_depth = 64; in ubd_add()
930 ubd_dev->tag_set.numa_node = NUMA_NO_NODE; in ubd_add()
931 ubd_dev->tag_set.flags = BLK_MQ_F_SHOULD_MERGE; in ubd_add()
932 ubd_dev->tag_set.driver_data = ubd_dev; in ubd_add()
933 ubd_dev->tag_set.nr_hw_queues = 1; in ubd_add()
935 err = blk_mq_alloc_tag_set(&ubd_dev->tag_set); in ubd_add()
939 ubd_dev->queue = blk_mq_init_queue(&ubd_dev->tag_set); in ubd_add()
940 if (IS_ERR(ubd_dev->queue)) { in ubd_add()
941 err = PTR_ERR(ubd_dev->queue); in ubd_add()
945 ubd_dev->queue->queuedata = ubd_dev; in ubd_add()
946 blk_queue_write_cache(ubd_dev->queue, true, false); in ubd_add()
948 blk_queue_max_segments(ubd_dev->queue, MAX_SG); in ubd_add()
949 err = ubd_disk_register(UBD_MAJOR, ubd_dev->size, n, &ubd_gendisk[n]); in ubd_add()
956 ubd_disk_register(fake_major, ubd_dev->size, n, in ubd_add()
971 blk_mq_free_tag_set(&ubd_dev->tag_set); in ubd_add()
972 if (!(IS_ERR(ubd_dev->queue))) in ubd_add()
973 blk_cleanup_queue(ubd_dev->queue); in ubd_add()
1016 struct ubd *ubd_dev; in ubd_get_config() local
1025 ubd_dev = &ubd_devs[n]; in ubd_get_config()
1028 if(ubd_dev->file == NULL){ in ubd_get_config()
1033 CONFIG_CHUNK(str, size, len, ubd_dev->file, 0); in ubd_get_config()
1035 if(ubd_dev->cow.file != NULL){ in ubd_get_config()
1037 CONFIG_CHUNK(str, size, len, ubd_dev->cow.file, 1); in ubd_get_config()
1059 struct ubd *ubd_dev; in ubd_remove() local
1064 ubd_dev = &ubd_devs[n]; in ubd_remove()
1066 if(ubd_dev->file == NULL) in ubd_remove()
1071 if(ubd_dev->count > 0) in ubd_remove()
1087 platform_device_unregister(&ubd_dev->pdev); in ubd_remove()
1115 struct ubd *ubd_dev = &ubd_devs[0]; in ubd0_init() local
1118 if(ubd_dev->file == NULL) in ubd0_init()
1119 ubd_dev->file = "root_fs"; in ubd0_init()
1217 struct ubd *ubd_dev = disk->private_data; in ubd_open() local
1221 if(ubd_dev->count == 0){ in ubd_open()
1222 err = ubd_open_dev(ubd_dev); in ubd_open()
1225 disk->disk_name, ubd_dev->file, -err); in ubd_open()
1229 ubd_dev->count++; in ubd_open()
1230 set_disk_ro(disk, !ubd_dev->openflags.w); in ubd_open()
1245 struct ubd *ubd_dev = disk->private_data; in ubd_release() local
1248 if(--ubd_dev->count == 0) in ubd_release()
1249 ubd_close_dev(ubd_dev); in ubd_release()
1379 struct ubd *ubd_dev = hctx->queue->queuedata; in ubd_queue_rq() local
1385 spin_lock_irq(&ubd_dev->lock); in ubd_queue_rq()
1405 spin_unlock_irq(&ubd_dev->lock); in ubd_queue_rq()
1419 struct ubd *ubd_dev = bdev->bd_disk->private_data; in ubd_getgeo() local
1423 geo->cylinders = ubd_dev->size / (128 * 32 * 512); in ubd_getgeo()
1430 struct ubd *ubd_dev = bdev->bd_disk->private_data; in ubd_ioctl() local
1437 ubd_id[ATA_ID_CYLS] = ubd_dev->size / (128 * 32 * 512); in ubd_ioctl()