Lines Matching refs:ubd_dev

265 	struct ubd *ubd_dev;  in ubd_setup_common()  local
318 ubd_dev = &ubd_devs[n]; in ubd_setup_common()
319 if(ubd_dev->file != NULL){ in ubd_setup_common()
337 ubd_dev->no_cow = 1; in ubd_setup_common()
340 ubd_dev->shared = 1; in ubd_setup_common()
343 ubd_dev->no_trim = 1; in ubd_setup_common()
369 if(ubd_dev->no_cow){ in ubd_setup_common()
379 ubd_dev->file = str; in ubd_setup_common()
380 ubd_dev->cow.file = backing_file; in ubd_setup_common()
381 ubd_dev->boot_openflags = flags; in ubd_setup_common()
555 static inline int ubd_file_size(struct ubd *ubd_dev, __u64 *size_out) in ubd_file_size() argument
569 if (ubd_dev->file && ubd_dev->cow.file) { in ubd_file_size()
570 file = ubd_dev->cow.file; in ubd_file_size()
575 fd = os_open_file(ubd_dev->file, of_read(OPENFLAGS()), 0); in ubd_file_size()
584 file = ubd_dev->file; in ubd_file_size()
774 static void ubd_close_dev(struct ubd *ubd_dev) in ubd_close_dev() argument
776 os_close_file(ubd_dev->fd); in ubd_close_dev()
777 if(ubd_dev->cow.file == NULL) in ubd_close_dev()
780 os_close_file(ubd_dev->cow.fd); in ubd_close_dev()
781 vfree(ubd_dev->cow.bitmap); in ubd_close_dev()
782 ubd_dev->cow.bitmap = NULL; in ubd_close_dev()
785 static int ubd_open_dev(struct ubd *ubd_dev) in ubd_open_dev() argument
792 ubd_dev->openflags = ubd_dev->boot_openflags; in ubd_open_dev()
794 create_ptr = (ubd_dev->cow.file != NULL) ? &create_cow : NULL; in ubd_open_dev()
795 back_ptr = ubd_dev->no_cow ? NULL : &ubd_dev->cow.file; in ubd_open_dev()
797 fd = open_ubd_file(ubd_dev->file, &ubd_dev->openflags, ubd_dev->shared, in ubd_open_dev()
798 back_ptr, &ubd_dev->cow.bitmap_offset, in ubd_open_dev()
799 &ubd_dev->cow.bitmap_len, &ubd_dev->cow.data_offset, in ubd_open_dev()
803 fd = create_cow_file(ubd_dev->file, ubd_dev->cow.file, in ubd_open_dev()
804 ubd_dev->openflags, SECTOR_SIZE, PAGE_SIZE, in ubd_open_dev()
805 &ubd_dev->cow.bitmap_offset, in ubd_open_dev()
806 &ubd_dev->cow.bitmap_len, in ubd_open_dev()
807 &ubd_dev->cow.data_offset); in ubd_open_dev()
810 "\"%s\"\n", ubd_dev->file, ubd_dev->cow.file); in ubd_open_dev()
815 printk("Failed to open '%s', errno = %d\n", ubd_dev->file, in ubd_open_dev()
819 ubd_dev->fd = fd; in ubd_open_dev()
821 if(ubd_dev->cow.file != NULL){ in ubd_open_dev()
822 blk_queue_max_hw_sectors(ubd_dev->queue, 8 * sizeof(long)); in ubd_open_dev()
825 ubd_dev->cow.bitmap = vmalloc(ubd_dev->cow.bitmap_len); in ubd_open_dev()
826 if(ubd_dev->cow.bitmap == NULL){ in ubd_open_dev()
832 err = read_cow_bitmap(ubd_dev->fd, ubd_dev->cow.bitmap, in ubd_open_dev()
833 ubd_dev->cow.bitmap_offset, in ubd_open_dev()
834 ubd_dev->cow.bitmap_len); in ubd_open_dev()
838 flags = ubd_dev->openflags; in ubd_open_dev()
840 err = open_ubd_file(ubd_dev->cow.file, &flags, ubd_dev->shared, NULL, in ubd_open_dev()
843 ubd_dev->cow.fd = err; in ubd_open_dev()
845 if (ubd_dev->no_trim == 0) { in ubd_open_dev()
846 ubd_dev->queue->limits.discard_granularity = SECTOR_SIZE; in ubd_open_dev()
847 ubd_dev->queue->limits.discard_alignment = SECTOR_SIZE; in ubd_open_dev()
848 blk_queue_max_discard_sectors(ubd_dev->queue, UBD_MAX_REQUEST); in ubd_open_dev()
849 blk_queue_max_write_zeroes_sectors(ubd_dev->queue, UBD_MAX_REQUEST); in ubd_open_dev()
850 blk_queue_flag_set(QUEUE_FLAG_DISCARD, ubd_dev->queue); in ubd_open_dev()
852 blk_queue_flag_set(QUEUE_FLAG_NONROT, ubd_dev->queue); in ubd_open_dev()
855 os_close_file(ubd_dev->fd); in ubd_open_dev()
861 struct ubd *ubd_dev = dev_get_drvdata(dev); in ubd_device_release() local
863 blk_cleanup_queue(ubd_dev->queue); in ubd_device_release()
864 blk_mq_free_tag_set(&ubd_dev->tag_set); in ubd_device_release()
865 *ubd_dev = ((struct ubd) DEFAULT_UBD); in ubd_device_release()
913 struct ubd *ubd_dev = &ubd_devs[n]; in ubd_add() local
916 if(ubd_dev->file == NULL) in ubd_add()
919 err = ubd_file_size(ubd_dev, &ubd_dev->size); in ubd_add()
925 ubd_dev->size = ROUND_BLOCK(ubd_dev->size); in ubd_add()
927 ubd_dev->tag_set.ops = &ubd_mq_ops; in ubd_add()
928 ubd_dev->tag_set.queue_depth = 64; in ubd_add()
929 ubd_dev->tag_set.numa_node = NUMA_NO_NODE; in ubd_add()
930 ubd_dev->tag_set.flags = BLK_MQ_F_SHOULD_MERGE; in ubd_add()
931 ubd_dev->tag_set.driver_data = ubd_dev; in ubd_add()
932 ubd_dev->tag_set.nr_hw_queues = 1; in ubd_add()
934 err = blk_mq_alloc_tag_set(&ubd_dev->tag_set); in ubd_add()
938 ubd_dev->queue = blk_mq_init_queue(&ubd_dev->tag_set); in ubd_add()
939 if (IS_ERR(ubd_dev->queue)) { in ubd_add()
940 err = PTR_ERR(ubd_dev->queue); in ubd_add()
944 ubd_dev->queue->queuedata = ubd_dev; in ubd_add()
945 blk_queue_write_cache(ubd_dev->queue, true, false); in ubd_add()
947 blk_queue_max_segments(ubd_dev->queue, MAX_SG); in ubd_add()
948 err = ubd_disk_register(UBD_MAJOR, ubd_dev->size, n, &ubd_gendisk[n]); in ubd_add()
955 ubd_disk_register(fake_major, ubd_dev->size, n, in ubd_add()
970 blk_mq_free_tag_set(&ubd_dev->tag_set); in ubd_add()
971 if (!(IS_ERR(ubd_dev->queue))) in ubd_add()
972 blk_cleanup_queue(ubd_dev->queue); in ubd_add()
1015 struct ubd *ubd_dev; in ubd_get_config() local
1024 ubd_dev = &ubd_devs[n]; in ubd_get_config()
1027 if(ubd_dev->file == NULL){ in ubd_get_config()
1032 CONFIG_CHUNK(str, size, len, ubd_dev->file, 0); in ubd_get_config()
1034 if(ubd_dev->cow.file != NULL){ in ubd_get_config()
1036 CONFIG_CHUNK(str, size, len, ubd_dev->cow.file, 1); in ubd_get_config()
1058 struct ubd *ubd_dev; in ubd_remove() local
1063 ubd_dev = &ubd_devs[n]; in ubd_remove()
1065 if(ubd_dev->file == NULL) in ubd_remove()
1070 if(ubd_dev->count > 0) in ubd_remove()
1086 platform_device_unregister(&ubd_dev->pdev); in ubd_remove()
1114 struct ubd *ubd_dev = &ubd_devs[0]; in ubd0_init() local
1117 if(ubd_dev->file == NULL) in ubd0_init()
1118 ubd_dev->file = "root_fs"; in ubd0_init()
1216 struct ubd *ubd_dev = disk->private_data; in ubd_open() local
1220 if(ubd_dev->count == 0){ in ubd_open()
1221 err = ubd_open_dev(ubd_dev); in ubd_open()
1224 disk->disk_name, ubd_dev->file, -err); in ubd_open()
1228 ubd_dev->count++; in ubd_open()
1229 set_disk_ro(disk, !ubd_dev->openflags.w); in ubd_open()
1244 struct ubd *ubd_dev = disk->private_data; in ubd_release() local
1247 if(--ubd_dev->count == 0) in ubd_release()
1248 ubd_close_dev(ubd_dev); in ubd_release()
1378 struct ubd *ubd_dev = hctx->queue->queuedata; in ubd_queue_rq() local
1384 spin_lock_irq(&ubd_dev->lock); in ubd_queue_rq()
1404 spin_unlock_irq(&ubd_dev->lock); in ubd_queue_rq()
1418 struct ubd *ubd_dev = bdev->bd_disk->private_data; in ubd_getgeo() local
1422 geo->cylinders = ubd_dev->size / (128 * 32 * 512); in ubd_getgeo()
1429 struct ubd *ubd_dev = bdev->bd_disk->private_data; in ubd_ioctl() local
1436 ubd_id[ATA_ID_CYLS] = ubd_dev->size / (128 * 32 * 512); in ubd_ioctl()