Lines Matching refs:ldev

758 		dc = rcu_dereference(peer_device->device->ldev->disk_conf);  in drbd_send_sync_param()
854 spin_lock_irq(&device->ldev->md.uuid_lock); in _drbd_send_uuids()
856 p->uuid[i] = cpu_to_be64(device->ldev->md.uuid[i]); in _drbd_send_uuids()
857 spin_unlock_irq(&device->ldev->md.uuid_lock); in _drbd_send_uuids()
885 u64 *uuid = device->ldev->md.uuid; in drbd_print_uuids()
909 uuid = device->ldev->md.uuid[UI_BITMAP]; in drbd_gen_and_send_sync_uuid()
972 struct request_queue *q = bdev_get_queue(device->ldev->backing_bdev); in drbd_send_sizes()
973 d_size = drbd_get_max_capacity(device->ldev); in drbd_send_sizes()
975 u_size = rcu_dereference(device->ldev->disk_conf)->disk_size; in drbd_send_sizes()
1288 if (drbd_md_test_flag(device->ldev, MDF_FULL_SYNC)) { in _drbd_send_bitmap()
2069 drbd_backing_dev_free(device, device->ldev); in drbd_device_cleanup()
2070 device->ldev = NULL; in drbd_device_cleanup()
2232 drbd_backing_dev_free(device, device->ldev); in drbd_destroy_device()
2233 device->ldev = NULL; in drbd_destroy_device()
2446 q = bdev_get_queue(device->ldev->backing_bdev); in drbd_congested()
3101 buffer->uuid[i] = cpu_to_be64(device->ldev->md.uuid[i]); in drbd_md_write()
3102 buffer->flags = cpu_to_be32(device->ldev->md.flags); in drbd_md_write()
3105 buffer->md_size_sect = cpu_to_be32(device->ldev->md.md_size_sect); in drbd_md_write()
3106 buffer->al_offset = cpu_to_be32(device->ldev->md.al_offset); in drbd_md_write()
3109 buffer->device_uuid = cpu_to_be64(device->ldev->md.device_uuid); in drbd_md_write()
3111 buffer->bm_offset = cpu_to_be32(device->ldev->md.bm_offset); in drbd_md_write()
3114 buffer->al_stripes = cpu_to_be32(device->ldev->md.al_stripes); in drbd_md_write()
3115 buffer->al_stripe_size_4k = cpu_to_be32(device->ldev->md.al_stripe_size_4k); in drbd_md_write()
3117 D_ASSERT(device, drbd_md_ss(device->ldev) == device->ldev->md.md_offset); in drbd_md_write()
3118 sector = device->ldev->md.md_offset; in drbd_md_write()
3120 if (drbd_md_sync_page_io(device, device->ldev, sector, REQ_OP_WRITE)) { in drbd_md_write()
3157 device->ldev->md.la_size_sect = drbd_get_capacity(device->this_bdev); in drbd_md_sync()
3431 device->ldev->md.uuid[i+1] = device->ldev->md.uuid[i]; in drbd_uuid_move_history()
3445 device->ldev->md.uuid[idx] = val; in __drbd_uuid_set()
3452 spin_lock_irqsave(&device->ldev->md.uuid_lock, flags); in _drbd_uuid_set()
3454 spin_unlock_irqrestore(&device->ldev->md.uuid_lock, flags); in _drbd_uuid_set()
3460 spin_lock_irqsave(&device->ldev->md.uuid_lock, flags); in drbd_uuid_set()
3461 if (device->ldev->md.uuid[idx]) { in drbd_uuid_set()
3463 device->ldev->md.uuid[UI_HISTORY_START] = device->ldev->md.uuid[idx]; in drbd_uuid_set()
3466 spin_unlock_irqrestore(&device->ldev->md.uuid_lock, flags); in drbd_uuid_set()
3483 spin_lock_irq(&device->ldev->md.uuid_lock); in drbd_uuid_new_current()
3484 bm_uuid = device->ldev->md.uuid[UI_BITMAP]; in drbd_uuid_new_current()
3489 device->ldev->md.uuid[UI_BITMAP] = device->ldev->md.uuid[UI_CURRENT]; in drbd_uuid_new_current()
3491 spin_unlock_irq(&device->ldev->md.uuid_lock); in drbd_uuid_new_current()
3501 if (device->ldev->md.uuid[UI_BITMAP] == 0 && val == 0) in drbd_uuid_set_bm()
3504 spin_lock_irqsave(&device->ldev->md.uuid_lock, flags); in drbd_uuid_set_bm()
3507 device->ldev->md.uuid[UI_HISTORY_START] = device->ldev->md.uuid[UI_BITMAP]; in drbd_uuid_set_bm()
3508 device->ldev->md.uuid[UI_BITMAP] = 0; in drbd_uuid_set_bm()
3510 unsigned long long bm_uuid = device->ldev->md.uuid[UI_BITMAP]; in drbd_uuid_set_bm()
3514 device->ldev->md.uuid[UI_BITMAP] = val & ~((u64)1); in drbd_uuid_set_bm()
3516 spin_unlock_irqrestore(&device->ldev->md.uuid_lock, flags); in drbd_uuid_set_bm()
3671 if ((device->ldev->md.flags & flag) != flag) { in drbd_md_set_flag()
3673 device->ldev->md.flags |= flag; in drbd_md_set_flag()
3679 if ((device->ldev->md.flags & flag) != 0) { in drbd_md_clear_flag()
3681 device->ldev->md.flags &= ~flag; in drbd_md_clear_flag()