Lines Matching refs:backup
891 &res->backup, in vmw_surface_define_ioctl()
1238 if (res->backup->dirty && res->backup_dirty) { in vmw_gb_surface_bind()
1550 &res->backup, in vmw_gb_surface_define_internal()
1553 if (res->backup->base.num_pages * PAGE_SIZE < in vmw_gb_surface_define_internal()
1556 vmw_bo_unreference(&res->backup); in vmw_gb_surface_define_internal()
1571 &res->backup, in vmw_gb_surface_define_internal()
1580 struct vmw_buffer_object *backup = res->backup; in vmw_gb_surface_define_internal() local
1582 ttm_bo_reserve(&backup->base, false, false, NULL); in vmw_gb_surface_define_internal()
1586 ret = vmw_bo_dirty_add(backup); in vmw_gb_surface_define_internal()
1591 ttm_bo_unreserve(&backup->base); in vmw_gb_surface_define_internal()
1614 if (res->backup) { in vmw_gb_surface_define_internal()
1616 drm_vma_node_offset_addr(&res->backup->base.base.vma_node); in vmw_gb_surface_define_internal()
1617 rep->buffer_size = res->backup->base.num_pages * PAGE_SIZE; in vmw_gb_surface_define_internal()
1663 if (!srf->res.backup) { in vmw_gb_surface_reference_internal()
1670 ret = vmw_user_bo_reference(tfile, srf->res.backup, &backup_handle); in vmw_gb_surface_reference_internal()
1694 drm_vma_node_offset_addr(&srf->res.backup->base.base.vma_node); in vmw_gb_surface_reference_internal()
1695 rep->crep.buffer_size = srf->res.backup->base.num_pages * PAGE_SIZE; in vmw_gb_surface_reference_internal()