Lines Matching full:v0

49 		struct nvif_vmm_pfnclr_v0 v0;  in nvkm_uvmm_mthd_pfnclr()  member
55 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { in nvkm_uvmm_mthd_pfnclr()
56 addr = args->v0.addr; in nvkm_uvmm_mthd_pfnclr()
57 size = args->v0.size; in nvkm_uvmm_mthd_pfnclr()
74 struct nvif_vmm_pfnmap_v0 v0; in nvkm_uvmm_mthd_pfnmap() member
81 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, true))) { in nvkm_uvmm_mthd_pfnmap()
82 page = args->v0.page; in nvkm_uvmm_mthd_pfnmap()
83 addr = args->v0.addr; in nvkm_uvmm_mthd_pfnmap()
84 size = args->v0.size; in nvkm_uvmm_mthd_pfnmap()
85 phys = args->v0.phys; in nvkm_uvmm_mthd_pfnmap()
86 if (argc != (size >> page) * sizeof(args->v0.phys[0])) in nvkm_uvmm_mthd_pfnmap()
104 struct nvif_vmm_unmap_v0 v0; in nvkm_uvmm_mthd_unmap() member
111 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { in nvkm_uvmm_mthd_unmap()
112 addr = args->v0.addr; in nvkm_uvmm_mthd_unmap()
146 struct nvif_vmm_map_v0 v0; in nvkm_uvmm_mthd_map() member
154 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, true))) { in nvkm_uvmm_mthd_map()
155 addr = args->v0.addr; in nvkm_uvmm_mthd_map()
156 size = args->v0.size; in nvkm_uvmm_mthd_map()
157 handle = args->v0.memory; in nvkm_uvmm_mthd_map()
158 offset = args->v0.offset; in nvkm_uvmm_mthd_map()
223 struct nvif_vmm_put_v0 v0; in nvkm_uvmm_mthd_put() member
230 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { in nvkm_uvmm_mthd_put()
231 addr = args->v0.addr; in nvkm_uvmm_mthd_put()
236 vma = nvkm_vmm_node_search(vmm, args->v0.addr); in nvkm_uvmm_mthd_put()
259 struct nvif_vmm_get_v0 v0; in nvkm_uvmm_mthd_get() member
268 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { in nvkm_uvmm_mthd_get()
269 getref = args->v0.type == NVIF_VMM_GET_V0_PTES; in nvkm_uvmm_mthd_get()
270 mapref = args->v0.type == NVIF_VMM_GET_V0_ADDR; in nvkm_uvmm_mthd_get()
271 sparse = args->v0.sparse; in nvkm_uvmm_mthd_get()
272 page = args->v0.page; in nvkm_uvmm_mthd_get()
273 align = args->v0.align; in nvkm_uvmm_mthd_get()
274 size = args->v0.size; in nvkm_uvmm_mthd_get()
285 args->v0.addr = vma->addr; in nvkm_uvmm_mthd_get()
293 struct nvif_vmm_page_v0 v0; in nvkm_uvmm_mthd_page() member
302 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { in nvkm_uvmm_mthd_page()
303 if ((index = args->v0.index) >= nr) in nvkm_uvmm_mthd_page()
306 args->v0.shift = page[index].shift; in nvkm_uvmm_mthd_page()
307 args->v0.sparse = !!(type & NVKM_VMM_PAGE_SPARSE); in nvkm_uvmm_mthd_page()
308 args->v0.vram = !!(type & NVKM_VMM_PAGE_VRAM); in nvkm_uvmm_mthd_page()
309 args->v0.host = !!(type & NVKM_VMM_PAGE_HOST); in nvkm_uvmm_mthd_page()
310 args->v0.comp = !!(type & NVKM_VMM_PAGE_COMP); in nvkm_uvmm_mthd_page()
363 struct nvif_vmm_v0 v0; in nvkm_uvmm_new() member
371 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, more))) { in nvkm_uvmm_new()
372 managed = args->v0.managed != 0; in nvkm_uvmm_new()
373 addr = args->v0.addr; in nvkm_uvmm_new()
374 size = args->v0.size; in nvkm_uvmm_new()
398 args->v0.page_nr = 0; in nvkm_uvmm_new()
400 args->v0.page_nr++; in nvkm_uvmm_new()
401 args->v0.addr = uvmm->vmm->start; in nvkm_uvmm_new()
402 args->v0.size = uvmm->vmm->limit; in nvkm_uvmm_new()