/Linux-v5.4/drivers/infiniband/hw/qib/ |
D | qib_user_pages.c | 40 static void __qib_release_user_pages(struct page **p, size_t num_pages, in __qib_release_user_pages() argument 43 put_user_pages_dirty_lock(p, num_pages, dirty); in __qib_release_user_pages() 94 int qib_get_user_pages(unsigned long start_page, size_t num_pages, in qib_get_user_pages() argument 102 locked = atomic64_add_return(num_pages, ¤t->mm->pinned_vm); in qib_get_user_pages() 110 for (got = 0; got < num_pages; got += ret) { in qib_get_user_pages() 112 num_pages - got, in qib_get_user_pages() 126 atomic64_sub(num_pages, ¤t->mm->pinned_vm); in qib_get_user_pages() 130 void qib_release_user_pages(struct page **p, size_t num_pages) in qib_release_user_pages() argument 132 __qib_release_user_pages(p, num_pages, 1); in qib_release_user_pages() 136 atomic64_sub(num_pages, ¤t->mm->pinned_vm); in qib_release_user_pages()
|
/Linux-v5.4/drivers/staging/vc04_services/interface/vchiq_arm/ |
D | vchiq_2835_arm.c | 41 unsigned int num_pages; member 333 pagelistinfo->num_pages, pagelistinfo->dma_dir); in cleanup_pagelistinfo() 339 for (i = 0; i < pagelistinfo->num_pages; i++) in cleanup_pagelistinfo() 362 unsigned int num_pages, offset, i, k; in create_pagelist() local 373 num_pages = DIV_ROUND_UP(count + offset, PAGE_SIZE); in create_pagelist() 375 if (num_pages > (SIZE_MAX - sizeof(struct pagelist) - in create_pagelist() 382 (num_pages * sizeof(u32)) + in create_pagelist() 383 (num_pages * sizeof(pages[0]) + in create_pagelist() 384 (num_pages * sizeof(struct scatterlist))) + in create_pagelist() 399 pages = (struct page **)(addrs + num_pages); in create_pagelist() [all …]
|
/Linux-v5.4/drivers/staging/gasket/ |
D | gasket_page_table.c | 468 uint num_pages, int is_simple_mapping) in gasket_perform_mapping() argument 477 for (i = 0; i < num_pages; i++) { in gasket_perform_mapping() 576 ulong dev_addr, uint num_pages) in gasket_alloc_simple_entries() argument 580 num_pages)) in gasket_alloc_simple_entries() 592 u64 __iomem *slots, uint num_pages, in gasket_perform_unmapping() argument 600 for (i = 0; i < num_pages; i++) { in gasket_perform_unmapping() 630 ulong dev_addr, uint num_pages) in gasket_unmap_simple_pages() argument 635 pg_tbl->base_slot + slot, num_pages, 1); in gasket_unmap_simple_pages() 643 ulong dev_addr, uint num_pages) in gasket_unmap_extended_pages() argument 649 remain = num_pages; in gasket_unmap_extended_pages() [all …]
|
/Linux-v5.4/drivers/gpu/drm/vmwgfx/ |
D | vmwgfx_gmr.c | 40 unsigned long num_pages, in vmw_gmr2_bind() argument 48 uint32_t remap_num = num_pages / VMW_PPN_PER_REMAP + ((num_pages % VMW_PPN_PER_REMAP) > 0); in vmw_gmr2_bind() 49 uint32_t remap_size = VMW_PPN_SIZE * num_pages + (sizeof(remap_cmd) + sizeof(*cmd)) * remap_num; in vmw_gmr2_bind() 59 define_cmd.numPages = num_pages; in vmw_gmr2_bind() 74 while (num_pages > 0) { in vmw_gmr2_bind() 75 unsigned long nr = min(num_pages, (unsigned long)VMW_PPN_PER_REMAP); in vmw_gmr2_bind() 95 num_pages -= nr; in vmw_gmr2_bind() 129 unsigned long num_pages, in vmw_gmr_bind() argument 142 return vmw_gmr2_bind(dev_priv, &data_iter, num_pages, gmr_id); in vmw_gmr_bind()
|
D | vmwgfx_ttm_buffer.c | 264 return ++(viter->i) < viter->num_pages; in __vmw_piter_non_sg_next() 328 viter->num_pages = vsgt->num_pages; in vmw_piter_start() 428 vsgt->num_pages = vmw_tt->dma_ttm.ttm.num_pages; in vmw_ttm_map_dma() 439 vmw_tt->sg_alloc_size = sgt_size + sgl_size * vsgt->num_pages; in vmw_ttm_map_dma() 445 (&vmw_tt->sgt, vsgt->pages, vsgt->num_pages, 0, in vmw_ttm_map_dma() 446 (unsigned long) vsgt->num_pages << PAGE_SHIFT, in vmw_ttm_map_dma() 452 if (vsgt->num_pages > vmw_tt->sgt.nents) { in vmw_ttm_map_dma() 454 sgl_size * (vsgt->num_pages - in vmw_ttm_map_dma() 595 ttm->num_pages, vmw_be->gmr_id); in vmw_ttm_bind() 599 vmw_mob_create(ttm->num_pages); in vmw_ttm_bind() [all …]
|
D | vmwgfx_gmrid_manager.c | 65 gman->used_gmr_pages += bo->num_pages; in vmw_gmrid_man_get_node() 72 mem->num_pages = bo->num_pages; in vmw_gmrid_man_get_node() 78 gman->used_gmr_pages -= bo->num_pages; in vmw_gmrid_man_get_node() 93 gman->used_gmr_pages -= mem->num_pages; in vmw_gmrid_man_put_node()
|
/Linux-v5.4/drivers/gpu/drm/xen/ |
D | xen_drm_front_gem.c | 28 size_t num_pages; member 47 xen_obj->num_pages = DIV_ROUND_UP(buf_size, PAGE_SIZE); in gem_alloc_pages_array() 48 xen_obj->pages = kvmalloc_array(xen_obj->num_pages, in gem_alloc_pages_array() 102 ret = alloc_xenballooned_pages(xen_obj->num_pages, in gem_create() 106 xen_obj->num_pages, ret); in gem_create() 118 xen_obj->num_pages = DIV_ROUND_UP(size, PAGE_SIZE); in gem_create() 155 free_xenballooned_pages(xen_obj->num_pages, in xen_drm_front_gem_free_object_unlocked() 182 return drm_prime_pages_to_sg(xen_obj->pages, xen_obj->num_pages); in xen_drm_front_gem_get_sg_table() 207 NULL, xen_obj->num_pages); in xen_drm_front_gem_import_sg_table() 253 ret = vm_map_pages(vma, xen_obj->pages, xen_obj->num_pages); in gem_mmap_obj() [all …]
|
/Linux-v5.4/drivers/xen/ |
D | xen-front-pgdir-shbuf.c | 169 return DIV_ROUND_UP(buf->num_pages, XEN_NUM_GREFS_PER_PAGE); in get_num_pages_dir() 196 buf->num_grefs = get_num_pages_dir(buf) + buf->num_pages; in guest_calc_num_grefs() 217 unmap_ops = kcalloc(buf->num_pages, sizeof(*unmap_ops), in backend_unmap() 222 for (i = 0; i < buf->num_pages; i++) { in backend_unmap() 231 buf->num_pages); in backend_unmap() 233 for (i = 0; i < buf->num_pages; i++) { in backend_unmap() 262 map_ops = kcalloc(buf->num_pages, sizeof(*map_ops), GFP_KERNEL); in backend_map() 266 buf->backend_map_handles = kcalloc(buf->num_pages, in backend_map() 280 grefs_left = buf->num_pages; in backend_map() 304 ret = gnttab_map_refs(map_ops, NULL, buf->pages, buf->num_pages); in backend_map() [all …]
|
/Linux-v5.4/drivers/infiniband/sw/siw/ |
D | siw_mem.c | 63 static void siw_free_plist(struct siw_page_chunk *chunk, int num_pages, in siw_free_plist() argument 66 put_user_pages_dirty_lock(chunk->plist, num_pages, dirty); in siw_free_plist() 72 int i, num_pages = umem->num_pages; in siw_umem_release() local 74 for (i = 0; num_pages; i++) { in siw_umem_release() 75 int to_free = min_t(int, PAGES_PER_CHUNK, num_pages); in siw_umem_release() 80 num_pages -= to_free; in siw_umem_release() 82 atomic64_sub(umem->num_pages, &mm_s->pinned_vm); in siw_umem_release() 375 int num_pages, num_chunks, i, rv = 0; in siw_umem_get() local 384 num_pages = PAGE_ALIGN(start + len - first_page_va) >> PAGE_SHIFT; in siw_umem_get() 385 num_chunks = (num_pages >> CHUNK_SHIFT) + 1; in siw_umem_get() [all …]
|
/Linux-v5.4/drivers/media/common/videobuf2/ |
D | videobuf2-dma-sg.c | 49 unsigned int num_pages; member 106 int num_pages; in vb2_dma_sg_alloc() local 120 buf->num_pages = size >> PAGE_SHIFT; in vb2_dma_sg_alloc() 123 buf->pages = kvmalloc_array(buf->num_pages, sizeof(struct page *), in vb2_dma_sg_alloc() 133 buf->num_pages, 0, size, GFP_KERNEL); in vb2_dma_sg_alloc() 157 __func__, buf->num_pages); in vb2_dma_sg_alloc() 164 num_pages = buf->num_pages; in vb2_dma_sg_alloc() 165 while (num_pages--) in vb2_dma_sg_alloc() 166 __free_page(buf->pages[num_pages]); in vb2_dma_sg_alloc() 178 int i = buf->num_pages; in vb2_dma_sg_put() [all …]
|
/Linux-v5.4/drivers/gpu/drm/ttm/ |
D | ttm_tt.c | 87 ttm->pages = kvmalloc_array(ttm->num_pages, sizeof(void*), in ttm_tt_alloc_page_directory() 96 ttm->ttm.pages = kvmalloc_array(ttm->ttm.num_pages, in ttm_dma_tt_alloc_page_directory() 102 ttm->dma_address = (void *) (ttm->ttm.pages + ttm->ttm.num_pages); in ttm_dma_tt_alloc_page_directory() 108 ttm->dma_address = kvmalloc_array(ttm->ttm.num_pages, in ttm_sg_tt_alloc_page_directory() 164 drm_clflush_pages(ttm->pages, ttm->num_pages); in ttm_tt_set_caching() 166 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_set_caching() 230 ttm->num_pages = bo->num_pages; in ttm_tt_init_fields() 361 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_swapin() 403 ttm->num_pages << PAGE_SHIFT, in ttm_tt_swapout() 415 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_swapout() [all …]
|
/Linux-v5.4/net/ceph/ |
D | pagevec.c | 13 void ceph_put_page_vector(struct page **pages, int num_pages, bool dirty) in ceph_put_page_vector() argument 17 for (i = 0; i < num_pages; i++) { in ceph_put_page_vector() 26 void ceph_release_page_vector(struct page **pages, int num_pages) in ceph_release_page_vector() argument 30 for (i = 0; i < num_pages; i++) in ceph_release_page_vector() 39 struct page **ceph_alloc_page_vector(int num_pages, gfp_t flags) in ceph_alloc_page_vector() argument 44 pages = kmalloc_array(num_pages, sizeof(*pages), flags); in ceph_alloc_page_vector() 47 for (i = 0; i < num_pages; i++) { in ceph_alloc_page_vector()
|
/Linux-v5.4/drivers/gpu/drm/ |
D | drm_cache.c | 61 unsigned long num_pages) in drm_cache_flush_clflush() argument 66 for (i = 0; i < num_pages; i++) in drm_cache_flush_clflush() 81 drm_clflush_pages(struct page *pages[], unsigned long num_pages) in drm_clflush_pages() argument 86 drm_cache_flush_clflush(pages, num_pages); in drm_clflush_pages() 95 for (i = 0; i < num_pages; i++) { in drm_clflush_pages()
|
/Linux-v5.4/drivers/firmware/efi/ |
D | memmap.c | 224 end = start + (md->num_pages << EFI_PAGE_SHIFT) - 1; in efi_memmap_split_count() 289 end = md->phys_addr + (md->num_pages << EFI_PAGE_SHIFT) - 1; in efi_memmap_insert() 298 md->num_pages = (m_end - md->phys_addr + 1) >> in efi_memmap_insert() 305 md->num_pages = (end - md->phys_addr + 1) >> in efi_memmap_insert() 311 md->num_pages = (m_start - md->phys_addr) >> in efi_memmap_insert() 319 md->num_pages = (m_end - m_start + 1) >> in efi_memmap_insert() 326 md->num_pages = (end - m_end) >> in efi_memmap_insert() 333 md->num_pages = (m_start - md->phys_addr) >> in efi_memmap_insert() 340 md->num_pages = (end - md->phys_addr + 1) >> in efi_memmap_insert()
|
/Linux-v5.4/arch/x86/platform/efi/ |
D | efi.c | 133 unsigned long long size = md->num_pages << EFI_PAGE_SHIFT; in efi_find_mirror() 158 unsigned long long size = md->num_pages << EFI_PAGE_SHIFT; in do_add_efi_memmap() 245 u64 end = (md->num_pages << EFI_PAGE_SHIFT) + md->phys_addr - 1; in efi_memmap_entry_valid() 249 if (md->num_pages == 0) { in efi_memmap_entry_valid() 251 } else if (md->num_pages > EFI_PAGES_MAX || in efi_memmap_entry_valid() 252 EFI_PAGES_MAX - md->num_pages < in efi_memmap_entry_valid() 254 end_hi = (md->num_pages & OVERFLOW_ADDR_MASK) in efi_memmap_entry_valid() 314 md->phys_addr + (md->num_pages << EFI_PAGE_SHIFT) - 1, in efi_print_memmap() 315 (md->num_pages >> (20 - EFI_PAGE_SHIFT))); in efi_print_memmap() 576 npages = md->num_pages; in efi_set_executable() [all …]
|
/Linux-v5.4/drivers/gpu/drm/gma500/ |
D | mmu.c | 500 uint32_t num_pages, uint32_t desired_tile_stride, in psb_mmu_flush_ptes() argument 518 rows = num_pages / desired_tile_stride; in psb_mmu_flush_ptes() 520 desired_tile_stride = num_pages; in psb_mmu_flush_ptes() 548 uint32_t num_pages, uint32_t desired_tile_stride, in psb_mmu_flush_ptes() argument 556 unsigned long address, uint32_t num_pages) in psb_mmu_remove_pfn_sequence() argument 567 end = addr + (num_pages << PAGE_SHIFT); in psb_mmu_remove_pfn_sequence() 584 psb_mmu_flush_ptes(pd, f_address, num_pages, 1, 1); in psb_mmu_remove_pfn_sequence() 595 uint32_t num_pages, uint32_t desired_tile_stride, in psb_mmu_remove_pages() argument 609 rows = num_pages / desired_tile_stride; in psb_mmu_remove_pages() 611 desired_tile_stride = num_pages; in psb_mmu_remove_pages() [all …]
|
D | mmu.h | 69 uint32_t num_pages); 73 uint32_t num_pages, int type); 78 unsigned long address, uint32_t num_pages, 82 unsigned long address, uint32_t num_pages,
|
/Linux-v5.4/drivers/hv/ |
D | hv_balloon.c | 278 __u32 num_pages; member 450 __u32 num_pages; member 1193 int num_pages = range_array->finfo.page_cnt; in free_balloon_pages() local 1198 for (i = 0; i < num_pages; i++) { in free_balloon_pages() 1209 unsigned int num_pages, in alloc_balloon_pages() argument 1216 if (num_pages < alloc_unit) in alloc_balloon_pages() 1219 for (i = 0; (i * alloc_unit) < num_pages; i++) { in alloc_balloon_pages() 1257 return num_pages; in alloc_balloon_pages() 1262 unsigned int num_pages = dm_device.balloon_wrk.num_pages; in balloon_up() local 1273 WARN_ON_ONCE(num_pages % PAGES_IN_2M != 0); in balloon_up() [all …]
|
/Linux-v5.4/drivers/tee/optee/ |
D | call.c | 456 void optee_fill_pages_list(u64 *dst, struct page **pages, int num_pages, in optee_fill_pages_list() argument 501 if (!--num_pages) in optee_fill_pages_list() 552 static int check_mem_type(unsigned long start, size_t num_pages) in check_mem_type() argument 559 start + num_pages * PAGE_SIZE); in check_mem_type() 566 struct page **pages, size_t num_pages, in optee_shm_register() argument 575 if (!num_pages) in optee_shm_register() 578 rc = check_mem_type(start, num_pages); in optee_shm_register() 582 pages_list = optee_allocate_pages_list(num_pages); in optee_shm_register() 592 optee_fill_pages_list(pages_list, pages, num_pages, in optee_shm_register() 613 optee_free_pages_list(pages_list, num_pages); in optee_shm_register() [all …]
|
/Linux-v5.4/drivers/misc/mic/scif/ |
D | scif_rma.c | 153 window->num_pages = scif_zalloc(nr_pages * sizeof(*window->num_pages)); in scif_create_window() 154 if (!window->num_pages) in scif_create_window() 214 scif_free(window->num_pages, nr_pages * sizeof(*window->num_pages)); in scif_destroy_incomplete_window() 243 window->num_pages[j] << in scif_unmap_window() 335 scif_free(window->num_pages, nr_pages * sizeof(*window->num_pages)); in scif_destroy_window() 389 vmalloc_num_pages = is_vmalloc_addr(&window->num_pages[0]); in scif_create_remote_lookup() 402 vmalloc_to_page(&window->num_pages[i]) : in scif_create_remote_lookup() 403 virt_to_page(&window->num_pages[i]), in scif_create_remote_lookup() 480 window->num_pages = scif_zalloc(nr_pages * in scif_create_remote_window() 481 sizeof(*window->num_pages)); in scif_create_remote_window() [all …]
|
/Linux-v5.4/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_ttm.c | 60 struct ttm_mem_reg *mem, unsigned num_pages, 442 new_mem->num_pages << PAGE_SHIFT, in amdgpu_move_blit() 604 if (nodes->size != mem->num_pages) in amdgpu_mem_visible() 695 atomic64_add((u64)bo->num_pages << PAGE_SHIFT, &adev->num_bytes_moved); in amdgpu_bo_move() 712 mem->bus.size = mem->num_pages << PAGE_SHIFT; in amdgpu_ttm_io_mem_reserve() 733 (mm_node->size == mem->num_pages)) in amdgpu_ttm_io_mem_reserve() 825 pfns = kvmalloc_array(ttm->num_pages, sizeof(*pfns), GFP_KERNEL); in amdgpu_ttm_tt_get_user_pages() 838 range->end = start + ttm->num_pages * PAGE_SIZE; in amdgpu_ttm_tt_get_user_pages() 856 for (i = 0; i < ttm->num_pages; i++) { in amdgpu_ttm_tt_get_user_pages() 895 gtt->userptr, ttm->num_pages); in amdgpu_ttm_tt_get_user_pages_done() [all …]
|
D | amdgpu_vram_mgr.c | 220 unsigned pages = mem->num_pages; in amdgpu_vram_mgr_bo_visible_size() 250 if (start > mem->num_pages) in amdgpu_vram_mgr_virt_start() 251 start -= mem->num_pages; in amdgpu_vram_mgr_virt_start() 287 mem_bytes = (u64)mem->num_pages << PAGE_SHIFT; in amdgpu_vram_mgr_new() 305 num_nodes = DIV_ROUND_UP(mem->num_pages, pages_per_node); in amdgpu_vram_mgr_new() 320 pages_left = mem->num_pages; in amdgpu_vram_mgr_new() 368 atomic64_sub(mem->num_pages << PAGE_SHIFT, &mgr->usage); in amdgpu_vram_mgr_new() 391 unsigned pages = mem->num_pages; in amdgpu_vram_mgr_del()
|
/Linux-v5.4/drivers/gpu/drm/radeon/ |
D | radeon_ttm.c | 209 unsigned num_pages; in radeon_move_blit() local 246 num_pages = new_mem->num_pages * (PAGE_SIZE / RADEON_GPU_PAGE_SIZE); in radeon_move_blit() 247 fence = radeon_copy(rdev, old_start, new_start, num_pages, bo->base.resv); in radeon_move_blit() 398 atomic64_add((u64)bo->num_pages << PAGE_SHIFT, &rdev->num_bytes_moved); in radeon_bo_move() 409 mem->bus.size = mem->num_pages << PAGE_SHIFT; in radeon_ttm_io_mem_reserve() 502 unsigned long end = gtt->userptr + ttm->num_pages * PAGE_SIZE; in radeon_ttm_tt_pin_userptr() 510 unsigned num_pages = ttm->num_pages - pinned; in radeon_ttm_tt_pin_userptr() local 514 r = get_user_pages(userptr, num_pages, write ? FOLL_WRITE : 0, in radeon_ttm_tt_pin_userptr() 521 } while (pinned < ttm->num_pages); in radeon_ttm_tt_pin_userptr() 523 r = sg_alloc_table_from_pages(ttm->sg, ttm->pages, ttm->num_pages, 0, in radeon_ttm_tt_pin_userptr() [all …]
|
/Linux-v5.4/drivers/virtio/ |
D | virtio_balloon.c | 95 unsigned int num_pages; member 203 vb->num_pages += VIRTIO_BALLOON_PAGES_PER_PAGE; in fill_balloon() 245 num = min(num, (size_t)vb->num_pages); in leak_balloon() 253 vb->num_pages -= VIRTIO_BALLOON_PAGES_PER_PAGE; in leak_balloon() 357 u32 num_pages; in towards_target() local 359 virtio_cread(vb->vdev, struct virtio_balloon_config, num_pages, in towards_target() 360 &num_pages); in towards_target() 364 num_pages = le32_to_cpu((__force __le32)num_pages); in towards_target() 366 target = num_pages; in towards_target() 367 return target - vb->num_pages; in towards_target() [all …]
|
/Linux-v5.4/fs/fuse/ |
D | file.c | 376 if (idx_from < curr_index + wpa->ia.ap.num_pages && in fuse_find_writeback() 576 for (i = 0; i < ap->num_pages; i++) { in fuse_release_user_pages() 769 for (i = start_idx; i < ap->num_pages; i++) { in fuse_short_read() 788 .ap.num_pages = 1, in fuse_do_readpage() 845 for (i = 0; mapping == NULL && i < ap->num_pages; i++) in fuse_readpages_end() 860 for (i = 0; i < ap->num_pages; i++) { in fuse_readpages_end() 882 size_t count = ap->num_pages << PAGE_SHIFT; in fuse_send_readpages() 920 if (ap->num_pages && in fuse_readpages_fill() 921 (ap->num_pages == fc->max_pages || in fuse_readpages_fill() 922 (ap->num_pages + 1) * PAGE_SIZE > fc->max_read || in fuse_readpages_fill() [all …]
|