/Linux-v4.19/mm/ |
D | page_counter.c | 55 void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages) in page_counter_cancel() argument 59 new = atomic_long_sub_return(nr_pages, &counter->usage); in page_counter_cancel() 72 void page_counter_charge(struct page_counter *counter, unsigned long nr_pages) in page_counter_charge() argument 79 new = atomic_long_add_return(nr_pages, &c->usage); in page_counter_charge() 100 unsigned long nr_pages, in page_counter_try_charge() argument 121 new = atomic_long_add_return(nr_pages, &c->usage); in page_counter_try_charge() 123 atomic_long_sub(nr_pages, &c->usage); in page_counter_try_charge() 145 page_counter_cancel(c, nr_pages); in page_counter_try_charge() 155 void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages) in page_counter_uncharge() argument 160 page_counter_cancel(c, nr_pages); in page_counter_uncharge() [all …]
|
D | percpu-km.c | 50 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; in pcpu_create_chunk() local 59 pages = alloc_pages(gfp, order_base_2(nr_pages)); in pcpu_create_chunk() 65 for (i = 0; i < nr_pages; i++) in pcpu_create_chunk() 72 pcpu_chunk_populated(chunk, 0, nr_pages, false); in pcpu_create_chunk() 83 const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; in pcpu_destroy_chunk() local 92 __free_pages(chunk->data, order_base_2(nr_pages)); in pcpu_destroy_chunk() 103 size_t nr_pages, alloc_pages; in pcpu_verify_alloc_info() local 111 nr_pages = (ai->groups[0].nr_units * ai->unit_size) >> PAGE_SHIFT; in pcpu_verify_alloc_info() 112 alloc_pages = roundup_pow_of_two(nr_pages); in pcpu_verify_alloc_info() 114 if (alloc_pages > nr_pages) in pcpu_verify_alloc_info() [all …]
|
D | memory_hotplug.c | 222 unsigned long i, pfn, end_pfn, nr_pages; in register_page_bootmem_info_node() local 226 nr_pages = PAGE_ALIGN(sizeof(struct pglist_data)) >> PAGE_SHIFT; in register_page_bootmem_info_node() 229 for (i = 0; i < nr_pages; i++, page++) in register_page_bootmem_info_node() 274 unsigned long nr_pages, struct vmem_altmap *altmap, in __add_pages() argument 283 end_sec = pfn_to_section_nr(phys_start_pfn + nr_pages - 1); in __add_pages() 290 || vmem_altmap_offset(altmap) > nr_pages) { in __add_pages() 508 int nr_pages = PAGES_PER_SECTION; in __remove_zone() local 512 shrink_zone_span(zone, start_pfn, start_pfn + nr_pages); in __remove_zone() 513 shrink_pgdat_span(pgdat, start_pfn, start_pfn + nr_pages); in __remove_zone() 552 unsigned long nr_pages, struct vmem_altmap *altmap) in __remove_pages() argument [all …]
|
D | hugetlb_cgroup.c | 128 unsigned int nr_pages; in hugetlb_cgroup_move_parent() local 142 nr_pages = 1 << compound_order(page); in hugetlb_cgroup_move_parent() 146 page_counter_charge(&parent->hugepage[idx], nr_pages); in hugetlb_cgroup_move_parent() 150 page_counter_cancel(counter, nr_pages); in hugetlb_cgroup_move_parent() 181 int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, in hugetlb_cgroup_charge_cgroup() argument 205 if (!page_counter_try_charge(&h_cg->hugepage[idx], nr_pages, &counter)) in hugetlb_cgroup_charge_cgroup() 214 void hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, in hugetlb_cgroup_commit_charge() argument 228 void hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, in hugetlb_cgroup_uncharge_page() argument 240 page_counter_uncharge(&h_cg->hugepage[idx], nr_pages); in hugetlb_cgroup_uncharge_page() 244 void hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, in hugetlb_cgroup_uncharge_cgroup() argument [all …]
|
D | page_ext.c | 147 unsigned long nr_pages; in alloc_node_page_ext() local 149 nr_pages = NODE_DATA(nid)->node_spanned_pages; in alloc_node_page_ext() 150 if (!nr_pages) in alloc_node_page_ext() 160 nr_pages += MAX_ORDER_NR_PAGES; in alloc_node_page_ext() 162 table_size = get_entry_size() * nr_pages; in alloc_node_page_ext() 294 unsigned long nr_pages, in online_page_ext() argument 301 end = SECTION_ALIGN_UP(start_pfn + nr_pages); in online_page_ext() 329 unsigned long nr_pages, int nid) in offline_page_ext() argument 334 end = SECTION_ALIGN_UP(start_pfn + nr_pages); in offline_page_ext() 351 mn->nr_pages, mn->status_change_nid); in page_ext_callback() [all …]
|
D | page_isolation.c | 41 arg.nr_pages = pageblock_nr_pages; in set_migratetype_isolate() 74 unsigned long nr_pages; in set_migratetype_isolate() local 79 nr_pages = move_freepages_block(zone, page, MIGRATE_ISOLATE, in set_migratetype_isolate() 82 __mod_zone_freepage_state(zone, -nr_pages, mt); in set_migratetype_isolate() 94 unsigned long flags, nr_pages; in unset_migratetype_isolate() local 134 nr_pages = move_freepages_block(zone, page, migratetype, NULL); in unset_migratetype_isolate() 135 __mod_zone_freepage_state(zone, nr_pages, migratetype); in unset_migratetype_isolate() 148 __first_valid_page(unsigned long pfn, unsigned long nr_pages) in __first_valid_page() argument 152 for (i = 0; i < nr_pages; i++) { in __first_valid_page()
|
D | memcontrol.c | 586 unsigned long nr_pages = page_counter_read(&memcg->memory); in soft_limit_excess() local 590 if (nr_pages > soft_limit) in soft_limit_excess() 591 excess = nr_pages - soft_limit; in soft_limit_excess() 691 bool compound, int nr_pages) in mem_cgroup_charge_statistics() argument 698 __mod_memcg_state(memcg, MEMCG_RSS, nr_pages); in mem_cgroup_charge_statistics() 700 __mod_memcg_state(memcg, MEMCG_CACHE, nr_pages); in mem_cgroup_charge_statistics() 702 __mod_memcg_state(memcg, NR_SHMEM, nr_pages); in mem_cgroup_charge_statistics() 707 __mod_memcg_state(memcg, MEMCG_RSS_HUGE, nr_pages); in mem_cgroup_charge_statistics() 711 if (nr_pages > 0) in mem_cgroup_charge_statistics() 715 nr_pages = -nr_pages; /* for event */ in mem_cgroup_charge_statistics() [all …]
|
D | process_vm_access.c | 88 unsigned long nr_pages; in process_vm_rw_single_vec() local 97 nr_pages = (addr + len - 1) / PAGE_SIZE - addr / PAGE_SIZE + 1; in process_vm_rw_single_vec() 102 while (!rc && nr_pages && iov_iter_count(iter)) { in process_vm_rw_single_vec() 103 int pages = min(nr_pages, max_pages_per_loop); in process_vm_rw_single_vec() 129 nr_pages -= pages; in process_vm_rw_single_vec() 166 unsigned long nr_pages = 0; in process_vm_rw_core() local 182 nr_pages = max(nr_pages, nr_pages_iov); in process_vm_rw_core() 186 if (nr_pages == 0) in process_vm_rw_core() 189 if (nr_pages > PVM_MAX_PP_ARRAY_COUNT) { in process_vm_rw_core() 193 sizeof(struct pages *)*nr_pages), in process_vm_rw_core()
|
D | gup.c | 658 unsigned long start, unsigned long nr_pages, in __get_user_pages() argument 666 if (!nr_pages) in __get_user_pages() 702 &start, &nr_pages, i, in __get_user_pages() 754 if (page_increm > nr_pages) in __get_user_pages() 755 page_increm = nr_pages; in __get_user_pages() 758 nr_pages -= page_increm; in __get_user_pages() 759 } while (nr_pages); in __get_user_pages() 867 unsigned long nr_pages, in __get_user_pages_locked() argument 889 ret = __get_user_pages(tsk, mm, start, nr_pages, flags, pages, in __get_user_pages_locked() 898 BUG_ON(ret >= nr_pages); in __get_user_pages_locked() [all …]
|
/Linux-v4.19/drivers/media/v4l2-core/ |
D | videobuf-dma-sg.c | 66 int nr_pages) in videobuf_vmalloc_to_sg() argument 72 sglist = vzalloc(array_size(nr_pages, sizeof(*sglist))); in videobuf_vmalloc_to_sg() 75 sg_init_table(sglist, nr_pages); in videobuf_vmalloc_to_sg() 76 for (i = 0; i < nr_pages; i++, virt += PAGE_SIZE) { in videobuf_vmalloc_to_sg() 96 int nr_pages, int offset, size_t size) in videobuf_pages_to_sg() argument 103 sglist = vmalloc(array_size(nr_pages, sizeof(*sglist))); in videobuf_pages_to_sg() 106 sg_init_table(sglist, nr_pages); in videobuf_pages_to_sg() 114 for (i = 1; i < nr_pages; i++) { in videobuf_pages_to_sg() 177 dma->nr_pages = last-first+1; in videobuf_dma_init_user_locked() 178 dma->pages = kmalloc_array(dma->nr_pages, sizeof(struct page *), in videobuf_dma_init_user_locked() [all …]
|
/Linux-v4.19/include/linux/ |
D | page_counter.h | 50 void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages); 51 void page_counter_charge(struct page_counter *counter, unsigned long nr_pages); 53 unsigned long nr_pages, 55 void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages); 56 void page_counter_set_min(struct page_counter *counter, unsigned long nr_pages); 57 void page_counter_set_low(struct page_counter *counter, unsigned long nr_pages); 58 int page_counter_set_max(struct page_counter *counter, unsigned long nr_pages); 60 unsigned long *nr_pages);
|
D | hugetlb_cgroup.h | 54 extern int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, 56 extern void hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, 59 extern void hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, 61 extern void hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, 85 hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, in hugetlb_cgroup_charge_cgroup() argument 92 hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, in hugetlb_cgroup_commit_charge() argument 99 hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, struct page *page) in hugetlb_cgroup_uncharge_page() argument 104 hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, in hugetlb_cgroup_uncharge_cgroup() argument
|
D | memory_hotplug.h | 82 extern int zone_grow_waitqueues(struct zone *zone, unsigned long nr_pages); 113 unsigned long nr_pages, struct vmem_altmap *altmap); 117 extern int __add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, 122 unsigned long nr_pages, struct vmem_altmap *altmap, in add_pages() argument 125 return __add_pages(nid, start_pfn, nr_pages, altmap, want_memblock); in add_pages() 128 int add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, 300 extern bool is_mem_section_removable(unsigned long pfn, unsigned long nr_pages); 302 extern int offline_pages(unsigned long start_pfn, unsigned long nr_pages); 307 unsigned long nr_pages) in is_mem_section_removable() argument 314 static inline int offline_pages(unsigned long start_pfn, unsigned long nr_pages) in offline_pages() argument [all …]
|
D | mm_inline.h | 28 int nr_pages) in __update_lru_size() argument 32 __mod_node_page_state(pgdat, NR_LRU_BASE + lru, nr_pages); in __update_lru_size() 34 NR_ZONE_LRU_BASE + lru, nr_pages); in __update_lru_size() 39 int nr_pages) in update_lru_size() argument 41 __update_lru_size(lruvec, lru, zid, nr_pages); in update_lru_size() 43 mem_cgroup_update_lru_size(lruvec, lru, zid, nr_pages); in update_lru_size()
|
/Linux-v4.19/drivers/xen/ |
D | balloon.c | 455 static enum bp_state increase_reservation(unsigned long nr_pages) in increase_reservation() argument 461 if (nr_pages > ARRAY_SIZE(frame_list)) in increase_reservation() 462 nr_pages = ARRAY_SIZE(frame_list); in increase_reservation() 465 for (i = 0; i < nr_pages; i++) { in increase_reservation() 467 nr_pages = i; in increase_reservation() 475 rc = xenmem_reservation_increase(nr_pages, frame_list); in increase_reservation() 494 static enum bp_state decrease_reservation(unsigned long nr_pages, gfp_t gfp) in decrease_reservation() argument 502 if (nr_pages > ARRAY_SIZE(frame_list)) in decrease_reservation() 503 nr_pages = ARRAY_SIZE(frame_list); in decrease_reservation() 505 for (i = 0; i < nr_pages; i++) { in decrease_reservation() [all …]
|
/Linux-v4.19/drivers/misc/mic/scif/ |
D | scif_rma.c | 90 scif_create_pinned_pages(int nr_pages, int prot) in scif_create_pinned_pages() argument 99 pin->pages = scif_zalloc(nr_pages * sizeof(*pin->pages)); in scif_create_pinned_pages() 125 for (j = 0; j < pin->nr_pages; j++) { in scif_destroy_pinned_pages() 134 pin->nr_pages * sizeof(*pin->pages)); in scif_destroy_pinned_pages() 148 struct scif_window *scif_create_window(struct scif_endpt *ep, int nr_pages, in scif_create_window() argument 158 window->dma_addr = scif_zalloc(nr_pages * sizeof(*window->dma_addr)); in scif_create_window() 162 window->num_pages = scif_zalloc(nr_pages * sizeof(*window->num_pages)); in scif_create_window() 180 nr_pages * sizeof(*window->dma_addr)); in scif_create_window() 197 int nr_pages = window->nr_pages; in scif_destroy_incomplete_window() local 222 scif_free(window->dma_addr, nr_pages * sizeof(*window->dma_addr)); in scif_destroy_incomplete_window() [all …]
|
D | scif_mmap.c | 43 req.nr_bytes = recv_window->nr_pages << PAGE_SHIFT; in scif_recv_munmap() 57 scif_put_window(window, window->nr_pages); in scif_recv_munmap() 139 scif_put_window(window, window->nr_pages); in __scif_cleanup_rma_for_zombies() 227 int nr_pages, err, i; in scif_get_pages() local 242 nr_pages = len >> PAGE_SHIFT; in scif_get_pages() 268 (*pages)->phys_addr = scif_zalloc(nr_pages * sizeof(dma_addr_t)); in scif_get_pages() 276 ((*pages)->va = scif_zalloc(nr_pages * sizeof(void *))); in scif_get_pages() 284 (*pages)->nr_pages = nr_pages; in scif_get_pages() 287 for (i = 0; i < nr_pages; i++) { in scif_get_pages() 300 scif_get_window(window, nr_pages); in scif_get_pages() [all …]
|
D | scif_rma.h | 187 s64 nr_pages; member 243 s64 nr_pages; member 310 s64 offset, int nr_pages, s64 *out_offset); 315 struct scif_window *scif_create_window(struct scif_endpt *ep, int nr_pages, 426 static inline void scif_get_window(struct scif_window *window, int nr_pages) in scif_get_window() argument 428 window->ref_count += nr_pages; in scif_get_window() 431 static inline void scif_put_window(struct scif_window *window, int nr_pages) in scif_put_window() argument 433 window->ref_count -= nr_pages; in scif_put_window() 436 static inline void scif_set_window_ref(struct scif_window *window, int nr_pages) in scif_set_window_ref() argument 438 window->ref_count = nr_pages; in scif_set_window_ref()
|
/Linux-v4.19/net/rds/ |
D | info.c | 163 unsigned long nr_pages = 0; in rds_info_getsockopt() local 188 nr_pages = (PAGE_ALIGN(start + len) - (start & PAGE_MASK)) in rds_info_getsockopt() 191 pages = kmalloc_array(nr_pages, sizeof(struct page *), GFP_KERNEL); in rds_info_getsockopt() 196 ret = get_user_pages_fast(start, nr_pages, 1, pages); in rds_info_getsockopt() 197 if (ret != nr_pages) { in rds_info_getsockopt() 199 nr_pages = ret; in rds_info_getsockopt() 201 nr_pages = 0; in rds_info_getsockopt() 206 rdsdebug("len %d nr_pages %lu\n", len, nr_pages); in rds_info_getsockopt() 238 for (i = 0; pages && i < nr_pages; i++) in rds_info_getsockopt()
|
D | rdma.c | 156 static int rds_pin_pages(unsigned long user_addr, unsigned int nr_pages, in rds_pin_pages() argument 161 ret = get_user_pages_fast(user_addr, nr_pages, write, pages); in rds_pin_pages() 163 if (ret >= 0 && ret < nr_pages) { in rds_pin_pages() 177 unsigned int nr_pages; in __rds_rdma_map() local 197 nr_pages = rds_pages_in_vec(&args->vec); in __rds_rdma_map() 198 if (nr_pages == 0) { in __rds_rdma_map() 206 if ((nr_pages - 1) > (RDS_MAX_MSG_SIZE >> PAGE_SHIFT)) { in __rds_rdma_map() 212 args->vec.addr, args->vec.bytes, nr_pages); in __rds_rdma_map() 215 pages = kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL); in __rds_rdma_map() 249 ret = rds_pin_pages(args->vec.addr, nr_pages, pages, 1); in __rds_rdma_map() [all …]
|
/Linux-v4.19/kernel/events/ |
D | ring_buffer.c | 143 if (rb->nr_pages) in __perf_output_begin() 203 handle->page = (offset >> page_shift) & (rb->nr_pages - 1); in __perf_output_begin() 297 if (!rb->nr_pages) in ring_buffer_init() 588 pgoff_t pgoff, int nr_pages, long watermark, int flags) in rb_alloc_aux() argument 602 max_order = ilog2(nr_pages); in rb_alloc_aux() 617 rb->aux_pages = kcalloc_node(nr_pages, sizeof(void *), GFP_KERNEL, in rb_alloc_aux() 623 for (rb->aux_nr_pages = 0; rb->aux_nr_pages < nr_pages;) { in rb_alloc_aux() 627 order = min(max_order, ilog2(nr_pages - rb->aux_nr_pages)); in rb_alloc_aux() 651 rb->aux_priv = event->pmu->setup_aux(event->cpu, rb->aux_pages, nr_pages, in rb_alloc_aux() 670 rb->aux_watermark = nr_pages << (PAGE_SHIFT - 1); in rb_alloc_aux() [all …]
|
/Linux-v4.19/arch/powerpc/platforms/powernv/ |
D | memtrace.c | 73 static bool memtrace_offline_pages(u32 nid, u64 start_pfn, u64 nr_pages) in memtrace_offline_pages() argument 75 u64 end_pfn = start_pfn + nr_pages - 1; in memtrace_offline_pages() 84 if (offline_pages(start_pfn, nr_pages)) { in memtrace_offline_pages() 94 remove_memory(nid, start_pfn << PAGE_SHIFT, nr_pages << PAGE_SHIFT); in memtrace_offline_pages() 102 u64 start_pfn, end_pfn, nr_pages; in memtrace_alloc_node() local 110 nr_pages = size >> PAGE_SHIFT; in memtrace_alloc_node() 113 end_pfn = round_down(end_pfn - nr_pages, nr_pages); in memtrace_alloc_node() 115 for (base_pfn = end_pfn; base_pfn > start_pfn; base_pfn -= nr_pages) { in memtrace_alloc_node() 116 if (memtrace_offline_pages(nid, base_pfn, nr_pages) == true) in memtrace_alloc_node()
|
/Linux-v4.19/tools/testing/selftests/vm/ |
D | userfaultfd.c | 76 static unsigned long nr_cpus, nr_pages, nr_pages_per_cpu, page_size; variable 122 if (madvise(rel_area, nr_pages * page_size, MADV_DONTNEED)) { in anon_release_pages() 132 if (posix_memalign(alloc_area, page_size, nr_pages * page_size)) { in anon_allocate_area() 149 nr_pages * page_size, in hugetlb_release_pages() 150 nr_pages * page_size)) { in hugetlb_release_pages() 163 *alloc_area = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, in hugetlb_allocate_area() 167 nr_pages * page_size); in hugetlb_allocate_area() 174 area_alias = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, in hugetlb_allocate_area() 177 nr_pages * page_size); in hugetlb_allocate_area() 179 if (munmap(*alloc_area, nr_pages * page_size) < 0) in hugetlb_allocate_area() [all …]
|
/Linux-v4.19/fs/ |
D | fs-writeback.c | 46 long nr_pages; member 770 static long wb_split_bdi_pages(struct bdi_writeback *wb, long nr_pages) in wb_split_bdi_pages() argument 775 if (nr_pages == LONG_MAX) in wb_split_bdi_pages() 784 return nr_pages; in wb_split_bdi_pages() 786 return DIV_ROUND_UP_ULL((u64)nr_pages * this_bw, tot_bw); in wb_split_bdi_pages() 815 long nr_pages; in bdi_split_work_to_wbs() local 830 nr_pages = wb_split_bdi_pages(wb, base_work->nr_pages); in bdi_split_work_to_wbs() 835 work->nr_pages = nr_pages; in bdi_split_work_to_wbs() 844 work->nr_pages = nr_pages; in bdi_split_work_to_wbs() 918 static long wb_split_bdi_pages(struct bdi_writeback *wb, long nr_pages) in wb_split_bdi_pages() argument [all …]
|
/Linux-v4.19/drivers/staging/android/ion/ |
D | ion_cma_heap.c | 35 unsigned long nr_pages = size >> PAGE_SHIFT; in ion_cma_allocate() local 42 pages = cma_alloc(cma_heap->cma, nr_pages, align, false); in ion_cma_allocate() 47 unsigned long nr_clear_pages = nr_pages; in ion_cma_allocate() 79 cma_release(cma_heap->cma, pages, nr_pages); in ion_cma_allocate() 87 unsigned long nr_pages = PAGE_ALIGN(buffer->size) >> PAGE_SHIFT; in ion_cma_free() local 90 cma_release(cma_heap->cma, pages, nr_pages); in ion_cma_free()
|