Lines Matching refs:nr_pages
65 static unsigned long nr_cpus, nr_pages, nr_pages_per_cpu, page_size; variable
207 if (madvise(rel_area, nr_pages * page_size, MADV_DONTNEED)) in anon_release_pages()
213 *alloc_area = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, in anon_allocate_area()
226 rel_area == huge_fd_off0 ? 0 : nr_pages * page_size, in hugetlb_release_pages()
227 nr_pages * page_size)) in hugetlb_release_pages()
236 *alloc_area = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, in hugetlb_allocate_area()
240 nr_pages * page_size); in hugetlb_allocate_area()
245 area_alias = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, in hugetlb_allocate_area()
248 nr_pages * page_size); in hugetlb_allocate_area()
278 if (madvise(rel_area, nr_pages * page_size, MADV_REMOVE)) in shmem_release_pages()
286 unsigned long offset = is_src ? 0 : nr_pages * page_size; in shmem_allocate_area()
288 *alloc_area = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, in shmem_allocate_area()
293 area_alias = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, in shmem_allocate_area()
371 if (munmap(*area, nr_pages * page_size)) in munmap_area()
419 count_verify = malloc(nr_pages * sizeof(unsigned long long)); in uffd_test_ctx_init_ext()
423 for (nr = 0; nr < nr_pages; nr++) { in uffd_test_ctx_init_ext()
555 page_nr %= nr_pages; in locking_thread()
602 if (offset >= nr_pages * page_size) in __copy_page()
921 split_nr_pages = (nr_pages + 1) / 2; in faulting_process()
923 split_nr_pages = nr_pages; in faulting_process()
983 area_dst = mremap(area_dst, nr_pages * page_size, nr_pages * page_size, in faulting_process()
990 for (; nr < nr_pages; nr++) { in faulting_process()
1005 for (nr = 0; nr < nr_pages; nr++) in faulting_process()
1038 if (offset >= nr_pages * page_size) in __uffdio_zeropage()
1085 uffdio_register.range.len = nr_pages * page_size; in userfaultfd_zeropage_test()
1124 uffdio_register.range.len = nr_pages * page_size; in userfaultfd_events_test()
1155 return stats.missing_faults != nr_pages; in userfaultfd_events_test()
1178 uffdio_register.range.len = nr_pages * page_size; in userfaultfd_sig_test()
1254 uffdio_register.range.len = nr_pages * page_size; in userfaultfd_minor_test()
1268 for (p = 0; p < nr_pages; ++p) { in userfaultfd_minor_test()
1286 for (p = 0; p < nr_pages; ++p) { in userfaultfd_minor_test()
1301 return stats.missing_faults != 0 || stats.minor_faults != nr_pages; in userfaultfd_minor_test()
1374 if (test_pgsize > nr_pages * page_size) in userfaultfd_pagemap_test()
1385 if (madvise(area_dst, nr_pages * page_size, MADV_HUGEPAGE)) in userfaultfd_pagemap_test()
1389 if (madvise(area_dst, nr_pages * page_size, MADV_NOHUGEPAGE)) in userfaultfd_pagemap_test()
1394 uffdio_register.range.len = nr_pages * page_size; in userfaultfd_pagemap_test()
1482 uffdio_register.range.len = nr_pages * page_size; in userfaultfd_stress()
1535 nr_pages * page_size, false); in userfaultfd_stress()
1549 for (nr = 0; nr < nr_pages; nr++) in userfaultfd_stress()
1680 nr_pages = nr_pages_per_cpu * nr_cpus; in main()
1694 if (ftruncate(shm_fd, nr_pages * page_size * 2)) in main()
1698 nr_pages * page_size * 2)) in main()
1702 nr_pages, nr_pages_per_cpu); in main()