Home
last modified time | relevance | path

Searched refs:page_size (Results 1 – 25 of 348) sorted by relevance

12345678910>>...14

/Linux-v5.15/tools/testing/selftests/vm/
Dmremap_dontunmap.c22 unsigned long page_size; variable
49 void *source_mapping = mmap(NULL, num_pages * page_size, PROT_NONE, in kernel_support_for_mremap_dontunmap()
56 mremap(source_mapping, num_pages * page_size, num_pages * page_size, in kernel_support_for_mremap_dontunmap()
61 BUG_ON(munmap(dest_mapping, num_pages * page_size) == -1, in kernel_support_for_mremap_dontunmap()
65 BUG_ON(munmap(source_mapping, num_pages * page_size) == -1, in kernel_support_for_mremap_dontunmap()
74 BUG_ON(size & (page_size - 1), in check_region_contains_byte()
76 BUG_ON((unsigned long)addr & (page_size - 1), in check_region_contains_byte()
79 memset(page_buffer, byte, page_size); in check_region_contains_byte()
81 unsigned long num_pages = size / page_size; in check_region_contains_byte()
87 memcmp(addr + (i * page_size), page_buffer, page_size); in check_region_contains_byte()
[all …]
Dmlock2-tests.c195 unsigned long page_size = getpagesize(); in test_mlock_lock() local
197 map = mmap(NULL, 2 * page_size, PROT_READ | PROT_WRITE, in test_mlock_lock()
204 if (mlock2_(map, 2 * page_size, 0)) { in test_mlock_lock()
217 if (munlock(map, 2 * page_size)) { in test_mlock_lock()
225 munmap(map, 2 * page_size); in test_mlock_lock()
243 unsigned long page_size = getpagesize(); in unlock_onfault_check() local
246 is_vma_lock_on_fault((unsigned long)map + page_size)) { in unlock_onfault_check()
258 unsigned long page_size = getpagesize(); in test_mlock_onfault() local
260 map = mmap(NULL, 2 * page_size, PROT_READ | PROT_WRITE, in test_mlock_onfault()
267 if (mlock2_(map, 2 * page_size, MLOCK_ONFAULT)) { in test_mlock_onfault()
[all …]
Dksm_tests.c242 static int check_ksm_merge(int mapping, int prot, long page_count, int timeout, size_t page_size) in check_ksm_merge() argument
253 map_ptr = allocate_memory(NULL, prot, mapping, '*', page_size * page_count); in check_ksm_merge()
257 if (ksm_merge_pages(map_ptr, page_size * page_count, start_time, timeout)) in check_ksm_merge()
263 munmap(map_ptr, page_size * page_count); in check_ksm_merge()
269 munmap(map_ptr, page_size * page_count); in check_ksm_merge()
273 static int check_ksm_unmerge(int mapping, int prot, int timeout, size_t page_size) in check_ksm_unmerge() argument
285 map_ptr = allocate_memory(NULL, prot, mapping, '*', page_size * page_count); in check_ksm_unmerge()
289 if (ksm_merge_pages(map_ptr, page_size * page_count, start_time, timeout)) in check_ksm_unmerge()
294 memset(map_ptr + page_size, '+', 1); in check_ksm_unmerge()
303 munmap(map_ptr, page_size * page_count); in check_ksm_unmerge()
[all …]
Dmap_fixed_noreplace.c33 unsigned long flags, addr, size, page_size; in main() local
36 page_size = sysconf(_SC_PAGE_SIZE); in main()
43 size = 5 * page_size; in main()
55 if (munmap((void *)addr, 5 * page_size) != 0) { in main()
63 addr = BASE_ADDRESS + page_size; in main()
64 size = 3 * page_size; in main()
84 size = 5 * page_size; in main()
104 addr = BASE_ADDRESS + (2 * page_size); in main()
105 size = page_size; in main()
124 addr = BASE_ADDRESS + (3 * page_size); in main()
[all …]
Duserfaultfd.c65 static unsigned long nr_cpus, nr_pages, nr_pages_per_cpu, page_size; variable
108 ((pthread_mutex_t *) ((___area) + (___nr)*page_size))
115 ((___area) + (___nr)*page_size + \
207 if (madvise(rel_area, nr_pages * page_size, MADV_DONTNEED)) in anon_release_pages()
213 *alloc_area = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, in anon_allocate_area()
226 rel_area == huge_fd_off0 ? 0 : nr_pages * page_size, in hugetlb_release_pages()
227 nr_pages * page_size)) in hugetlb_release_pages()
236 *alloc_area = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, in hugetlb_allocate_area()
240 nr_pages * page_size); in hugetlb_allocate_area()
245 area_alias = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, in hugetlb_allocate_area()
[all …]
Dkhugepaged.c20 static unsigned long page_size; variable
455 for (i = start / page_size; i < end / page_size; i++) in fill_memory()
456 p[i * page_size / sizeof(*p)] = i + 0xdead0000; in fill_memory()
463 for (i = start / page_size; i < end / page_size; i++) { in validate_memory()
464 if (p[i * page_size / sizeof(*p)] != i + 0xdead0000) { in validate_memory()
466 i, p[i * page_size / sizeof(*p)]); in validate_memory()
522 madvise(p, page_size, MADV_DONTNEED); in alloc_at_fault()
566 fill_memory(p, 0, page_size); in collapse_single_pte_entry()
573 validate_memory(p, 0, page_size); in collapse_single_pte_entry()
588 fill_memory(p, 0, (hpage_pmd_nr - max_ptes_none - 1) * page_size); in collapse_max_ptes_none()
[all …]
Dmremap_test.c52 #define PTE page_size
267 int page_size; in main() local
278 page_size = sysconf(_SC_PAGESIZE); in main()
281 test_cases[0] = MAKE_TEST(page_size, page_size, page_size, in main()
285 test_cases[1] = MAKE_TEST(page_size, page_size/4, page_size, in main()
288 test_cases[2] = MAKE_TEST(page_size/4, page_size, page_size, in main()
321 perf_test_cases[0] = MAKE_TEST(page_size, page_size, _1GB, NON_OVERLAPPING, EXPECT_SUCCESS, in main()
Dmemfd_secret.c37 static unsigned long page_size; variable
181 mem = mmap(NULL, page_size, prot, mode, fd, 0); in test_remote_access()
187 ftruncate(fd, page_size); in test_remote_access()
188 memset(mem, PATTERN, page_size); in test_remote_access()
233 page_size = sysconf(_SC_PAGE_SIZE); in prepare()
234 if (!page_size) in prepare()
246 page_size, mlock_limit_cur, mlock_limit_max); in prepare()
248 if (page_size > mlock_limit_cur) in prepare()
249 mlock_limit_cur = page_size; in prepare()
250 if (page_size > mlock_limit_max) in prepare()
[all …]
Dhmm-tests.c59 unsigned int page_size; in FIXTURE() local
67 unsigned int page_size; in FIXTURE() local
86 self->page_size = sysconf(_SC_PAGE_SIZE); in FIXTURE_SETUP()
87 self->page_shift = ffs(self->page_size) - 1; in FIXTURE_SETUP()
95 self->page_size = sysconf(_SC_PAGE_SIZE); in FIXTURE_SETUP()
96 self->page_shift = ffs(self->page_size) - 1; in FIXTURE_SETUP()
234 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift; in TEST_F()
256 i = 2 * self->page_size / sizeof(*ptr); in TEST_F()
265 val = *(int *)(buffer->ptr + self->page_size); in TEST_F()
276 for (i = 0; i < 2 * self->page_size / sizeof(*ptr); ++i) in TEST_F()
[all …]
/Linux-v5.15/tools/testing/selftests/mincore/
Dmincore_selftest.c33 int page_size; in TEST() local
37 page_size = sysconf(_SC_PAGESIZE); in TEST()
45 retval = mincore(NULL, page_size, vec); in TEST()
50 addr = mmap(NULL, page_size, PROT_READ | PROT_WRITE, in TEST()
58 retval = mincore(addr + 1, page_size, vec); in TEST()
70 retval = mincore(addr, page_size, NULL); in TEST()
73 munmap(addr, page_size); in TEST()
88 int page_size; in TEST() local
90 page_size = sysconf(_SC_PAGESIZE); in TEST()
94 addr = mmap(NULL, page_size, PROT_READ | PROT_WRITE, in TEST()
[all …]
/Linux-v5.15/drivers/misc/habanalabs/common/mmu/
Dmmu.c16 return hl_mem_area_inside_range(virt_addr, prop->dmmu.page_size, in hl_is_dram_va()
145 int hl_mmu_unmap_page(struct hl_ctx *ctx, u64 virt_addr, u32 page_size, in hl_mmu_unmap_page() argument
163 else if ((page_size % prop->pmmu_huge.page_size) == 0) in hl_mmu_unmap_page()
173 if ((page_size % mmu_prop->page_size) == 0) { in hl_mmu_unmap_page()
174 real_page_size = mmu_prop->page_size; in hl_mmu_unmap_page()
182 if (is_dram_addr && ((page_size % prop->dram_page_size) == 0)) { in hl_mmu_unmap_page()
187 page_size, mmu_prop->page_size >> 10); in hl_mmu_unmap_page()
193 npages = page_size / real_page_size; in hl_mmu_unmap_page()
234 u32 page_size, bool flush_pte) in hl_mmu_map_page() argument
252 else if ((page_size % prop->pmmu_huge.page_size) == 0) in hl_mmu_map_page()
[all …]
/Linux-v5.15/tools/testing/selftests/bpf/prog_tests/
Dringbuf.c90 int page_size = getpagesize(); in test_ringbuf() local
97 skel->maps.ringbuf.max_entries = page_size; in test_ringbuf()
105 mmap_ptr = mmap(NULL, page_size, PROT_READ | PROT_WRITE, MAP_SHARED, rb_fd, 0); in test_ringbuf()
107 tmp_ptr = mremap(mmap_ptr, page_size, 2 * page_size, MREMAP_MAYMOVE); in test_ringbuf()
110 ASSERT_ERR(mprotect(mmap_ptr, page_size, PROT_EXEC), "exec_cons_pos_protect"); in test_ringbuf()
111 ASSERT_OK(munmap(mmap_ptr, page_size), "unmap_rw"); in test_ringbuf()
114 mmap_ptr = mmap(NULL, page_size, PROT_WRITE, MAP_SHARED, rb_fd, page_size); in test_ringbuf()
120 mmap_ptr = mmap(NULL, page_size, PROT_WRITE, MAP_SHARED, rb_fd, 2 * page_size); in test_ringbuf()
124 mmap_ptr = mmap(NULL, page_size, PROT_WRITE, MAP_SHARED, rb_fd, 3 * page_size); in test_ringbuf()
126 mmap_ptr = mmap(NULL, 2 * page_size, PROT_WRITE, MAP_SHARED, rb_fd, 2 * page_size); in test_ringbuf()
[all …]
Dmmap.c12 long page_size = sysconf(_SC_PAGE_SIZE); in roundup_page() local
13 return (sz + page_size - 1) / page_size * page_size; in roundup_page()
21 const long page_size = sysconf(_SC_PAGE_SIZE); in test_mmap() local
36 err = bpf_map__set_max_entries(skel->maps.rdonly_map, page_size); in test_mmap()
42 4 * (page_size / sizeof(u64))); in test_mmap()
55 tmp1 = mmap(NULL, page_size, PROT_READ | PROT_WRITE, MAP_SHARED, rdmap_fd, 0); in test_mmap()
57 munmap(tmp1, page_size); in test_mmap()
61 tmp1 = mmap(NULL, page_size, PROT_READ, MAP_SHARED, rdmap_fd, 0); in test_mmap()
200 tmp0 = mmap(NULL, 4 * page_size, PROT_READ, MAP_SHARED | MAP_ANONYMOUS, in test_mmap()
206 tmp1 = mmap(tmp0, 3 * page_size, PROT_READ, MAP_SHARED | MAP_FIXED, in test_mmap()
[all …]
/Linux-v5.15/drivers/pci/endpoint/
Dpci-epc-mem.c26 unsigned int page_shift = ilog2(mem->window.page_size); in pci_epc_mem_get_order()
54 size_t page_size; in pci_epc_multi_mem_init() local
70 page_size = windows[i].page_size; in pci_epc_multi_mem_init()
71 if (page_size < PAGE_SIZE) in pci_epc_multi_mem_init()
72 page_size = PAGE_SIZE; in pci_epc_multi_mem_init()
73 page_shift = ilog2(page_size); in pci_epc_multi_mem_init()
94 mem->window.page_size = page_size; in pci_epc_multi_mem_init()
119 size_t size, size_t page_size) in pci_epc_mem_init() argument
125 mem_window.page_size = page_size; in pci_epc_mem_init()
182 align_size = ALIGN(size, mem->window.page_size); in pci_epc_mem_alloc_addr()
[all …]
/Linux-v5.15/arch/powerpc/mm/
Dinit_64.c186 unsigned long page_size) in altmap_cross_boundary() argument
188 unsigned long nr_pfn = page_size / sizeof(struct page); in altmap_cross_boundary()
204 unsigned long page_size = 1 << mmu_psize_defs[mmu_vmemmap_psize].shift; in vmemmap_populate() local
207 start = ALIGN_DOWN(start, page_size); in vmemmap_populate()
211 for (; start < end; start += page_size) { in vmemmap_populate()
221 if (vmemmap_populated(start, page_size)) in vmemmap_populate()
229 if (altmap && !altmap_cross_boundary(altmap, start, page_size)) { in vmemmap_populate()
230 p = vmemmap_alloc_block_buf(page_size, node, altmap); in vmemmap_populate()
237 p = vmemmap_alloc_block_buf(page_size, node, NULL); in vmemmap_populate()
250 int nr_pfns = page_size >> PAGE_SHIFT; in vmemmap_populate()
[all …]
/Linux-v5.15/tools/testing/selftests/powerpc/primitives/
Dload_unaligned_zeropad.c38 static int page_size; variable
43 if (mprotect(mem_region + page_size, page_size, PROT_NONE)) { in protect_region()
53 if (mprotect(mem_region + page_size, page_size, PROT_READ|PROT_WRITE)) { in unprotect_region()
125 page_size = getpagesize(); in test_body()
126 mem_region = mmap(NULL, page_size * 2, PROT_READ|PROT_WRITE, in test_body()
131 for (i = 0; i < page_size; i++) in test_body()
134 memset(mem_region+page_size, 0, page_size); in test_body()
138 for (i = 0; i < page_size; i++) in test_body()
/Linux-v5.15/drivers/mtd/spi-nor/
Dxilinx.c33 offset = addr % nor->page_size; in s3an_convert_addr()
34 page = addr / nor->page_size; in s3an_convert_addr()
35 page <<= (nor->page_size > 512) ? 10 : 9; in s3an_convert_addr()
67 nor->page_size = (nor->page_size == 264) ? 256 : 512; in xilinx_nor_setup()
68 nor->mtd.writebufsize = nor->page_size; in xilinx_nor_setup()
69 nor->mtd.size = 8 * nor->page_size * nor->info->n_sectors; in xilinx_nor_setup()
70 nor->mtd.erasesize = 8 * nor->page_size; in xilinx_nor_setup()
/Linux-v5.15/tools/testing/selftests/kvm/lib/s390x/
Dprocessor.c18 TEST_ASSERT(vm->page_size == 4096, "Unsupported page size: 0x%x", in virt_pgd_alloc()
19 vm->page_size); in virt_pgd_alloc()
26 memset(addr_gpa2hva(vm, paddr), 0xff, PAGES_PER_REGION * vm->page_size); in virt_pgd_alloc()
43 memset(addr_gpa2hva(vm, taddr), 0xff, PAGES_PER_REGION * vm->page_size); in virt_alloc_region()
55 TEST_ASSERT((gva % vm->page_size) == 0, in virt_pg_map()
58 gva, vm->page_size); in virt_pg_map()
63 TEST_ASSERT((gpa % vm->page_size) == 0, in virt_pg_map()
66 gva, vm->page_size); in virt_pg_map()
70 gva, vm->max_gfn, vm->page_size); in virt_pg_map()
94 TEST_ASSERT(vm->page_size == 4096, "Unsupported page size: 0x%x", in addr_gva2gpa()
[all …]
/Linux-v5.15/drivers/misc/
Dvmw_balloon.c247 enum vmballoon_page_size_type page_size; member
576 unsigned int vmballoon_page_order(enum vmballoon_page_size_type page_size) in vmballoon_page_order() argument
578 return page_size == VMW_BALLOON_2M_PAGE ? VMW_BALLOON_2M_ORDER : 0; in vmballoon_page_order()
588 vmballoon_page_in_frames(enum vmballoon_page_size_type page_size) in vmballoon_page_in_frames() argument
590 return 1 << vmballoon_page_order(page_size); in vmballoon_page_in_frames()
600 enum vmballoon_page_size_type page_size) in vmballoon_mark_page_offline() argument
604 for (i = 0; i < vmballoon_page_in_frames(page_size); i++) in vmballoon_mark_page_offline()
615 enum vmballoon_page_size_type page_size) in vmballoon_mark_page_online() argument
619 for (i = 0; i < vmballoon_page_in_frames(page_size); i++) in vmballoon_mark_page_online()
679 if (ctl->page_size == VMW_BALLOON_2M_PAGE) in vmballoon_alloc_page_list()
[all …]
/Linux-v5.15/drivers/net/ethernet/qlogic/qed/
Dqed_chain.c22 params->page_size); in qed_chain_init()
24 params->page_size, in qed_chain_init()
33 chain->page_size = params->page_size; in qed_chain_init()
88 dma_free_coherent(dev, chain->page_size, virt, phys); in qed_chain_free_next_ptr()
101 dma_free_coherent(&cdev->pdev->dev, chain->page_size, in qed_chain_free_single()
119 dma_free_coherent(dev, chain->page_size, entry->virt_addr, in qed_chain_free_pbl()
164 chain_size = ELEMS_PER_PAGE(params->elem_size, params->page_size); in qed_chain_alloc_sanity_check()
207 virt = dma_alloc_coherent(dev, chain->page_size, &phys, in qed_chain_alloc_next_ptr()
238 virt = dma_alloc_coherent(&cdev->pdev->dev, chain->page_size, in qed_chain_alloc_single()
290 virt = dma_alloc_coherent(dev, chain->page_size, &phys, in qed_chain_alloc_pbl()
[all …]
/Linux-v5.15/tools/power/acpi/os_specific/service_layers/
Dosunixmap.c67 acpi_size page_size; in acpi_os_map_memory() local
78 page_size = acpi_os_get_page_size(); in acpi_os_map_memory()
79 offset = where % page_size; in acpi_os_map_memory()
112 acpi_size page_size; in acpi_os_unmap_memory() local
114 page_size = acpi_os_get_page_size(); in acpi_os_unmap_memory()
115 offset = ACPI_TO_INTEGER(where) % page_size; in acpi_os_unmap_memory()
/Linux-v5.15/drivers/infiniband/hw/mlx5/
Dmem.c43 void mlx5_ib_populate_pas(struct ib_umem *umem, size_t page_size, __be64 *pas, in mlx5_ib_populate_pas() argument
48 rdma_umem_for_each_dma_block (umem, &biter, page_size) { in mlx5_ib_populate_pas()
66 unsigned long page_size; in __mlx5_umem_find_best_quantized_pgoff() local
69 page_size = ib_umem_find_best_pgoff(umem, pgsz_bitmap, pgoff_bitmask); in __mlx5_umem_find_best_quantized_pgoff()
70 if (!page_size) in __mlx5_umem_find_best_quantized_pgoff()
80 page_offset = ib_umem_dma_offset(umem, page_size); in __mlx5_umem_find_best_quantized_pgoff()
81 while (page_offset & ~(u64)(page_offset_mask * (page_size / scale))) { in __mlx5_umem_find_best_quantized_pgoff()
82 page_size /= 2; in __mlx5_umem_find_best_quantized_pgoff()
83 page_offset = ib_umem_dma_offset(umem, page_size); in __mlx5_umem_find_best_quantized_pgoff()
90 if (!(pgsz_bitmap & page_size)) in __mlx5_umem_find_best_quantized_pgoff()
[all …]
/Linux-v5.15/tools/testing/selftests/powerpc/copyloops/
Dexc_validate.c81 int page_size; in test_copy_exception() local
85 page_size = getpagesize(); in test_copy_exception()
86 p = mmap(NULL, page_size * 2, PROT_READ|PROT_WRITE, in test_copy_exception()
94 memset(p, 0, page_size); in test_copy_exception()
98 if (mprotect(p + page_size, page_size, PROT_NONE)) { in test_copy_exception()
103 q = p + page_size - MAX_LEN; in test_copy_exception()
/Linux-v5.15/drivers/misc/habanalabs/common/
Dmemory.c60 u32 num_curr_pgs, page_size; in alloc_device_memory() local
65 page_size = hdev->asic_prop.dram_page_size; in alloc_device_memory()
66 num_pgs = DIV_ROUND_UP_ULL(args->alloc.mem_size, page_size); in alloc_device_memory()
67 total_size = num_pgs * page_size; in alloc_device_memory()
95 phys_pg_pack->page_size = page_size; in alloc_device_memory()
108 phys_pg_pack->pages[i] = paddr + i * page_size; in alloc_device_memory()
113 page_size); in alloc_device_memory()
153 page_size); in alloc_device_memory()
302 phys_pg_pack->page_size); in free_phys_pg_pack()
309 phys_pg_pack->page_size); in free_phys_pg_pack()
[all …]
/Linux-v5.15/arch/um/os-Linux/
Delf_aux.c28 long page_size = 0; in scan_elf_aux() local
59 page_size = auxv->a_un.a_val; in scan_elf_aux()
65 ! page_size || (vsyscall_ehdr % page_size) ) { in scan_elf_aux()
72 vsyscall_end = vsyscall_ehdr + page_size; in scan_elf_aux()

12345678910>>...14