/Linux-v4.19/arch/xtensa/mm/ |
D | kasan_init.c | 27 for (i = 0; i < PTRS_PER_PTE; ++i) in kasan_early_init() 41 unsigned long n_pmds = n_pages / PTRS_PER_PTE; in populate() 53 for (k = 0; k < PTRS_PER_PTE; ++k, ++j) { in populate() 62 for (i = 0; i < n_pmds ; ++i, pte += PTRS_PER_PTE) in populate() 85 for (i = 0; i < PTRS_PER_PTE; ++i) in kasan_init()
|
D | mmu.c | 29 n_pages = ALIGN(n_pages, PTRS_PER_PTE); in init_pmd() 39 for (i = 0; i < n_pages; i += PTRS_PER_PTE, ++pmd) { in init_pmd()
|
/Linux-v4.19/arch/s390/mm/ |
D | pgalloc.c | 164 memset64(table, _PAGE_INVALID, PTRS_PER_PTE); in page_table_alloc_pgste() 165 memset64(table + PTRS_PER_PTE, 0, PTRS_PER_PTE); in page_table_alloc_pgste() 199 table += PTRS_PER_PTE; in page_table_alloc() 223 memset64((u64 *)table, _PAGE_INVALID, PTRS_PER_PTE); in page_table_alloc() 224 memset64((u64 *)table + PTRS_PER_PTE, 0, PTRS_PER_PTE); in page_table_alloc() 228 memset64((u64 *)table, _PAGE_INVALID, 2 * PTRS_PER_PTE); in page_table_alloc() 244 bit = (__pa(table) & ~PAGE_MASK)/(PTRS_PER_PTE*sizeof(pte_t)); in page_table_free() 278 bit = (__pa(table) & ~PAGE_MASK) / (PTRS_PER_PTE*sizeof(pte_t)); in page_table_free_rcu() 391 memset64(table, _PAGE_INVALID, PTRS_PER_PTE); in base_pgt_alloc()
|
D | pageattr.c | 72 mask = ~(PTRS_PER_PTE * sizeof(pte_t) - 1); in pgt_set() 127 for (i = 0; i < PTRS_PER_PTE; i++) { in split_pmd_page() 134 update_page_count(PG_DIRECT_MAP_4K, PTRS_PER_PTE); in split_pmd_page() 355 nr = PTRS_PER_PTE - (nr & (PTRS_PER_PTE - 1)); in __kernel_map_pages()
|
/Linux-v4.19/arch/m68k/include/asm/ |
D | pgtable_mm.h | 57 #define PTRS_PER_PTE 16 macro 62 #define PTRS_PER_PTE 512 macro 67 #define PTRS_PER_PTE 1024 macro
|
D | motorola_pgtable.h | 116 ptbl += (sizeof(pte_t)*PTRS_PER_PTE/16); in pmd_set() 220 return (pte_t *)__pmd_page(*pmdp) + ((address >> PAGE_SHIFT) & (PTRS_PER_PTE - 1)); in pte_offset_kernel() 223 …t_map(pmdp,address) ((pte_t *)__pmd_page(*pmdp) + (((address) >> PAGE_SHIFT) & (PTRS_PER_PTE - 1)))
|
/Linux-v4.19/arch/hexagon/include/asm/ |
D | pgtable.h | 91 #define PTRS_PER_PTE 1024 macro 95 #define PTRS_PER_PTE 256 macro 99 #define PTRS_PER_PTE 64 macro 103 #define PTRS_PER_PTE 16 macro 107 #define PTRS_PER_PTE 4 macro 445 #define __pte_offset(address) (((address) >> PAGE_SHIFT) & (PTRS_PER_PTE - 1))
|
/Linux-v4.19/arch/arm/include/asm/ |
D | pgtable-2level.h | 73 #define PTRS_PER_PTE 512 macro 77 #define PTE_HWTABLE_PTRS (PTRS_PER_PTE) 79 #define PTE_HWTABLE_SIZE (PTRS_PER_PTE * sizeof(u32))
|
/Linux-v4.19/arch/m68k/mm/ |
D | sun3mmu.c | 66 next_pgtable += PTRS_PER_PTE * sizeof (pte_t); in paging_init() 72 for (i=0; i<PTRS_PER_PTE; ++i, ++pg_table) { in paging_init()
|
D | mcfmmu.c | 63 next_pgtable += PTRS_PER_PTE * sizeof(pte_t); in paging_init() 68 for (i = 0; i < PTRS_PER_PTE; ++i, ++pg_table) { in paging_init()
|
/Linux-v4.19/arch/powerpc/mm/ |
D | subpage-prot.c | 120 i = (addr >> PAGE_SHIFT) & (PTRS_PER_PTE - 1); in subpage_prot_clear() 121 nw = PTRS_PER_PTE - i; in subpage_prot_clear() 250 i = (addr >> PAGE_SHIFT) & (PTRS_PER_PTE - 1); in SYSCALL_DEFINE3() 251 nw = PTRS_PER_PTE - i; in SYSCALL_DEFINE3()
|
D | hash64_64k.c | 89 rpte = __real_pte(__pte(old_pte), ptep, PTRS_PER_PTE); in __hash_page_4K() 216 new_pte |= pte_set_hidx(ptep, rpte, subpg_index, slot, PTRS_PER_PTE); in __hash_page_4K() 264 rpte = __real_pte(__pte(old_pte), ptep, PTRS_PER_PTE); in __hash_page_64K() 329 new_pte |= pte_set_hidx(ptep, rpte, 0, slot, PTRS_PER_PTE); in __hash_page_64K()
|
D | hash64_4k.c | 58 rpte = __real_pte(__pte(old_pte), ptep, PTRS_PER_PTE); in __hash_page_4K() 120 new_pte |= pte_set_hidx(ptep, rpte, 0, slot, PTRS_PER_PTE); in __hash_page_4K()
|
D | tlb_hash64.c | 85 offset = PTRS_PER_PTE; in hpte_need_flush() 251 for (pte = start_pte; pte < start_pte + PTRS_PER_PTE; pte++) { in flush_tlb_pmd_range()
|
/Linux-v4.19/arch/unicore32/include/asm/ |
D | pgalloc.h | 43 clean_dcache_area(pte, PTRS_PER_PTE * sizeof(pte_t)); in pte_alloc_one_kernel() 58 clean_dcache_area(page, PTRS_PER_PTE * sizeof(pte_t)); in pte_alloc_one()
|
/Linux-v4.19/arch/arc/mm/ |
D | highmem.c | 136 BUILD_BUG_ON(KM_TYPE_NR > PTRS_PER_PTE); in kmap_init() 139 BUILD_BUG_ON(LAST_PKMAP > PTRS_PER_PTE); in kmap_init()
|
/Linux-v4.19/arch/mips/include/asm/ |
D | pgtable-64.h | 136 #define PTRS_PER_PTE ((PAGE_SIZE << PTE_ORDER) / sizeof(pte_t)) macro 149 min(PTRS_PER_PGD * PTRS_PER_PUD * PTRS_PER_PMD * PTRS_PER_PTE * PAGE_SIZE, \ 172 extern pte_t invalid_pte_table[PTRS_PER_PTE]; 342 (((address) >> PAGE_SHIFT) & (PTRS_PER_PTE - 1))
|
/Linux-v4.19/arch/arc/include/asm/ |
D | pgalloc.h | 90 return get_order(PTRS_PER_PTE * sizeof(pte_t)); in __get_order_pte() 113 memzero((void *)pte_pg, PTRS_PER_PTE * sizeof(pte_t)); in pte_alloc_one()
|
/Linux-v4.19/arch/x86/power/ |
D | hibernate_32.c | 114 pfn += PTRS_PER_PTE; in resume_physical_mapping_init() 122 max_pte = pte + PTRS_PER_PTE; in resume_physical_mapping_init()
|
/Linux-v4.19/arch/arm64/include/asm/ |
D | pgtable-hwdef.h | 54 #define PTRS_PER_PTE (1 << (PAGE_SHIFT - 3)) macro 63 #define PTRS_PER_PMD PTRS_PER_PTE 73 #define PTRS_PER_PUD PTRS_PER_PTE
|
/Linux-v4.19/arch/nios2/include/asm/ |
D | pgtable.h | 78 #define PTRS_PER_PTE ((PAGE_SIZE << PTE_ORDER) / sizeof(pte_t)) macro 256 (((addr) >> PAGE_SHIFT) & (PTRS_PER_PTE - 1))) 264 (((addr) >> PAGE_SHIFT) & (PTRS_PER_PTE - 1)))
|
/Linux-v4.19/arch/x86/include/asm/ |
D | pgtable-2level_types.h | 36 #define PTRS_PER_PTE 1024 macro
|
D | pgtable-3level_types.h | 46 #define PTRS_PER_PTE 512 macro
|
/Linux-v4.19/Documentation/xtensa/ |
D | mmu.txt | 86 | KMAP area | PKMAP_BASE PTRS_PER_PTE * 128 | KMAP area | PKMAP_BASE PTRS_PER_PTE * 171 | KMAP area | PKMAP_BASE PTRS_PER_PTE *
|
/Linux-v4.19/arch/sparc/include/asm/ |
D | pgtable_32.h | 44 #define PTRS_PER_PTE 1024 macro 49 #define PTE_SIZE (PTRS_PER_PTE*4) 183 for (i = 0; i < PTRS_PER_PTE/SRMMU_REAL_PTRS_PER_PTE; i++) in pmd_clear()
|