/Linux-v5.10/fs/verity/ |
D | verify.c | 43 static void extract_hash(struct page *hpage, unsigned int hoffset, in extract_hash() argument 46 void *virt = kmap_atomic(hpage); in extract_hash() 114 struct page *hpage; in verify_page() local 121 hpage = inode->i_sb->s_vop->read_merkle_tree_page(inode, hindex, in verify_page() 123 if (IS_ERR(hpage)) { in verify_page() 124 err = PTR_ERR(hpage); in verify_page() 131 if (PageChecked(hpage)) { in verify_page() 132 extract_hash(hpage, hoffset, hsize, _want_hash); in verify_page() 134 put_page(hpage); in verify_page() 141 hpages[level] = hpage; in verify_page() [all …]
|
/Linux-v5.10/mm/ |
D | khugepaged.c | 848 static bool khugepaged_prealloc_page(struct page **hpage, bool *wait) in khugepaged_prealloc_page() argument 850 if (IS_ERR(*hpage)) { in khugepaged_prealloc_page() 855 *hpage = NULL; in khugepaged_prealloc_page() 857 } else if (*hpage) { in khugepaged_prealloc_page() 858 put_page(*hpage); in khugepaged_prealloc_page() 859 *hpage = NULL; in khugepaged_prealloc_page() 866 khugepaged_alloc_page(struct page **hpage, gfp_t gfp, int node) in khugepaged_alloc_page() argument 868 VM_BUG_ON_PAGE(*hpage, *hpage); in khugepaged_alloc_page() 870 *hpage = __alloc_pages_node(node, gfp, HPAGE_PMD_ORDER); in khugepaged_alloc_page() 871 if (unlikely(!*hpage)) { in khugepaged_alloc_page() [all …]
|
D | hwpoison-inject.c | 18 struct page *hpage; in hwpoison_inject() local 28 hpage = compound_head(p); in hwpoison_inject() 33 shake_page(hpage, 0); in hwpoison_inject() 37 if (!PageLRU(hpage) && !PageHuge(p)) in hwpoison_inject() 45 err = hwpoison_filter(hpage); in hwpoison_inject()
|
D | memory-failure.c | 813 struct page *hpage = compound_head(p); in me_huge_page() local 816 if (!PageHuge(hpage)) in me_huge_page() 819 mapping = page_mapping(hpage); in me_huge_page() 821 res = truncate_error_page(hpage, pfn, mapping); in me_huge_page() 823 unlock_page(hpage); in me_huge_page() 829 if (PageAnon(hpage)) in me_huge_page() 830 put_page(hpage); in me_huge_page() 833 lock_page(hpage); in me_huge_page() 997 struct page *hpage = *hpagep; in hwpoison_user_mappings() local 998 bool mlocked = PageMlocked(hpage); in hwpoison_user_mappings() [all …]
|
D | migrate.c | 1276 struct page *hpage, int force, in unmap_and_move_huge_page() argument 1292 if (!hugepage_migration_supported(page_hstate(hpage))) { in unmap_and_move_huge_page() 1293 putback_active_hugepage(hpage); in unmap_and_move_huge_page() 1297 new_hpage = get_new_page(hpage, private); in unmap_and_move_huge_page() 1301 if (!trylock_page(hpage)) { in unmap_and_move_huge_page() 1311 lock_page(hpage); in unmap_and_move_huge_page() 1319 if (page_private(hpage) && !page_mapping(hpage)) { in unmap_and_move_huge_page() 1324 if (PageAnon(hpage)) in unmap_and_move_huge_page() 1325 anon_vma = page_get_anon_vma(hpage); in unmap_and_move_huge_page() 1330 if (page_mapped(hpage)) { in unmap_and_move_huge_page() [all …]
|
D | hugetlb.c | 1578 struct address_space *hugetlb_page_mapping_lock_write(struct page *hpage) in hugetlb_page_mapping_lock_write() argument 1580 struct address_space *mapping = page_mapping(hpage); in hugetlb_page_mapping_lock_write()
|
/Linux-v5.10/arch/powerpc/kvm/ |
D | book3s_pr.c | 663 struct page *hpage; in kvmppc_patch_dcbz() local 668 hpage = gfn_to_page(vcpu->kvm, pte->raddr >> PAGE_SHIFT); in kvmppc_patch_dcbz() 669 if (is_error_page(hpage)) in kvmppc_patch_dcbz() 676 get_page(hpage); in kvmppc_patch_dcbz() 677 page = kmap_atomic(hpage); in kvmppc_patch_dcbz() 685 put_page(hpage); in kvmppc_patch_dcbz()
|
/Linux-v5.10/include/linux/ |
D | hugetlb.h | 157 struct address_space *hugetlb_page_mapping_lock_write(struct page *hpage); 203 struct page *hpage) in hugetlb_page_mapping_lock_write() argument
|
/Linux-v5.10/Documentation/virt/kvm/ |
D | running-nested-guests.rst | 169 .. note:: On s390x, the kernel parameter ``hpage`` is mutually exclusive 171 ``nested``, the ``hpage`` parameter *must* be disabled.
|
D | api.rst | 5856 :Returns: 0 on success, -EINVAL if hpage module parameter was not set 5864 hpage module parameter is not set to 1, -EINVAL is returned.
|
/Linux-v5.10/arch/s390/kvm/ |
D | kvm-s390.c | 178 static int hpage; variable 179 module_param(hpage, int, 0444); 180 MODULE_PARM_DESC(hpage, "1m huge page backing support"); 553 if (hpage && !kvm_is_ucontrol(kvm)) in kvm_vm_ioctl_check_extension() 763 else if (!hpage || kvm->arch.use_cmma || kvm_is_ucontrol(kvm)) in kvm_vm_ioctl_enable_cap() 5043 if (nested && hpage) { in kvm_s390_init()
|
/Linux-v5.10/fs/ |
D | io_uring.c | 8038 int nr_pages, struct page *hpage) in headpage_already_acct() argument 8046 if (compound_head(pages[i]) == hpage) in headpage_already_acct() 8057 if (compound_head(imu->bvec[j].bv_page) == hpage) in headpage_already_acct() 8075 struct page *hpage; in io_buffer_account_pin() local 8077 hpage = compound_head(pages[i]); in io_buffer_account_pin() 8078 if (hpage == *last_hpage) in io_buffer_account_pin() 8080 *last_hpage = hpage; in io_buffer_account_pin() 8081 if (headpage_already_acct(ctx, pages, i, hpage)) in io_buffer_account_pin() 8083 imu->acct_pages += page_size(hpage) >> PAGE_SHIFT; in io_buffer_account_pin()
|