Home
last modified time | relevance | path

Searched refs:new_page (Results 1 – 25 of 29) sorted by relevance

12

/Linux-v4.19/mm/
Dkhugepaged.c942 struct page *new_page; in collapse_huge_page() local
963 new_page = khugepaged_alloc_page(hpage, gfp, node); in collapse_huge_page()
964 if (!new_page) { in collapse_huge_page()
969 if (unlikely(mem_cgroup_try_charge(new_page, mm, gfp, &memcg, true))) { in collapse_huge_page()
977 mem_cgroup_cancel_charge(new_page, memcg, true); in collapse_huge_page()
985 mem_cgroup_cancel_charge(new_page, memcg, true); in collapse_huge_page()
996 mem_cgroup_cancel_charge(new_page, memcg, true); in collapse_huge_page()
1060 __collapse_huge_page_copy(pte, new_page, vma, address, pte_ptl); in collapse_huge_page()
1062 __SetPageUptodate(new_page); in collapse_huge_page()
1065 _pmd = mk_huge_pmd(new_page, vma->vm_page_prot); in collapse_huge_page()
[all …]
Dswap_state.c381 struct page *found_page, *new_page = NULL; in __read_swap_cache_async() local
410 if (!new_page) { in __read_swap_cache_async()
411 new_page = alloc_page_vma(gfp_mask, vma, addr); in __read_swap_cache_async()
412 if (!new_page) in __read_swap_cache_async()
443 __SetPageLocked(new_page); in __read_swap_cache_async()
444 __SetPageSwapBacked(new_page); in __read_swap_cache_async()
445 err = __add_to_swap_cache(new_page, entry); in __read_swap_cache_async()
451 lru_cache_add_anon(new_page); in __read_swap_cache_async()
453 return new_page; in __read_swap_cache_async()
456 __ClearPageLocked(new_page); in __read_swap_cache_async()
[all …]
Dmigrate.c1974 struct page *new_page = NULL; in migrate_misplaced_transhuge_page() local
1979 new_page = alloc_pages_node(node, in migrate_misplaced_transhuge_page()
1982 if (!new_page) in migrate_misplaced_transhuge_page()
1984 prep_transhuge_page(new_page); in migrate_misplaced_transhuge_page()
1988 put_page(new_page); in migrate_misplaced_transhuge_page()
1993 __SetPageLocked(new_page); in migrate_misplaced_transhuge_page()
1995 __SetPageSwapBacked(new_page); in migrate_misplaced_transhuge_page()
1998 new_page->mapping = page->mapping; in migrate_misplaced_transhuge_page()
1999 new_page->index = page->index; in migrate_misplaced_transhuge_page()
2000 migrate_page_copy(new_page, page); in migrate_misplaced_transhuge_page()
[all …]
Dhuge_memory.c1243 struct page *page = NULL, *new_page; in do_huge_pmd_wp_page() local
1294 new_page = alloc_hugepage_vma(huge_gfp, vma, haddr, HPAGE_PMD_ORDER); in do_huge_pmd_wp_page()
1296 new_page = NULL; in do_huge_pmd_wp_page()
1298 if (likely(new_page)) { in do_huge_pmd_wp_page()
1299 prep_transhuge_page(new_page); in do_huge_pmd_wp_page()
1316 if (unlikely(mem_cgroup_try_charge_delay(new_page, vma->vm_mm, in do_huge_pmd_wp_page()
1318 put_page(new_page); in do_huge_pmd_wp_page()
1330 clear_huge_page(new_page, vmf->address, HPAGE_PMD_NR); in do_huge_pmd_wp_page()
1332 copy_user_huge_page(new_page, page, vmf->address, in do_huge_pmd_wp_page()
1334 __SetPageUptodate(new_page); in do_huge_pmd_wp_page()
[all …]
Dmemory.c2491 struct page *new_page = NULL; in wp_page_copy() local
2502 new_page = alloc_zeroed_user_highpage_movable(vma, in wp_page_copy()
2504 if (!new_page) in wp_page_copy()
2507 new_page = alloc_page_vma(GFP_HIGHUSER_MOVABLE, vma, in wp_page_copy()
2509 if (!new_page) in wp_page_copy()
2511 cow_user_page(new_page, old_page, vmf->address, vma); in wp_page_copy()
2514 if (mem_cgroup_try_charge_delay(new_page, mm, GFP_KERNEL, &memcg, false)) in wp_page_copy()
2517 __SetPageUptodate(new_page); in wp_page_copy()
2536 entry = mk_pte(new_page, vma->vm_page_prot); in wp_page_copy()
2545 page_add_new_anon_rmap(new_page, vma, vmf->address, false); in wp_page_copy()
[all …]
Dksm.c2552 struct page *new_page; in ksm_might_need_to_copy() local
2567 new_page = alloc_page_vma(GFP_HIGHUSER_MOVABLE, vma, address); in ksm_might_need_to_copy()
2568 if (new_page) { in ksm_might_need_to_copy()
2569 copy_user_highpage(new_page, page, address, vma); in ksm_might_need_to_copy()
2571 SetPageDirty(new_page); in ksm_might_need_to_copy()
2572 __SetPageUptodate(new_page); in ksm_might_need_to_copy()
2573 __SetPageLocked(new_page); in ksm_might_need_to_copy()
2576 return new_page; in ksm_might_need_to_copy()
Dhugetlb.c3530 struct page *old_page, *new_page; in hugetlb_cow() local
3569 new_page = alloc_huge_page(vma, haddr, outside_reserve); in hugetlb_cow()
3571 if (IS_ERR(new_page)) { in hugetlb_cow()
3596 ret = vmf_error(PTR_ERR(new_page)); in hugetlb_cow()
3609 copy_user_huge_page(new_page, old_page, address, vma, in hugetlb_cow()
3611 __SetPageUptodate(new_page); in hugetlb_cow()
3612 set_page_huge_active(new_page); in hugetlb_cow()
3625 ClearPagePrivate(new_page); in hugetlb_cow()
3631 make_huge_pte(vma, new_page, 1)); in hugetlb_cow()
3633 hugepage_add_new_anon_rmap(new_page, vma, haddr); in hugetlb_cow()
[all …]
Dmemory-failure.c1611 static struct page *new_page(struct page *p, unsigned long private) in new_page() function
1712 ret = migrate_pages(&pagelist, new_page, NULL, MPOL_MF_MOVE_ALL, in soft_offline_huge_page()
1800 ret = migrate_pages(&pagelist, new_page, NULL, MPOL_MF_MOVE_ALL, in __soft_offline_page()
/Linux-v4.19/include/linux/
Dmigrate.h39 struct page *new_page = NULL; in new_page_nodemask() local
53 new_page = __alloc_pages_nodemask(gfp_mask, order, in new_page_nodemask()
56 if (new_page && PageTransHuge(new_page)) in new_page_nodemask()
57 prep_transhuge_page(new_page); in new_page_nodemask()
59 return new_page; in new_page_nodemask()
/Linux-v4.19/arch/s390/mm/
Dvmem.c266 void *new_page; in vmemmap_populate() local
268 new_page = vmemmap_alloc_block(PMD_SIZE, node); in vmemmap_populate()
269 if (!new_page) in vmemmap_populate()
271 pmd_val(*pm_dir) = __pa(new_page) | sgt_prot; in vmemmap_populate()
286 void *new_page; in vmemmap_populate() local
288 new_page = vmemmap_alloc_block(PAGE_SIZE, node); in vmemmap_populate()
289 if (!new_page) in vmemmap_populate()
291 pte_val(*pt_dir) = __pa(new_page) | pgt_prot; in vmemmap_populate()
/Linux-v4.19/fs/f2fs/
Dnamei.c821 struct page *old_page, *new_page = NULL; in f2fs_rename() local
880 &new_page); in f2fs_rename()
882 if (IS_ERR(new_page)) in f2fs_rename()
883 err = PTR_ERR(new_page); in f2fs_rename()
895 f2fs_set_link(new_dir, new_entry, new_page, old_inode); in f2fs_rename()
990 if (new_page) in f2fs_rename()
991 f2fs_put_page(new_page, 0); in f2fs_rename()
1011 struct page *old_page, *new_page; in f2fs_cross_rename() local
1043 new_entry = f2fs_find_entry(new_dir, &new_dentry->d_name, &new_page); in f2fs_cross_rename()
1045 if (IS_ERR(new_page)) in f2fs_cross_rename()
[all …]
/Linux-v4.19/kernel/events/
Duprobes.c155 struct page *old_page, struct page *new_page) in __replace_page() argument
171 err = mem_cgroup_try_charge(new_page, vma->vm_mm, GFP_KERNEL, &memcg, in __replace_page()
182 mem_cgroup_cancel_charge(new_page, memcg, false); in __replace_page()
187 get_page(new_page); in __replace_page()
188 page_add_new_anon_rmap(new_page, vma, addr, false); in __replace_page()
189 mem_cgroup_commit_charge(new_page, memcg, false, false); in __replace_page()
190 lru_cache_add_active_or_unevictable(new_page, vma); in __replace_page()
200 mk_pte(new_page, vma->vm_page_prot)); in __replace_page()
305 struct page *old_page, *new_page; in uprobe_write_opcode() local
325 new_page = alloc_page_vma(GFP_HIGHUSER_MOVABLE, vma, vaddr); in uprobe_write_opcode()
[all …]
/Linux-v4.19/fs/minix/
Dnamei.c211 struct page * new_page; in minix_rename() local
219 new_de = minix_find_entry(new_dentry, &new_page); in minix_rename()
222 minix_set_link(new_de, new_page, old_inode); in minix_rename()
/Linux-v4.19/fs/sysv/
Dnamei.c231 struct page * new_page; in sysv_rename() local
239 new_de = sysv_find_entry(new_dentry, &new_page); in sysv_rename()
242 sysv_set_link(new_de, new_page, old_inode); in sysv_rename()
/Linux-v4.19/fs/exofs/
Dnamei.c256 struct page *new_page; in exofs_rename() local
264 new_de = exofs_find_entry(new_dir, new_dentry, &new_page); in exofs_rename()
267 err = exofs_set_link(new_dir, new_de, new_page, old_inode); in exofs_rename()
/Linux-v4.19/fs/ufs/
Dnamei.c270 struct page *new_page; in ufs_rename() local
278 new_de = ufs_find_entry(new_dir, &new_dentry->d_name, &new_page); in ufs_rename()
281 ufs_set_link(new_dir, new_de, new_page, old_inode, 1); in ufs_rename()
/Linux-v4.19/fs/ubifs/
Dbudget.c379 znodes = req->new_ino + (req->new_page << UBIFS_BLOCKS_PER_PAGE_SHIFT) + in calc_idx_growth()
396 if (req->new_page) in calc_data_growth()
442 ubifs_assert(c, req->new_page <= 1); in ubifs_budget_space()
529 ubifs_assert(c, req->new_page <= 1); in ubifs_release_budget()
Dfile.c210 struct ubifs_budget_req req = { .recalculate = 1, .new_page = 1 }; in release_new_page_budget()
236 struct ubifs_budget_req req = { .new_page = 1 }; in write_begin_slow()
377 req.new_page = 1; in allocate_budget()
1529 struct ubifs_budget_req req = { .new_page = 1 }; in ubifs_vm_page_mkwrite()
/Linux-v4.19/fs/ext2/
Dnamei.c349 struct page *new_page; in ext2_rename() local
357 new_de = ext2_find_entry (new_dir, &new_dentry->d_name, &new_page); in ext2_rename()
360 ext2_set_link(new_dir, new_de, new_page, old_inode, 1); in ext2_rename()
/Linux-v4.19/arch/powerpc/mm/
Dmmu_context_iommu.c87 struct page *new_page; in new_iommu_non_cma_page() local
98 new_page = alloc_page(gfp_mask | __GFP_NORETRY | __GFP_NOWARN); in new_iommu_non_cma_page()
99 return new_page; in new_iommu_non_cma_page()
/Linux-v4.19/drivers/net/wireless/intel/iwlwifi/fw/
Ddbg.c530 struct page *new_page; in alloc_sgtable() local
541 new_page = alloc_page(GFP_KERNEL); in alloc_sgtable()
542 if (!new_page) { in alloc_sgtable()
546 new_page = sg_page(iter); in alloc_sgtable()
547 if (new_page) in alloc_sgtable()
548 __free_page(new_page); in alloc_sgtable()
554 sg_set_page(iter, new_page, alloc_size, 0); in alloc_sgtable()
/Linux-v4.19/fs/nilfs2/
Dnamei.c374 struct page *new_page; in nilfs_rename() local
382 new_de = nilfs_find_entry(new_dir, &new_dentry->d_name, &new_page); in nilfs_rename()
385 nilfs_set_link(new_dir, new_de, new_page, old_inode); in nilfs_rename()
/Linux-v4.19/fs/jbd2/
Djournal.c364 struct page *new_page; in jbd2_journal_write_metadata_buffer() local
393 new_page = virt_to_page(jh_in->b_frozen_data); in jbd2_journal_write_metadata_buffer()
396 new_page = jh2bh(jh_in)->b_page; in jbd2_journal_write_metadata_buffer()
400 mapped_data = kmap_atomic(new_page); in jbd2_journal_write_metadata_buffer()
440 mapped_data = kmap_atomic(new_page); in jbd2_journal_write_metadata_buffer()
444 new_page = virt_to_page(tmp); in jbd2_journal_write_metadata_buffer()
461 mapped_data = kmap_atomic(new_page); in jbd2_journal_write_metadata_buffer()
466 set_bh_page(new_bh, new_page, new_offset); in jbd2_journal_write_metadata_buffer()
/Linux-v4.19/drivers/tty/serial/
Dicom.c346 unsigned char *new_page = NULL; in load_code() local
419 new_page = pci_alloc_consistent(dev, 4096, &temp_pci); in load_code()
421 if (!new_page) { in load_code()
441 new_page[index] = fw->data[index]; in load_code()
498 if (new_page != NULL) in load_code()
499 pci_free_consistent(dev, 4096, new_page, temp_pci); in load_code()
/Linux-v4.19/drivers/staging/rts5208/
Dxd.c1160 u32 old_page, new_page; in xd_copy_page() local
1176 new_page = (new_blk << xd_card->block_shift) + start_page; in xd_copy_page()
1246 xd_assign_phy_addr(chip, new_page, XD_RW_ADDR); in xd_copy_page()
1269 new_page++; in xd_copy_page()

12