/Linux-v5.4/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_init_ops.h | 622 BNX2X_ILT_FREE(line->page, line->page_mapping, line->size); in bnx2x_ilt_line_mem_op() 625 BNX2X_ILT_ZALLOC(line->page, &line->page_mapping, size); in bnx2x_ilt_line_mem_op() 677 dma_addr_t page_mapping) in bnx2x_ilt_line_wr() argument 686 bnx2x_wr_64(bp, reg, ILT_ADDR1(page_mapping), ILT_ADDR2(page_mapping)); in bnx2x_ilt_line_wr() 700 bnx2x_ilt_line_wr(bp, abs_idx, ilt->lines[idx].page_mapping); in bnx2x_ilt_line_init_op()
|
D | bnx2x_init.h | 504 dma_addr_t page_mapping; member
|
D | bnx2x_sriov.c | 1665 ilt->lines[line+i].page_mapping = hw_cxt->mapping; in bnx2x_iov_init_ilt()
|
/Linux-v5.4/mm/ |
D | memory-failure.c | 96 mapping = page_mapping(p); in hwpoison_filter_dev() 668 mapping = page_mapping(p); in me_pagecache_clean() 691 struct address_space *mapping = page_mapping(p); in me_pagecache_dirty() 792 mapping = page_mapping(hpage); in me_huge_page() 1009 mapping = page_mapping(hpage); in hwpoison_user_mappings() 1565 if (page_mapping(page)) { in unpoison_memory()
|
D | page-writeback.c | 2469 struct address_space *mapping = page_mapping(page); in __set_page_dirty_nobuffers() 2478 BUG_ON(page_mapping(page) != mapping); in __set_page_dirty_nobuffers() 2551 struct address_space *mapping = page_mapping(page); in set_page_dirty() 2618 struct address_space *mapping = page_mapping(page); in __cancel_dirty_page() 2655 struct address_space *mapping = page_mapping(page); in clear_page_dirty_for_io() 2716 struct address_space *mapping = page_mapping(page); in test_clear_page_writeback() 2766 struct address_space *mapping = page_mapping(page); in __test_set_page_writeback() 2821 trace_wait_on_page_writeback(page, page_mapping(page)); in wait_on_page_writeback()
|
D | util.c | 648 struct address_space *page_mapping(struct page *page) in page_mapping() function 671 EXPORT_SYMBOL(page_mapping); 680 return page_mapping(page); in page_mapping_file()
|
D | debug.c | 61 mapping = page_mapping(page); in __dump_page()
|
D | migrate.c | 125 mapping = page_mapping(page); in isolate_movable_page() 155 mapping = page_mapping(page); in putback_movable_page() 937 mapping = page_mapping(page); in move_to_new_page() 1324 if (page_private(hpage) && !page_mapping(hpage)) { in unmap_and_move_huge_page() 2392 if (page_mapping(page)) in migrate_vma_check_page() 2849 mapping = page_mapping(page); in migrate_vma_pages()
|
D | vmscan.c | 805 if (page_mapping(page) == mapping) in handle_write_error() 908 BUG_ON(mapping != page_mapping(page)); in __remove_mapping() 1111 mapping = page_mapping(page); in page_check_dirty_writeback() 1185 mapping = page_mapping(page); in shrink_page_list() 1323 mapping = page_mapping(page); in shrink_page_list() 1416 mapping = page_mapping(page); in shrink_page_list() 1617 mapping = page_mapping(page); in __isolate_lru_page() 4321 ret = !mapping_unevictable(page_mapping(page)) && !PageMlocked(page); in page_evictable()
|
D | rmap.c | 981 mapping = page_mapping(page); in page_mkclean() 1198 if (PageTransCompound(page) && page_mapping(page)) { in page_add_file_rmap() 1867 struct address_space *mapping = page_mapping(page); in rmap_walk_file()
|
D | compaction.c | 104 mapping = page_mapping(page); in PageMovable() 940 if (!page_mapping(page) && in isolate_migratepages_block() 948 if (!(cc->gfp_mask & __GFP_FS) && page_mapping(page)) in isolate_migratepages_block()
|
D | zsmalloc.c | 1912 __SetPageMovable(newpage, page_mapping(oldpage)); in replace_sub_page() 1939 mapping = page_mapping(page); in zs_page_isolate() 2109 mapping = page_mapping(page); in zs_page_putback()
|
D | truncate.c | 257 struct address_space *mapping = page_mapping(page); in invalidate_inode_page()
|
D | filemap.c | 268 struct address_space *mapping = page_mapping(page); in delete_from_page_cache() 1366 mapping = page_mapping(page); in page_endio() 1588 if (unlikely(page_mapping(page) != mapping)) { in find_lock_entry()
|
D | shmem.c | 862 if (page_mapping(page) == mapping) { in shmem_undo_range() 967 if (page_mapping(page) == mapping) { in shmem_undo_range() 1565 swap_mapping = page_mapping(oldpage); in shmem_replace_page()
|
D | z3fold.c | 1410 new_mapping = page_mapping(page); in z3fold_page_migrate()
|
D | khugepaged.c | 1635 if (page_mapping(page) != mapping) { in collapse_file()
|
/Linux-v5.4/arch/nds32/mm/ |
D | cacheflush.c | 242 mapping = page_mapping(page); in flush_dcache_page()
|
/Linux-v5.4/include/trace/events/ |
D | writeback.h | 255 struct address_space *mapping = page_mapping(page);
|
/Linux-v5.4/include/linux/ |
D | mm.h | 1349 extern struct address_space *page_mapping(struct page *page); 1376 struct address_space *page_mapping(struct page *page);
|
/Linux-v5.4/Documentation/vm/ |
D | page_migration.rst | 223 use page_mapping which mask off the low two bits of page->mapping under
|
D | unevictable-lru.rst | 313 populate_vma_page_range() checks page_mapping() after acquiring the page lock. 426 zeros out the page_mapping of the old page before unlocking it, so m[un]lock
|
/Linux-v5.4/fs/iomap/ |
D | buffered-io.c | 635 struct address_space *mapping = page_mapping(page); in iomap_set_page_dirty()
|
/Linux-v5.4/fs/ |
D | buffer.c | 619 struct address_space *mapping = page_mapping(page); in __set_page_dirty_buffers() 1110 mapping = page_mapping(page); in mark_buffer_dirty()
|
D | splice.c | 55 mapping = page_mapping(page); in page_cache_pipe_buf_steal()
|