/Linux-v5.10/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_init_ops.h | 622 BNX2X_ILT_FREE(line->page, line->page_mapping, line->size); in bnx2x_ilt_line_mem_op() 625 BNX2X_ILT_ZALLOC(line->page, &line->page_mapping, size); in bnx2x_ilt_line_mem_op() 677 dma_addr_t page_mapping) in bnx2x_ilt_line_wr() argument 686 bnx2x_wr_64(bp, reg, ILT_ADDR1(page_mapping), ILT_ADDR2(page_mapping)); in bnx2x_ilt_line_wr() 700 bnx2x_ilt_line_wr(bp, abs_idx, ilt->lines[idx].page_mapping); in bnx2x_ilt_line_init_op()
|
D | bnx2x_init.h | 503 dma_addr_t page_mapping; member
|
/Linux-v5.10/mm/ |
D | memory-failure.c | 123 mapping = page_mapping(p); in hwpoison_filter_dev() 695 mapping = page_mapping(p); in me_pagecache_clean() 718 struct address_space *mapping = page_mapping(p); in me_pagecache_dirty() 819 mapping = page_mapping(hpage); in me_huge_page() 1033 mapping = page_mapping(hpage); in hwpoison_user_mappings() 1626 if (page_mapping(page)) { in unpoison_memory()
|
D | page-writeback.c | 2477 struct address_space *mapping = page_mapping(page); in __set_page_dirty_nobuffers() 2486 BUG_ON(page_mapping(page) != mapping); in __set_page_dirty_nobuffers() 2559 struct address_space *mapping = page_mapping(page); in set_page_dirty() 2626 struct address_space *mapping = page_mapping(page); in __cancel_dirty_page() 2663 struct address_space *mapping = page_mapping(page); in clear_page_dirty_for_io() 2724 struct address_space *mapping = page_mapping(page); in test_clear_page_writeback() 2768 struct address_space *mapping = page_mapping(page); in __test_set_page_writeback() 2830 trace_wait_on_page_writeback(page, page_mapping(page)); in wait_on_page_writeback()
|
D | util.c | 677 struct address_space *page_mapping(struct page *page) in page_mapping() function 700 EXPORT_SYMBOL(page_mapping); 709 return page_mapping(page); in page_mapping_file()
|
D | debug.c | 88 mapping = page_mapping(page); in __dump_page()
|
D | vmscan.c | 759 if (page_mapping(page) == mapping) in handle_write_error() 862 BUG_ON(mapping != page_mapping(page)); in __remove_mapping() 1064 mapping = page_mapping(page); in page_check_dirty_writeback() 1137 mapping = page_mapping(page); in shrink_page_list() 1275 mapping = page_mapping(page); in shrink_page_list() 1374 mapping = page_mapping(page); in shrink_page_list() 1587 mapping = page_mapping(page); in __isolate_lru_page()
|
D | migrate.c | 126 mapping = page_mapping(page); in isolate_movable_page() 156 mapping = page_mapping(page); in putback_movable_page() 952 mapping = page_mapping(page); in move_to_new_page() 1319 if (page_private(hpage) && !page_mapping(hpage)) { in unmap_and_move_huge_page() 2561 if (page_mapping(page)) in migrate_vma_check_page() 3019 mapping = page_mapping(page); in migrate_vma_pages()
|
D | rmap.c | 994 mapping = page_mapping(page); in page_mkclean() 1219 if (PageTransCompound(page) && page_mapping(page)) { in page_add_file_rmap() 1902 struct address_space *mapping = page_mapping(page); in rmap_walk_file()
|
D | internal.h | 84 ret = !mapping_unevictable(page_mapping(page)) && !PageMlocked(page); in page_evictable()
|
D | compaction.c | 122 mapping = page_mapping(page); in PageMovable() 963 if (!page_mapping(page) && in isolate_migratepages_block() 971 if (!(cc->gfp_mask & __GFP_FS) && page_mapping(page)) in isolate_migratepages_block()
|
D | zsmalloc.c | 1866 __SetPageMovable(newpage, page_mapping(oldpage)); in replace_sub_page() 1893 mapping = page_mapping(page); in zs_page_isolate() 2068 mapping = page_mapping(page); in zs_page_putback()
|
D | truncate.c | 257 struct address_space *mapping = page_mapping(page); in invalidate_inode_page()
|
D | filemap.c | 269 struct address_space *mapping = page_mapping(page); in delete_from_page_cache() 1522 mapping = page_mapping(page); in page_endio()
|
D | shmem.c | 946 page_mapping(page) == mapping) { in shmem_undo_range() 1024 if (page_mapping(page) != mapping) { in shmem_undo_range() 1632 swap_mapping = page_mapping(oldpage); in shmem_replace_page()
|
D | z3fold.c | 1643 new_mapping = page_mapping(page); in z3fold_page_migrate()
|
/Linux-v5.10/fs/crypto/ |
D | inline_crypt.c | 270 mapping = page_mapping(page); in bh_get_inode_and_lblk_num()
|
/Linux-v5.10/arch/s390/kernel/ |
D | uv.c | 176 } else if (page_mapping(page)) { in expected_page_refs()
|
/Linux-v5.10/arch/nds32/mm/ |
D | cacheflush.c | 241 mapping = page_mapping(page); in flush_dcache_page()
|
/Linux-v5.10/include/trace/events/ |
D | writeback.h | 252 struct address_space *mapping = page_mapping(page);
|
/Linux-v5.10/include/linux/ |
D | mm.h | 1549 extern struct address_space *page_mapping(struct page *page); 1576 struct address_space *page_mapping(struct page *page);
|
/Linux-v5.10/fs/ |
D | buffer.c | 644 struct address_space *mapping = page_mapping(page); in __set_page_dirty_buffers() 1135 mapping = page_mapping(page); in mark_buffer_dirty()
|
D | splice.c | 54 mapping = page_mapping(page); in page_cache_pipe_buf_try_steal()
|
/Linux-v5.10/Documentation/vm/ |
D | page_migration.rst | 227 use page_mapping() which masks off the low two bits of page->mapping under
|
D | unevictable-lru.rst | 313 populate_vma_page_range() checks page_mapping() after acquiring the page lock. 426 zeros out the page_mapping of the old page before unlocking it, so m[un]lock
|