/Linux-v4.19/fs/nilfs2/ |
D | btnode.c | 187 xa_lock_irq(&btnc->i_pages); in nilfs_btnode_prepare_change_key() 188 err = radix_tree_insert(&btnc->i_pages, newkey, obh->b_page); in nilfs_btnode_prepare_change_key() 189 xa_unlock_irq(&btnc->i_pages); in nilfs_btnode_prepare_change_key() 245 xa_lock_irq(&btnc->i_pages); in nilfs_btnode_commit_change_key() 246 radix_tree_delete(&btnc->i_pages, oldkey); in nilfs_btnode_commit_change_key() 247 radix_tree_tag_set(&btnc->i_pages, newkey, in nilfs_btnode_commit_change_key() 249 xa_unlock_irq(&btnc->i_pages); in nilfs_btnode_commit_change_key() 277 xa_lock_irq(&btnc->i_pages); in nilfs_btnode_abort_change_key() 278 radix_tree_delete(&btnc->i_pages, newkey); in nilfs_btnode_abort_change_key() 279 xa_unlock_irq(&btnc->i_pages); in nilfs_btnode_abort_change_key()
|
D | page.c | 325 xa_lock_irq(&smap->i_pages); in nilfs_copy_back_pages() 326 page2 = radix_tree_delete(&smap->i_pages, offset); in nilfs_copy_back_pages() 330 xa_unlock_irq(&smap->i_pages); in nilfs_copy_back_pages() 332 xa_lock_irq(&dmap->i_pages); in nilfs_copy_back_pages() 333 err = radix_tree_insert(&dmap->i_pages, offset, page); in nilfs_copy_back_pages() 342 radix_tree_tag_set(&dmap->i_pages, in nilfs_copy_back_pages() 346 xa_unlock_irq(&dmap->i_pages); in nilfs_copy_back_pages() 468 xa_lock_irq(&mapping->i_pages); in __nilfs_clear_page_dirty() 470 radix_tree_tag_clear(&mapping->i_pages, in __nilfs_clear_page_dirty() 473 xa_unlock_irq(&mapping->i_pages); in __nilfs_clear_page_dirty() [all …]
|
/Linux-v4.19/mm/ |
D | memfd.c | 42 radix_tree_for_each_slot(slot, &mapping->i_pages, &iter, start) { in memfd_tag_pins() 50 xa_lock_irq(&mapping->i_pages); in memfd_tag_pins() 51 radix_tree_tag_set(&mapping->i_pages, iter.index, in memfd_tag_pins() 53 xa_unlock_irq(&mapping->i_pages); in memfd_tag_pins() 85 if (!radix_tree_tagged(&mapping->i_pages, MEMFD_TAG_PINNED)) in memfd_wait_for_pins() 95 radix_tree_for_each_tagged(slot, &mapping->i_pages, &iter, in memfd_wait_for_pins() 121 xa_lock_irq(&mapping->i_pages); in memfd_wait_for_pins() 122 radix_tree_tag_clear(&mapping->i_pages, in memfd_wait_for_pins() 124 xa_unlock_irq(&mapping->i_pages); in memfd_wait_for_pins()
|
D | truncate.c | 39 if (!__radix_tree_lookup(&mapping->i_pages, index, &node, &slot)) in __clear_shadow_entry() 43 __radix_tree_replace(&mapping->i_pages, node, slot, NULL, in __clear_shadow_entry() 51 xa_lock_irq(&mapping->i_pages); in clear_shadow_entry() 53 xa_unlock_irq(&mapping->i_pages); in clear_shadow_entry() 82 xa_lock_irq(&mapping->i_pages); in truncate_exceptional_pvec_entries() 105 xa_unlock_irq(&mapping->i_pages); in truncate_exceptional_pvec_entries() 521 xa_lock_irq(&mapping->i_pages); in truncate_inode_pages_final() 522 xa_unlock_irq(&mapping->i_pages); in truncate_inode_pages_final() 630 xa_lock_irqsave(&mapping->i_pages, flags); in invalidate_complete_page2() 636 xa_unlock_irqrestore(&mapping->i_pages, flags); in invalidate_complete_page2() [all …]
|
D | khugepaged.c | 1348 xa_lock_irq(&mapping->i_pages); in collapse_shmem() 1349 radix_tree_for_each_slot(slot, &mapping->i_pages, &iter, start) { in collapse_shmem() 1362 radix_tree_insert(&mapping->i_pages, index, in collapse_shmem() 1371 &mapping->i_pages.xa_lock); in collapse_shmem() 1373 xa_unlock_irq(&mapping->i_pages); in collapse_shmem() 1380 xa_lock_irq(&mapping->i_pages); in collapse_shmem() 1400 xa_unlock_irq(&mapping->i_pages); in collapse_shmem() 1410 xa_lock_irq(&mapping->i_pages); in collapse_shmem() 1412 slot = radix_tree_lookup_slot(&mapping->i_pages, index); in collapse_shmem() 1414 &mapping->i_pages.xa_lock), page); in collapse_shmem() [all …]
|
D | filemap.c | 121 error = __radix_tree_create(&mapping->i_pages, page->index, 0, in page_cache_tree_insert() 129 &mapping->i_pages.xa_lock); in page_cache_tree_insert() 137 __radix_tree_replace(&mapping->i_pages, node, slot, page, in page_cache_tree_insert() 159 __radix_tree_lookup(&mapping->i_pages, page->index + i, in page_cache_tree_delete() 164 radix_tree_clear_tags(&mapping->i_pages, node, slot); in page_cache_tree_delete() 165 __radix_tree_replace(&mapping->i_pages, node, slot, shadow, in page_cache_tree_delete() 300 xa_lock_irqsave(&mapping->i_pages, flags); in delete_from_page_cache() 302 xa_unlock_irqrestore(&mapping->i_pages, flags); in delete_from_page_cache() 334 radix_tree_for_each_slot(slot, &mapping->i_pages, &iter, start) { in page_cache_tree_delete_batch() 338 &mapping->i_pages.xa_lock); in page_cache_tree_delete_batch() [all …]
|
D | workingset.c | 433 mapping = container_of(node->root, struct address_space, i_pages); in shadow_lru_isolate() 436 if (!xa_trylock(&mapping->i_pages)) { in shadow_lru_isolate() 471 __radix_tree_delete_node(&mapping->i_pages, node, in shadow_lru_isolate() 475 xa_unlock_irq(&mapping->i_pages); in shadow_lru_isolate()
|
D | swap_state.c | 127 xa_lock_irq(&address_space->i_pages); in __add_to_swap_cache() 130 error = radix_tree_insert(&address_space->i_pages, in __add_to_swap_cache() 148 radix_tree_delete(&address_space->i_pages, idx + i); in __add_to_swap_cache() 154 xa_unlock_irq(&address_space->i_pages); in __add_to_swap_cache() 191 radix_tree_delete(&address_space->i_pages, idx + i); in __delete_from_swap_cache() 272 xa_lock_irq(&address_space->i_pages); in delete_from_swap_cache() 274 xa_unlock_irq(&address_space->i_pages); in delete_from_swap_cache() 628 INIT_RADIX_TREE(&space->i_pages, GFP_ATOMIC|__GFP_NOWARN); in init_swap_address_space()
|
D | page-writeback.c | 2112 xa_lock_irq(&mapping->i_pages); in tag_pages_for_writeback() 2113 radix_tree_for_each_tagged(slot, &mapping->i_pages, &iter, start, in tag_pages_for_writeback() 2117 radix_tree_iter_tag_set(&mapping->i_pages, &iter, in tag_pages_for_writeback() 2123 xa_unlock_irq(&mapping->i_pages); in tag_pages_for_writeback() 2125 xa_lock_irq(&mapping->i_pages); in tag_pages_for_writeback() 2127 xa_unlock_irq(&mapping->i_pages); in tag_pages_for_writeback() 2470 xa_lock_irqsave(&mapping->i_pages, flags); in __set_page_dirty_nobuffers() 2474 radix_tree_tag_set(&mapping->i_pages, page_index(page), in __set_page_dirty_nobuffers() 2476 xa_unlock_irqrestore(&mapping->i_pages, flags); in __set_page_dirty_nobuffers() 2721 xa_lock_irqsave(&mapping->i_pages, flags); in test_clear_page_writeback() [all …]
|
D | migrate.c | 473 xa_lock_irq(&mapping->i_pages); in migrate_page_move_mapping() 475 pslot = radix_tree_lookup_slot(&mapping->i_pages, in migrate_page_move_mapping() 481 &mapping->i_pages.xa_lock) != page) { in migrate_page_move_mapping() 482 xa_unlock_irq(&mapping->i_pages); in migrate_page_move_mapping() 487 xa_unlock_irq(&mapping->i_pages); in migrate_page_move_mapping() 501 xa_unlock_irq(&mapping->i_pages); in migrate_page_move_mapping() 529 radix_tree_replace_slot(&mapping->i_pages, pslot, newpage); in migrate_page_move_mapping() 535 pslot = radix_tree_lookup_slot(&mapping->i_pages, in migrate_page_move_mapping() 537 radix_tree_replace_slot(&mapping->i_pages, pslot, in migrate_page_move_mapping() 549 xa_unlock(&mapping->i_pages); in migrate_page_move_mapping() [all …]
|
D | shmem.c | 336 item = __radix_tree_lookup(&mapping->i_pages, index, &node, &pslot); in shmem_radix_tree_replace() 341 __radix_tree_replace(&mapping->i_pages, node, pslot, in shmem_radix_tree_replace() 359 item = radix_tree_lookup(&mapping->i_pages, index); in shmem_confirm_swap() 603 xa_lock_irq(&mapping->i_pages); in shmem_add_to_page_cache() 610 if (radix_tree_gang_lookup_slot(&mapping->i_pages, in shmem_add_to_page_cache() 618 error = radix_tree_insert(&mapping->i_pages, in shmem_add_to_page_cache() 625 error = radix_tree_insert(&mapping->i_pages, index, page); in shmem_add_to_page_cache() 637 xa_unlock_irq(&mapping->i_pages); in shmem_add_to_page_cache() 640 xa_unlock_irq(&mapping->i_pages); in shmem_add_to_page_cache() 656 xa_lock_irq(&mapping->i_pages); in shmem_delete_from_page_cache() [all …]
|
D | readahead.c | 180 page = radix_tree_lookup(&mapping->i_pages, page_offset); in __do_page_cache_readahead()
|
D | huge_memory.c | 2452 xa_unlock(&head->mapping->i_pages); in __split_huge_page() 2660 xa_lock(&mapping->i_pages); in split_huge_page_to_list() 2661 pslot = radix_tree_lookup_slot(&mapping->i_pages, in split_huge_page_to_list() 2668 &mapping->i_pages.xa_lock) != head) in split_huge_page_to_list() 2702 xa_unlock(&mapping->i_pages); in split_huge_page_to_list()
|
D | vmscan.c | 884 xa_lock_irqsave(&mapping->i_pages, flags); in __remove_mapping() 926 xa_unlock_irqrestore(&mapping->i_pages, flags); in __remove_mapping() 953 xa_unlock_irqrestore(&mapping->i_pages, flags); in __remove_mapping() 962 xa_unlock_irqrestore(&mapping->i_pages, flags); in __remove_mapping()
|
/Linux-v4.19/fs/ |
D | dax.c | 190 radix_tree_deref_slot_protected(slot, &mapping->i_pages.xa_lock); in slot_locked() 200 radix_tree_deref_slot_protected(slot, &mapping->i_pages.xa_lock); in lock_slot() 203 radix_tree_replace_slot(&mapping->i_pages, slot, (void *)entry); in lock_slot() 213 radix_tree_deref_slot_protected(slot, &mapping->i_pages.xa_lock); in unlock_slot() 216 radix_tree_replace_slot(&mapping->i_pages, slot, (void *)entry); in unlock_slot() 242 entry = __radix_tree_lookup(&mapping->i_pages, index, NULL, in __get_unlocked_mapping_entry() 255 xa_unlock_irq(&mapping->i_pages); in __get_unlocked_mapping_entry() 258 xa_lock_irq(&mapping->i_pages); in __get_unlocked_mapping_entry() 284 xa_lock_irq(&mapping->i_pages); in unlock_mapping_entry() 285 entry = __radix_tree_lookup(&mapping->i_pages, index, NULL, &slot); in unlock_mapping_entry() [all …]
|
D | fs-writeback.c | 364 xa_lock_irq(&mapping->i_pages); in inode_switch_wbs_work_fn() 378 radix_tree_for_each_tagged(slot, &mapping->i_pages, &iter, 0, in inode_switch_wbs_work_fn() 381 &mapping->i_pages.xa_lock); in inode_switch_wbs_work_fn() 388 radix_tree_for_each_tagged(slot, &mapping->i_pages, &iter, 0, in inode_switch_wbs_work_fn() 391 &mapping->i_pages.xa_lock); in inode_switch_wbs_work_fn() 433 xa_unlock_irq(&mapping->i_pages); in inode_switch_wbs_work_fn()
|
D | inode.c | 352 INIT_RADIX_TREE(&mapping->i_pages, GFP_ATOMIC | __GFP_ACCOUNT); in __address_space_init_once() 511 xa_lock_irq(&inode->i_data.i_pages); in clear_inode() 514 xa_unlock_irq(&inode->i_data.i_pages); in clear_inode()
|
D | buffer.c | 578 xa_lock_irqsave(&mapping->i_pages, flags); in __set_page_dirty() 582 radix_tree_tag_set(&mapping->i_pages, in __set_page_dirty() 585 xa_unlock_irqrestore(&mapping->i_pages, flags); in __set_page_dirty()
|
/Linux-v4.19/arch/nios2/include/asm/ |
D | cacheflush.h | 49 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages) 50 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages)
|
/Linux-v4.19/arch/nds32/include/asm/ |
D | cacheflush.h | 39 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&(mapping)->i_pages) 40 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&(mapping)->i_pages)
|
/Linux-v4.19/arch/parisc/include/asm/ |
D | cacheflush.h | 58 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages) 59 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages)
|
/Linux-v4.19/include/linux/ |
D | backing-dev.h | 341 !lockdep_is_held(&inode->i_mapping->i_pages.xa_lock) && in inode_to_wb() 374 xa_lock_irqsave(&inode->i_mapping->i_pages, cookie->flags); in unlocked_inode_to_wb_begin() 392 xa_unlock_irqrestore(&inode->i_mapping->i_pages, cookie->flags); in unlocked_inode_to_wb_end()
|
/Linux-v4.19/arch/arm/include/asm/ |
D | cacheflush.h | 321 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages) 322 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages)
|
/Linux-v4.19/Documentation/vm/ |
D | page_migration.rst | 106 5. The i_pages lock is taken. This will cause all processes trying 124 11. The i_pages lock is dropped. With that lookups in the mapping
|
/Linux-v4.19/fs/f2fs/ |
D | data.c | 2705 xa_lock_irqsave(&mapping->i_pages, flags); in f2fs_clear_radix_tree_dirty_tag() 2706 radix_tree_tag_clear(&mapping->i_pages, page_index(page), in f2fs_clear_radix_tree_dirty_tag() 2708 xa_unlock_irqrestore(&mapping->i_pages, flags); in f2fs_clear_radix_tree_dirty_tag()
|