Searched refs:xa_lock (Results 1 – 23 of 23) sorted by relevance
293 spinlock_t xa_lock; member300 .xa_lock = __SPIN_LOCK_UNLOCKED(name.xa_lock), \376 spin_lock_init(&xa->xa_lock); in xa_init_flags()527 #define xa_trylock(xa) spin_trylock(&(xa)->xa_lock)528 #define xa_lock(xa) spin_lock(&(xa)->xa_lock) macro529 #define xa_unlock(xa) spin_unlock(&(xa)->xa_lock)530 #define xa_lock_bh(xa) spin_lock_bh(&(xa)->xa_lock)531 #define xa_unlock_bh(xa) spin_unlock_bh(&(xa)->xa_lock)532 #define xa_lock_irq(xa) spin_lock_irq(&(xa)->xa_lock)533 #define xa_unlock_irq(xa) spin_unlock_irq(&(xa)->xa_lock)[all …]
101 #define idr_lock(idr) xa_lock(&(idr)->idr_rt)
285 !lockdep_is_held(&inode->i_mapping->i_pages.xa_lock) && in inode_to_wb()
128 xa_lock(&mt_tree); in regression1_fn()133 xa_lock(&mt_tree); in regression1_fn()137 xa_lock(&mt_tree); in regression1_fn()145 xa_lock(&mt_tree); in regression1_fn()
125 xa_lock(&ucontext->mmap_xa); in rdma_user_mmap_entry_get_pgoff()190 xa_lock(&ucontext->mmap_xa); in rdma_user_mmap_entry_free()235 xa_lock(&entry->ucontext->mmap_xa); in rdma_user_mmap_entry_remove()287 xa_lock(&ucontext->mmap_xa); in rdma_user_mmap_entry_insert_range()
118 xa_lock(&rt->xa); in rdma_restrack_count()275 xa_lock(&rt->xa); in rdma_restrack_get_byid()
231 xa_lock(&rt->xa); in rdma_get_counter_auto_mode()360 xa_lock(&rt->xa); in get_running_counters_hwstat_sum()375 xa_lock(&rt->xa); in get_running_counters_hwstat_sum()
144 xa_lock(&ctx_table); in ucma_get_ctx()597 xa_lock(&ctx_table); in ucma_destroy_id()1556 xa_lock(&multicast_table); in ucma_leave_multicast()1629 xa_lock(&ctx_table); in ucma_migrate_id()
759 xa_lock(&rt->xa); in fill_stat_counter_qps()1390 xa_lock(&rt->xa); in res_get_common_dumpit()1424 again: xa_lock(&rt->xa); in res_get_common_dumpit()
203 Takes xa_lock internally:227 Assumes xa_lock held on entry:237 that you are storing in the XArray, you can call xa_lock()272 xa_lock(&foo->array);283 coverage of the xa_lock on the store side to protect some statistics288 context, or xa_lock_irq() in process context and xa_lock()297 the xa_lock; the xa_lock is used for lockdep validation and will be used311 to use the xa_lock while modifying the array. You can choose whether312 to use the xa_lock or the RCU lock while doing read-only operations on336 the xa_lock, attempt the operation and drop the lock. The operation[all …]
77 xa_lock(&mte_pages); in mte_invalidate_tags_area()
365 xa_lock(&sriov->xa_rej_tmout); in alloc_rej_tmout()410 xa_lock(&sriov->xa_rej_tmout); in lookup_rej_tmout_slave()500 xa_lock(&sriov->xa_rej_tmout); in rej_tmout_xa_cleanup()
318 xa_lock(&mte_pages); in swsusp_mte_free_storage()369 xa_lock(&mte_pages); in swsusp_mte_restore_tags()
94 xa_lock(&sbi->managed_pslots); in erofs_insert_workgroup()
325 __must_hold(xas->xa->xa_lock) in __xas_nomem()1509 xa_lock(xa); in xa_erase()1573 xa_lock(xa); in xa_store()1964 xa_lock(xa); in xa_set_mark()1982 xa_lock(xa); in xa_clear_mark()
1643 xa_lock(xa); in shadow_remove()
17 xa_lock(&srq_table->xa); in hns_roce_srq_event()
88 xa_lock(&hr_dev->qp_table_xa); in hns_roce_qp_event()
522 xa_lock(&dev->qps); in qedr_iw_load_qp()
267 xa_lock(&imr->implicit_children); in destroy_unused_implicit_child_mr()620 xa_lock(&imr->implicit_children); in mlx5_ib_free_implicit_mr()
668 xa_lock(&table->array); in srq_event_notifier()
549 xa_lock(&dev->sig_mrs); in mlx5_poll_one()
2434 xa_lock(&swap_cache->i_pages); in __split_huge_page()2696 xa_lock(&mapping->i_pages); in split_huge_page_to_list()