Searched refs:xa_lock (Results 1 – 19 of 19) sorted by relevance
293 spinlock_t xa_lock; member300 .xa_lock = __SPIN_LOCK_UNLOCKED(name.xa_lock), \376 spin_lock_init(&xa->xa_lock); in xa_init_flags()500 #define xa_trylock(xa) spin_trylock(&(xa)->xa_lock)501 #define xa_lock(xa) spin_lock(&(xa)->xa_lock) macro502 #define xa_unlock(xa) spin_unlock(&(xa)->xa_lock)503 #define xa_lock_bh(xa) spin_lock_bh(&(xa)->xa_lock)504 #define xa_unlock_bh(xa) spin_unlock_bh(&(xa)->xa_lock)505 #define xa_lock_irq(xa) spin_lock_irq(&(xa)->xa_lock)506 #define xa_unlock_irq(xa) spin_unlock_irq(&(xa)->xa_lock)[all …]
101 #define idr_lock(idr) xa_lock(&(idr)->idr_rt)
347 !lockdep_is_held(&inode->i_mapping->i_pages.xa_lock) && in inode_to_wb()
128 xa_lock(&mt_tree); in regression1_fn()133 xa_lock(&mt_tree); in regression1_fn()137 xa_lock(&mt_tree); in regression1_fn()145 xa_lock(&mt_tree); in regression1_fn()
188 Takes xa_lock internally:212 Assumes xa_lock held on entry:222 that you are storing in the XArray, you can call xa_lock()257 xa_lock(&foo->array);268 coverage of the xa_lock on the store side to protect some statistics273 context, or xa_lock_irq() in process context and xa_lock()282 the xa_lock; the xa_lock is used for lockdep validation and will be used296 to use the xa_lock while modifying the array. You can choose whether297 to use the xa_lock or the RCU lock while doing read-only operations on321 the xa_lock, attempt the operation and drop the lock. The operation[all …]
104 xa_lock(&sbi->workstn_tree); in erofs_register_workgroup()200 xa_lock(&sbi->workstn_tree); in erofs_shrink_workstation()
51 xa_lock(&rnicp->qps); in post_qp_event()117 xa_lock(&rnicp->qps); in iwch_ev_dispatch()
118 xa_lock(&rt->xa); in rdma_restrack_count()296 xa_lock(&rt->xa); in rdma_restrack_get_byid()
230 xa_lock(&rt->xa); in rdma_get_counter_auto_mode()364 xa_lock(&rt->xa); in get_running_counters_hwstat_sum()379 xa_lock(&rt->xa); in get_running_counters_hwstat_sum()
149 xa_lock(&ctx_table); in ucma_get_ctx()317 xa_lock(&ctx_table); in ucma_removal_event_handler()610 xa_lock(&ctx_table); in ucma_destroy_id()626 xa_lock(&ctx_table); in ucma_destroy_id()1499 xa_lock(&multicast_table); in ucma_leave_multicast()1606 xa_lock(&ctx_table); in ucma_migrate_id()1754 xa_lock(&ctx_table); in ucma_close()
699 xa_lock(&rt->xa); in fill_stat_counter_qps()1330 xa_lock(&rt->xa); in res_get_common_dumpit()1366 again: xa_lock(&rt->xa); in res_get_common_dumpit()
322 __must_hold(xas->xa->xa_lock) in __xas_nomem()1350 xa_lock(xa); in xa_erase()1414 xa_lock(xa); in xa_store()1765 xa_lock(xa); in xa_set_mark()1783 xa_lock(xa); in xa_clear_mark()
1471 xa_lock(xa); in shadow_remove()
86 xa_lock(&table->array); in mlx5_cmd_get_srq()675 xa_lock(&table->array); in srq_event_notifier()
525 xa_lock(&dev->mdev->priv.mkey_table); in mlx5_poll_one()
17 xa_lock(&srq_table->xa); in hns_roce_srq_event()
51 xa_lock(&hr_dev->qp_table_xa); in hns_roce_qp_event()
2517 xa_lock(&swap_cache->i_pages); in __split_huge_page()2778 xa_lock(&mapping->i_pages); in split_huge_page_to_list()
246 xa_lock(&ucontext->mmap_xa); in mmap_entry_insert()