Lines Matching refs:slots

810 		void __rcu **slots, unsigned char offset)  in mt_slot()  argument
812 return rcu_dereference_check(slots[offset], mt_locked(mt)); in mt_slot()
815 static inline void *mt_slot_locked(struct maple_tree *mt, void __rcu **slots, in mt_slot_locked() argument
818 return rcu_dereference_protected(slots[offset], mt_write_locked(mt)); in mt_slot_locked()
828 static inline void *mas_slot_locked(struct ma_state *mas, void __rcu **slots, in mas_slot_locked() argument
831 return mt_slot_locked(mas->tree, slots, offset); in mas_slot_locked()
842 static inline void *mas_slot(struct ma_state *mas, void __rcu **slots, in mas_slot() argument
845 return mt_slot(mas->tree, slots, offset); in mas_slot()
915 void __rcu **slots; in mt_clear_meta() local
922 slots = mn->mr64.slot; in mt_clear_meta()
923 next = mt_slot_locked(mt, slots, in mt_clear_meta()
1037 void __rcu **slots; in mas_descend() local
1042 slots = ma_slots(node, type); in mas_descend()
1047 mas->node = mas_slot(mas, slots, mas->offset); in mas_descend()
1251 void **slots = NULL; in mas_alloc_nodes() local
1285 slots = (void **)&node->slot[node->node_count]; in mas_alloc_nodes()
1287 count = mt_alloc_bulk(gfp, max_req, slots); in mas_alloc_nodes()
1306 memset(slots, 0, max_req * sizeof(unsigned long)); in mas_alloc_nodes()
1498 void __rcu **slots; in mas_leaf_max_gap() local
1504 slots = ma_slots(mn, mt); in mas_leaf_max_gap()
1509 if (slots[i]) { in mas_leaf_max_gap()
1527 if (likely(!slots[0])) { in mas_leaf_max_gap()
1540 if (unlikely(mas->max == ULONG_MAX) && !slots[max_piv + 1]) { in mas_leaf_max_gap()
1548 if (likely(slots[i])) in mas_leaf_max_gap()
1709 void __rcu **slots = ma_slots(node, type); in mas_adopt_children() local
1716 child = mas_slot_locked(mas, slots, offset); in mas_adopt_children()
1732 void __rcu **slots; in mas_put_in_tree() local
1741 slots = ma_slots(mte_parent(mas->node), in mas_put_in_tree()
1743 rcu_assign_pointer(slots[offset], mas->node); in mas_put_in_tree()
1778 void __rcu **slots; in mas_find_child() local
1782 slots = ma_slots(node, mt); in mas_find_child()
1786 entry = mas_slot_locked(mas, slots, offset); in mas_find_child()
1948 void __rcu **slots; in mas_mab_cp() local
1979 slots = ma_slots(node, mt); in mas_mab_cp()
1980 memcpy(b_node->slot + mab_start, slots + mas_start, sizeof(void *) * j); in mas_mab_cp()
2029 void __rcu **slots = ma_slots(node, mt); in mab_mas_cp() local
2038 slots[mt_pivots[mt]] = NULL; in mab_mas_cp()
2045 memcpy(slots, b_node->slot + mab_start, in mab_mas_cp()
2145 wr_mas->content = mas_slot_locked(mas, wr_mas->slots, in mas_store_b_node()
2790 void __rcu **slots; in mtree_range_walk() local
2826 slots = ma_slots(node, type); in mtree_range_walk()
2827 next = mt_slot(mas->tree, slots, offset); in mtree_range_walk()
3055 void __rcu **l_slots, **slots; in mas_destroy_rebalance() local
3078 slots = ma_slots(newnode, mt); in mas_destroy_rebalance()
3087 memcpy(slots, l_slots + split + 1, sizeof(void *) * tmp); in mas_destroy_rebalance()
3090 memcpy(slots + tmp, ma_slots(node, mt), sizeof(void *) * end); in mas_destroy_rebalance()
3107 memset(slots + tmp, 0, sizeof(void *) * (max_s - tmp)); in mas_destroy_rebalance()
3131 slots = ma_slots(new_left, mt); in mas_destroy_rebalance()
3133 memcpy(slots, l_slots, sizeof(void *) * split); in mas_destroy_rebalance()
3142 slots = ma_slots(parent, mt); in mas_destroy_rebalance()
3145 rcu_assign_pointer(slots[offset], mas->node); in mas_destroy_rebalance()
3146 rcu_assign_pointer(slots[offset - 1], l_mas.node); in mas_destroy_rebalance()
3454 memset(wr_mas->slots + bn->b_end, 0, sizeof(void *) * clear--); in mas_reuse_node()
3511 void __rcu **slots; in mas_root_expand() local
3521 slots = ma_slots(node, type); in mas_root_expand()
3527 rcu_assign_pointer(slots[slot], contents); in mas_root_expand()
3534 rcu_assign_pointer(slots[slot], entry); in mas_root_expand()
3607 wr_mas->slots = ma_slots(wr_mas->node, wr_mas->type); in mas_wr_walk_descend()
3635 wr_mas->content = mas_slot_locked(mas, wr_mas->slots, in mas_wr_walk()
3652 wr_mas->content = mas_slot_locked(mas, wr_mas->slots, in mas_wr_walk_index()
3679 !mas_slot_locked(l_mas, l_wr_mas->slots, l_slot - 1))) { in mas_extend_spanning_null()
3694 !mas_slot_locked(r_mas, r_wr_mas->slots, r_mas->offset + 1)) { in mas_extend_spanning_null()
3731 void __rcu **slots; in mtree_lookup_walk() local
3752 slots = ma_slots(node, type); in mtree_lookup_walk()
3753 next = mt_slot(mas->tree, slots, offset); in mtree_lookup_walk()
3781 void __rcu **slots; in mas_new_root() local
3798 slots = ma_slots(node, type); in mas_new_root()
3801 rcu_assign_pointer(slots[0], entry); in mas_new_root()
3958 memcpy(dst_slots, wr_mas->slots, sizeof(void *) * mas->offset); in mas_wr_node_store()
3981 memcpy(dst_slots + dst_offset, wr_mas->slots + offset_end, in mas_wr_node_store()
4014 void __rcu **slots = wr_mas->slots; in mas_wr_slot_store() local
4017 gap |= !mt_slot_locked(mas->tree, slots, offset); in mas_wr_slot_store()
4018 gap |= !mt_slot_locked(mas->tree, slots, offset + 1); in mas_wr_slot_store()
4023 rcu_assign_pointer(slots[offset], wr_mas->entry); in mas_wr_slot_store()
4027 rcu_assign_pointer(slots[offset + 1], wr_mas->entry); in mas_wr_slot_store()
4036 gap |= !mt_slot_locked(mas->tree, slots, offset + 2); in mas_wr_slot_store()
4037 rcu_assign_pointer(slots[offset + 1], wr_mas->entry); in mas_wr_slot_store()
4060 if (!wr_mas->slots[wr_mas->offset_end]) { in mas_wr_extend_null()
4067 !wr_mas->slots[wr_mas->offset_end + 1]) { in mas_wr_extend_null()
4083 !wr_mas->slots[mas->offset - 1]) { in mas_wr_extend_null()
4137 void __rcu **slots; in mas_wr_append() local
4156 slots = wr_mas->slots; in mas_wr_append()
4160 rcu_assign_pointer(slots[new_end], wr_mas->entry); in mas_wr_append()
4165 rcu_assign_pointer(slots[new_end], wr_mas->content); in mas_wr_append()
4167 rcu_assign_pointer(slots[end], wr_mas->entry); in mas_wr_append()
4171 rcu_assign_pointer(slots[new_end], wr_mas->content); in mas_wr_append()
4173 rcu_assign_pointer(slots[end + 1], wr_mas->entry); in mas_wr_append()
4208 rcu_assign_pointer(wr_mas->slots[mas->offset], wr_mas->entry); in mas_wr_modify()
4363 void __rcu **slots; in mas_prev_node() local
4393 slots = ma_slots(node, mt); in mas_prev_node()
4394 mas->node = mas_slot(mas, slots, offset); in mas_prev_node()
4406 slots = ma_slots(node, mt); in mas_prev_node()
4407 mas->node = mas_slot(mas, slots, offset); in mas_prev_node()
4443 void __rcu **slots; in mas_prev_slot() local
4488 slots = ma_slots(node, type); in mas_prev_slot()
4489 entry = mas_slot(mas, slots, mas->offset); in mas_prev_slot()
4528 void __rcu **slots; in mas_next_node() local
4553 slots = ma_slots(node, mt); in mas_next_node()
4555 enode = mas_slot(mas, slots, mas->offset); in mas_next_node()
4567 slots = ma_slots(node, mt); in mas_next_node()
4568 enode = mas_slot(mas, slots, 0); in mas_next_node()
4606 void __rcu **slots; in mas_next_slot() local
4664 slots = ma_slots(node, type); in mas_next_slot()
4665 entry = mt_slot(mas->tree, slots, mas->offset); in mas_next_slot()
4727 void __rcu **slots; in mas_rev_awalk() local
4742 slots = ma_slots(node, type); in mas_rev_awalk()
4755 else if (!mas_slot(mas, slots, offset)) in mas_rev_awalk()
4793 mas->node = mas_slot(mas, slots, offset); in mas_rev_awalk()
4814 void __rcu **slots; in mas_anode_descend() local
4825 slots = ma_slots(node, type); in mas_anode_descend()
4839 else if (!mas_slot(mas, slots, offset)) in mas_anode_descend()
4850 mas->node = mas_slot(mas, slots, offset); in mas_anode_descend()
5141 void __rcu **slots) in mte_dead_leaves() argument
5149 entry = mt_slot(mt, slots, offset); in mte_dead_leaves()
5158 rcu_assign_pointer(slots[offset], node); in mte_dead_leaves()
5174 void __rcu **slots = NULL; in mte_dead_walk() local
5180 slots = ma_slots(node, node->type); in mte_dead_walk()
5181 next = rcu_dereference_protected(slots[offset], in mte_dead_walk()
5186 return slots; in mte_dead_walk()
5197 void __rcu **slots; in mt_free_walk() local
5210 slots = mte_dead_walk(&enode, 0); in mt_free_walk()
5213 mt_free_bulk(node->slot_len, slots); in mt_free_walk()
5220 slots = ma_slots(mte_to_node(enode), type); in mt_free_walk()
5222 rcu_dereference_protected(slots[offset], in mt_free_walk()
5224 slots = mte_dead_walk(&enode, offset); in mt_free_walk()
5228 slots = ma_slots(node, node->type); in mt_free_walk()
5229 mt_free_bulk(node->slot_len, slots); in mt_free_walk()
5240 void __rcu **slots = NULL; in mte_destroy_descend() local
5248 slots = ma_slots(node, type); in mte_destroy_descend()
5249 next = mt_slot_locked(mt, slots, next_offset); in mte_destroy_descend()
5251 next = mt_slot_locked(mt, slots, ++next_offset); in mte_destroy_descend()
5262 return slots; in mte_destroy_descend()
5268 void __rcu **slots; in mt_destroy_walk() local
5278 slots = mte_destroy_descend(&enode, mt, start, 0); in mt_destroy_walk()
5285 node->slot_len = mte_dead_leaves(enode, mt, slots); in mt_destroy_walk()
5287 mt_free_bulk(node->slot_len, slots); in mt_destroy_walk()
5294 slots = ma_slots(mte_to_node(enode), type); in mt_destroy_walk()
5298 tmp = mt_slot_locked(mt, slots, offset); in mt_destroy_walk()
5302 slots = mte_destroy_descend(&enode, mt, parent, offset); in mt_destroy_walk()
5309 node->slot_len = mte_dead_leaves(enode, mt, slots); in mt_destroy_walk()
5311 mt_free_bulk(node->slot_len, slots); in mt_destroy_walk()
7009 void __rcu **slots; in mas_validate_parent_slot() local
7018 slots = ma_slots(parent, p_type); in mas_validate_parent_slot()
7024 node = mas_slot(mas, slots, i); in mas_validate_parent_slot()
7041 void __rcu **slots = ma_slots(mte_to_node(mas->node), type); in mas_validate_child_slot() local
7050 child = mas_slot(mas, slots, i); in mas_validate_child_slot()
7087 void __rcu **slots = ma_slots(mte_to_node(mas->node), type); in mas_validate_limits() local
7129 void *entry = mas_slot(mas, slots, i); in mas_validate_limits()
7154 void __rcu **slots; in mt_validate_nulls() local
7164 slots = ma_slots(mte_to_node(mas.node), mte_node_type(mas.node)); in mt_validate_nulls()
7166 entry = mas_slot(&mas, slots, offset); in mt_validate_nulls()
7178 slots = ma_slots(mte_to_node(mas.node), in mt_validate_nulls()