Lines Matching refs:slots

815 		void __rcu **slots, unsigned char offset)  in mt_slot()  argument
817 return rcu_dereference_check(slots[offset], mt_locked(mt)); in mt_slot()
828 static inline void *mas_slot_locked(struct ma_state *mas, void __rcu **slots, in mas_slot_locked() argument
831 return rcu_dereference_protected(slots[offset], mt_locked(mas->tree)); in mas_slot_locked()
842 static inline void *mas_slot(struct ma_state *mas, void __rcu **slots, in mas_slot() argument
845 return mt_slot(mas->tree, slots, offset); in mas_slot()
1013 void __rcu **slots; in mas_descend() local
1018 slots = ma_slots(node, type); in mas_descend()
1023 mas->node = mas_slot(mas, slots, mas->offset); in mas_descend()
1219 void **slots = NULL; in mas_alloc_nodes() local
1256 slots = (void **)&node->slot[offset]; in mas_alloc_nodes()
1259 slots = (void **)&node->slot; in mas_alloc_nodes()
1263 count = mt_alloc_bulk(gfp, max_req, slots); in mas_alloc_nodes()
1279 memset(slots, 0, max_req * sizeof(unsigned long)); in mas_alloc_nodes()
1461 void __rcu **slots; in mas_leaf_max_gap() local
1467 slots = ma_slots(mn, mt); in mas_leaf_max_gap()
1472 if (slots[i]) { in mas_leaf_max_gap()
1490 if (likely(!slots[0])) { in mas_leaf_max_gap()
1503 if (unlikely(mas->max == ULONG_MAX) && !slots[max_piv + 1]) { in mas_leaf_max_gap()
1511 if (likely(slots[i])) in mas_leaf_max_gap()
1679 void __rcu **slots = ma_slots(node, type); in mas_adopt_children() local
1686 child = mas_slot_locked(mas, slots, offset); in mas_adopt_children()
1704 void __rcu **slots = NULL; in mas_replace() local
1710 slots = ma_slots(mte_parent(mas->node), in mas_replace()
1712 old_enode = mas_slot_locked(mas, slots, offset); in mas_replace()
1724 rcu_assign_pointer(slots[offset], mas->node); in mas_replace()
1745 void __rcu **slots; in mas_new_child() local
1749 slots = ma_slots(node, mt); in mas_new_child()
1753 entry = mas_slot_locked(mas, slots, offset); in mas_new_child()
1914 void __rcu **slots; in mas_mab_cp() local
1945 slots = ma_slots(node, mt); in mas_mab_cp()
1946 memcpy(b_node->slot + mab_start, slots + mas_start, sizeof(void *) * j); in mas_mab_cp()
1995 void __rcu **slots = ma_slots(node, mt); in mab_mas_cp() local
2004 slots[mt_pivots[mt]] = NULL; in mab_mas_cp()
2011 memcpy(slots, b_node->slot + mab_start, in mab_mas_cp()
2161 wr_mas->content = mas_slot_locked(mas, wr_mas->slots, in mas_store_b_node()
2310 void __rcu **slots; in mas_topiary_range() local
2314 slots = ma_slots(mas_mn(mas), mte_node_type(mas->node)); in mas_topiary_range()
2316 struct maple_enode *enode = mas_slot_locked(mas, slots, offset); in mas_topiary_range()
2903 void __rcu **slots; in mtree_range_walk() local
2939 slots = ma_slots(node, type); in mtree_range_walk()
2940 next = mt_slot(mas->tree, slots, offset); in mtree_range_walk()
3175 void __rcu **l_slots, **slots; in mas_destroy_rebalance() local
3198 slots = ma_slots(newnode, mt); in mas_destroy_rebalance()
3207 memcpy(slots, l_slots + split + 1, sizeof(void *) * tmp); in mas_destroy_rebalance()
3210 memcpy(slots + tmp, ma_slots(node, mt), sizeof(void *) * end); in mas_destroy_rebalance()
3227 memset(slots + tmp, 0, sizeof(void *) * (max_s - tmp)); in mas_destroy_rebalance()
3250 slots = ma_slots(new_left, mt); in mas_destroy_rebalance()
3252 memcpy(slots, l_slots, sizeof(void *) * split); in mas_destroy_rebalance()
3261 slots = ma_slots(parent, mt); in mas_destroy_rebalance()
3264 rcu_assign_pointer(slots[offset], mas->node); in mas_destroy_rebalance()
3265 rcu_assign_pointer(slots[offset - 1], l_mas.node); in mas_destroy_rebalance()
3576 memset(wr_mas->slots + bn->b_end, 0, sizeof(void *) * clear--); in mas_reuse_node()
3631 void __rcu **slots; in mas_root_expand() local
3641 slots = ma_slots(node, type); in mas_root_expand()
3648 rcu_assign_pointer(slots[slot], contents); in mas_root_expand()
3655 rcu_assign_pointer(slots[slot], entry); in mas_root_expand()
3740 wr_mas->slots = ma_slots(wr_mas->node, wr_mas->type); in mas_wr_walk_descend()
3768 wr_mas->content = mas_slot_locked(mas, wr_mas->slots, in mas_wr_walk()
3785 wr_mas->content = mas_slot_locked(mas, wr_mas->slots, in mas_wr_walk_index()
3812 !mas_slot_locked(l_mas, l_wr_mas->slots, l_slot - 1))) { in mas_extend_spanning_null()
3827 !mas_slot_locked(r_mas, r_wr_mas->slots, r_mas->offset + 1)) { in mas_extend_spanning_null()
3864 void __rcu **slots; in mtree_lookup_walk() local
3890 slots = ma_slots(node, type); in mtree_lookup_walk()
3891 next = mt_slot(mas->tree, slots, offset); in mtree_lookup_walk()
3918 void __rcu **slots; in mas_new_root() local
3935 slots = ma_slots(node, type); in mas_new_root()
3939 rcu_assign_pointer(slots[0], entry); in mas_new_root()
4122 memcpy(dst_slots, wr_mas->slots, sizeof(void *) * (offset + 1)); in mas_wr_node_store()
4150 memcpy(dst_slots + dst_offset, wr_mas->slots + wr_mas->offset_end, in mas_wr_node_store()
4208 rcu_assign_pointer(wr_mas->slots[offset], wr_mas->entry); in mas_wr_slot_store()
4222 rcu_assign_pointer(wr_mas->slots[offset + 1], wr_mas->entry); in mas_wr_slot_store()
4246 if (mas->last < wr_mas->end_piv && !wr_mas->slots[wr_mas->offset_end]) in mas_wr_extend_null()
4252 !wr_mas->slots[wr_mas->offset_end + 1]) { in mas_wr_extend_null()
4267 !wr_mas->slots[mas->offset - 1]) { in mas_wr_extend_null()
4290 rcu_assign_pointer(wr_mas->slots[new_end], wr_mas->entry); in mas_wr_append()
4301 rcu_assign_pointer(wr_mas->slots[new_end], wr_mas->content); in mas_wr_append()
4306 rcu_assign_pointer(wr_mas->slots[end], wr_mas->entry); in mas_wr_append()
4337 rcu_assign_pointer(wr_mas->slots[mas->offset], wr_mas->entry); in mas_wr_modify()
4482 void __rcu **slots; in mas_prev_node() local
4506 slots = ma_slots(node, mt); in mas_prev_node()
4519 enode = mas_slot(mas, slots, offset); in mas_prev_node()
4526 slots = ma_slots(node, mt); in mas_prev_node()
4539 mas->node = mas_slot(mas, slots, offset); in mas_prev_node()
4578 void __rcu **slots; in mas_next_node() local
4602 slots = ma_slots(node, mt); in mas_next_node()
4606 enode = mas_slot(mas, slots, offset); in mas_next_node()
4614 slots = ma_slots(node, mt); in mas_next_node()
4620 enode = mas_slot(mas, slots, offset); in mas_next_node()
4654 void __rcu **slots; in mas_next_nentry() local
4663 slots = ma_slots(node, type); in mas_next_nentry()
4677 entry = mas_slot(mas, slots, mas->offset); in mas_next_nentry()
4697 entry = mas_slot(mas, slots, mas->offset); in mas_next_nentry()
4805 void __rcu **slots; in mas_prev_nentry() local
4818 slots = ma_slots(mn, mt); in mas_prev_nentry()
4830 while (offset && ((!mas_slot(mas, slots, offset) && pivot >= limit) || in mas_prev_nentry()
4835 entry = mas_slot(mas, slots, offset); in mas_prev_nentry()
4895 void __rcu **slots; in mas_rev_awalk() local
4910 slots = ma_slots(node, type); in mas_rev_awalk()
4923 else if (!mas_slot(mas, slots, offset)) in mas_rev_awalk()
4961 mas->node = mas_slot(mas, slots, offset); in mas_rev_awalk()
4983 void __rcu **slots = ma_slots(mas_mn(mas), type); in mas_anode_descend() local
5005 else if (!mas_slot(mas, slots, offset)) in mas_anode_descend()
5016 mas->node = mas_slot(mas, slots, offset); in mas_anode_descend()
5426 unsigned char mas_dead_leaves(struct ma_state *mas, void __rcu **slots) in mas_dead_leaves() argument
5434 entry = mas_slot_locked(mas, slots, offset); in mas_dead_leaves()
5444 rcu_assign_pointer(slots[offset], node); in mas_dead_leaves()
5453 void __rcu **slots = NULL; in mas_dead_walk() local
5459 slots = ma_slots(node, node->type); in mas_dead_walk()
5460 next = mas_slot_locked(mas, slots, offset); in mas_dead_walk()
5464 return slots; in mas_dead_walk()
5469 void __rcu **slots; in mt_free_walk() local
5485 slots = mas_dead_walk(&mas, 0); in mt_free_walk()
5488 mt_free_bulk(node->slot_len, slots); in mt_free_walk()
5495 slots = ma_slots(mte_to_node(mas.node), type); in mt_free_walk()
5496 if ((offset < mt_slots[type]) && (slots[offset])) in mt_free_walk()
5497 slots = mas_dead_walk(&mas, offset); in mt_free_walk()
5502 slots = ma_slots(node, node->type); in mt_free_walk()
5503 mt_free_bulk(node->slot_len, slots); in mt_free_walk()
5516 void __rcu **slots = NULL; in mas_destroy_descend() local
5521 slots = ma_slots(node, mte_node_type(mas->node)); in mas_destroy_descend()
5522 next = mas_slot_locked(mas, slots, 0); in mas_destroy_descend()
5524 next = mas_slot_locked(mas, slots, 1); in mas_destroy_descend()
5534 return slots; in mas_destroy_descend()
5540 void __rcu **slots; in mt_destroy_walk() local
5554 slots = mas_destroy_descend(&mas, start, 0); in mt_destroy_walk()
5561 node->slot_len = mas_dead_leaves(&mas, slots); in mt_destroy_walk()
5563 mt_free_bulk(node->slot_len, slots); in mt_destroy_walk()
5570 slots = ma_slots(mte_to_node(mas.node), type); in mt_destroy_walk()
5574 tmp = mas_slot_locked(&mas, slots, offset); in mt_destroy_walk()
5578 slots = mas_destroy_descend(&mas, parent, offset); in mt_destroy_walk()
5585 node->slot_len = mas_dead_leaves(&mas, slots); in mt_destroy_walk()
5587 mt_free_bulk(node->slot_len, slots); in mt_destroy_walk()
6622 void __rcu **slots; in mas_first_entry() local
6633 slots = ma_slots(mn, mt); in mas_first_entry()
6636 entry = mas_slot(mas, slots, 0); in mas_first_entry()
6646 slots = ma_slots(mn, mt); in mas_first_entry()
6647 entry = mas_slot(mas, slots, 0); in mas_first_entry()
6661 entry = mas_slot(mas, slots, 1); in mas_first_entry()
6961 void __rcu **slots; in mas_validate_parent_slot() local
6968 slots = ma_slots(parent, p_type); in mas_validate_parent_slot()
6974 node = mas_slot(mas, slots, i); in mas_validate_parent_slot()
6991 void __rcu **slots = ma_slots(mte_to_node(mas->node), type); in mas_validate_child_slot() local
7000 child = mas_slot(mas, slots, i); in mas_validate_child_slot()
7031 void __rcu **slots = ma_slots(mte_to_node(mas->node), type); in mas_validate_limits() local
7047 void *entry = mas_slot(mas, slots, i); in mas_validate_limits()
7077 void *entry = mas_slot(mas, slots, i); in mas_validate_limits()
7102 void __rcu **slots; in mt_validate_nulls() local
7112 slots = ma_slots(mte_to_node(mas.node), mte_node_type(mas.node)); in mt_validate_nulls()
7114 entry = mas_slot(&mas, slots, offset); in mt_validate_nulls()
7126 slots = ma_slots(mte_to_node(mas.node), in mt_validate_nulls()