2 * Copyright (C) 2001 Momchil Velikov
3 * Portions Copyright (C) 2001 Christoph Hellwig
4 * Copyright (C) 2005 SGI, Christoph Lameter <clameter@sgi.com>
5 * Copyright (C) 2006 Nick Piggin
7 * This program is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU General Public License as
9 * published by the Free Software Foundation; either version 2, or (at
10 * your option) any later version.
12 * This program is distributed in the hope that it will be useful, but
13 * WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * General Public License for more details.
17 * You should have received a copy of the GNU General Public License
18 * along with this program; if not, write to the Free Software
19 * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
22 #include <linux/errno.h>
23 #include <linux/init.h>
24 #include <linux/kernel.h>
25 #include <linux/module.h>
26 #include <linux/radix-tree.h>
27 #include <linux/percpu.h>
28 #include <linux/slab.h>
29 #include <linux/notifier.h>
30 #include <linux/cpu.h>
31 #include <linux/gfp.h>
32 #include <linux/string.h>
33 #include <linux/bitops.h>
34 #include <linux/rcupdate.h>
38 #define RADIX_TREE_MAP_SHIFT (CONFIG_BASE_SMALL ? 4 : 6)
40 #define RADIX_TREE_MAP_SHIFT 3 /* For more stressful testing */
43 #define RADIX_TREE_MAP_SIZE (1UL << RADIX_TREE_MAP_SHIFT)
44 #define RADIX_TREE_MAP_MASK (RADIX_TREE_MAP_SIZE-1)
46 #define RADIX_TREE_TAG_LONGS \
47 ((RADIX_TREE_MAP_SIZE + BITS_PER_LONG - 1) / BITS_PER_LONG)
49 struct radix_tree_node
{
50 unsigned int height
; /* Height from the bottom */
52 struct rcu_head rcu_head
;
53 void *slots
[RADIX_TREE_MAP_SIZE
];
54 unsigned long tags
[RADIX_TREE_MAX_TAGS
][RADIX_TREE_TAG_LONGS
];
57 struct radix_tree_path
{
58 struct radix_tree_node
*node
;
62 #define RADIX_TREE_INDEX_BITS (8 /* CHAR_BIT */ * sizeof(unsigned long))
63 #define RADIX_TREE_MAX_PATH (RADIX_TREE_INDEX_BITS/RADIX_TREE_MAP_SHIFT + 2)
65 static unsigned long height_to_maxindex
[RADIX_TREE_MAX_PATH
] __read_mostly
;
68 * Radix tree node cache.
70 static struct kmem_cache
*radix_tree_node_cachep
;
73 * Per-cpu pool of preloaded nodes
75 struct radix_tree_preload
{
77 struct radix_tree_node
*nodes
[RADIX_TREE_MAX_PATH
];
79 DEFINE_PER_CPU(struct radix_tree_preload
, radix_tree_preloads
) = { 0, };
81 static inline gfp_t
root_gfp_mask(struct radix_tree_root
*root
)
83 return root
->gfp_mask
& __GFP_BITS_MASK
;
87 * This assumes that the caller has performed appropriate preallocation, and
88 * that the caller has pinned this thread of control to the current CPU.
90 static struct radix_tree_node
*
91 radix_tree_node_alloc(struct radix_tree_root
*root
)
93 struct radix_tree_node
*ret
;
94 gfp_t gfp_mask
= root_gfp_mask(root
);
96 ret
= kmem_cache_alloc(radix_tree_node_cachep
, gfp_mask
);
97 if (ret
== NULL
&& !(gfp_mask
& __GFP_WAIT
)) {
98 struct radix_tree_preload
*rtp
;
100 rtp
= &__get_cpu_var(radix_tree_preloads
);
102 ret
= rtp
->nodes
[rtp
->nr
- 1];
103 rtp
->nodes
[rtp
->nr
- 1] = NULL
;
107 BUG_ON(radix_tree_is_direct_ptr(ret
));
111 static void radix_tree_node_rcu_free(struct rcu_head
*head
)
113 struct radix_tree_node
*node
=
114 container_of(head
, struct radix_tree_node
, rcu_head
);
115 kmem_cache_free(radix_tree_node_cachep
, node
);
119 radix_tree_node_free(struct radix_tree_node
*node
)
121 call_rcu(&node
->rcu_head
, radix_tree_node_rcu_free
);
125 * Load up this CPU's radix_tree_node buffer with sufficient objects to
126 * ensure that the addition of a single element in the tree cannot fail. On
127 * success, return zero, with preemption disabled. On error, return -ENOMEM
128 * with preemption not disabled.
130 int radix_tree_preload(gfp_t gfp_mask
)
132 struct radix_tree_preload
*rtp
;
133 struct radix_tree_node
*node
;
137 rtp
= &__get_cpu_var(radix_tree_preloads
);
138 while (rtp
->nr
< ARRAY_SIZE(rtp
->nodes
)) {
140 node
= kmem_cache_alloc(radix_tree_node_cachep
, gfp_mask
);
144 rtp
= &__get_cpu_var(radix_tree_preloads
);
145 if (rtp
->nr
< ARRAY_SIZE(rtp
->nodes
))
146 rtp
->nodes
[rtp
->nr
++] = node
;
148 kmem_cache_free(radix_tree_node_cachep
, node
);
155 static inline void tag_set(struct radix_tree_node
*node
, unsigned int tag
,
158 __set_bit(offset
, node
->tags
[tag
]);
161 static inline void tag_clear(struct radix_tree_node
*node
, unsigned int tag
,
164 __clear_bit(offset
, node
->tags
[tag
]);
167 static inline int tag_get(struct radix_tree_node
*node
, unsigned int tag
,
170 return test_bit(offset
, node
->tags
[tag
]);
173 static inline void root_tag_set(struct radix_tree_root
*root
, unsigned int tag
)
175 root
->gfp_mask
|= (__force gfp_t
)(1 << (tag
+ __GFP_BITS_SHIFT
));
179 static inline void root_tag_clear(struct radix_tree_root
*root
, unsigned int tag
)
181 root
->gfp_mask
&= (__force gfp_t
)~(1 << (tag
+ __GFP_BITS_SHIFT
));
184 static inline void root_tag_clear_all(struct radix_tree_root
*root
)
186 root
->gfp_mask
&= __GFP_BITS_MASK
;
189 static inline int root_tag_get(struct radix_tree_root
*root
, unsigned int tag
)
191 return (__force
unsigned)root
->gfp_mask
& (1 << (tag
+ __GFP_BITS_SHIFT
));
195 * Returns 1 if any slot in the node has this tag set.
196 * Otherwise returns 0.
198 static inline int any_tag_set(struct radix_tree_node
*node
, unsigned int tag
)
201 for (idx
= 0; idx
< RADIX_TREE_TAG_LONGS
; idx
++) {
202 if (node
->tags
[tag
][idx
])
209 * Return the maximum key which can be store into a
210 * radix tree with height HEIGHT.
212 static inline unsigned long radix_tree_maxindex(unsigned int height
)
214 return height_to_maxindex
[height
];
218 * Extend a radix tree so it can store key @index.
220 static int radix_tree_extend(struct radix_tree_root
*root
, unsigned long index
)
222 struct radix_tree_node
*node
;
226 /* Figure out what the height should be. */
227 height
= root
->height
+ 1;
228 while (index
> radix_tree_maxindex(height
))
231 if (root
->rnode
== NULL
) {
232 root
->height
= height
;
237 unsigned int newheight
;
238 if (!(node
= radix_tree_node_alloc(root
)))
241 /* Increase the height. */
242 node
->slots
[0] = radix_tree_direct_to_ptr(root
->rnode
);
244 /* Propagate the aggregated tag info into the new root */
245 for (tag
= 0; tag
< RADIX_TREE_MAX_TAGS
; tag
++) {
246 if (root_tag_get(root
, tag
))
247 tag_set(node
, tag
, 0);
250 newheight
= root
->height
+1;
251 node
->height
= newheight
;
253 rcu_assign_pointer(root
->rnode
, node
);
254 root
->height
= newheight
;
255 } while (height
> root
->height
);
261 * radix_tree_insert - insert into a radix tree
262 * @root: radix tree root
264 * @item: item to insert
266 * Insert an item into the radix tree at position @index.
268 int radix_tree_insert(struct radix_tree_root
*root
,
269 unsigned long index
, void *item
)
271 struct radix_tree_node
*node
= NULL
, *slot
;
272 unsigned int height
, shift
;
276 BUG_ON(radix_tree_is_direct_ptr(item
));
278 /* Make sure the tree is high enough. */
279 if (index
> radix_tree_maxindex(root
->height
)) {
280 error
= radix_tree_extend(root
, index
);
286 height
= root
->height
;
287 shift
= (height
-1) * RADIX_TREE_MAP_SHIFT
;
289 offset
= 0; /* uninitialised var warning */
292 /* Have to add a child node. */
293 if (!(slot
= radix_tree_node_alloc(root
)))
295 slot
->height
= height
;
297 rcu_assign_pointer(node
->slots
[offset
], slot
);
300 rcu_assign_pointer(root
->rnode
, slot
);
303 /* Go a level down */
304 offset
= (index
>> shift
) & RADIX_TREE_MAP_MASK
;
306 slot
= node
->slots
[offset
];
307 shift
-= RADIX_TREE_MAP_SHIFT
;
316 rcu_assign_pointer(node
->slots
[offset
], item
);
317 BUG_ON(tag_get(node
, 0, offset
));
318 BUG_ON(tag_get(node
, 1, offset
));
320 rcu_assign_pointer(root
->rnode
, radix_tree_ptr_to_direct(item
));
321 BUG_ON(root_tag_get(root
, 0));
322 BUG_ON(root_tag_get(root
, 1));
327 EXPORT_SYMBOL(radix_tree_insert
);
330 * radix_tree_lookup_slot - lookup a slot in a radix tree
331 * @root: radix tree root
334 * Returns: the slot corresponding to the position @index in the
335 * radix tree @root. This is useful for update-if-exists operations.
337 * This function cannot be called under rcu_read_lock, it must be
338 * excluded from writers, as must the returned slot for subsequent
339 * use by radix_tree_deref_slot() and radix_tree_replace slot.
340 * Caller must hold tree write locked across slot lookup and
343 void **radix_tree_lookup_slot(struct radix_tree_root
*root
, unsigned long index
)
345 unsigned int height
, shift
;
346 struct radix_tree_node
*node
, **slot
;
352 if (radix_tree_is_direct_ptr(node
)) {
355 return (void **)&root
->rnode
;
358 height
= node
->height
;
359 if (index
> radix_tree_maxindex(height
))
362 shift
= (height
-1) * RADIX_TREE_MAP_SHIFT
;
365 slot
= (struct radix_tree_node
**)
366 (node
->slots
+ ((index
>>shift
) & RADIX_TREE_MAP_MASK
));
371 shift
-= RADIX_TREE_MAP_SHIFT
;
373 } while (height
> 0);
375 return (void **)slot
;
377 EXPORT_SYMBOL(radix_tree_lookup_slot
);
380 * radix_tree_lookup - perform lookup operation on a radix tree
381 * @root: radix tree root
384 * Lookup the item at the position @index in the radix tree @root.
386 * This function can be called under rcu_read_lock, however the caller
387 * must manage lifetimes of leaf nodes (eg. RCU may also be used to free
388 * them safely). No RCU barriers are required to access or modify the
389 * returned item, however.
391 void *radix_tree_lookup(struct radix_tree_root
*root
, unsigned long index
)
393 unsigned int height
, shift
;
394 struct radix_tree_node
*node
, **slot
;
396 node
= rcu_dereference(root
->rnode
);
400 if (radix_tree_is_direct_ptr(node
)) {
403 return radix_tree_direct_to_ptr(node
);
406 height
= node
->height
;
407 if (index
> radix_tree_maxindex(height
))
410 shift
= (height
-1) * RADIX_TREE_MAP_SHIFT
;
413 slot
= (struct radix_tree_node
**)
414 (node
->slots
+ ((index
>>shift
) & RADIX_TREE_MAP_MASK
));
415 node
= rcu_dereference(*slot
);
419 shift
-= RADIX_TREE_MAP_SHIFT
;
421 } while (height
> 0);
425 EXPORT_SYMBOL(radix_tree_lookup
);
428 * radix_tree_tag_set - set a tag on a radix tree node
429 * @root: radix tree root
433 * Set the search tag (which must be < RADIX_TREE_MAX_TAGS)
434 * corresponding to @index in the radix tree. From
435 * the root all the way down to the leaf node.
437 * Returns the address of the tagged item. Setting a tag on a not-present
440 void *radix_tree_tag_set(struct radix_tree_root
*root
,
441 unsigned long index
, unsigned int tag
)
443 unsigned int height
, shift
;
444 struct radix_tree_node
*slot
;
446 height
= root
->height
;
447 BUG_ON(index
> radix_tree_maxindex(height
));
450 shift
= (height
- 1) * RADIX_TREE_MAP_SHIFT
;
455 offset
= (index
>> shift
) & RADIX_TREE_MAP_MASK
;
456 if (!tag_get(slot
, tag
, offset
))
457 tag_set(slot
, tag
, offset
);
458 slot
= slot
->slots
[offset
];
459 BUG_ON(slot
== NULL
);
460 shift
-= RADIX_TREE_MAP_SHIFT
;
464 /* set the root's tag bit */
465 if (slot
&& !root_tag_get(root
, tag
))
466 root_tag_set(root
, tag
);
470 EXPORT_SYMBOL(radix_tree_tag_set
);
473 * radix_tree_tag_clear - clear a tag on a radix tree node
474 * @root: radix tree root
478 * Clear the search tag (which must be < RADIX_TREE_MAX_TAGS)
479 * corresponding to @index in the radix tree. If
480 * this causes the leaf node to have no tags set then clear the tag in the
481 * next-to-leaf node, etc.
483 * Returns the address of the tagged item on success, else NULL. ie:
484 * has the same return value and semantics as radix_tree_lookup().
486 void *radix_tree_tag_clear(struct radix_tree_root
*root
,
487 unsigned long index
, unsigned int tag
)
489 struct radix_tree_path path
[RADIX_TREE_MAX_PATH
], *pathp
= path
;
490 struct radix_tree_node
*slot
= NULL
;
491 unsigned int height
, shift
;
493 height
= root
->height
;
494 if (index
> radix_tree_maxindex(height
))
497 shift
= (height
- 1) * RADIX_TREE_MAP_SHIFT
;
507 offset
= (index
>> shift
) & RADIX_TREE_MAP_MASK
;
508 pathp
[1].offset
= offset
;
509 pathp
[1].node
= slot
;
510 slot
= slot
->slots
[offset
];
512 shift
-= RADIX_TREE_MAP_SHIFT
;
519 while (pathp
->node
) {
520 if (!tag_get(pathp
->node
, tag
, pathp
->offset
))
522 tag_clear(pathp
->node
, tag
, pathp
->offset
);
523 if (any_tag_set(pathp
->node
, tag
))
528 /* clear the root's tag bit */
529 if (root_tag_get(root
, tag
))
530 root_tag_clear(root
, tag
);
535 EXPORT_SYMBOL(radix_tree_tag_clear
);
537 #ifndef __KERNEL__ /* Only the test harness uses this at present */
539 * radix_tree_tag_get - get a tag on a radix tree node
540 * @root: radix tree root
542 * @tag: tag index (< RADIX_TREE_MAX_TAGS)
546 * 0: tag not present or not set
549 int radix_tree_tag_get(struct radix_tree_root
*root
,
550 unsigned long index
, unsigned int tag
)
552 unsigned int height
, shift
;
553 struct radix_tree_node
*node
;
554 int saw_unset_tag
= 0;
556 /* check the root's tag bit */
557 if (!root_tag_get(root
, tag
))
560 node
= rcu_dereference(root
->rnode
);
564 if (radix_tree_is_direct_ptr(node
))
567 height
= node
->height
;
568 if (index
> radix_tree_maxindex(height
))
571 shift
= (height
- 1) * RADIX_TREE_MAP_SHIFT
;
579 offset
= (index
>> shift
) & RADIX_TREE_MAP_MASK
;
582 * This is just a debug check. Later, we can bale as soon as
583 * we see an unset tag.
585 if (!tag_get(node
, tag
, offset
))
588 int ret
= tag_get(node
, tag
, offset
);
590 BUG_ON(ret
&& saw_unset_tag
);
593 node
= rcu_dereference(node
->slots
[offset
]);
594 shift
-= RADIX_TREE_MAP_SHIFT
;
598 EXPORT_SYMBOL(radix_tree_tag_get
);
602 __lookup(struct radix_tree_node
*slot
, void **results
, unsigned long index
,
603 unsigned int max_items
, unsigned long *next_index
)
605 unsigned int nr_found
= 0;
606 unsigned int shift
, height
;
609 height
= slot
->height
;
612 shift
= (height
-1) * RADIX_TREE_MAP_SHIFT
;
614 for ( ; height
> 1; height
--) {
615 i
= (index
>> shift
) & RADIX_TREE_MAP_MASK
;
617 if (slot
->slots
[i
] != NULL
)
619 index
&= ~((1UL << shift
) - 1);
620 index
+= 1UL << shift
;
622 goto out
; /* 32-bit wraparound */
624 if (i
== RADIX_TREE_MAP_SIZE
)
628 shift
-= RADIX_TREE_MAP_SHIFT
;
629 slot
= rcu_dereference(slot
->slots
[i
]);
634 /* Bottom level: grab some items */
635 for (i
= index
& RADIX_TREE_MAP_MASK
; i
< RADIX_TREE_MAP_SIZE
; i
++) {
636 struct radix_tree_node
*node
;
638 node
= slot
->slots
[i
];
640 results
[nr_found
++] = rcu_dereference(node
);
641 if (nr_found
== max_items
)
651 * radix_tree_gang_lookup - perform multiple lookup on a radix tree
652 * @root: radix tree root
653 * @results: where the results of the lookup are placed
654 * @first_index: start the lookup from this key
655 * @max_items: place up to this many items at *results
657 * Performs an index-ascending scan of the tree for present items. Places
658 * them at *@results and returns the number of items which were placed at
661 * The implementation is naive.
663 * Like radix_tree_lookup, radix_tree_gang_lookup may be called under
664 * rcu_read_lock. In this case, rather than the returned results being
665 * an atomic snapshot of the tree at a single point in time, the semantics
666 * of an RCU protected gang lookup are as though multiple radix_tree_lookups
667 * have been issued in individual locks, and results stored in 'results'.
670 radix_tree_gang_lookup(struct radix_tree_root
*root
, void **results
,
671 unsigned long first_index
, unsigned int max_items
)
673 unsigned long max_index
;
674 struct radix_tree_node
*node
;
675 unsigned long cur_index
= first_index
;
678 node
= rcu_dereference(root
->rnode
);
682 if (radix_tree_is_direct_ptr(node
)) {
685 node
= radix_tree_direct_to_ptr(node
);
686 results
[0] = rcu_dereference(node
);
690 max_index
= radix_tree_maxindex(node
->height
);
693 while (ret
< max_items
) {
694 unsigned int nr_found
;
695 unsigned long next_index
; /* Index of next search */
697 if (cur_index
> max_index
)
699 nr_found
= __lookup(node
, results
+ ret
, cur_index
,
700 max_items
- ret
, &next_index
);
704 cur_index
= next_index
;
709 EXPORT_SYMBOL(radix_tree_gang_lookup
);
712 * FIXME: the two tag_get()s here should use find_next_bit() instead of
713 * open-coding the search.
716 __lookup_tag(struct radix_tree_node
*slot
, void **results
, unsigned long index
,
717 unsigned int max_items
, unsigned long *next_index
, unsigned int tag
)
719 unsigned int nr_found
= 0;
720 unsigned int shift
, height
;
722 height
= slot
->height
;
725 shift
= (height
-1) * RADIX_TREE_MAP_SHIFT
;
728 unsigned long i
= (index
>> shift
) & RADIX_TREE_MAP_MASK
;
731 if (tag_get(slot
, tag
, i
))
733 index
&= ~((1UL << shift
) - 1);
734 index
+= 1UL << shift
;
736 goto out
; /* 32-bit wraparound */
738 if (i
== RADIX_TREE_MAP_SIZE
)
742 if (height
== 0) { /* Bottom level: grab some items */
743 unsigned long j
= index
& RADIX_TREE_MAP_MASK
;
745 for ( ; j
< RADIX_TREE_MAP_SIZE
; j
++) {
746 struct radix_tree_node
*node
;
748 if (!tag_get(slot
, tag
, j
))
750 node
= slot
->slots
[j
];
752 * Even though the tag was found set, we need to
753 * recheck that we have a non-NULL node, because
754 * if this lookup is lockless, it may have been
755 * subsequently deleted.
757 * Similar care must be taken in any place that
758 * lookup ->slots[x] without a lock (ie. can't
759 * rely on its value remaining the same).
762 node
= rcu_dereference(node
);
763 results
[nr_found
++] = node
;
764 if (nr_found
== max_items
)
769 shift
-= RADIX_TREE_MAP_SHIFT
;
770 slot
= rcu_dereference(slot
->slots
[i
]);
780 * radix_tree_gang_lookup_tag - perform multiple lookup on a radix tree
782 * @root: radix tree root
783 * @results: where the results of the lookup are placed
784 * @first_index: start the lookup from this key
785 * @max_items: place up to this many items at *results
786 * @tag: the tag index (< RADIX_TREE_MAX_TAGS)
788 * Performs an index-ascending scan of the tree for present items which
789 * have the tag indexed by @tag set. Places the items at *@results and
790 * returns the number of items which were placed at *@results.
793 radix_tree_gang_lookup_tag(struct radix_tree_root
*root
, void **results
,
794 unsigned long first_index
, unsigned int max_items
,
797 struct radix_tree_node
*node
;
798 unsigned long max_index
;
799 unsigned long cur_index
= first_index
;
802 /* check the root's tag bit */
803 if (!root_tag_get(root
, tag
))
806 node
= rcu_dereference(root
->rnode
);
810 if (radix_tree_is_direct_ptr(node
)) {
813 node
= radix_tree_direct_to_ptr(node
);
814 results
[0] = rcu_dereference(node
);
818 max_index
= radix_tree_maxindex(node
->height
);
821 while (ret
< max_items
) {
822 unsigned int nr_found
;
823 unsigned long next_index
; /* Index of next search */
825 if (cur_index
> max_index
)
827 nr_found
= __lookup_tag(node
, results
+ ret
, cur_index
,
828 max_items
- ret
, &next_index
, tag
);
832 cur_index
= next_index
;
837 EXPORT_SYMBOL(radix_tree_gang_lookup_tag
);
840 * radix_tree_shrink - shrink height of a radix tree to minimal
841 * @root radix tree root
843 static inline void radix_tree_shrink(struct radix_tree_root
*root
)
845 /* try to shrink tree height */
846 while (root
->height
> 0 &&
847 root
->rnode
->count
== 1 &&
848 root
->rnode
->slots
[0]) {
849 struct radix_tree_node
*to_free
= root
->rnode
;
853 * We don't need rcu_assign_pointer(), since we are simply
854 * moving the node from one part of the tree to another. If
855 * it was safe to dereference the old pointer to it
856 * (to_free->slots[0]), it will be safe to dereference the new
859 newptr
= to_free
->slots
[0];
860 if (root
->height
== 1)
861 newptr
= radix_tree_ptr_to_direct(newptr
);
862 root
->rnode
= newptr
;
864 /* must only free zeroed nodes into the slab */
865 tag_clear(to_free
, 0, 0);
866 tag_clear(to_free
, 1, 0);
867 to_free
->slots
[0] = NULL
;
869 radix_tree_node_free(to_free
);
874 * radix_tree_delete - delete an item from a radix tree
875 * @root: radix tree root
878 * Remove the item at @index from the radix tree rooted at @root.
880 * Returns the address of the deleted item, or NULL if it was not present.
882 void *radix_tree_delete(struct radix_tree_root
*root
, unsigned long index
)
884 struct radix_tree_path path
[RADIX_TREE_MAX_PATH
], *pathp
= path
;
885 struct radix_tree_node
*slot
= NULL
;
886 struct radix_tree_node
*to_free
;
887 unsigned int height
, shift
;
891 height
= root
->height
;
892 if (index
> radix_tree_maxindex(height
))
896 if (height
== 0 && root
->rnode
) {
897 slot
= radix_tree_direct_to_ptr(slot
);
898 root_tag_clear_all(root
);
903 shift
= (height
- 1) * RADIX_TREE_MAP_SHIFT
;
911 offset
= (index
>> shift
) & RADIX_TREE_MAP_MASK
;
912 pathp
->offset
= offset
;
914 slot
= slot
->slots
[offset
];
915 shift
-= RADIX_TREE_MAP_SHIFT
;
917 } while (height
> 0);
923 * Clear all tags associated with the just-deleted item
925 for (tag
= 0; tag
< RADIX_TREE_MAX_TAGS
; tag
++) {
926 if (tag_get(pathp
->node
, tag
, pathp
->offset
))
927 radix_tree_tag_clear(root
, index
, tag
);
931 /* Now free the nodes we do not need anymore */
932 while (pathp
->node
) {
933 pathp
->node
->slots
[pathp
->offset
] = NULL
;
934 pathp
->node
->count
--;
936 * Queue the node for deferred freeing after the
937 * last reference to it disappears (set NULL, above).
940 radix_tree_node_free(to_free
);
942 if (pathp
->node
->count
) {
943 if (pathp
->node
== root
->rnode
)
944 radix_tree_shrink(root
);
948 /* Node with zero slots in use so free it */
949 to_free
= pathp
->node
;
953 root_tag_clear_all(root
);
957 radix_tree_node_free(to_free
);
962 EXPORT_SYMBOL(radix_tree_delete
);
965 * radix_tree_tagged - test whether any items in the tree are tagged
966 * @root: radix tree root
969 int radix_tree_tagged(struct radix_tree_root
*root
, unsigned int tag
)
971 return root_tag_get(root
, tag
);
973 EXPORT_SYMBOL(radix_tree_tagged
);
976 radix_tree_node_ctor(void *node
, struct kmem_cache
*cachep
, unsigned long flags
)
978 memset(node
, 0, sizeof(struct radix_tree_node
));
981 static __init
unsigned long __maxindex(unsigned int height
)
983 unsigned int tmp
= height
* RADIX_TREE_MAP_SHIFT
;
984 unsigned long index
= (~0UL >> (RADIX_TREE_INDEX_BITS
- tmp
- 1)) >> 1;
986 if (tmp
>= RADIX_TREE_INDEX_BITS
)
991 static __init
void radix_tree_init_maxindex(void)
995 for (i
= 0; i
< ARRAY_SIZE(height_to_maxindex
); i
++)
996 height_to_maxindex
[i
] = __maxindex(i
);
999 static int radix_tree_callback(struct notifier_block
*nfb
,
1000 unsigned long action
,
1003 int cpu
= (long)hcpu
;
1004 struct radix_tree_preload
*rtp
;
1006 /* Free per-cpu pool of perloaded nodes */
1007 if (action
== CPU_DEAD
|| action
== CPU_DEAD_FROZEN
) {
1008 rtp
= &per_cpu(radix_tree_preloads
, cpu
);
1010 kmem_cache_free(radix_tree_node_cachep
,
1011 rtp
->nodes
[rtp
->nr
-1]);
1012 rtp
->nodes
[rtp
->nr
-1] = NULL
;
1019 void __init
radix_tree_init(void)
1021 radix_tree_node_cachep
= kmem_cache_create("radix_tree_node",
1022 sizeof(struct radix_tree_node
), 0,
1023 SLAB_PANIC
, radix_tree_node_ctor
, NULL
);
1024 radix_tree_init_maxindex();
1025 hotcpu_notifier(radix_tree_callback
, 0);