arch/parisc/mm/fault.c
129
while (tree != vm_avl_empty) {
arch/parisc/mm/fault.c
130
if (tree->vm_start > addr) {
arch/parisc/mm/fault.c
131
tree = tree->vm_avl_left;
arch/parisc/mm/fault.c
133
prev = tree;
arch/parisc/mm/fault.c
138
tree = tree->vm_avl_right;
block/bfq-iosched.h
165
struct rb_root *tree;
block/bfq-wf2q.c
1029
if (entity->tree)
block/bfq-wf2q.c
1058
entity->tree == &st->active)
block/bfq-wf2q.c
1141
if (entity->tree == &st->active)
block/bfq-wf2q.c
1143
else if (!is_in_service && entity->tree == &st->idle)
block/bfq-wf2q.c
23
static struct bfq_entity *bfq_root_active_entity(struct rb_root *tree)
block/bfq-wf2q.c
25
struct rb_node *node = tree->rb_node;
block/bfq-wf2q.c
340
entity->tree = NULL;
block/bfq-wf2q.c
398
entity->tree = root;
block/bfq-wf2q.c
941
if (entity->tree == &st->idle) {
drivers/firewire/core-topology.c
341
struct fw_node *tree;
drivers/firewire/core-topology.c
344
tree = node1->ports[port];
drivers/firewire/core-topology.c
345
node0->ports[port] = tree;
drivers/firewire/core-topology.c
346
for (i = 0; i < tree->port_count; i++) {
drivers/firewire/core-topology.c
347
if (tree->ports[i] == node1) {
drivers/firewire/core-topology.c
348
tree->ports[i] = node0;
drivers/gpu/drm/amd/display/dc/core/dc_hw_sequencer.c
3053
struct mpc_tree *tree = params->mpc_insert_plane_params.mpc_tree_params;
drivers/gpu/drm/amd/display/dc/core/dc_hw_sequencer.c
3061
mpc->funcs->insert_plane(mpc, tree, blnd_cfg, sm_cfg, insert_above_mpcc,
drivers/gpu/drm/amd/display/dc/inc/hw/mpc.h
428
struct mpc_tree *tree,
drivers/gpu/drm/amd/display/dc/inc/hw/mpc.h
452
struct mpc_tree *tree,
drivers/gpu/drm/amd/display/dc/inc/hw/mpc.h
549
struct mpc_tree *tree,
drivers/gpu/drm/amd/display/dc/inc/hw/mpc.h
573
struct mpc_tree *tree,
drivers/gpu/drm/amd/display/dc/inc/hw/mpc.h
591
struct mpc_tree *tree,
drivers/gpu/drm/amd/display/dc/inc/hw/mpc.h
609
struct mpc_tree *tree,
drivers/gpu/drm/amd/display/dc/inc/hw/mpc.h
658
struct mpc_tree *tree);
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
127
struct mpcc *mpc1_get_mpcc_for_dpp(struct mpc_tree *tree, int dpp_id)
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
129
struct mpcc *tmp_mpcc = tree->opp_list;
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
181
struct mpc_tree *tree,
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
197
struct mpcc *temp_mpcc = tree->opp_list;
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
221
REG_SET(MPCC_OPP_ID[mpcc_id], 0, MPCC_OPP_ID, tree->opp_id);
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
224
REG_SET(MPCC_UPDATE_LOCK_SEL[mpcc_id], 0, MPCC_UPDATE_LOCK_SEL, tree->opp_id);
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
227
if (tree->opp_list == insert_above_mpcc) {
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
229
tree->opp_list = new_mpcc;
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
230
REG_UPDATE(MUX[tree->opp_id], MPC_OUT_MUX, mpcc_id);
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
233
struct mpcc *temp_mpcc = tree->opp_list;
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
273
struct mpc_tree *tree,
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
280
if (tree->opp_list == mpcc_to_remove) {
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
285
tree->opp_list = mpcc_to_remove->mpcc_bot;
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
286
REG_UPDATE(MUX[tree->opp_id], MPC_OUT_MUX, tree->opp_list->mpcc_id);
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
289
tree->opp_list = NULL;
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
290
REG_UPDATE(MUX[tree->opp_id], MPC_OUT_MUX, 0xf);
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
294
struct mpcc *temp_mpcc = tree->opp_list;
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
398
struct mpc_tree *tree)
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
409
REG_GET(MUX[tree->opp_id], MPC_OUT_MUX, &out_mux);
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
420
if ((opp_id == tree->opp_id) && (top_sel != 0xf)) {
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
426
tree->opp_list = mpcc;
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
431
if ((opp_id == tree->opp_id) && (top_sel != 0xf)) {
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.h
143
struct mpc_tree *tree,
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.h
152
struct mpc_tree *tree,
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.h
182
struct mpc_tree *tree);
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.h
189
struct mpc_tree *tree,
drivers/gpu/drm/amd/display/dc/mpc/dcn20/dcn20_mpc.c
525
static struct mpcc *mpc2_get_mpcc_for_dpp(struct mpc_tree *tree, int dpp_id)
drivers/gpu/drm/amd/display/dc/mpc/dcn20/dcn20_mpc.c
527
struct mpcc *tmp_mpcc = tree->opp_list;
drivers/gpu/drm/drm_buddy.c
106
enum drm_buddy_free_tree tree;
drivers/gpu/drm/drm_buddy.c
109
tree = get_block_tree(block);
drivers/gpu/drm/drm_buddy.c
110
root = &mm->free_trees[tree][order];
drivers/gpu/drm/drm_buddy.c
1289
unsigned int tree;
drivers/gpu/drm/drm_buddy.c
1291
for_each_free_tree(tree) {
drivers/gpu/drm/drm_buddy.c
1292
root = &mm->free_trees[tree][order];
drivers/gpu/drm/drm_buddy.c
138
enum drm_buddy_free_tree tree;
drivers/gpu/drm/drm_buddy.c
143
tree = get_block_tree(block);
drivers/gpu/drm/drm_buddy.c
144
rbtree_insert(mm, block, tree);
drivers/gpu/drm/drm_buddy.c
230
unsigned int tree, order;
drivers/gpu/drm/drm_buddy.c
239
for_each_free_tree(tree) {
drivers/gpu/drm/drm_buddy.c
24
#define for_each_free_tree(tree) \
drivers/gpu/drm/drm_buddy.c
241
struct rb_node *iter = rb_last(&mm->free_trees[tree][i]);
drivers/gpu/drm/drm_buddy.c
25
for ((tree) = 0; (tree) < DRM_BUDDY_MAX_FREE_TREES; (tree)++)
drivers/gpu/drm/drm_buddy.c
709
enum drm_buddy_free_tree tree)
drivers/gpu/drm/drm_buddy.c
716
root = &mm->free_trees[tree][i];
drivers/gpu/drm/drm_buddy.c
742
enum drm_buddy_free_tree tree;
drivers/gpu/drm/drm_buddy.c
746
tree = (flags & DRM_BUDDY_CLEAR_ALLOCATION) ?
drivers/gpu/drm/drm_buddy.c
750
block = get_maxblock(mm, order, tree);
drivers/gpu/drm/drm_buddy.c
757
root = &mm->free_trees[tree][tmp];
drivers/gpu/drm/drm_buddy.c
766
tree = (tree == DRM_BUDDY_CLEAR_TREE) ?
drivers/gpu/drm/drm_buddy.c
770
root = &mm->free_trees[tree][tmp];
drivers/gpu/drm/drm_buddy.c
916
unsigned int tree, order;
drivers/gpu/drm/drm_buddy.c
928
for_each_free_tree(tree) {
drivers/gpu/drm/drm_buddy.c
932
root = &mm->free_trees[tree][order];
drivers/gpu/drm/drm_buddy.c
95
enum drm_buddy_free_tree tree)
drivers/gpu/drm/drm_buddy.c
98
&mm->free_trees[tree][drm_buddy_block_order(block)],
drivers/gpu/drm/drm_gpuvm.c
1101
gpuvm->rb.tree = RB_ROOT_CACHED;
drivers/gpu/drm/drm_gpuvm.c
1146
drm_WARN(gpuvm->drm, !RB_EMPTY_ROOT(&gpuvm->rb.tree.rb_root),
drivers/gpu/drm/drm_gpuvm.c
1988
if (drm_gpuva_it_iter_first(&gpuvm->rb.tree,
drivers/gpu/drm/drm_gpuvm.c
1995
drm_gpuva_it_insert(va, &gpuvm->rb.tree);
drivers/gpu/drm/drm_gpuvm.c
2049
drm_gpuva_it_remove(va, &va->vm->rb.tree);
drivers/gpu/drm/drm_gpuvm.c
2193
return drm_gpuva_it_iter_first(&gpuvm->rb.tree, addr, last);
drivers/gpu/drm/drm_gpuvm.c
2244
return drm_gpuva_it_iter_first(&gpuvm->rb.tree, start - 1, start);
drivers/gpu/drm/drm_gpuvm.c
2266
return drm_gpuva_it_iter_first(&gpuvm->rb.tree, end, end + 1);
drivers/gpu/drm/i915/gem/i915_gem_object_types.h
278
struct rb_root tree;
drivers/gpu/drm/i915/i915_active.c
143
ref->cache = fetch_node(ref->tree.rb_node);
drivers/gpu/drm/i915/i915_active.c
148
rb_erase(&ref->cache->node, &ref->tree);
drivers/gpu/drm/i915/i915_active.c
149
root = ref->tree;
drivers/gpu/drm/i915/i915_active.c
152
rb_link_node(&ref->cache->node, NULL, &ref->tree.rb_node);
drivers/gpu/drm/i915/i915_active.c
153
rb_insert_color(&ref->cache->node, &ref->tree);
drivers/gpu/drm/i915/i915_active.c
154
GEM_BUG_ON(ref->tree.rb_node != &ref->cache->node);
drivers/gpu/drm/i915/i915_active.c
271
it = fetch_node(ref->tree.rb_node);
drivers/gpu/drm/i915/i915_active.c
301
p = &ref->tree.rb_node;
drivers/gpu/drm/i915/i915_active.c
328
rb_insert_color(&node->node, &ref->tree);
drivers/gpu/drm/i915/i915_active.c
351
ref->tree = RB_ROOT;
drivers/gpu/drm/i915/i915_active.c
571
rbtree_postorder_for_each_entry_safe(it, n, &ref->tree, node) {
drivers/gpu/drm/i915/i915_active.c
692
rbtree_postorder_for_each_entry_safe(it, n, &ref->tree, node) {
drivers/gpu/drm/i915/i915_active.c
759
if (RB_EMPTY_ROOT(&ref->tree))
drivers/gpu/drm/i915/i915_active.c
777
p = ref->tree.rb_node;
drivers/gpu/drm/i915/i915_active.c
830
rb_erase(p, &ref->tree); /* Hide from waits and sibling allocations */
drivers/gpu/drm/i915/i915_active.c
941
p = &ref->tree.rb_node;
drivers/gpu/drm/i915/i915_active.c
954
rb_insert_color(&node->node, &ref->tree);
drivers/gpu/drm/i915/i915_active_types.h
31
struct rb_root tree;
drivers/gpu/drm/i915/i915_gem.c
896
rb_erase(&vma->obj_node, &obj->vma.tree);
drivers/gpu/drm/i915/i915_vma.c
1802
rb_erase(&vma->obj_node, &obj->vma.tree);
drivers/gpu/drm/i915/i915_vma.c
238
p = &obj->vma.tree.rb_node;
drivers/gpu/drm/i915/i915_vma.c
259
rb_insert_color(&vma->obj_node, &obj->vma.tree);
drivers/gpu/drm/i915/i915_vma.c
293
rb = obj->vma.tree.rb_node;
drivers/gpu/drm/i915/selftests/i915_active.c
291
rbtree_postorder_for_each_entry_safe(it, n, &ref->tree, node) {
drivers/gpu/drm/i915/selftests/i915_active.c
342
rbtree_postorder_for_each_entry_safe(it, n, &ref->tree, node)
drivers/gpu/drm/nouveau/include/nvkm/core/object.h
18
struct list_head tree;
drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h
9
struct rb_node tree;
drivers/gpu/drm/nouveau/nvkm/core/ioctl.c
137
list_add_tail(&object->head, &parent->tree);
drivers/gpu/drm/nouveau/nvkm/core/object.c
166
list_for_each_entry_reverse(child, &object->tree, head) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
193
list_for_each_entry_continue_reverse(child, &object->tree, head) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
215
list_for_each_entry(child, &object->tree, head) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
226
list_for_each_entry_continue_reverse(child, &object->tree, head)
drivers/gpu/drm/nouveau/nvkm/core/object.c
244
list_for_each_entry_safe(child, ctemp, &object->tree, head) {
drivers/gpu/drm/nouveau/nvkm/core/object.c
282
INIT_LIST_HEAD(&object->tree);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1064
struct nvkm_vma *vma = rb_entry(node, typeof(*vma), tree);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1778
struct nvkm_vma *this = rb_entry(temp, typeof(*this), tree);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1794
struct nvkm_vma *this = rb_entry(node, typeof(*this), tree);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
862
rb_erase(&vma->tree, &vmm->free);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
880
struct nvkm_vma *this = rb_entry(*ptr, typeof(*this), tree);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
897
rb_link_node(&vma->tree, parent, ptr);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
898
rb_insert_color(&vma->tree, &vmm->free);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
904
rb_erase(&vma->tree, &vmm->root);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
922
struct nvkm_vma *this = rb_entry(*ptr, typeof(*this), tree);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
933
rb_link_node(&vma->tree, parent, ptr);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
934
rb_insert_color(&vma->tree, &vmm->root);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
942
struct nvkm_vma *vma = rb_entry(node, typeof(*vma), tree);
drivers/gpu/drm/xe/xe_range_fence.c
103
retry = __xe_range_fence_tree_cleanup(tree);
drivers/gpu/drm/xe/xe_range_fence.c
110
void xe_range_fence_tree_init(struct xe_range_fence_tree *tree)
drivers/gpu/drm/xe/xe_range_fence.c
112
memset(tree, 0, sizeof(*tree));
drivers/gpu/drm/xe/xe_range_fence.c
119
void xe_range_fence_tree_fini(struct xe_range_fence_tree *tree)
drivers/gpu/drm/xe/xe_range_fence.c
121
xe_range_fence_tree_remove_all(tree);
drivers/gpu/drm/xe/xe_range_fence.c
122
XE_WARN_ON(!RB_EMPTY_ROOT(&tree->root.rb_root));
drivers/gpu/drm/xe/xe_range_fence.c
134
xe_range_fence_tree_first(struct xe_range_fence_tree *tree, u64 start,
drivers/gpu/drm/xe/xe_range_fence.c
137
return xe_range_fence_tree_iter_first(&tree->root, start, last);
drivers/gpu/drm/xe/xe_range_fence.c
24
struct xe_range_fence_tree *tree = rfence->tree;
drivers/gpu/drm/xe/xe_range_fence.c
26
llist_add(&rfence->link, &tree->list);
drivers/gpu/drm/xe/xe_range_fence.c
29
static bool __xe_range_fence_tree_cleanup(struct xe_range_fence_tree *tree)
drivers/gpu/drm/xe/xe_range_fence.c
31
struct llist_node *node = llist_del_all(&tree->list);
drivers/gpu/drm/xe/xe_range_fence.c
35
xe_range_fence_tree_remove(rfence, &tree->root);
drivers/gpu/drm/xe/xe_range_fence.c
54
int xe_range_fence_insert(struct xe_range_fence_tree *tree,
drivers/gpu/drm/xe/xe_range_fence.c
61
__xe_range_fence_tree_cleanup(tree);
drivers/gpu/drm/xe/xe_range_fence.c
69
rfence->tree = tree;
drivers/gpu/drm/xe/xe_range_fence.c
78
xe_range_fence_tree_insert(rfence, &tree->root);
drivers/gpu/drm/xe/xe_range_fence.c
89
static void xe_range_fence_tree_remove_all(struct xe_range_fence_tree *tree)
drivers/gpu/drm/xe/xe_range_fence.c
94
rfence = xe_range_fence_tree_iter_first(&tree->root, 0, U64_MAX);
drivers/gpu/drm/xe/xe_range_fence.c
98
llist_add(&rfence->link, &tree->list);
drivers/gpu/drm/xe/xe_range_fence.h
37
struct xe_range_fence_tree *tree;
drivers/gpu/drm/xe/xe_range_fence.h
59
xe_range_fence_tree_first(struct xe_range_fence_tree *tree, u64 start,
drivers/gpu/drm/xe/xe_range_fence.h
65
void xe_range_fence_tree_init(struct xe_range_fence_tree *tree);
drivers/gpu/drm/xe/xe_range_fence.h
67
void xe_range_fence_tree_fini(struct xe_range_fence_tree *tree);
drivers/gpu/drm/xe/xe_range_fence.h
69
int xe_range_fence_insert(struct xe_range_fence_tree *tree,
drivers/infiniband/hw/mlx5/cq.c
498
mqp = radix_tree_lookup(&dev->qp_table.tree, qpn);
drivers/infiniband/hw/mlx5/qp.h
18
struct radix_tree_root tree;
drivers/infiniband/hw/mlx5/qpc.c
169
err = radix_tree_insert(&table->tree,
drivers/infiniband/hw/mlx5/qpc.c
202
radix_tree_delete(&table->tree,
drivers/infiniband/hw/mlx5/qpc.c
23
common = radix_tree_lookup(&table->tree, rsn);
drivers/infiniband/hw/mlx5/qpc.c
521
INIT_RADIX_TREE(&table->tree, GFP_ATOMIC);
drivers/infiniband/sw/rxe/rxe_mcast.c
101
rb_insert_color(&mcg->node, tree);
drivers/infiniband/sw/rxe/rxe_mcast.c
126
struct rb_root *tree = &rxe->mcg_tree;
drivers/infiniband/sw/rxe/rxe_mcast.c
131
node = tree->rb_node;
drivers/infiniband/sw/rxe/rxe_mcast.c
83
struct rb_root *tree = &mcg->rxe->mcg_tree;
drivers/infiniband/sw/rxe/rxe_mcast.c
84
struct rb_node **link = &tree->rb_node;
drivers/md/bcache/bset.c
1006
f = &t->tree[inorder_prev(j, t->size)];
drivers/md/bcache/bset.c
1383
if (t->tree[j].exponent == 127)
drivers/md/bcache/bset.c
303
kfree(t->tree);
drivers/md/bcache/bset.c
305
free_pages((unsigned long) t->tree,
drivers/md/bcache/bset.c
311
t->tree = NULL;
drivers/md/bcache/bset.c
329
t->tree = bset_tree_bytes(b) < PAGE_SIZE
drivers/md/bcache/bset.c
332
if (!t->tree)
drivers/md/bcache/bset.c
546
return cacheline_to_bkey(t, to_inorder(j, t), t->tree[j].m);
drivers/md/bcache/bset.c
594
struct bkey_float *f = &t->tree[j];
drivers/md/bcache/bset.c
643
t->tree = t[-1].tree + j;
drivers/md/bcache/bset.c
660
if (t->tree != b->set->tree + btree_keys_cachelines(b)) {
drivers/md/bcache/bset.c
702
b->set->tree + btree_keys_cachelines(b) - t->tree);
drivers/md/bcache/bset.c
721
t->tree[j].m = bkey_to_cacheline_offset(t, cacheline++, k);
drivers/md/bcache/bset.c
820
if (t->size == b->set->tree + btree_keys_cachelines(b) - t->tree)
drivers/md/bcache/bset.c
970
prefetch(&t->tree[p]);
drivers/md/bcache/bset.c
973
f = &t->tree[j];
drivers/md/bcache/bset.c
998
f = &t->tree[inorder_next(j, t->size)];
drivers/md/bcache/bset.h
173
struct bkey_float *tree;
drivers/md/bcache/btree.c
1052
prefetch(b->keys.set[i].tree);
drivers/md/dm-bufio.c
1752
struct buffer_tree *tree, struct dm_buffer *b)
drivers/md/dm-bufio.c
1757
cache_read_lock(&c->cache, tree);
drivers/md/dm-bufio.c
1760
cache_read_unlock(&c->cache, tree);
drivers/md/dm-bufio.c
1775
struct buffer_tree *tree, sector_t block,
drivers/md/dm-bufio.c
1795
b = cache_get(&c->cache, tree, block);
drivers/md/dm-bufio.c
1823
cache_insert(&c->cache, tree, b);
drivers/md/dm-bufio.c
1829
cache_put_and_wake(c, tree, b);
drivers/md/dm-bufio.c
1841
cache_put_and_wake(c, tree, b);
drivers/md/dm-bufio.c
1875
struct buffer_tree *tree;
drivers/md/dm-bufio.c
1887
tree = cache_get_tree(&c->cache, block);
drivers/md/dm-bufio.c
1888
b = cache_get(&c->cache, tree, block);
drivers/md/dm-bufio.c
1891
cache_put_and_wake(c, tree, b);
drivers/md/dm-bufio.c
1903
cache_put_and_wake(c, tree, b);
drivers/md/dm-bufio.c
1913
b = __bufio_new(c, tree, block, nf, &need_submit, &write_list);
drivers/md/dm-bufio.c
2000
struct buffer_tree *tree;
drivers/md/dm-bufio.c
2004
tree = cache_get_tree(&c->cache, block);
drivers/md/dm-bufio.c
2005
b = cache_get(&c->cache, tree, block);
drivers/md/dm-bufio.c
2008
cache_put_and_wake(c, tree, b);
drivers/md/dm-bufio.c
2013
b = __bufio_new(c, tree, block, NF_PREFETCH, &need_submit,
drivers/md/dm-bufio.c
2058
struct buffer_tree *tree = cache_get_tree(&c->cache, b->block);
drivers/md/dm-bufio.c
2072
if (cache_remove(&c->cache, tree, b)) {
drivers/md/dm-bufio.c
2081
cache_put_and_wake(c, tree, b);
drivers/md/dm-bufio.c
2166
struct buffer_tree *tree;
drivers/md/dm-bufio.c
2180
tree = cache_get_tree(&c->cache, b->block);
drivers/md/dm-bufio.c
2183
cache_mark(&c->cache, tree, b, LIST_CLEAN);
drivers/md/dm-bufio.c
2185
cache_put_and_wake(c, tree, b);
drivers/md/dm-bufio.c
2253
struct buffer_tree *tree = cache_get_tree(&c->cache, block);
drivers/md/dm-bufio.c
2256
b = cache_get(&c->cache, tree, block);
drivers/md/dm-bufio.c
2259
if (cache_remove(&c->cache, tree, b))
drivers/md/dm-bufio.c
2262
cache_put_and_wake(c, tree, b);
drivers/md/dm-bufio.c
2264
cache_put_and_wake(c, tree, b);
drivers/md/dm-bufio.c
413
struct buffer_tree *tree)
drivers/md/dm-bufio.c
416
read_lock_bh(&tree->u.spinlock);
drivers/md/dm-bufio.c
418
down_read(&tree->u.lock);
drivers/md/dm-bufio.c
423
struct buffer_tree *tree)
drivers/md/dm-bufio.c
426
read_unlock_bh(&tree->u.spinlock);
drivers/md/dm-bufio.c
428
up_read(&tree->u.lock);
drivers/md/dm-bufio.c
433
struct buffer_tree *tree)
drivers/md/dm-bufio.c
436
write_lock_bh(&tree->u.spinlock);
drivers/md/dm-bufio.c
438
down_write(&tree->u.lock);
drivers/md/dm-bufio.c
443
struct buffer_tree *tree)
drivers/md/dm-bufio.c
446
write_unlock_bh(&tree->u.spinlock);
drivers/md/dm-bufio.c
448
up_write(&tree->u.lock);
drivers/md/dm-bufio.c
621
struct buffer_tree *tree, sector_t block)
drivers/md/dm-bufio.c
626
cache_read_lock(bc, tree);
drivers/md/dm-bufio.c
627
b = __cache_get(&tree->root, block);
drivers/md/dm-bufio.c
632
cache_read_unlock(bc, tree);
drivers/md/dm-bufio.c
853
r = __cache_insert(&tree->root, b);
drivers/md/dm-bufio.c
881
rb_erase(&b->node, &tree->root);
drivers/md/dm-dust.c
240
static bool __dust_clear_badblocks(struct rb_root *tree,
drivers/md/dm-dust.c
245
nnode = rb_first(tree);
drivers/md/dm-dust.c
254
rb_erase(node, tree);
drivers/md/dm-dust.c
259
BUG_ON(tree->rb_node != NULL);
drivers/md/dm-vdo/block-map.c
1383
struct block_map_tree *tree = &forest->trees[root_index];
drivers/md/dm-vdo/block-map.c
1385
return &(tree->segments[segment].levels[height - 1][page_index - offset]);
drivers/md/dm-vdo/block-map.c
2400
struct block_map_tree *tree = &(forest->trees[root]);
drivers/md/dm-vdo/block-map.c
2405
"tree root segments", &tree->segments);
drivers/md/dm-vdo/block-map.c
2410
memcpy(tree->segments, old_forest->trees[root].segments,
drivers/md/dm-vdo/block-map.c
2414
segment = &(tree->segments[index]);
drivers/md/dm-vdo/block-map.c
2556
&(cursor->tree->segments[0].levels[height][level->page_index]);
drivers/md/dm-vdo/block-map.c
2586
&(cursor->tree->segments[0].levels[height][level->page_index]);
drivers/md/dm-vdo/block-map.c
2726
.tree = &map->forest->trees[root],
drivers/md/dm-vdo/block-map.c
89
struct block_map_tree *tree;
drivers/md/dm-writecache.c
1021
wc->tree = RB_ROOT;
drivers/md/dm-writecache.c
118
struct rb_root tree;
drivers/md/dm-writecache.c
2009
e = container_of(rb_first(&wc->tree), struct wc_entry, rb_node);
drivers/md/dm-writecache.c
610
struct rb_node *node = wc->tree.rb_node;
drivers/md/dm-writecache.c
656
struct rb_node **node = &wc->tree.rb_node, *parent = NULL;
drivers/md/dm-writecache.c
667
rb_insert_color(&ins->rb_node, &wc->tree);
drivers/md/dm-writecache.c
675
rb_erase(&e->rb_node, &wc->tree);
drivers/md/dm-writecache.c
774
memset(&wc->tree, -1, sizeof(wc->tree));
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
338
memcpy(rkprobs->tree, seg->tree_probs, sizeof(rkprobs->tree));
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
75
u8 tree[7];
drivers/net/ethernet/mellanox/mlx4/cq.c
109
cq = radix_tree_lookup(&mlx4_priv(dev)->cq_table.tree,
drivers/net/ethernet/mellanox/mlx4/cq.c
132
cq = radix_tree_lookup(&cq_table->tree, cqn & (dev->caps.num_cqs - 1));
drivers/net/ethernet/mellanox/mlx4/cq.c
365
err = radix_tree_insert(&cq_table->tree, cq->cqn, cq);
drivers/net/ethernet/mellanox/mlx4/cq.c
426
radix_tree_delete(&cq_table->tree, cq->cqn);
drivers/net/ethernet/mellanox/mlx4/cq.c
447
radix_tree_delete(&cq_table->tree, cq->cqn);
drivers/net/ethernet/mellanox/mlx4/cq.c
468
INIT_RADIX_TREE(&cq_table->tree, GFP_ATOMIC);
drivers/net/ethernet/mellanox/mlx4/mlx4.h
685
struct radix_tree_root tree;
drivers/net/ethernet/mellanox/mlx4/mlx4.h
706
struct radix_tree_root tree;
drivers/net/ethernet/mellanox/mlx4/srq.c
176
err = radix_tree_insert(&srq_table->tree, srq->srqn, srq);
drivers/net/ethernet/mellanox/mlx4/srq.c
213
radix_tree_delete(&srq_table->tree, srq->srqn);
drivers/net/ethernet/mellanox/mlx4/srq.c
232
radix_tree_delete(&srq_table->tree, srq->srqn);
drivers/net/ethernet/mellanox/mlx4/srq.c
277
INIT_RADIX_TREE(&srq_table->tree, GFP_ATOMIC);
drivers/net/ethernet/mellanox/mlx4/srq.c
298
srq = radix_tree_lookup(&srq_table->tree,
drivers/net/ethernet/mellanox/mlx4/srq.c
49
srq = radix_tree_lookup(&srq_table->tree, srqn & (dev->caps.num_srqs - 1));
drivers/net/ethernet/mellanox/mlx5/core/eq.c
273
INIT_RADIX_TREE(&cq_table->tree, GFP_ATOMIC);
drivers/net/ethernet/mellanox/mlx5/core/eq.c
402
err = radix_tree_insert(&table->tree, cq->cqn, cq);
drivers/net/ethernet/mellanox/mlx5/core/eq.c
414
tmp = radix_tree_delete(&table->tree, cq->cqn);
drivers/net/ethernet/mellanox/mlx5/core/eq.c
98
cq = radix_tree_lookup(&table->tree, cqn);
drivers/net/ethernet/mellanox/mlx5/core/lib/eq.h
21
struct radix_tree_root tree;
drivers/of/of_private.h
97
int of_resolve_phandles(struct device_node *tree);
drivers/of/overlay.c
1082
static int find_node(const struct device_node *tree, struct device_node *np)
drivers/of/overlay.c
1084
if (tree == np)
drivers/of/overlay.c
1087
for_each_child_of_node_scoped(tree, child) {
drivers/sh/intc/core.c
214
INIT_RADIX_TREE(&d->tree, GFP_ATOMIC);
drivers/sh/intc/core.c
84
radix_tree_insert(&d->tree, enum_id, intc_irq_xlate_get(irq));
drivers/sh/intc/internals.h
57
struct radix_tree_root tree;
drivers/sh/intc/virq.c
154
mapped = radix_tree_lookup(&d->tree, subgroup->parent_id);
drivers/sh/intc/virq.c
179
err = radix_tree_insert(&d->tree, entry->enum_id, entry);
drivers/sh/intc/virq.c
183
radix_tree_tag_set(&d->tree, entry->enum_id,
drivers/sh/intc/virq.c
211
nr_found = radix_tree_gang_lookup_tag_slot(&d->tree,
drivers/sh/intc/virq.c
253
radix_tree_tag_clear(&d->tree, entry->enum_id,
drivers/sh/intc/virq.c
255
radix_tree_replace_slot(&d->tree, (void **)entries[i],
drivers/sh/intc/virq.c
267
if (radix_tree_tagged(&d->tree, INTC_TAG_VIRQ_NEEDS_ALLOC))
drivers/sh/intc/virq.c
68
tagged = radix_tree_tag_get(&d->tree, enum_id,
drivers/sh/intc/virq.c
73
ptr = radix_tree_lookup(&d->tree, enum_id);
fs/btrfs/backref.c
3127
void btrfs_backref_drop_node(struct btrfs_backref_cache *tree,
fs/btrfs/backref.c
3136
rb_erase(&node->rb_node, &tree->rb_root);
fs/btrfs/backref.c
3137
btrfs_backref_free_node(tree, node);
fs/btrfs/backref.c
434
struct preftree *tree = &preftrees->indirect;
fs/btrfs/backref.c
437
tree = &preftrees->indirect_missing_keys;
fs/btrfs/backref.c
438
return add_prelim_ref(fs_info, tree, root_id, key, level, 0,
fs/btrfs/backref.c
840
struct preftree *tree = &preftrees->indirect_missing_keys;
fs/btrfs/backref.c
843
while ((node = rb_first_cached(&tree->root))) {
fs/btrfs/backref.c
847
rb_erase_cached(node, &tree->root);
fs/btrfs/backref.h
444
void btrfs_backref_drop_node(struct btrfs_backref_cache *tree,
fs/btrfs/compression.c
384
struct extent_io_tree *tree;
fs/btrfs/compression.c
388
tree = &BTRFS_I(inode)->io_tree;
fs/btrfs/compression.c
463
btrfs_lock_extent(tree, cur, page_end, NULL);
fs/btrfs/compression.c
478
btrfs_unlock_extent(tree, cur, page_end, NULL);
fs/btrfs/compression.c
485
btrfs_unlock_extent(tree, cur, page_end, NULL);
fs/btrfs/disk-io.c
491
struct extent_io_tree *tree;
fs/btrfs/disk-io.c
493
tree = &folio_to_inode(folio)->io_tree;
fs/btrfs/disk-io.c
494
extent_invalidate_folio(tree, folio, offset);
fs/btrfs/extent-io-tree.c
100
struct extent_io_tree *tree, unsigned int owner)
fs/btrfs/extent-io-tree.c
102
tree->state = RB_ROOT;
fs/btrfs/extent-io-tree.c
1025
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
103
spin_lock_init(&tree->lock);
fs/btrfs/extent-io-tree.c
104
tree->fs_info = fs_info;
fs/btrfs/extent-io-tree.c
1043
static int set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
fs/btrfs/extent-io-tree.c
105
tree->owner = owner;
fs/btrfs/extent-io-tree.c
1060
btrfs_debug_check_extent_io_range(tree, start, end);
fs/btrfs/extent-io-tree.c
1061
trace_btrfs_set_extent_bit(tree, start, end - start + 1, bits);
fs/btrfs/extent-io-tree.c
1082
spin_lock(&tree->lock);
fs/btrfs/extent-io-tree.c
1093
state = tree_search_for_insert(tree, start, &p, &parent);
fs/btrfs/extent-io-tree.c
1100
insert_state_fast(tree, prealloc, p, parent, bits, changeset);
fs/btrfs/extent-io-tree.c
1123
set_state_bits(tree, state, bits, changeset);
fs/btrfs/extent-io-tree.c
1125
merge_state(tree, state);
fs/btrfs/extent-io-tree.c
115
void btrfs_extent_io_tree_release(struct extent_io_tree *tree)
fs/btrfs/extent-io-tree.c
1171
ret = split_state(tree, state, prealloc, start);
fs/btrfs/extent-io-tree.c
1173
extent_io_tree_panic(tree, state, "split", ret);
fs/btrfs/extent-io-tree.c
1179
set_state_bits(tree, state, bits, changeset);
fs/btrfs/extent-io-tree.c
1181
merge_state(tree, state);
fs/btrfs/extent-io-tree.c
121
spin_lock(&tree->lock);
fs/btrfs/extent-io-tree.c
1215
inserted_state = insert_state(tree, prealloc, bits, changeset);
fs/btrfs/extent-io-tree.c
1218
extent_io_tree_panic(tree, prealloc, "insert", ret);
fs/btrfs/extent-io-tree.c
122
root = tree->state;
fs/btrfs/extent-io-tree.c
123
tree->state = RB_ROOT;
fs/btrfs/extent-io-tree.c
1261
ret = split_state(tree, state, prealloc, end + 1);
fs/btrfs/extent-io-tree.c
1263
extent_io_tree_panic(tree, state, "split", ret);
fs/btrfs/extent-io-tree.c
1268
set_state_bits(tree, prealloc, bits, changeset);
fs/btrfs/extent-io-tree.c
1270
merge_state(tree, prealloc);
fs/btrfs/extent-io-tree.c
1278
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
1284
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
1291
int btrfs_set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
fs/btrfs/extent-io-tree.c
1294
return set_extent_bit(tree, start, end, bits, NULL, NULL, cached_state, NULL);
fs/btrfs/extent-io-tree.c
1315
int btrfs_convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
fs/btrfs/extent-io-tree.c
1328
btrfs_debug_check_extent_io_range(tree, start, end);
fs/btrfs/extent-io-tree.c
1329
trace_btrfs_convert_extent_bit(tree, start, end - start + 1, bits,
fs/btrfs/extent-io-tree.c
1346
spin_lock(&tree->lock);
fs/btrfs/extent-io-tree.c
135
cond_resched_lock(&tree->lock);
fs/btrfs/extent-io-tree.c
1358
state = tree_search_for_insert(tree, start, &p, &parent);
fs/btrfs/extent-io-tree.c
1367
insert_state_fast(tree, prealloc, p, parent, bits, NULL);
fs/btrfs/extent-io-tree.c
1383
set_state_bits(tree, state, bits, NULL);
fs/btrfs/extent-io-tree.c
1385
state = clear_state_bit(tree, state, clear_bits, 0, end, NULL);
fs/btrfs/extent-io-tree.c
141
ASSERT(RB_EMPTY_ROOT(&tree->state));
fs/btrfs/extent-io-tree.c
1415
ret = split_state(tree, state, prealloc, start);
fs/btrfs/extent-io-tree.c
1418
extent_io_tree_panic(tree, state, "split", ret);
fs/btrfs/extent-io-tree.c
142
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
1422
set_state_bits(tree, state, bits, NULL);
fs/btrfs/extent-io-tree.c
1424
state = clear_state_bit(tree, state, clear_bits, 0, end, NULL);
fs/btrfs/extent-io-tree.c
1459
inserted_state = insert_state(tree, prealloc, bits, NULL);
fs/btrfs/extent-io-tree.c
1462
extent_io_tree_panic(tree, prealloc, "insert", ret);
fs/btrfs/extent-io-tree.c
1500
ret = split_state(tree, state, prealloc, end + 1);
fs/btrfs/extent-io-tree.c
1502
extent_io_tree_panic(tree, state, "split", ret);
fs/btrfs/extent-io-tree.c
1507
set_state_bits(tree, prealloc, bits, NULL);
fs/btrfs/extent-io-tree.c
1509
clear_state_bit(tree, prealloc, clear_bits, 0, end, NULL);
fs/btrfs/extent-io-tree.c
1517
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
1523
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
1544
void btrfs_find_first_clear_extent_bit(struct extent_io_tree *tree, u64 start,
fs/btrfs/extent-io-tree.c
1550
spin_lock(&tree->lock);
fs/btrfs/extent-io-tree.c
1554
state = tree_search_prev_next(tree, start, &prev, &next);
fs/btrfs/extent-io-tree.c
1634
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
1662
u64 btrfs_count_range_bits(struct extent_io_tree *tree,
fs/btrfs/extent-io-tree.c
1677
spin_lock(&tree->lock);
fs/btrfs/extent-io-tree.c
1712
state = tree_search(tree, cur_start);
fs/btrfs/extent-io-tree.c
1742
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
1750
bool btrfs_test_range_bit_exists(struct extent_io_tree *tree, u64 start, u64 end, u32 bit)
fs/btrfs/extent-io-tree.c
1757
spin_lock(&tree->lock);
fs/btrfs/extent-io-tree.c
1758
state = tree_search(tree, start);
fs/btrfs/extent-io-tree.c
1772
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
1776
void btrfs_get_range_bits(struct extent_io_tree *tree, u64 start, u64 end, u32 *bits,
fs/btrfs/extent-io-tree.c
1790
spin_lock(&tree->lock);
fs/btrfs/extent-io-tree.c
1791
state = tree_search(tree, start);
fs/btrfs/extent-io-tree.c
1807
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
1813
bool btrfs_test_range_bit(struct extent_io_tree *tree, u64 start, u64 end, u32 bit,
fs/btrfs/extent-io-tree.c
1822
spin_lock(&tree->lock);
fs/btrfs/extent-io-tree.c
1827
state = tree_search(tree, start);
fs/btrfs/extent-io-tree.c
1850
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
1855
int btrfs_set_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
fs/btrfs/extent-io-tree.c
1865
return set_extent_bit(tree, start, end, bits, NULL, NULL, NULL, changeset);
fs/btrfs/extent-io-tree.c
1868
int btrfs_clear_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
fs/btrfs/extent-io-tree.c
1877
return btrfs_clear_extent_bit_changeset(tree, start, end, bits, NULL, changeset);
fs/btrfs/extent-io-tree.c
1880
bool btrfs_try_lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
fs/btrfs/extent-io-tree.c
1886
ret = set_extent_bit(tree, start, end, bits, &failed_start, NULL, cached, NULL);
fs/btrfs/extent-io-tree.c
1889
btrfs_clear_extent_bit(tree, start, failed_start - 1,
fs/btrfs/extent-io-tree.c
1900
int btrfs_lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, u32 bits,
fs/btrfs/extent-io-tree.c
1907
ret = set_extent_bit(tree, start, end, bits, &failed_start,
fs/btrfs/extent-io-tree.c
1911
btrfs_clear_extent_bit(tree, start, failed_start - 1,
fs/btrfs/extent-io-tree.c
1914
wait_extent_bit(tree, failed_start, end, bits, &failed_state);
fs/btrfs/extent-io-tree.c
1915
ret = set_extent_bit(tree, start, end, bits, &failed_start,
fs/btrfs/extent-io-tree.c
1926
struct extent_state *btrfs_next_extent_state(struct extent_io_tree *tree,
fs/btrfs/extent-io-tree.c
1931
spin_lock(&tree->lock);
fs/btrfs/extent-io-tree.c
1936
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
234
static inline struct extent_state *tree_search_for_insert(struct extent_io_tree *tree,
fs/btrfs/extent-io-tree.c
239
struct rb_root *root = &tree->state;
fs/btrfs/extent-io-tree.c
284
static struct extent_state *tree_search_prev_next(struct extent_io_tree *tree,
fs/btrfs/extent-io-tree.c
289
struct rb_root *root = &tree->state;
fs/btrfs/extent-io-tree.c
324
static inline struct extent_state *tree_search(struct extent_io_tree *tree, u64 offset)
fs/btrfs/extent-io-tree.c
326
return tree_search_for_insert(tree, offset, NULL, NULL);
fs/btrfs/extent-io-tree.c
329
static void __cold extent_io_tree_panic(const struct extent_io_tree *tree,
fs/btrfs/extent-io-tree.c
334
btrfs_panic(btrfs_extent_io_tree_to_fs_info(tree), err,
fs/btrfs/extent-io-tree.c
339
static void merge_prev_state(struct extent_io_tree *tree, struct extent_state *state)
fs/btrfs/extent-io-tree.c
345
if (tree->owner == IO_TREE_INODE_IO)
fs/btrfs/extent-io-tree.c
346
btrfs_merge_delalloc_extent(tree->inode, state, prev);
fs/btrfs/extent-io-tree.c
348
rb_erase(&prev->rb_node, &tree->state);
fs/btrfs/extent-io-tree.c
354
static void merge_next_state(struct extent_io_tree *tree, struct extent_state *state)
fs/btrfs/extent-io-tree.c
360
if (tree->owner == IO_TREE_INODE_IO)
fs/btrfs/extent-io-tree.c
361
btrfs_merge_delalloc_extent(tree->inode, state, next);
fs/btrfs/extent-io-tree.c
363
rb_erase(&next->rb_node, &tree->state);
fs/btrfs/extent-io-tree.c
378
static void merge_state(struct extent_io_tree *tree, struct extent_state *state)
fs/btrfs/extent-io-tree.c
383
merge_prev_state(tree, state);
fs/btrfs/extent-io-tree.c
384
merge_next_state(tree, state);
fs/btrfs/extent-io-tree.c
387
static void set_state_bits(struct extent_io_tree *tree,
fs/btrfs/extent-io-tree.c
394
if (tree->owner == IO_TREE_INODE_IO)
fs/btrfs/extent-io-tree.c
395
btrfs_set_delalloc_extent(tree->inode, state, bits);
fs/btrfs/extent-io-tree.c
417
static struct extent_state *insert_state(struct extent_io_tree *tree,
fs/btrfs/extent-io-tree.c
428
set_state_bits(tree, state, bits, changeset);
fs/btrfs/extent-io-tree.c
430
node = &tree->state.rb_node;
fs/btrfs/extent-io-tree.c
440
if (tree->owner == IO_TREE_INODE_IO)
fs/btrfs/extent-io-tree.c
441
btrfs_merge_delalloc_extent(tree->inode,
fs/btrfs/extent-io-tree.c
444
merge_prev_state(tree, entry);
fs/btrfs/extent-io-tree.c
452
if (tree->owner == IO_TREE_INODE_IO)
fs/btrfs/extent-io-tree.c
453
btrfs_merge_delalloc_extent(tree->inode,
fs/btrfs/extent-io-tree.c
456
merge_next_state(tree, entry);
fs/btrfs/extent-io-tree.c
467
rb_insert_color(&state->rb_node, &tree->state);
fs/btrfs/extent-io-tree.c
475
static void insert_state_fast(struct extent_io_tree *tree,
fs/btrfs/extent-io-tree.c
480
set_state_bits(tree, state, bits, changeset);
fs/btrfs/extent-io-tree.c
482
rb_insert_color(&state->rb_node, &tree->state);
fs/btrfs/extent-io-tree.c
483
merge_state(tree, state);
fs/btrfs/extent-io-tree.c
500
static int split_state(struct extent_io_tree *tree, struct extent_state *orig,
fs/btrfs/extent-io-tree.c
506
if (tree->owner == IO_TREE_INODE_IO)
fs/btrfs/extent-io-tree.c
507
btrfs_split_delalloc_extent(tree->inode, orig, split);
fs/btrfs/extent-io-tree.c
533
rb_insert_color(&prealloc->rb_node, &tree->state);
fs/btrfs/extent-io-tree.c
557
static struct extent_state *clear_state_bit(struct extent_io_tree *tree,
fs/btrfs/extent-io-tree.c
566
if (tree->owner == IO_TREE_INODE_IO)
fs/btrfs/extent-io-tree.c
567
btrfs_clear_delalloc_extent(tree->inode, state, bits);
fs/btrfs/extent-io-tree.c
57
#define btrfs_debug_check_extent_io_range(tree, start, end) \
fs/btrfs/extent-io-tree.c
577
rb_erase(&state->rb_node, &tree->state);
fs/btrfs/extent-io-tree.c
58
__btrfs_debug_check_extent_io_range(__func__, (tree), (start), (end))
fs/btrfs/extent-io-tree.c
584
merge_state(tree, state);
fs/btrfs/extent-io-tree.c
60
struct extent_io_tree *tree,
fs/btrfs/extent-io-tree.c
609
int btrfs_clear_extent_bit_changeset(struct extent_io_tree *tree, u64 start, u64 end,
fs/btrfs/extent-io-tree.c
624
btrfs_debug_check_extent_io_range(tree, start, end);
fs/btrfs/extent-io-tree.c
625
trace_btrfs_clear_extent_bit(tree, start, end - start + 1, bits);
fs/btrfs/extent-io-tree.c
63
const struct btrfs_inode *inode = tree->inode;
fs/btrfs/extent-io-tree.c
647
spin_lock(&tree->lock);
fs/btrfs/extent-io-tree.c
66
if (tree->owner != IO_TREE_INODE_IO)
fs/btrfs/extent-io-tree.c
668
state = tree_search(tree, start);
fs/btrfs/extent-io-tree.c
702
ret = split_state(tree, state, prealloc, start);
fs/btrfs/extent-io-tree.c
705
extent_io_tree_panic(tree, state, "split", ret);
fs/btrfs/extent-io-tree.c
709
state = clear_state_bit(tree, state, bits, wake, end,
fs/btrfs/extent-io-tree.c
730
ret = split_state(tree, state, prealloc, end + 1);
fs/btrfs/extent-io-tree.c
732
extent_io_tree_panic(tree, state, "split", ret);
fs/btrfs/extent-io-tree.c
740
clear_state_bit(tree, prealloc, bits, wake, end, changeset);
fs/btrfs/extent-io-tree.c
746
state = clear_state_bit(tree, state, bits, wake, end, changeset);
fs/btrfs/extent-io-tree.c
755
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
761
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
773
static void wait_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
fs/btrfs/extent-io-tree.c
778
btrfs_debug_check_extent_io_range(tree, start, end);
fs/btrfs/extent-io-tree.c
780
spin_lock(&tree->lock);
fs/btrfs/extent-io-tree.c
797
state = tree_search(tree, start);
fs/btrfs/extent-io-tree.c
810
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
812
spin_lock(&tree->lock);
fs/btrfs/extent-io-tree.c
822
if (!cond_resched_lock(&tree->lock)) {
fs/btrfs/extent-io-tree.c
834
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
84
const struct btrfs_inode *btrfs_extent_io_tree_to_inode(const struct extent_io_tree *tree)
fs/btrfs/extent-io-tree.c
86
if (tree->owner == IO_TREE_INODE_IO)
fs/btrfs/extent-io-tree.c
860
static struct extent_state *find_first_extent_bit_state(struct extent_io_tree *tree,
fs/btrfs/extent-io-tree.c
869
state = tree_search(tree, start);
fs/btrfs/extent-io-tree.c
87
return tree->inode;
fs/btrfs/extent-io-tree.c
886
bool btrfs_find_first_extent_bit(struct extent_io_tree *tree, u64 start,
fs/btrfs/extent-io-tree.c
893
spin_lock(&tree->lock);
fs/btrfs/extent-io-tree.c
918
state = find_first_extent_bit_state(tree, start, bits);
fs/btrfs/extent-io-tree.c
92
const struct btrfs_fs_info *btrfs_extent_io_tree_to_fs_info(const struct extent_io_tree *tree)
fs/btrfs/extent-io-tree.c
927
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
94
if (tree->owner == IO_TREE_INODE_IO)
fs/btrfs/extent-io-tree.c
95
return tree->inode->root->fs_info;
fs/btrfs/extent-io-tree.c
950
bool btrfs_find_contiguous_extent_bit(struct extent_io_tree *tree, u64 start,
fs/btrfs/extent-io-tree.c
956
ASSERT(!btrfs_fs_incompat(btrfs_extent_io_tree_to_fs_info(tree), NO_HOLES));
fs/btrfs/extent-io-tree.c
958
spin_lock(&tree->lock);
fs/btrfs/extent-io-tree.c
959
state = find_first_extent_bit_state(tree, start, bits);
fs/btrfs/extent-io-tree.c
96
return tree->fs_info;
fs/btrfs/extent-io-tree.c
970
spin_unlock(&tree->lock);
fs/btrfs/extent-io-tree.c
980
bool btrfs_find_delalloc_range(struct extent_io_tree *tree, u64 *start,
fs/btrfs/extent-io-tree.c
989
spin_lock(&tree->lock);
fs/btrfs/extent-io-tree.c
995
state = tree_search(tree, cur_start);
fs/btrfs/extent-io-tree.h
138
const struct btrfs_inode *btrfs_extent_io_tree_to_inode(const struct extent_io_tree *tree);
fs/btrfs/extent-io-tree.h
139
const struct btrfs_fs_info *btrfs_extent_io_tree_to_fs_info(const struct extent_io_tree *tree);
fs/btrfs/extent-io-tree.h
142
struct extent_io_tree *tree, unsigned int owner);
fs/btrfs/extent-io-tree.h
143
void btrfs_extent_io_tree_release(struct extent_io_tree *tree);
fs/btrfs/extent-io-tree.h
144
int btrfs_lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, u32 bits,
fs/btrfs/extent-io-tree.h
146
bool btrfs_try_lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
fs/btrfs/extent-io-tree.h
149
static inline int btrfs_lock_extent(struct extent_io_tree *tree, u64 start, u64 end,
fs/btrfs/extent-io-tree.h
152
return btrfs_lock_extent_bits(tree, start, end, EXTENT_LOCKED, cached);
fs/btrfs/extent-io-tree.h
155
static inline bool btrfs_try_lock_extent(struct extent_io_tree *tree, u64 start,
fs/btrfs/extent-io-tree.h
158
return btrfs_try_lock_extent_bits(tree, start, end, EXTENT_LOCKED, cached);
fs/btrfs/extent-io-tree.h
164
u64 btrfs_count_range_bits(struct extent_io_tree *tree,
fs/btrfs/extent-io-tree.h
170
bool btrfs_test_range_bit(struct extent_io_tree *tree, u64 start, u64 end, u32 bit,
fs/btrfs/extent-io-tree.h
172
bool btrfs_test_range_bit_exists(struct extent_io_tree *tree, u64 start, u64 end, u32 bit);
fs/btrfs/extent-io-tree.h
173
void btrfs_get_range_bits(struct extent_io_tree *tree, u64 start, u64 end, u32 *bits,
fs/btrfs/extent-io-tree.h
175
int btrfs_clear_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
fs/btrfs/extent-io-tree.h
177
int btrfs_clear_extent_bit_changeset(struct extent_io_tree *tree, u64 start, u64 end,
fs/btrfs/extent-io-tree.h
181
static inline int btrfs_clear_extent_bit(struct extent_io_tree *tree, u64 start,
fs/btrfs/extent-io-tree.h
185
return btrfs_clear_extent_bit_changeset(tree, start, end, bits, cached, NULL);
fs/btrfs/extent-io-tree.h
188
static inline int btrfs_unlock_extent(struct extent_io_tree *tree, u64 start, u64 end,
fs/btrfs/extent-io-tree.h
191
return btrfs_clear_extent_bit_changeset(tree, start, end, EXTENT_LOCKED,
fs/btrfs/extent-io-tree.h
195
int btrfs_set_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
fs/btrfs/extent-io-tree.h
197
int btrfs_set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
fs/btrfs/extent-io-tree.h
200
static inline int btrfs_clear_extent_dirty(struct extent_io_tree *tree, u64 start,
fs/btrfs/extent-io-tree.h
203
return btrfs_clear_extent_bit(tree, start, end,
fs/btrfs/extent-io-tree.h
208
int btrfs_convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
fs/btrfs/extent-io-tree.h
212
bool btrfs_find_first_extent_bit(struct extent_io_tree *tree, u64 start,
fs/btrfs/extent-io-tree.h
215
void btrfs_find_first_clear_extent_bit(struct extent_io_tree *tree, u64 start,
fs/btrfs/extent-io-tree.h
217
bool btrfs_find_contiguous_extent_bit(struct extent_io_tree *tree, u64 start,
fs/btrfs/extent-io-tree.h
219
bool btrfs_find_delalloc_range(struct extent_io_tree *tree, u64 *start,
fs/btrfs/extent-io-tree.h
222
static inline int btrfs_lock_dio_extent(struct extent_io_tree *tree, u64 start,
fs/btrfs/extent-io-tree.h
225
return btrfs_lock_extent_bits(tree, start, end, EXTENT_DIO_LOCKED, cached);
fs/btrfs/extent-io-tree.h
228
static inline bool btrfs_try_lock_dio_extent(struct extent_io_tree *tree, u64 start,
fs/btrfs/extent-io-tree.h
231
return btrfs_try_lock_extent_bits(tree, start, end, EXTENT_DIO_LOCKED, cached);
fs/btrfs/extent-io-tree.h
234
static inline int btrfs_unlock_dio_extent(struct extent_io_tree *tree, u64 start,
fs/btrfs/extent-io-tree.h
237
return btrfs_clear_extent_bit_changeset(tree, start, end, EXTENT_DIO_LOCKED,
fs/btrfs/extent-io-tree.h
241
struct extent_state *btrfs_next_extent_state(struct extent_io_tree *tree,
fs/btrfs/extent_io.c
2753
int extent_invalidate_folio(struct extent_io_tree *tree,
fs/btrfs/extent_io.c
2762
ASSERT(tree->owner == IO_TREE_BTREE_INODE_IO);
fs/btrfs/extent_io.c
2768
btrfs_lock_extent(tree, start, end, &cached_state);
fs/btrfs/extent_io.c
2776
btrfs_unlock_extent(tree, start, end, &cached_state);
fs/btrfs/extent_io.c
2785
static bool try_release_extent_state(struct extent_io_tree *tree,
fs/btrfs/extent_io.c
2796
btrfs_get_range_bits(tree, start, end, &range_bits, &cached_state);
fs/btrfs/extent_io.c
2814
ret2 = btrfs_clear_extent_bit(tree, start, end, clear_bits, &cached_state);
fs/btrfs/extent_io.c
374
struct extent_io_tree *tree = &BTRFS_I(inode)->io_tree;
fs/btrfs/extent_io.c
402
found = btrfs_find_delalloc_range(tree, &delalloc_start, &delalloc_end,
fs/btrfs/extent_io.c
448
btrfs_lock_extent(tree, delalloc_start, delalloc_end, &cached_state);
fs/btrfs/extent_io.c
451
ret = btrfs_test_range_bit(tree, delalloc_start, delalloc_end,
fs/btrfs/extent_io.c
454
btrfs_unlock_extent(tree, delalloc_start, delalloc_end, &cached_state);
fs/btrfs/extent_io.h
360
int extent_invalidate_folio(struct extent_io_tree *tree,
fs/btrfs/extent_map.c
1018
struct extent_map_tree *tree = &inode->extent_tree;
fs/btrfs/extent_map.c
1033
write_lock(&tree->lock);
fs/btrfs/extent_map.c
1035
write_unlock(&tree->lock);
fs/btrfs/extent_map.c
1138
struct extent_map_tree *tree = &inode->extent_tree;
fs/btrfs/extent_map.c
1142
lockdep_assert_held_write(&tree->lock);
fs/btrfs/extent_map.c
1161
node = rb_first(&tree->root);
fs/btrfs/extent_map.c
1196
if (need_resched() || rwlock_needbreak(&tree->lock) ||
fs/btrfs/extent_map.c
1214
struct extent_map_tree *tree;
fs/btrfs/extent_map.c
1220
tree = &inode->extent_tree;
fs/btrfs/extent_map.c
1227
if (!write_trylock(&tree->lock))
fs/btrfs/extent_map.c
1238
if (RB_EMPTY_ROOT(&tree->root)) {
fs/btrfs/extent_map.c
1239
write_unlock(&tree->lock);
fs/btrfs/extent_map.c
1246
write_unlock(&tree->lock);
fs/btrfs/extent_map.c
34
void btrfs_extent_map_tree_init(struct extent_map_tree *tree)
fs/btrfs/extent_map.c
36
tree->root = RB_ROOT;
fs/btrfs/extent_map.c
37
INIT_LIST_HEAD(&tree->modified_extents);
fs/btrfs/extent_map.c
38
rwlock_init(&tree->lock);
fs/btrfs/extent_map.c
427
struct extent_map_tree *tree = &inode->extent_tree;
fs/btrfs/extent_map.c
431
write_lock(&tree->lock);
fs/btrfs/extent_map.c
432
em = btrfs_lookup_extent_mapping(tree, start, len);
fs/btrfs/extent_map.c
458
write_unlock(&tree->lock);
fs/btrfs/extent_map.c
503
struct extent_map_tree *tree = &inode->extent_tree;
fs/btrfs/extent_map.c
508
lockdep_assert_held_write(&tree->lock);
fs/btrfs/extent_map.c
511
ret = tree_insert(&tree->root, em);
fs/btrfs/extent_map.c
523
static struct extent_map *lookup_extent_mapping(struct extent_map_tree *tree,
fs/btrfs/extent_map.c
531
rb_node = tree_search(&tree->root, start, &prev_or_next);
fs/btrfs/extent_map.c
560
struct extent_map *btrfs_lookup_extent_mapping(struct extent_map_tree *tree,
fs/btrfs/extent_map.c
563
return lookup_extent_mapping(tree, start, len, true);
fs/btrfs/extent_map.c
578
struct extent_map *btrfs_search_extent_mapping(struct extent_map_tree *tree,
fs/btrfs/extent_map.c
581
return lookup_extent_mapping(tree, start, len, false);
fs/btrfs/extent_map.c
595
struct extent_map_tree *tree = &inode->extent_tree;
fs/btrfs/extent_map.c
597
lockdep_assert_held_write(&tree->lock);
fs/btrfs/extent_map.c
612
struct extent_map_tree *tree = &inode->extent_tree;
fs/btrfs/extent_map.c
614
lockdep_assert_held_write(&tree->lock);
fs/btrfs/extent_map.c
622
rb_replace_node(&cur->rb_node, &new->rb_node, &tree->root);
fs/btrfs/extent_map.c
775
struct extent_map_tree *tree = &inode->extent_tree;
fs/btrfs/extent_map.c
778
write_lock(&tree->lock);
fs/btrfs/extent_map.c
779
node = rb_first(&tree->root);
fs/btrfs/extent_map.c
789
if (cond_resched_rwlock_write(&tree->lock))
fs/btrfs/extent_map.c
790
node = rb_first(&tree->root);
fs/btrfs/extent_map.c
794
write_unlock(&tree->lock);
fs/btrfs/extent_map.h
169
void btrfs_extent_map_tree_init(struct extent_map_tree *tree);
fs/btrfs/extent_map.h
170
struct extent_map *btrfs_lookup_extent_mapping(struct extent_map_tree *tree,
fs/btrfs/extent_map.h
182
struct extent_map *btrfs_search_extent_mapping(struct extent_map_tree *tree,
fs/btrfs/inode.c
7677
struct extent_io_tree *tree = &inode->io_tree;
fs/btrfs/inode.c
7718
btrfs_lock_extent(tree, page_start, page_end, &cached_state);
fs/btrfs/inode.c
7774
btrfs_clear_extent_bit(tree, cur, range_end,
fs/btrfs/inode.c
7820
btrfs_clear_extent_bit(tree, cur, range_end, EXTENT_LOCKED |
fs/btrfs/tests/extent-io-tests.c
560
struct extent_io_tree tree;
fs/btrfs/tests/extent-io-tests.c
566
btrfs_extent_io_tree_init(NULL, &tree, IO_TREE_SELFTEST);
fs/btrfs/tests/extent-io-tests.c
569
btrfs_find_first_clear_extent_bit(&tree, 0, &start, &end, CHUNK_TRIMMED);
fs/btrfs/tests/extent-io-tests.c
580
btrfs_set_extent_bit(&tree, SZ_1M, SZ_4M - 1,
fs/btrfs/tests/extent-io-tests.c
583
btrfs_find_first_clear_extent_bit(&tree, SZ_512K, &start, &end,
fs/btrfs/tests/extent-io-tests.c
593
btrfs_set_extent_bit(&tree, SZ_32M, SZ_64M - 1,
fs/btrfs/tests/extent-io-tests.c
599
btrfs_find_first_clear_extent_bit(&tree, 12 * SZ_1M, &start, &end,
fs/btrfs/tests/extent-io-tests.c
612
btrfs_find_first_clear_extent_bit(&tree, SZ_2M, &start, &end,
fs/btrfs/tests/extent-io-tests.c
625
btrfs_set_extent_bit(&tree, SZ_64M, SZ_64M + SZ_8M - 1, CHUNK_ALLOCATED, NULL);
fs/btrfs/tests/extent-io-tests.c
626
btrfs_find_first_clear_extent_bit(&tree, SZ_64M + SZ_1M, &start, &end,
fs/btrfs/tests/extent-io-tests.c
635
btrfs_find_first_clear_extent_bit(&tree, SZ_64M - SZ_8M, &start, &end,
fs/btrfs/tests/extent-io-tests.c
652
btrfs_find_first_clear_extent_bit(&tree, -1, &start, &end, CHUNK_TRIMMED);
fs/btrfs/tests/extent-io-tests.c
663
dump_extent_io_tree(&tree);
fs/btrfs/tests/extent-io-tests.c
664
btrfs_clear_extent_bit(&tree, 0, (u64)-1, CHUNK_TRIMMED | CHUNK_ALLOCATED, NULL);
fs/btrfs/tests/extent-io-tests.c
91
static void dump_extent_io_tree(const struct extent_io_tree *tree)
fs/btrfs/tests/extent-io-tests.c
96
node = rb_first(&tree->state);
fs/btrfs/tree-log.c
5370
struct extent_map_tree *tree = &inode->extent_tree;
fs/btrfs/tree-log.c
5374
write_lock(&tree->lock);
fs/btrfs/tree-log.c
5376
list_for_each_entry_safe(em, n, &tree->modified_extents, list) {
fs/btrfs/tree-log.c
5385
list_del_init(&tree->modified_extents);
fs/btrfs/tree-log.c
5422
write_unlock(&tree->lock);
fs/btrfs/tree-log.c
5425
write_lock(&tree->lock);
fs/btrfs/tree-log.c
5430
write_unlock(&tree->lock);
fs/erofs/data.c
197
dif = idr_find(&devs->tree, map->m_deviceid - 1);
fs/erofs/data.c
211
idr_for_each_entry(&devs->tree, dif, id) {
fs/erofs/internal.h
66
struct idr tree;
fs/erofs/super.c
222
idr_for_each_entry(&sbi->devs->tree, dif, id) {
fs/erofs/super.c
235
err = idr_alloc(&sbi->devs->tree, dif, 0, 0, GFP_KERNEL);
fs/erofs/super.c
499
ret = idr_alloc(&sbi->devs->tree, dif, 0, 0, GFP_KERNEL);
fs/erofs/super.c
858
idr_for_each(&devs->tree, &erofs_release_device_info, NULL);
fs/erofs/super.c
859
idr_destroy(&devs->tree);
fs/erofs/super.c
904
idr_init(&sbi->devs->tree);
fs/ext4/extents_status.c
1096
struct ext4_es_tree *tree;
fs/ext4/extents_status.c
1108
tree = &EXT4_I(inode)->i_es_tree;
fs/ext4/extents_status.c
1113
es1 = READ_ONCE(tree->cache_es);
fs/ext4/extents_status.c
1121
node = tree->root.rb_node;
fs/ext4/extents_status.c
1355
struct ext4_pending_tree *tree = &EXT4_I(inode)->i_pending_tree;
fs/ext4/extents_status.c
1447
pr = __pr_tree_search(&tree->root, first_lclu);
fs/ext4/extents_status.c
1451
rb_erase(&pr->rb_node, &tree->root);
fs/ext4/extents_status.c
1486
struct ext4_es_tree *tree = &EXT4_I(inode)->i_es_tree;
fs/ext4/extents_status.c
1501
es = __es_tree_search(&tree->root, lblk);
fs/ext4/extents_status.c
1512
tree->cache_es = NULL;
fs/ext4/extents_status.c
1579
rb_erase(&es->rb_node, &tree->root);
fs/ext4/extents_status.c
1904
struct ext4_es_tree *tree = &ei->i_es_tree;
fs/ext4/extents_status.c
1908
es = __es_tree_search(&tree->root, ei->i_es_shrink_lblk);
fs/ext4/extents_status.c
1928
rb_erase(&es->rb_node, &tree->root);
fs/ext4/extents_status.c
1975
struct ext4_es_tree *tree;
fs/ext4/extents_status.c
1979
tree = &EXT4_I(inode)->i_es_tree;
fs/ext4/extents_status.c
1980
tree->cache_es = NULL;
fs/ext4/extents_status.c
1981
node = rb_first(&tree->root);
fs/ext4/extents_status.c
1986
rb_erase(&es->rb_node, &tree->root);
fs/ext4/extents_status.c
1997
struct ext4_pending_tree *tree;
fs/ext4/extents_status.c
2002
tree = &EXT4_I(inode)->i_pending_tree;
fs/ext4/extents_status.c
2003
node = rb_first(&tree->root);
fs/ext4/extents_status.c
2028
void ext4_init_pending_tree(struct ext4_pending_tree *tree)
fs/ext4/extents_status.c
2030
tree->root = RB_ROOT;
fs/ext4/extents_status.c
2045
struct ext4_pending_tree *tree;
fs/ext4/extents_status.c
2049
tree = &EXT4_I(inode)->i_pending_tree;
fs/ext4/extents_status.c
205
void ext4_es_init_tree(struct ext4_es_tree *tree)
fs/ext4/extents_status.c
2050
node = (&tree->root)->rb_node;
fs/ext4/extents_status.c
207
tree->root = RB_ROOT;
fs/ext4/extents_status.c
2079
struct ext4_pending_tree *tree = &EXT4_I(inode)->i_pending_tree;
fs/ext4/extents_status.c
208
tree->cache_es = NULL;
fs/ext4/extents_status.c
2080
struct rb_node **p = &tree->root.rb_node;
fs/ext4/extents_status.c
2115
rb_insert_color(&pr->rb_node, &tree->root);
fs/ext4/extents_status.c
2135
struct ext4_pending_tree *tree;
fs/ext4/extents_status.c
2139
tree = &EXT4_I(inode)->i_pending_tree;
fs/ext4/extents_status.c
214
struct ext4_es_tree *tree;
fs/ext4/extents_status.c
2140
rb_erase(&pr->rb_node, &tree->root);
fs/ext4/extents_status.c
218
tree = &EXT4_I(inode)->i_es_tree;
fs/ext4/extents_status.c
219
node = rb_first(&tree->root);
fs/ext4/extents_status.c
317
struct ext4_es_tree *tree = NULL;
fs/ext4/extents_status.c
324
tree = &EXT4_I(inode)->i_es_tree;
fs/ext4/extents_status.c
328
es1 = READ_ONCE(tree->cache_es);
fs/ext4/extents_status.c
336
es1 = __es_tree_search(&tree->root, lblk);
fs/ext4/extents_status.c
352
WRITE_ONCE(tree->cache_es, es1);
fs/ext4/extents_status.c
626
struct ext4_es_tree *tree = &EXT4_I(inode)->i_es_tree;
fs/ext4/extents_status.c
639
rb_erase(&es->rb_node, &tree->root);
fs/ext4/extents_status.c
650
struct ext4_es_tree *tree = &EXT4_I(inode)->i_es_tree;
fs/ext4/extents_status.c
663
rb_erase(node, &tree->root);
fs/ext4/extents_status.c
845
struct ext4_es_tree *tree = &EXT4_I(inode)->i_es_tree;
fs/ext4/extents_status.c
846
struct rb_node **p = &tree->root.rb_node;
fs/ext4/extents_status.c
893
rb_insert_color(&es->rb_node, &tree->root);
fs/ext4/extents_status.c
896
tree->cache_es = es;
fs/ext4/extents_status.h
134
extern void ext4_es_init_tree(struct ext4_es_tree *tree);
fs/ext4/extents_status.h
246
extern void ext4_init_pending_tree(struct ext4_pending_tree *tree);
fs/fuse/dax.c
1045
node = interval_tree_iter_first(&fi->dax->tree, start_idx, start_idx);
fs/fuse/dax.c
1285
fi->dax->tree = RB_ROOT_CACHED;
fs/fuse/dax.c
222
interval_tree_insert(&dmap->itn, &fi->dax->tree);
fs/fuse/dax.c
322
node = interval_tree_iter_first(&fi->dax->tree, start_idx,
fs/fuse/dax.c
329
interval_tree_remove(&dmap->itn, &fi->dax->tree);
fs/fuse/dax.c
477
node = interval_tree_iter_first(&fi->dax->tree, start_idx, start_idx);
fs/fuse/dax.c
514
node = interval_tree_iter_first(&fi->dax->tree, idx, idx);
fs/fuse/dax.c
589
node = interval_tree_iter_first(&fi->dax->tree, start_idx, start_idx);
fs/fuse/dax.c
67
struct rb_root_cached tree;
fs/fuse/dax.c
870
interval_tree_remove(&dmap->itn, &fi->dax->tree);
fs/fuse/dax.c
894
for (node = interval_tree_iter_first(&fi->dax->tree, 0, -1); node;
fs/fuse/dax.c
949
node = interval_tree_iter_first(&fi->dax->tree, start_idx, start_idx);
fs/fuse/dir.c
111
rb_erase(&fd->node, &bucket->tree);
fs/fuse/dir.c
142
p = &bucket->tree.rb_node;
fs/fuse/dir.c
152
rb_insert_color(&fd->node, &bucket->tree);
fs/fuse/dir.c
169
node = rb_first(&dentry_hash[i].tree);
fs/fuse/dir.c
175
rb_erase(&fd->node, &dentry_hash[i].tree);
fs/fuse/dir.c
187
node = rb_first(&dentry_hash[i].tree);
fs/fuse/dir.c
224
dentry_hash[i].tree = RB_ROOT;
fs/fuse/dir.c
237
WARN_ON_ONCE(!RB_EMPTY_ROOT(&dentry_hash[i].tree));
fs/fuse/dir.c
31
struct rb_root tree;
fs/hfs/bfind.c
112
struct hfs_btree *tree;
fs/hfs/bfind.c
124
tree = fd->tree;
fs/hfs/bfind.c
128
nidx = tree->root;
fs/hfs/bfind.c
131
height = tree->depth;
fs/hfs/bfind.c
135
bnode = hfs_bnode_find(tree, nidx);
fs/hfs/bfind.c
15
int hfs_find_init(struct hfs_btree *tree, struct hfs_find_data *fd)
fs/hfs/bfind.c
185
struct hfs_btree *tree;
fs/hfs/bfind.c
19
if (!tree || !fd)
fs/hfs/bfind.c
191
tree = bnode->tree;
fs/hfs/bfind.c
204
bnode = hfs_bnode_find(tree, idx);
fs/hfs/bfind.c
22
fd->tree = tree;
fs/hfs/bfind.c
222
bnode = hfs_bnode_find(tree, idx);
fs/hfs/bfind.c
24
ptr = kzalloc(tree->max_key_len * 2 + 4, GFP_KERNEL);
fs/hfs/bfind.c
28
fd->key = ptr + tree->max_key_len + 2;
fs/hfs/bfind.c
30
tree->cnid, __builtin_return_address(0));
fs/hfs/bfind.c
31
switch (tree->cnid) {
fs/hfs/bfind.c
33
mutex_lock_nested(&tree->tree_lock, CATALOG_BTREE_MUTEX);
fs/hfs/bfind.c
36
mutex_lock_nested(&tree->tree_lock, EXTENTS_BTREE_MUTEX);
fs/hfs/bfind.c
39
mutex_lock_nested(&tree->tree_lock, ATTR_BTREE_MUTEX);
fs/hfs/bfind.c
52
fd->tree->cnid, __builtin_return_address(0));
fs/hfs/bfind.c
53
mutex_unlock(&fd->tree->tree_lock);
fs/hfs/bfind.c
54
fd->tree = NULL;
fs/hfs/bfind.c
78
cmpval = bnode->tree->keycmp(fd->key, fd->search_key);
fs/hfs/bnode.c
116
struct hfs_btree *tree;
fs/hfs/bnode.c
119
tree = node->tree;
fs/hfs/bnode.c
121
tree->attributes & HFS_TREE_VARIDXKEYS)
fs/hfs/bnode.c
124
key_len = tree->max_key_len + 1;
fs/hfs/bnode.c
147
node->tree->node_size, off, len);
fs/hfs/bnode.c
185
node->tree->node_size, off, len);
fs/hfs/bnode.c
21
bool is_valid = off < node->tree->node_size;
fs/hfs/bnode.c
252
off = node->tree->node_size - 2;
fs/hfs/bnode.c
259
if (node->tree->attributes & HFS_TREE_VARIDXKEYS)
fs/hfs/bnode.c
262
tmp = node->tree->max_key_len + 1;
fs/hfs/bnode.c
279
struct hfs_btree *tree;
fs/hfs/bnode.c
28
node->tree->node_size, off);
fs/hfs/bnode.c
283
tree = node->tree;
fs/hfs/bnode.c
285
tmp = hfs_bnode_find(tree, node->prev);
fs/hfs/bnode.c
293
tree->leaf_head = node->next;
fs/hfs/bnode.c
296
tmp = hfs_bnode_find(tree, node->next);
fs/hfs/bnode.c
304
tree->leaf_tail = node->prev;
fs/hfs/bnode.c
311
tree->root = 0;
fs/hfs/bnode.c
312
tree->depth = 0;
fs/hfs/bnode.c
324
struct hfs_bnode *hfs_bnode_findhash(struct hfs_btree *tree, u32 cnid)
fs/hfs/bnode.c
328
if (cnid >= tree->node_count) {
fs/hfs/bnode.c
333
for (node = tree->node_hash[hfs_bnode_hash(cnid)];
fs/hfs/bnode.c
342
static struct hfs_bnode *__hfs_bnode_create(struct hfs_btree *tree, u32 cnid)
fs/hfs/bnode.c
350
if (cnid >= tree->node_count) {
fs/hfs/bnode.c
355
size = sizeof(struct hfs_bnode) + tree->pages_per_bnode *
fs/hfs/bnode.c
360
node->tree = tree;
fs/hfs/bnode.c
365
node->tree->cnid, node->this);
fs/hfs/bnode.c
367
spin_lock(&tree->hash_lock);
fs/hfs/bnode.c
368
node2 = hfs_bnode_findhash(tree, cnid);
fs/hfs/bnode.c
371
node->next_hash = tree->node_hash[hash];
fs/hfs/bnode.c
372
tree->node_hash[hash] = node;
fs/hfs/bnode.c
373
tree->node_hash_cnt++;
fs/hfs/bnode.c
376
spin_unlock(&tree->hash_lock);
fs/hfs/bnode.c
381
spin_unlock(&tree->hash_lock);
fs/hfs/bnode.c
383
mapping = tree->inode->i_mapping;
fs/hfs/bnode.c
384
off = (loff_t)cnid * tree->node_size;
fs/hfs/bnode.c
387
for (i = 0; i < tree->pages_per_bnode; i++) {
fs/hfs/bnode.c
405
node->tree->cnid, node->this, atomic_read(&node->refcnt));
fs/hfs/bnode.c
406
for (p = &node->tree->node_hash[hfs_bnode_hash(node->this)];
fs/hfs/bnode.c
411
node->tree->node_hash_cnt--;
fs/hfs/bnode.c
415
struct hfs_bnode *hfs_bnode_find(struct hfs_btree *tree, u32 num)
fs/hfs/bnode.c
42
node_size = node->tree->node_size;
fs/hfs/bnode.c
422
spin_lock(&tree->hash_lock);
fs/hfs/bnode.c
423
node = hfs_bnode_findhash(tree, num);
fs/hfs/bnode.c
426
spin_unlock(&tree->hash_lock);
fs/hfs/bnode.c
432
spin_unlock(&tree->hash_lock);
fs/hfs/bnode.c
433
node = __hfs_bnode_create(tree, num);
fs/hfs/bnode.c
461
if (node->height <= 1 || node->height > tree->depth)
fs/hfs/bnode.c
468
rec_off = tree->node_size - 2;
fs/hfs/bnode.c
476
next_off > tree->node_size ||
fs/hfs/bnode.c
503
for (i = 0; i < node->tree->pages_per_bnode; i++)
fs/hfs/bnode.c
509
struct hfs_bnode *hfs_bnode_create(struct hfs_btree *tree, u32 num)
fs/hfs/bnode.c
515
spin_lock(&tree->hash_lock);
fs/hfs/bnode.c
516
node = hfs_bnode_findhash(tree, num);
fs/hfs/bnode.c
517
spin_unlock(&tree->hash_lock);
fs/hfs/bnode.c
52
node->tree->node_size, off, len, new_len);
fs/hfs/bnode.c
523
node = __hfs_bnode_create(tree, num);
fs/hfs/bnode.c
533
min((int)PAGE_SIZE, (int)tree->node_size));
fs/hfs/bnode.c
535
for (i = 1; i < tree->pages_per_bnode; i++) {
fs/hfs/bnode.c
550
node->tree->cnid, node->this,
fs/hfs/bnode.c
559
struct hfs_btree *tree = node->tree;
fs/hfs/bnode.c
563
node->tree->cnid, node->this,
fs/hfs/bnode.c
566
if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock))
fs/hfs/bnode.c
568
for (i = 0; i < tree->pages_per_bnode; i++) {
fs/hfs/bnode.c
576
spin_unlock(&tree->hash_lock);
fs/hfs/bnode.c
577
hfs_bnode_clear(node, 0, tree->node_size);
fs/hfs/bnode.c
582
spin_unlock(&tree->hash_lock);
fs/hfs/bnode.c
75
node->tree->node_size, off, len);
fs/hfs/bnode.c
86
if (pagenum >= node->tree->pages_per_bnode)
fs/hfs/brec.c
108
tree->leaf_count++;
fs/hfs/brec.c
109
mark_inode_dirty(tree->inode);
fs/hfs/brec.c
117
idx_rec_off = tree->node_size - (rec + 1) * 2;
fs/hfs/brec.c
148
hfs_btree_inc_height(tree);
fs/hfs/brec.c
149
new_node->parent = tree->root;
fs/hfs/brec.c
151
fd->bnode = hfs_bnode_find(tree, new_node->parent);
fs/hfs/brec.c
16
static int hfs_btree_inc_height(struct hfs_btree *tree);
fs/hfs/brec.c
165
if (tree->attributes & HFS_TREE_VARIDXKEYS)
fs/hfs/brec.c
168
fd->search_key->key_len = tree->max_key_len;
fs/hfs/brec.c
169
key_len = tree->max_key_len + 1;
fs/hfs/brec.c
179
struct hfs_btree *tree;
fs/hfs/brec.c
184
tree = fd->tree;
fs/hfs/brec.c
187
rec_off = tree->node_size - (fd->record + 2) * 2;
fs/hfs/brec.c
188
end_off = tree->node_size - (node->num_recs + 1) * 2;
fs/hfs/brec.c
191
tree->leaf_count--;
fs/hfs/brec.c
192
mark_inode_dirty(tree->inode);
fs/hfs/brec.c
201
parent = hfs_bnode_find(tree, node->parent);
fs/hfs/brec.c
24
dataoff = node->tree->node_size - (rec + 2) * 2;
fs/hfs/brec.c
252
struct hfs_btree *tree;
fs/hfs/brec.c
258
tree = fd->tree;
fs/hfs/brec.c
260
new_node = hfs_bmap_alloc(tree);
fs/hfs/brec.c
273
next_node = hfs_bnode_find(tree, node->next);
fs/hfs/brec.c
283
size = tree->node_size / 2 - node->num_recs * 2 - 14;
fs/hfs/brec.c
284
old_rec_off = tree->node_size - 4;
fs/hfs/brec.c
319
new_rec_off = tree->node_size - 2;
fs/hfs/brec.c
358
} else if (node->this == tree->leaf_tail) {
fs/hfs/brec.c
360
tree->leaf_tail = new_node->this;
fs/hfs/brec.c
361
mark_inode_dirty(tree->inode);
fs/hfs/brec.c
373
struct hfs_btree *tree;
fs/hfs/brec.c
379
tree = fd->tree;
fs/hfs/brec.c
386
parent = hfs_bnode_find(tree, node->parent);
fs/hfs/brec.c
39
!(node->tree->attributes & HFS_TREE_VARIDXKEYS)) {
fs/hfs/brec.c
396
if (tree->attributes & HFS_TREE_VARIDXKEYS)
fs/hfs/brec.c
399
fd->keylength = newkeylen = tree->max_key_len + 1;
fs/hfs/brec.c
40
if (node->tree->attributes & HFS_TREE_BIGKEYS)
fs/hfs/brec.c
403
rec_off = tree->node_size - (rec + 2) * 2;
fs/hfs/brec.c
404
end_rec_off = tree->node_size - (parent->num_recs + 1) * 2;
fs/hfs/brec.c
41
retval = node->tree->max_key_len + 2;
fs/hfs/brec.c
419
rec_off = tree->node_size - (rec + 2) * 2;
fs/hfs/brec.c
420
end_rec_off = tree->node_size - (parent->num_recs + 1) * 2;
fs/hfs/brec.c
43
retval = node->tree->max_key_len + 1;
fs/hfs/brec.c
437
if (!(tree->attributes & HFS_TREE_VARIDXKEYS))
fs/hfs/brec.c
448
hfs_btree_inc_height(tree);
fs/hfs/brec.c
449
new_node->parent = tree->root;
fs/hfs/brec.c
45
recoff = hfs_bnode_read_u16(node, node->tree->node_size - (rec + 1) * 2);
fs/hfs/brec.c
451
fd->bnode = hfs_bnode_find(tree, new_node->parent);
fs/hfs/brec.c
477
static int hfs_btree_inc_height(struct hfs_btree *tree)
fs/hfs/brec.c
48
if (node->tree->attributes & HFS_TREE_BIGKEYS) {
fs/hfs/brec.c
485
if (tree->root) {
fs/hfs/brec.c
486
node = hfs_bnode_find(tree, tree->root);
fs/hfs/brec.c
490
new_node = hfs_bmap_alloc(tree);
fs/hfs/brec.c
496
tree->root = new_node->this;
fs/hfs/brec.c
497
if (!tree->depth) {
fs/hfs/brec.c
498
tree->leaf_head = tree->leaf_tail = new_node->this;
fs/hfs/brec.c
50
if (retval > node->tree->max_key_len + 2) {
fs/hfs/brec.c
508
new_node->height = ++tree->depth;
fs/hfs/brec.c
518
rec = tree->node_size - 2;
fs/hfs/brec.c
523
node->parent = tree->root;
fs/hfs/brec.c
525
tree->attributes & HFS_TREE_VARIDXKEYS)
fs/hfs/brec.c
528
key_size = tree->max_key_len + 1;
fs/hfs/brec.c
531
if (!(tree->attributes & HFS_TREE_VARIDXKEYS)) {
fs/hfs/brec.c
532
key_size = tree->max_key_len + 1;
fs/hfs/brec.c
533
hfs_bnode_write_u8(new_node, 14, tree->max_key_len);
fs/hfs/brec.c
545
mark_inode_dirty(tree->inode);
fs/hfs/brec.c
56
if (retval > node->tree->max_key_len + 1) {
fs/hfs/brec.c
67
struct hfs_btree *tree;
fs/hfs/brec.c
74
tree = fd->tree;
fs/hfs/brec.c
76
if (!tree->root)
fs/hfs/brec.c
77
hfs_btree_inc_height(tree);
fs/hfs/brec.c
78
node = hfs_bnode_find(tree, tree->leaf_head);
fs/hfs/brec.c
94
end_rec_off = tree->node_size - (node->num_recs + 1) * 2;
fs/hfs/btree.c
117
tree->root = be32_to_cpu(head->root);
fs/hfs/btree.c
118
tree->leaf_count = be32_to_cpu(head->leaf_count);
fs/hfs/btree.c
119
tree->leaf_head = be32_to_cpu(head->leaf_head);
fs/hfs/btree.c
120
tree->leaf_tail = be32_to_cpu(head->leaf_tail);
fs/hfs/btree.c
121
tree->node_count = be32_to_cpu(head->node_count);
fs/hfs/btree.c
122
tree->free_nodes = be32_to_cpu(head->free_nodes);
fs/hfs/btree.c
123
tree->attributes = be32_to_cpu(head->attributes);
fs/hfs/btree.c
124
tree->node_size = be16_to_cpu(head->node_size);
fs/hfs/btree.c
125
tree->max_key_len = be16_to_cpu(head->max_key_len);
fs/hfs/btree.c
126
tree->depth = be16_to_cpu(head->depth);
fs/hfs/btree.c
128
size = tree->node_size;
fs/hfs/btree.c
131
if (!tree->node_count)
fs/hfs/btree.c
135
if (tree->max_key_len != HFS_MAX_EXT_KEYLEN) {
fs/hfs/btree.c
137
tree->max_key_len);
fs/hfs/btree.c
142
if (tree->max_key_len != HFS_MAX_CAT_KEYLEN) {
fs/hfs/btree.c
144
tree->max_key_len);
fs/hfs/btree.c
152
tree->node_size_shift = ffs(size) - 1;
fs/hfs/btree.c
153
tree->pages_per_bnode = (tree->node_size + PAGE_SIZE - 1) >> PAGE_SHIFT;
fs/hfs/btree.c
158
return tree;
fs/hfs/btree.c
166
tree->inode->i_mapping->a_ops = &hfs_aops;
fs/hfs/btree.c
167
iput(tree->inode);
fs/hfs/btree.c
169
kfree(tree);
fs/hfs/btree.c
174
void hfs_btree_close(struct hfs_btree *tree)
fs/hfs/btree.c
179
if (!tree)
fs/hfs/btree.c
183
while ((node = tree->node_hash[i])) {
fs/hfs/btree.c
184
tree->node_hash[i] = node->next_hash;
fs/hfs/btree.c
187
node->tree->cnid, node->this,
fs/hfs/btree.c
190
tree->node_hash_cnt--;
fs/hfs/btree.c
193
iput(tree->inode);
fs/hfs/btree.c
194
kfree(tree);
fs/hfs/btree.c
197
void hfs_btree_write(struct hfs_btree *tree)
fs/hfs/btree.c
203
node = hfs_bnode_find(tree, 0);
fs/hfs/btree.c
21
struct hfs_btree *tree;
fs/hfs/btree.c
212
head->root = cpu_to_be32(tree->root);
fs/hfs/btree.c
213
head->leaf_count = cpu_to_be32(tree->leaf_count);
fs/hfs/btree.c
214
head->leaf_head = cpu_to_be32(tree->leaf_head);
fs/hfs/btree.c
215
head->leaf_tail = cpu_to_be32(tree->leaf_tail);
fs/hfs/btree.c
216
head->node_count = cpu_to_be32(tree->node_count);
fs/hfs/btree.c
217
head->free_nodes = cpu_to_be32(tree->free_nodes);
fs/hfs/btree.c
218
head->attributes = cpu_to_be32(tree->attributes);
fs/hfs/btree.c
219
head->depth = cpu_to_be16(tree->depth);
fs/hfs/btree.c
228
struct hfs_btree *tree = prev->tree;
fs/hfs/btree.c
233
node = hfs_bnode_create(tree, idx);
fs/hfs/btree.c
237
if (!tree->free_nodes)
fs/hfs/btree.c
239
tree->free_nodes--;
fs/hfs/btree.c
246
hfs_bnode_clear(node, 0, tree->node_size);
fs/hfs/btree.c
255
hfs_bnode_write_u16(node, tree->node_size - 2, 14);
fs/hfs/btree.c
256
hfs_bnode_write_u16(node, tree->node_size - 4, tree->node_size - 6);
fs/hfs/btree.c
262
int hfs_bmap_reserve(struct hfs_btree *tree, u32 rsvd_nodes)
fs/hfs/btree.c
264
struct inode *inode = tree->inode;
fs/hfs/btree.c
268
while (tree->free_nodes < rsvd_nodes) {
fs/hfs/btree.c
274
HFS_SB(tree->sb)->alloc_blksz;
fs/hfs/btree.c
276
tree->sb->s_blocksize_bits;
fs/hfs/btree.c
278
count = inode->i_size >> tree->node_size_shift;
fs/hfs/btree.c
279
tree->free_nodes += count - tree->node_count;
fs/hfs/btree.c
280
tree->node_count = count;
fs/hfs/btree.c
285
struct hfs_bnode *hfs_bmap_alloc(struct hfs_btree *tree)
fs/hfs/btree.c
296
res = hfs_bmap_reserve(tree, 1);
fs/hfs/btree.c
301
node = hfs_bnode_find(tree, nidx);
fs/hfs/btree.c
31
tree = kzalloc_obj(*tree);
fs/hfs/btree.c
32
if (!tree)
fs/hfs/btree.c
323
tree->free_nodes--;
fs/hfs/btree.c
324
mark_inode_dirty(tree->inode);
fs/hfs/btree.c
326
return hfs_bnode_create(tree, idx);
fs/hfs/btree.c
344
next_node = hfs_bnode_find(tree, nidx);
fs/hfs/btree.c
35
mutex_init(&tree->tree_lock);
fs/hfs/btree.c
36
spin_lock_init(&tree->hash_lock);
fs/hfs/btree.c
361
struct hfs_btree *tree;
fs/hfs/btree.c
368
tree = node->tree;
fs/hfs/btree.c
370
node = hfs_bnode_find(tree, 0);
fs/hfs/btree.c
38
tree->sb = sb;
fs/hfs/btree.c
387
node = hfs_bnode_find(tree, i);
fs/hfs/btree.c
39
tree->cnid = id;
fs/hfs/btree.c
40
tree->keycmp = keycmp;
fs/hfs/btree.c
416
tree->free_nodes++;
fs/hfs/btree.c
417
mark_inode_dirty(tree->inode);
fs/hfs/btree.c
42
tree->inode = iget_locked(sb, id);
fs/hfs/btree.c
43
if (!tree->inode)
fs/hfs/btree.c
45
BUG_ON(!(inode_state_read_once(tree->inode) & I_NEW));
fs/hfs/btree.c
48
HFS_I(tree->inode)->flags = 0;
fs/hfs/btree.c
49
mutex_init(&HFS_I(tree->inode)->extents_lock);
fs/hfs/btree.c
52
hfs_inode_read_fork(tree->inode, mdb->drXTExtRec, mdb->drXTFlSize,
fs/hfs/btree.c
54
if (HFS_I(tree->inode)->alloc_blocks >
fs/hfs/btree.c
55
HFS_I(tree->inode)->first_blocks) {
fs/hfs/btree.c
57
unlock_new_inode(tree->inode);
fs/hfs/btree.c
61
tree->inode->i_mapping->a_ops = &hfs_btree_aops;
fs/hfs/btree.c
64
hfs_inode_read_fork(tree->inode, mdb->drCTExtRec, mdb->drCTFlSize,
fs/hfs/btree.c
67
if (!HFS_I(tree->inode)->first_blocks) {
fs/hfs/btree.c
69
unlock_new_inode(tree->inode);
fs/hfs/btree.c
73
tree->inode->i_mapping->a_ops = &hfs_btree_aops;
fs/hfs/btree.c
79
unlock_new_inode(tree->inode);
fs/hfs/btree.c
81
mapping = tree->inode->i_mapping;
fs/hfs/btree.c
88
dblock = hfs_ext_find_block(HFS_I(tree->inode)->first_extents, 0);
fs/hfs/btree.h
111
extern struct hfs_bnode *hfs_bnode_findhash(struct hfs_btree *tree, u32 cnid);
fs/hfs/btree.h
112
extern struct hfs_bnode *hfs_bnode_find(struct hfs_btree *tree, u32 num);
fs/hfs/btree.h
115
extern struct hfs_bnode *hfs_bnode_create(struct hfs_btree *tree, u32 num);
fs/hfs/btree.h
126
extern int hfs_find_init(struct hfs_btree *tree, struct hfs_find_data *fd);
fs/hfs/btree.h
54
struct hfs_btree *tree;
fs/hfs/btree.h
80
struct hfs_btree *tree;
fs/hfs/btree.h
91
extern void hfs_btree_close(struct hfs_btree *tree);
fs/hfs/btree.h
92
extern void hfs_btree_write(struct hfs_btree *tree);
fs/hfs/btree.h
93
extern int hfs_bmap_reserve(struct hfs_btree *tree, u32 rsvd_nodes);
fs/hfs/btree.h
94
extern struct hfs_bnode *hfs_bmap_alloc(struct hfs_btree *tree);
fs/hfs/catalog.c
104
err = hfs_bmap_reserve(fd.tree, 2 * fd.tree->depth);
fs/hfs/catalog.c
372
if (fd.tree->keycmp(fd.search_key, (void *)&rd->key) < 0)
fs/hfs/catalog.c
433
err = hfs_bmap_reserve(src_fd.tree, 2 * src_fd.tree->depth);
fs/hfs/extent.c
121
res = hfs_bmap_reserve(fd->tree, fd->tree->depth + 1);
fs/hfs/inode.c
104
spin_unlock(&tree->hash_lock);
fs/hfs/inode.c
106
nidx = folio->index << (PAGE_SHIFT - tree->node_size_shift);
fs/hfs/inode.c
107
i = 1 << (PAGE_SHIFT - tree->node_size_shift);
fs/hfs/inode.c
108
spin_lock(&tree->hash_lock);
fs/hfs/inode.c
110
node = hfs_bnode_findhash(tree, nidx++);
fs/hfs/inode.c
119
} while (--i && nidx < tree->node_count);
fs/hfs/inode.c
120
spin_unlock(&tree->hash_lock);
fs/hfs/inode.c
71
struct hfs_btree *tree;
fs/hfs/inode.c
79
tree = HFS_SB(sb)->ext_tree;
fs/hfs/inode.c
82
tree = HFS_SB(sb)->cat_tree;
fs/hfs/inode.c
89
if (!tree)
fs/hfs/inode.c
92
if (tree->node_size >= PAGE_SIZE) {
fs/hfs/inode.c
93
nidx = folio->index >> (tree->node_size_shift - PAGE_SHIFT);
fs/hfs/inode.c
94
spin_lock(&tree->hash_lock);
fs/hfs/inode.c
95
node = hfs_bnode_findhash(tree, nidx);
fs/hfsplus/attributes.c
275
err = hfs_bmap_reserve(fd.tree, fd.tree->depth + 1);
fs/hfsplus/attributes.c
383
err = hfs_bmap_reserve(fd.tree, fd.tree->depth);
fs/hfsplus/attributes.c
456
err = hfs_bmap_reserve(fd.tree, fd.tree->depth + 1);
fs/hfsplus/bfind.c
15
int hfs_find_init(struct hfs_btree *tree, struct hfs_find_data *fd)
fs/hfsplus/bfind.c
155
struct hfs_btree *tree;
fs/hfsplus/bfind.c
167
tree = fd->tree;
fs/hfsplus/bfind.c
171
nidx = tree->root;
fs/hfsplus/bfind.c
174
height = tree->depth;
fs/hfsplus/bfind.c
178
bnode = hfs_bnode_find(tree, nidx);
fs/hfsplus/bfind.c
19
fd->tree = tree;
fs/hfsplus/bfind.c
21
ptr = kzalloc(tree->max_key_len * 2 + 4, GFP_KERNEL);
fs/hfsplus/bfind.c
228
struct hfs_btree *tree;
fs/hfsplus/bfind.c
234
tree = bnode->tree;
fs/hfsplus/bfind.c
247
bnode = hfs_bnode_find(tree, idx);
fs/hfsplus/bfind.c
25
fd->key = ptr + tree->max_key_len + 2;
fs/hfsplus/bfind.c
265
bnode = hfs_bnode_find(tree, idx);
fs/hfsplus/bfind.c
27
tree->cnid, __builtin_return_address(0));
fs/hfsplus/bfind.c
28
mutex_lock_nested(&tree->tree_lock,
fs/hfsplus/bfind.c
29
hfsplus_btree_lock_class(tree));
fs/hfsplus/bfind.c
38
fd->tree->cnid, __builtin_return_address(0));
fs/hfsplus/bfind.c
39
mutex_unlock(&fd->tree->tree_lock);
fs/hfsplus/bfind.c
40
fd->tree = NULL;
fs/hfsplus/bfind.c
52
if (bnode->tree->cnid == HFSPLUS_EXT_CNID) {
fs/hfsplus/bfind.c
55
} else if (bnode->tree->cnid == HFSPLUS_CAT_CNID) {
fs/hfsplus/bfind.c
58
} else if (bnode->tree->cnid == HFSPLUS_ATTR_CNID) {
fs/hfsplus/bfind.c
89
cmpval = bnode->tree->keycmp(fd->key, fd->search_key);
fs/hfsplus/bnode.c
107
node->tree->node_size, off, len);
fs/hfsplus/bnode.c
149
node->tree->node_size, off, len);
fs/hfsplus/bnode.c
36
node->tree->node_size, off, len);
fs/hfsplus/bnode.c
360
off = node->tree->node_size - 2;
fs/hfsplus/bnode.c
367
if (node->tree->attributes & HFS_TREE_VARIDXKEYS ||
fs/hfsplus/bnode.c
368
node->tree->cnid == HFSPLUS_ATTR_CNID)
fs/hfsplus/bnode.c
371
tmp = node->tree->max_key_len + 2;
fs/hfsplus/bnode.c
387
struct hfs_btree *tree;
fs/hfsplus/bnode.c
391
tree = node->tree;
fs/hfsplus/bnode.c
393
tmp = hfs_bnode_find(tree, node->prev);
fs/hfsplus/bnode.c
402
tree->leaf_head = node->next;
fs/hfsplus/bnode.c
405
tmp = hfs_bnode_find(tree, node->next);
fs/hfsplus/bnode.c
414
tree->leaf_tail = node->prev;
fs/hfsplus/bnode.c
420
tree->root = 0;
fs/hfsplus/bnode.c
421
tree->depth = 0;
fs/hfsplus/bnode.c
433
struct hfs_bnode *hfs_bnode_findhash(struct hfs_btree *tree, u32 cnid)
fs/hfsplus/bnode.c
437
if (cnid >= tree->node_count) {
fs/hfsplus/bnode.c
443
for (node = tree->node_hash[hfs_bnode_hash(cnid)];
fs/hfsplus/bnode.c
450
static struct hfs_bnode *__hfs_bnode_create(struct hfs_btree *tree, u32 cnid)
fs/hfsplus/bnode.c
458
if (cnid >= tree->node_count) {
fs/hfsplus/bnode.c
464
size = sizeof(struct hfs_bnode) + tree->pages_per_bnode *
fs/hfsplus/bnode.c
469
node->tree = tree;
fs/hfsplus/bnode.c
474
node->tree->cnid, node->this);
fs/hfsplus/bnode.c
476
spin_lock(&tree->hash_lock);
fs/hfsplus/bnode.c
477
node2 = hfs_bnode_findhash(tree, cnid);
fs/hfsplus/bnode.c
480
node->next_hash = tree->node_hash[hash];
fs/hfsplus/bnode.c
481
tree->node_hash[hash] = node;
fs/hfsplus/bnode.c
482
tree->node_hash_cnt++;
fs/hfsplus/bnode.c
485
spin_unlock(&tree->hash_lock);
fs/hfsplus/bnode.c
491
spin_unlock(&tree->hash_lock);
fs/hfsplus/bnode.c
493
mapping = tree->inode->i_mapping;
fs/hfsplus/bnode.c
494
off = (loff_t)cnid << tree->node_size_shift;
fs/hfsplus/bnode.c
497
for (i = 0; i < tree->pages_per_bnode; block++, i++) {
fs/hfsplus/bnode.c
515
node->tree->cnid, node->this, atomic_read(&node->refcnt));
fs/hfsplus/bnode.c
516
for (p = &node->tree->node_hash[hfs_bnode_hash(node->this)];
fs/hfsplus/bnode.c
521
node->tree->node_hash_cnt--;
fs/hfsplus/bnode.c
525
struct hfs_bnode *hfs_bnode_find(struct hfs_btree *tree, u32 num)
fs/hfsplus/bnode.c
532
spin_lock(&tree->hash_lock);
fs/hfsplus/bnode.c
533
node = hfs_bnode_findhash(tree, num);
fs/hfsplus/bnode.c
536
spin_unlock(&tree->hash_lock);
fs/hfsplus/bnode.c
543
spin_unlock(&tree->hash_lock);
fs/hfsplus/bnode.c
544
node = __hfs_bnode_create(tree, num);
fs/hfsplus/bnode.c
572
if (node->height <= 1 || node->height > tree->depth)
fs/hfsplus/bnode.c
579
rec_off = tree->node_size - 2;
fs/hfsplus/bnode.c
587
next_off > tree->node_size ||
fs/hfsplus/bnode.c
614
for (i = 0; i < node->tree->pages_per_bnode; i++)
fs/hfsplus/bnode.c
620
struct hfs_bnode *hfs_bnode_create(struct hfs_btree *tree, u32 num)
fs/hfsplus/bnode.c
626
spin_lock(&tree->hash_lock);
fs/hfsplus/bnode.c
627
node = hfs_bnode_findhash(tree, num);
fs/hfsplus/bnode.c
628
spin_unlock(&tree->hash_lock);
fs/hfsplus/bnode.c
634
node = __hfs_bnode_create(tree, num);
fs/hfsplus/bnode.c
644
min_t(int, PAGE_SIZE, tree->node_size));
fs/hfsplus/bnode.c
646
for (i = 1; i < tree->pages_per_bnode; i++) {
fs/hfsplus/bnode.c
661
node->tree->cnid, node->this,
fs/hfsplus/bnode.c
670
struct hfs_btree *tree = node->tree;
fs/hfsplus/bnode.c
674
node->tree->cnid, node->this,
fs/hfsplus/bnode.c
677
if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock))
fs/hfsplus/bnode.c
679
for (i = 0; i < tree->pages_per_bnode; i++) {
fs/hfsplus/bnode.c
687
spin_unlock(&tree->hash_lock);
fs/hfsplus/bnode.c
688
if (hfs_bnode_need_zeroout(tree))
fs/hfsplus/bnode.c
689
hfs_bnode_clear(node, 0, tree->node_size);
fs/hfsplus/bnode.c
694
spin_unlock(&tree->hash_lock);
fs/hfsplus/bnode.c
702
bool hfs_bnode_need_zeroout(struct hfs_btree *tree)
fs/hfsplus/bnode.c
704
struct super_block *sb = tree->inode->i_sb;
fs/hfsplus/bnode.c
74
struct hfs_btree *tree;
fs/hfsplus/bnode.c
77
tree = node->tree;
fs/hfsplus/bnode.c
79
tree->attributes & HFS_TREE_VARIDXKEYS ||
fs/hfsplus/bnode.c
80
node->tree->cnid == HFSPLUS_ATTR_CNID)
fs/hfsplus/bnode.c
83
key_len = tree->max_key_len + 2;
fs/hfsplus/brec.c
106
tree->leaf_count++;
fs/hfsplus/brec.c
107
mark_inode_dirty(tree->inode);
fs/hfsplus/brec.c
117
idx_rec_off = tree->node_size - (rec + 1) * 2;
fs/hfsplus/brec.c
148
hfs_btree_inc_height(tree);
fs/hfsplus/brec.c
149
new_node->parent = tree->root;
fs/hfsplus/brec.c
151
fd->bnode = hfs_bnode_find(tree, new_node->parent);
fs/hfsplus/brec.c
165
if ((tree->attributes & HFS_TREE_VARIDXKEYS) ||
fs/hfsplus/brec.c
166
(tree->cnid == HFSPLUS_ATTR_CNID))
fs/hfsplus/brec.c
170
cpu_to_be16(tree->max_key_len);
fs/hfsplus/brec.c
171
key_len = tree->max_key_len + 2;
fs/hfsplus/brec.c
181
struct hfs_btree *tree;
fs/hfsplus/brec.c
185
tree = fd->tree;
fs/hfsplus/brec.c
188
rec_off = tree->node_size - (fd->record + 2) * 2;
fs/hfsplus/brec.c
189
end_off = tree->node_size - (node->num_recs + 1) * 2;
fs/hfsplus/brec.c
192
tree->leaf_count--;
fs/hfsplus/brec.c
193
mark_inode_dirty(tree->inode);
fs/hfsplus/brec.c
202
parent = hfs_bnode_find(tree, node->parent);
fs/hfsplus/brec.c
237
struct hfs_btree *tree;
fs/hfsplus/brec.c
243
tree = fd->tree;
fs/hfsplus/brec.c
245
new_node = hfs_bmap_alloc(tree);
fs/hfsplus/brec.c
25
dataoff = node->tree->node_size - (rec + 2) * 2;
fs/hfsplus/brec.c
258
next_node = hfs_bnode_find(tree, node->next);
fs/hfsplus/brec.c
268
size = tree->node_size / 2 - node->num_recs * 2 - 14;
fs/hfsplus/brec.c
269
old_rec_off = tree->node_size - 4;
fs/hfsplus/brec.c
304
new_rec_off = tree->node_size - 2;
fs/hfsplus/brec.c
343
} else if (node->this == tree->leaf_tail) {
fs/hfsplus/brec.c
345
tree->leaf_tail = new_node->this;
fs/hfsplus/brec.c
346
mark_inode_dirty(tree->inode);
fs/hfsplus/brec.c
358
struct hfs_btree *tree;
fs/hfsplus/brec.c
364
tree = fd->tree;
fs/hfsplus/brec.c
371
parent = hfs_bnode_find(tree, node->parent);
fs/hfsplus/brec.c
381
if ((tree->attributes & HFS_TREE_VARIDXKEYS) ||
fs/hfsplus/brec.c
382
(tree->cnid == HFSPLUS_ATTR_CNID))
fs/hfsplus/brec.c
385
fd->keylength = newkeylen = tree->max_key_len + 2;
fs/hfsplus/brec.c
389
rec_off = tree->node_size - (rec + 2) * 2;
fs/hfsplus/brec.c
390
end_rec_off = tree->node_size - (parent->num_recs + 1) * 2;
fs/hfsplus/brec.c
40
!(node->tree->attributes & HFS_TREE_VARIDXKEYS) &&
fs/hfsplus/brec.c
405
rec_off = tree->node_size - (rec + 2) * 2;
fs/hfsplus/brec.c
406
end_rec_off = tree->node_size -
fs/hfsplus/brec.c
41
(node->tree->cnid != HFSPLUS_ATTR_CNID)) {
fs/hfsplus/brec.c
42
retval = node->tree->max_key_len + 2;
fs/hfsplus/brec.c
433
hfs_btree_inc_height(tree);
fs/hfsplus/brec.c
434
new_node->parent = tree->root;
fs/hfsplus/brec.c
436
fd->bnode = hfs_bnode_find(tree, new_node->parent);
fs/hfsplus/brec.c
45
node->tree->node_size - (rec + 1) * 2);
fs/hfsplus/brec.c
462
static int hfs_btree_inc_height(struct hfs_btree *tree)
fs/hfsplus/brec.c
470
if (tree->root) {
fs/hfsplus/brec.c
471
node = hfs_bnode_find(tree, tree->root);
fs/hfsplus/brec.c
475
new_node = hfs_bmap_alloc(tree);
fs/hfsplus/brec.c
48
if (recoff > node->tree->node_size - 2) {
fs/hfsplus/brec.c
481
tree->root = new_node->this;
fs/hfsplus/brec.c
482
if (!tree->depth) {
fs/hfsplus/brec.c
483
tree->leaf_head = tree->leaf_tail = new_node->this;
fs/hfsplus/brec.c
493
new_node->height = ++tree->depth;
fs/hfsplus/brec.c
503
rec = tree->node_size - 2;
fs/hfsplus/brec.c
508
node->parent = tree->root;
fs/hfsplus/brec.c
510
tree->attributes & HFS_TREE_VARIDXKEYS ||
fs/hfsplus/brec.c
511
tree->cnid == HFSPLUS_ATTR_CNID)
fs/hfsplus/brec.c
514
key_size = tree->max_key_len + 2;
fs/hfsplus/brec.c
517
if (!(tree->attributes & HFS_TREE_VARIDXKEYS) &&
fs/hfsplus/brec.c
518
(tree->cnid != HFSPLUS_ATTR_CNID)) {
fs/hfsplus/brec.c
519
key_size = tree->max_key_len + 2;
fs/hfsplus/brec.c
520
hfs_bnode_write_u16(new_node, 14, tree->max_key_len);
fs/hfsplus/brec.c
531
mark_inode_dirty(tree->inode);
fs/hfsplus/brec.c
54
if (retval > node->tree->max_key_len + 2) {
fs/hfsplus/brec.c
65
struct hfs_btree *tree;
fs/hfsplus/brec.c
72
tree = fd->tree;
fs/hfsplus/brec.c
74
if (!tree->root)
fs/hfsplus/brec.c
75
hfs_btree_inc_height(tree);
fs/hfsplus/brec.c
76
node = hfs_bnode_find(tree, tree->leaf_head);
fs/hfsplus/brec.c
92
end_rec_off = tree->node_size - (node->num_recs + 1) * 2;
fs/hfsplus/btree.c
135
struct hfs_btree *tree;
fs/hfsplus/btree.c
142
tree = kzalloc_obj(*tree);
fs/hfsplus/btree.c
143
if (!tree)
fs/hfsplus/btree.c
146
mutex_init(&tree->tree_lock);
fs/hfsplus/btree.c
147
spin_lock_init(&tree->hash_lock);
fs/hfsplus/btree.c
148
tree->sb = sb;
fs/hfsplus/btree.c
149
tree->cnid = id;
fs/hfsplus/btree.c
153
tree->inode = inode;
fs/hfsplus/btree.c
155
if (!HFSPLUS_I(tree->inode)->first_blocks) {
fs/hfsplus/btree.c
160
mapping = tree->inode->i_mapping;
fs/hfsplus/btree.c
168
tree->root = be32_to_cpu(head->root);
fs/hfsplus/btree.c
169
tree->leaf_count = be32_to_cpu(head->leaf_count);
fs/hfsplus/btree.c
170
tree->leaf_head = be32_to_cpu(head->leaf_head);
fs/hfsplus/btree.c
171
tree->leaf_tail = be32_to_cpu(head->leaf_tail);
fs/hfsplus/btree.c
172
tree->node_count = be32_to_cpu(head->node_count);
fs/hfsplus/btree.c
173
tree->free_nodes = be32_to_cpu(head->free_nodes);
fs/hfsplus/btree.c
174
tree->attributes = be32_to_cpu(head->attributes);
fs/hfsplus/btree.c
175
tree->node_size = be16_to_cpu(head->node_size);
fs/hfsplus/btree.c
176
tree->max_key_len = be16_to_cpu(head->max_key_len);
fs/hfsplus/btree.c
177
tree->depth = be16_to_cpu(head->depth);
fs/hfsplus/btree.c
182
if (tree->max_key_len != HFSPLUS_EXT_KEYLEN - sizeof(u16)) {
fs/hfsplus/btree.c
184
tree->max_key_len);
fs/hfsplus/btree.c
187
if (tree->attributes & HFS_TREE_VARIDXKEYS) {
fs/hfsplus/btree.c
192
tree->keycmp = hfsplus_ext_cmp_key;
fs/hfsplus/btree.c
195
if (tree->max_key_len != HFSPLUS_CAT_KEYLEN - sizeof(u16)) {
fs/hfsplus/btree.c
197
tree->max_key_len);
fs/hfsplus/btree.c
200
if (!(tree->attributes & HFS_TREE_VARIDXKEYS)) {
fs/hfsplus/btree.c
207
tree->keycmp = hfsplus_cat_bin_cmp_key;
fs/hfsplus/btree.c
209
tree->keycmp = hfsplus_cat_case_cmp_key;
fs/hfsplus/btree.c
214
if (tree->max_key_len != HFSPLUS_ATTR_KEYLEN - sizeof(u16)) {
fs/hfsplus/btree.c
216
tree->max_key_len);
fs/hfsplus/btree.c
219
tree->keycmp = hfsplus_attr_bin_cmp_key;
fs/hfsplus/btree.c
226
if (!(tree->attributes & HFS_TREE_BIGKEYS)) {
fs/hfsplus/btree.c
231
size = tree->node_size;
fs/hfsplus/btree.c
234
if (!tree->node_count)
fs/hfsplus/btree.c
237
tree->node_size_shift = ffs(size) - 1;
fs/hfsplus/btree.c
239
tree->pages_per_bnode =
fs/hfsplus/btree.c
240
(tree->node_size + PAGE_SIZE - 1) >>
fs/hfsplus/btree.c
245
return tree;
fs/hfsplus/btree.c
251
tree->inode->i_mapping->a_ops = &hfsplus_aops;
fs/hfsplus/btree.c
252
iput(tree->inode);
fs/hfsplus/btree.c
254
kfree(tree);
fs/hfsplus/btree.c
259
void hfs_btree_close(struct hfs_btree *tree)
fs/hfsplus/btree.c
264
if (!tree)
fs/hfsplus/btree.c
268
while ((node = tree->node_hash[i])) {
fs/hfsplus/btree.c
269
tree->node_hash[i] = node->next_hash;
fs/hfsplus/btree.c
273
node->tree->cnid, node->this,
fs/hfsplus/btree.c
276
tree->node_hash_cnt--;
fs/hfsplus/btree.c
279
iput(tree->inode);
fs/hfsplus/btree.c
280
kfree(tree);
fs/hfsplus/btree.c
283
int hfs_btree_write(struct hfs_btree *tree)
fs/hfsplus/btree.c
289
node = hfs_bnode_find(tree, 0);
fs/hfsplus/btree.c
298
head->root = cpu_to_be32(tree->root);
fs/hfsplus/btree.c
299
head->leaf_count = cpu_to_be32(tree->leaf_count);
fs/hfsplus/btree.c
300
head->leaf_head = cpu_to_be32(tree->leaf_head);
fs/hfsplus/btree.c
301
head->leaf_tail = cpu_to_be32(tree->leaf_tail);
fs/hfsplus/btree.c
302
head->node_count = cpu_to_be32(tree->node_count);
fs/hfsplus/btree.c
303
head->free_nodes = cpu_to_be32(tree->free_nodes);
fs/hfsplus/btree.c
304
head->attributes = cpu_to_be32(tree->attributes);
fs/hfsplus/btree.c
305
head->depth = cpu_to_be16(tree->depth);
fs/hfsplus/btree.c
315
struct hfs_btree *tree = prev->tree;
fs/hfsplus/btree.c
320
node = hfs_bnode_create(tree, idx);
fs/hfsplus/btree.c
324
tree->free_nodes--;
fs/hfsplus/btree.c
331
hfs_bnode_clear(node, 0, tree->node_size);
fs/hfsplus/btree.c
340
hfs_bnode_write_u16(node, tree->node_size - 2, 14);
fs/hfsplus/btree.c
341
hfs_bnode_write_u16(node, tree->node_size - 4, tree->node_size - 6);
fs/hfsplus/btree.c
347
int hfs_bmap_reserve(struct hfs_btree *tree, u32 rsvd_nodes)
fs/hfsplus/btree.c
349
struct inode *inode = tree->inode;
fs/hfsplus/btree.c
357
while (tree->free_nodes < rsvd_nodes) {
fs/hfsplus/btree.c
358
res = hfsplus_file_extend(inode, hfs_bnode_need_zeroout(tree));
fs/hfsplus/btree.c
363
HFSPLUS_SB(tree->sb)->alloc_blksz_shift;
fs/hfsplus/btree.c
365
hip->alloc_blocks << HFSPLUS_SB(tree->sb)->fs_shift;
fs/hfsplus/btree.c
367
count = inode->i_size >> tree->node_size_shift;
fs/hfsplus/btree.c
368
tree->free_nodes += count - tree->node_count;
fs/hfsplus/btree.c
369
tree->node_count = count;
fs/hfsplus/btree.c
374
struct hfs_bnode *hfs_bmap_alloc(struct hfs_btree *tree)
fs/hfsplus/btree.c
385
res = hfs_bmap_reserve(tree, 1);
fs/hfsplus/btree.c
390
node = hfs_bnode_find(tree, nidx);
fs/hfsplus/btree.c
418
tree->free_nodes--;
fs/hfsplus/btree.c
419
mark_inode_dirty(tree->inode);
fs/hfsplus/btree.c
421
return hfs_bnode_create(tree,
fs/hfsplus/btree.c
440
next_node = hfs_bnode_find(tree, nidx);
fs/hfsplus/btree.c
457
struct hfs_btree *tree;
fs/hfsplus/btree.c
465
tree = node->tree;
fs/hfsplus/btree.c
467
node = hfs_bnode_find(tree, 0);
fs/hfsplus/btree.c
485
node = hfs_bnode_find(tree, i);
fs/hfsplus/btree.c
516
tree->free_nodes++;
fs/hfsplus/btree.c
517
mark_inode_dirty(tree->inode);
fs/hfsplus/catalog.c
272
err = hfs_bmap_reserve(fd.tree, 2 * fd.tree->depth);
fs/hfsplus/catalog.c
348
err = hfs_bmap_reserve(fd.tree, 2 * (int)fd.tree->depth - 2);
fs/hfsplus/catalog.c
399
if (fd.tree->keycmp(fd.search_key, (void *)&rd->key) < 0)
fs/hfsplus/catalog.c
456
err = hfs_bmap_reserve(src_fd.tree, 4 * (int)src_fd.tree->depth - 1);
fs/hfsplus/extents.c
104
res = hfs_bmap_reserve(fd->tree, fd->tree->depth + 1);
fs/hfsplus/extents.c
424
mutex_unlock(&fd.tree->tree_lock);
fs/hfsplus/extents.c
428
mutex_lock_nested(&fd.tree->tree_lock,
fs/hfsplus/extents.c
429
hfsplus_btree_lock_class(fd.tree));
fs/hfsplus/extents.c
587
mutex_unlock(&fd.tree->tree_lock);
fs/hfsplus/extents.c
592
mutex_lock_nested(&fd.tree->tree_lock,
fs/hfsplus/extents.c
593
hfsplus_btree_lock_class(fd.tree));
fs/hfsplus/extents.c
603
mutex_unlock(&fd.tree->tree_lock);
fs/hfsplus/extents.c
607
mutex_lock_nested(&fd.tree->tree_lock,
fs/hfsplus/extents.c
608
hfsplus_btree_lock_class(fd.tree));
fs/hfsplus/hfsplus_fs.h
259
struct hfs_btree *tree;
fs/hfsplus/hfsplus_fs.h
360
void hfs_btree_close(struct hfs_btree *tree);
fs/hfsplus/hfsplus_fs.h
361
int hfs_btree_write(struct hfs_btree *tree);
fs/hfsplus/hfsplus_fs.h
362
int hfs_bmap_reserve(struct hfs_btree *tree, u32 rsvd_nodes);
fs/hfsplus/hfsplus_fs.h
363
struct hfs_bnode *hfs_bmap_alloc(struct hfs_btree *tree);
fs/hfsplus/hfsplus_fs.h
379
struct hfs_bnode *hfs_bnode_findhash(struct hfs_btree *tree, u32 cnid);
fs/hfsplus/hfsplus_fs.h
381
struct hfs_bnode *hfs_bnode_find(struct hfs_btree *tree, u32 num);
fs/hfsplus/hfsplus_fs.h
383
struct hfs_bnode *hfs_bnode_create(struct hfs_btree *tree, u32 num);
fs/hfsplus/hfsplus_fs.h
386
bool hfs_bnode_need_zeroout(struct hfs_btree *tree);
fs/hfsplus/hfsplus_fs.h
395
int hfs_find_init(struct hfs_btree *tree, struct hfs_find_data *fd);
fs/hfsplus/hfsplus_fs.h
535
hfsplus_btree_lock_class(struct hfs_btree *tree)
fs/hfsplus/hfsplus_fs.h
539
switch (tree->cnid) {
fs/hfsplus/hfsplus_fs.h
558
bool is_valid = off < node->tree->node_size;
fs/hfsplus/hfsplus_fs.h
565
node->tree->node_size, off);
fs/hfsplus/hfsplus_fs.h
579
node_size = node->tree->node_size;
fs/hfsplus/hfsplus_fs.h
589
node->tree->node_size, off, len, new_len);
fs/hfsplus/hfsplus_fs.h
71
struct hfs_btree *tree;
fs/hfsplus/inode.c
101
spin_unlock(&tree->hash_lock);
fs/hfsplus/inode.c
104
(PAGE_SHIFT - tree->node_size_shift);
fs/hfsplus/inode.c
105
i = 1 << (PAGE_SHIFT - tree->node_size_shift);
fs/hfsplus/inode.c
106
spin_lock(&tree->hash_lock);
fs/hfsplus/inode.c
108
node = hfs_bnode_findhash(tree, nidx++);
fs/hfsplus/inode.c
117
} while (--i && nidx < tree->node_count);
fs/hfsplus/inode.c
118
spin_unlock(&tree->hash_lock);
fs/hfsplus/inode.c
637
struct hfs_btree *tree = HFSPLUS_SB(inode->i_sb)->cat_tree;
fs/hfsplus/inode.c
650
if (hfs_find_init(tree, &fd))
fs/hfsplus/inode.c
66
struct hfs_btree *tree;
fs/hfsplus/inode.c
717
res = hfs_btree_write(tree);
fs/hfsplus/inode.c
74
tree = HFSPLUS_SB(sb)->ext_tree;
fs/hfsplus/inode.c
77
tree = HFSPLUS_SB(sb)->cat_tree;
fs/hfsplus/inode.c
80
tree = HFSPLUS_SB(sb)->attr_tree;
fs/hfsplus/inode.c
86
if (!tree)
fs/hfsplus/inode.c
88
if (tree->node_size >= PAGE_SIZE) {
fs/hfsplus/inode.c
90
(tree->node_size_shift - PAGE_SHIFT);
fs/hfsplus/inode.c
91
spin_lock(&tree->hash_lock);
fs/hfsplus/inode.c
92
node = hfs_bnode_findhash(tree, nidx);
fs/hfsplus/super.c
125
struct hfs_btree *tree = NULL;
fs/hfsplus/super.c
130
tree = sbi->ext_tree;
fs/hfsplus/super.c
134
tree = sbi->cat_tree;
fs/hfsplus/super.c
144
tree = sbi->attr_tree;
fs/hfsplus/super.c
155
if (tree) {
fs/hfsplus/super.c
156
int err = hfs_btree_write(tree);
fs/hfsplus/xattr.c
759
if (key_len == 0 || key_len > fd.tree->max_key_len) {
fs/jfs/jfs_dmap.c
1079
if (dp->tree.leafidx != cpu_to_le32(LEAFIND)) {
fs/jfs/jfs_dmap.c
1086
leaf = dp->tree.stree + le32_to_cpu(dp->tree.leafidx);
fs/jfs/jfs_dmap.c
1209
if (dp->tree.leafidx != cpu_to_le32(LEAFIND)) {
fs/jfs/jfs_dmap.c
1214
leaf = dp->tree.stree + le32_to_cpu(dp->tree.leafidx);
fs/jfs/jfs_dmap.c
1818
if (dp->tree.budmin < 0) {
fs/jfs/jfs_dmap.c
1854
if (dp->tree.stree[ROOT] != L2BPERDMAP) {
fs/jfs/jfs_dmap.c
1969
if (dbFindLeaf((dmtree_t *) &dp->tree, l2nb, &leafidx, false))
fs/jfs/jfs_dmap.c
1984
if (dp->tree.stree[leafidx + LEAFIND] < BUDMIN)
fs/jfs/jfs_dmap.c
2031
oldroot = dp->tree.stree[ROOT];
fs/jfs/jfs_dmap.c
2037
if (dp->tree.stree[ROOT] == oldroot)
fs/jfs/jfs_dmap.c
2044
if ((rc = dbAdjCtl(bmp, blkno, dp->tree.stree[ROOT], 1, 0)))
fs/jfs/jfs_dmap.c
2086
oldroot = dp->tree.stree[ROOT];
fs/jfs/jfs_dmap.c
2092
if (rc || (dp->tree.stree[ROOT] == oldroot))
fs/jfs/jfs_dmap.c
2099
if ((rc = dbAdjCtl(bmp, blkno, dp->tree.stree[ROOT], 0, 0))) {
fs/jfs/jfs_dmap.c
2107
if (dp->tree.stree[word] == NOFREE)
fs/jfs/jfs_dmap.c
2108
dbBackSplit((dmtree_t *)&dp->tree, word, false);
fs/jfs/jfs_dmap.c
2143
dmtree_t *tp = (dmtree_t *) & dp->tree;
fs/jfs/jfs_dmap.c
2148
leaf = dp->tree.stree + LEAFIND;
fs/jfs/jfs_dmap.c
2290
dmtree_t *tp = (dmtree_t *) & dp->tree;
fs/jfs/jfs_dmap.c
3252
struct dmaptree *tp = (struct dmaptree *) & dp->tree;
fs/jfs/jfs_dmap.c
3839
tp = &dp->tree;
fs/jfs/jfs_dmap.h
151
struct dmaptree tree; /* 360: dmap tree */
fs/ntfs3/frecord.c
20
static struct mft_inode *ni_ins_mi(struct ntfs_inode *ni, struct rb_root *tree,
fs/ntfs3/frecord.c
23
struct rb_node **p = &tree->rb_node;
fs/ntfs3/frecord.c
43
rb_insert_color(ins, tree);
fs/ocfs2/dlmglue.c
4151
struct ocfs2_refcount_tree *tree =
fs/ocfs2/dlmglue.c
4154
return ocfs2_ci_checkpointed(&tree->rf_ci, lockres, new_level);
fs/ocfs2/dlmglue.c
4160
struct ocfs2_refcount_tree *tree =
fs/ocfs2/dlmglue.c
4163
ocfs2_metadata_cache_purge(&tree->rf_ci);
fs/ocfs2/refcounttree.c
199
struct ocfs2_refcount_tree *tree = NULL;
fs/ocfs2/refcounttree.c
202
tree = rb_entry(n, struct ocfs2_refcount_tree, rf_node);
fs/ocfs2/refcounttree.c
204
if (blkno < tree->rf_blkno)
fs/ocfs2/refcounttree.c
206
else if (blkno > tree->rf_blkno)
fs/ocfs2/refcounttree.c
209
return tree;
fs/ocfs2/refcounttree.c
2290
struct ocfs2_refcount_tree *tree;
fs/ocfs2/refcounttree.c
2300
ret = ocfs2_get_refcount_tree(OCFS2_SB(inode->i_sb), ref_blkno, &tree);
fs/ocfs2/refcounttree.c
2306
ret = ocfs2_read_refcount_block(&tree->rf_ci, tree->rf_blkno,
fs/ocfs2/refcounttree.c
2313
ret = __ocfs2_decrease_refcount(handle, &tree->rf_ci, ref_root_bh,
fs/ocfs2/refcounttree.c
246
static void ocfs2_free_refcount_tree(struct ocfs2_refcount_tree *tree)
fs/ocfs2/refcounttree.c
248
ocfs2_metadata_cache_exit(&tree->rf_ci);
fs/ocfs2/refcounttree.c
249
ocfs2_simple_drop_lockres(OCFS2_SB(tree->rf_sb), &tree->rf_lockres);
fs/ocfs2/refcounttree.c
250
ocfs2_lock_res_free(&tree->rf_lockres);
fs/ocfs2/refcounttree.c
251
kfree(tree);
fs/ocfs2/refcounttree.c
2523
struct ocfs2_refcount_tree *tree;
fs/ocfs2/refcounttree.c
2535
refcount_loc, &tree);
fs/ocfs2/refcounttree.c
2541
ret = ocfs2_read_refcount_block(&tree->rf_ci, refcount_loc,
fs/ocfs2/refcounttree.c
2549
&tree->rf_ci,
fs/ocfs2/refcounttree.c
256
struct ocfs2_refcount_tree *tree)
fs/ocfs2/refcounttree.c
258
rb_erase(&tree->rf_node, &osb->osb_rf_lock_tree);
fs/ocfs2/refcounttree.c
259
if (osb->osb_ref_tree_lru && osb->osb_ref_tree_lru == tree)
fs/ocfs2/refcounttree.c
264
struct ocfs2_refcount_tree *tree)
fs/ocfs2/refcounttree.c
267
ocfs2_erase_refcount_tree_from_list_no_lock(osb, tree);
fs/ocfs2/refcounttree.c
273
struct ocfs2_refcount_tree *tree =
fs/ocfs2/refcounttree.c
276
ocfs2_free_refcount_tree(tree);
fs/ocfs2/refcounttree.c
280
ocfs2_refcount_tree_get(struct ocfs2_refcount_tree *tree)
fs/ocfs2/refcounttree.c
282
kref_get(&tree->rf_getcnt);
fs/ocfs2/refcounttree.c
286
ocfs2_refcount_tree_put(struct ocfs2_refcount_tree *tree)
fs/ocfs2/refcounttree.c
288
kref_put(&tree->rf_getcnt, ocfs2_kref_remove_refcount_tree);
fs/ocfs2/refcounttree.c
329
struct ocfs2_refcount_tree *tree, *new = NULL;
fs/ocfs2/refcounttree.c
336
tree = osb->osb_ref_tree_lru;
fs/ocfs2/refcounttree.c
338
tree = ocfs2_find_refcount_tree(osb, rf_blkno);
fs/ocfs2/refcounttree.c
339
if (tree)
fs/ocfs2/refcounttree.c
373
tree = ocfs2_find_refcount_tree(osb, rf_blkno);
fs/ocfs2/refcounttree.c
374
if (tree)
fs/ocfs2/refcounttree.c
379
tree = new;
fs/ocfs2/refcounttree.c
383
*ret_tree = tree;
fs/ocfs2/refcounttree.c
385
osb->osb_ref_tree_lru = tree;
fs/ocfs2/refcounttree.c
418
struct ocfs2_refcount_tree *tree, int rw)
fs/ocfs2/refcounttree.c
422
ret = ocfs2_refcount_lock(tree, rw);
fs/ocfs2/refcounttree.c
429
down_write(&tree->rf_sem);
fs/ocfs2/refcounttree.c
431
down_read(&tree->rf_sem);
fs/ocfs2/refcounttree.c
451
struct ocfs2_refcount_tree *tree = NULL;
fs/ocfs2/refcounttree.c
456
ret = ocfs2_get_refcount_tree(osb, ref_blkno, &tree);
fs/ocfs2/refcounttree.c
462
ocfs2_refcount_tree_get(tree);
fs/ocfs2/refcounttree.c
464
ret = __ocfs2_lock_refcount_tree(osb, tree, rw);
fs/ocfs2/refcounttree.c
467
ocfs2_refcount_tree_put(tree);
fs/ocfs2/refcounttree.c
471
ret = ocfs2_read_refcount_block(&tree->rf_ci, tree->rf_blkno,
fs/ocfs2/refcounttree.c
475
ocfs2_unlock_refcount_tree(osb, tree, rw);
fs/ocfs2/refcounttree.c
489
if (tree->rf_generation != le32_to_cpu(rb->rf_generation)) {
fs/ocfs2/refcounttree.c
490
if (!tree->rf_removed) {
fs/ocfs2/refcounttree.c
491
ocfs2_erase_refcount_tree_from_list(osb, tree);
fs/ocfs2/refcounttree.c
492
tree->rf_removed = 1;
fs/ocfs2/refcounttree.c
496
ocfs2_unlock_refcount_tree(osb, tree, rw);
fs/ocfs2/refcounttree.c
502
ocfs2_refcount_tree_put(tree);
fs/ocfs2/refcounttree.c
508
*ret_tree = tree;
fs/ocfs2/refcounttree.c
519
struct ocfs2_refcount_tree *tree, int rw)
fs/ocfs2/refcounttree.c
522
up_write(&tree->rf_sem);
fs/ocfs2/refcounttree.c
524
up_read(&tree->rf_sem);
fs/ocfs2/refcounttree.c
526
ocfs2_refcount_unlock(tree, rw);
fs/ocfs2/refcounttree.c
527
ocfs2_refcount_tree_put(tree);
fs/ocfs2/refcounttree.c
533
struct ocfs2_refcount_tree *tree;
fs/ocfs2/refcounttree.c
537
tree = rb_entry(node, struct ocfs2_refcount_tree, rf_node);
fs/ocfs2/refcounttree.c
540
(unsigned long long) tree->rf_blkno);
fs/ocfs2/refcounttree.c
542
rb_erase(&tree->rf_node, root);
fs/ocfs2/refcounttree.c
543
ocfs2_free_refcount_tree(tree);
fs/ocfs2/refcounttree.c
562
struct ocfs2_refcount_tree *new_tree = NULL, *tree = NULL;
fs/ocfs2/refcounttree.c
659
tree = ocfs2_find_refcount_tree(osb, first_blkno);
fs/ocfs2/refcounttree.c
667
BUG_ON(tree && tree->rf_generation == new_tree->rf_generation);
fs/ocfs2/refcounttree.c
668
if (tree)
fs/ocfs2/refcounttree.c
669
ocfs2_erase_refcount_tree_from_list_no_lock(osb, tree);
fs/ocfs2/refcounttree.c
673
if (tree)
fs/ocfs2/refcounttree.c
674
ocfs2_refcount_tree_put(tree);
fs/ocfs2/refcounttree.h
28
struct ocfs2_refcount_tree **tree,
fs/ocfs2/refcounttree.h
31
struct ocfs2_refcount_tree *tree,
fs/ocfs2/uptodate.c
157
unsigned int tree, to_purge, purged;
fs/ocfs2/uptodate.c
163
tree = !(ci->ci_flags & OCFS2_CACHE_FL_INLINE);
fs/ocfs2/uptodate.c
168
to_purge, tree);
fs/ocfs2/uptodate.c
173
if (tree)
fs/ocfs2/uptodate.c
183
if (tree && purged != to_purge)
fs/ocfs2/uptodate.c
350
struct ocfs2_meta_cache_item **tree)
fs/ocfs2/uptodate.c
365
tree[i]->c_block = ci->ci_cache.ci_array[i];
fs/ocfs2/uptodate.c
373
__ocfs2_insert_cache_tree(ci, tree[i]);
fs/ocfs2/uptodate.c
374
tree[i] = NULL;
fs/ocfs2/uptodate.c
390
struct ocfs2_meta_cache_item *tree[OCFS2_CACHE_INFO_MAX_ARRAY] =
fs/ocfs2/uptodate.c
408
tree[i] = kmem_cache_alloc(ocfs2_uptodate_cachep,
fs/ocfs2/uptodate.c
410
if (!tree[i]) {
fs/ocfs2/uptodate.c
429
ocfs2_expand_cache(ci, tree);
fs/ocfs2/uptodate.c
441
if (tree[0]) {
fs/ocfs2/uptodate.c
443
if (tree[i])
fs/ocfs2/uptodate.c
445
tree[i]);
fs/smb/client/cifsproto.h
64
void *page, const char *tree,
fs/smb/client/cifssmb.c
542
const char *tree, struct cifs_tcon *tcon,
fs/smb/client/cifssmb.c
592
cifs_strtoUTF16((__le16 *) bcc_ptr, tree,
fs/smb/client/cifssmb.c
598
strcpy(bcc_ptr, tree);
fs/smb/client/cifssmb.c
599
bcc_ptr += strlen(tree) + 1;
fs/smb/client/cifssmb.c
641
strscpy(tcon->tree_name, tree, sizeof(tcon->tree_name));
fs/smb/client/dfs.c
339
char *tree, bool islink,
fs/smb/client/dfs.c
372
scnprintf(tree, MAX_TREE_SIZE, "\\%s", share);
fs/smb/client/dfs.c
373
rc = ops->tree_connect(xid, tcon->ses, tree,
fs/smb/client/dfs.c
395
char *tree;
fs/smb/client/dfs.c
418
tree = kzalloc(MAX_TREE_SIZE, GFP_KERNEL);
fs/smb/client/dfs.c
419
if (!tree) {
fs/smb/client/dfs.c
426
scnprintf(tree, MAX_TREE_SIZE, "\\\\%s\\IPC$", server->hostname);
fs/smb/client/dfs.c
428
rc = ops->tree_connect(xid, tcon->ses, tree,
fs/smb/client/dfs.c
445
rc = tree_connect_dfs_target(xid, tcon, cifs_sb, tree, ref.server_type == DFS_TYPE_LINK,
fs/smb/client/dfs.c
450
kfree(tree);
fs/smb/client/dir.c
126
memcpy(s, tree, dfsplen);
fs/smb/client/dir.c
82
const char *tree, int tree_len,
fs/smb/client/dir.c
96
dfsplen = strnlen(tree, tree_len + 1);
fs/smb/client/smb1proto.h
35
int CIFSTCon(const unsigned int xid, struct cifs_ses *ses, const char *tree,
fs/smb/client/smb2pdu.c
2093
SMB2_tcon(const unsigned int xid, struct cifs_ses *ses, const char *tree,
fs/smb/client/smb2pdu.c
2111
if (!server || !tree)
fs/smb/client/smb2pdu.c
2118
unc_path_len = cifs_strtoUTF16(unc_path, tree, strlen(tree), cp);
fs/smb/client/smb2pdu.c
2176
trace_smb3_tcon(xid, tcon->tid, ses->Suid, tree, rc);
fs/smb/client/smb2pdu.c
2205
strscpy(tcon->tree_name, tree, sizeof(tcon->tree_name));
fs/smb/client/smb2pdu.c
2229
cifs_dbg(VFS | ONCE, "BAD_NETWORK_NAME: %s\n", tree);
fs/smb/client/smb2proto.h
130
int SMB2_tcon(const unsigned int xid, struct cifs_ses *ses, const char *tree,
fs/unicode/mkutf8data.c
1048
static int size_nodes(struct tree *tree)
fs/unicode/mkutf8data.c
1050
struct tree *next;
fs/unicode/mkutf8data.c
1070
printf("Sizing %s_%x\n", tree->type, tree->maxage);
fs/unicode/mkutf8data.c
1071
if (tree->childnode == LEAF)
fs/unicode/mkutf8data.c
1074
assert(tree->childnode == NODE);
fs/unicode/mkutf8data.c
1077
node = tree->root;
fs/unicode/mkutf8data.c
1094
next = tree->next;
fs/unicode/mkutf8data.c
1123
offset = *tree->leaf_index(tree, node->right);
fs/unicode/mkutf8data.c
1185
static void emit(struct tree *tree, unsigned char *data)
fs/unicode/mkutf8data.c
1204
index = tree->index;
fs/unicode/mkutf8data.c
1208
printf("Emitting %s_%x\n", tree->type, tree->maxage);
fs/unicode/mkutf8data.c
1209
if (tree->childnode == LEAF) {
fs/unicode/mkutf8data.c
1210
assert(tree->root);
fs/unicode/mkutf8data.c
1211
tree->leaf_emit(tree->root, data);
fs/unicode/mkutf8data.c
1212
size = tree->leaf_size(tree->root);
fs/unicode/mkutf8data.c
1218
assert(tree->childnode == NODE);
fs/unicode/mkutf8data.c
1219
node = tree->root;
fs/unicode/mkutf8data.c
1275
data = tree->leaf_emit(node->left,
fs/unicode/mkutf8data.c
1277
size = tree->leaf_size(node->left);
fs/unicode/mkutf8data.c
1292
data = tree->leaf_emit(node->right,
fs/unicode/mkutf8data.c
1294
size = tree->leaf_size(node->right);
fs/unicode/mkutf8data.c
1318
printf(" %d total\n", index - tree->index);
fs/unicode/mkutf8data.c
1354
struct tree *nfdi_tree;
fs/unicode/mkutf8data.c
1355
struct tree *nfdicf_tree;
fs/unicode/mkutf8data.c
1357
struct tree *trees;
fs/unicode/mkutf8data.c
1493
static int *nfdi_index(struct tree *tree, void *l)
fs/unicode/mkutf8data.c
1497
return &tree->leafindex[leaf->code];
fs/unicode/mkutf8data.c
1500
static int *nfdicf_index(struct tree *tree, void *l)
fs/unicode/mkutf8data.c
1504
return &tree->leafindex[leaf->code];
fs/unicode/mkutf8data.c
1623
trees = calloc(trees_count, sizeof(struct tree));
fs/unicode/mkutf8data.c
1745
static void verify(struct tree *tree)
fs/unicode/mkutf8data.c
1756
printf("Verifying %s_%x\n", tree->type, tree->maxage);
fs/unicode/mkutf8data.c
1757
nocf = strcmp(tree->type, "nfdicf");
fs/unicode/mkutf8data.c
1762
if (data->correction <= tree->maxage)
fs/unicode/mkutf8data.c
1765
leaf = utf8lookup(tree, hangul, key);
fs/unicode/mkutf8data.c
189
struct tree;
fs/unicode/mkutf8data.c
190
static utf8leaf_t *utf8nlookup(struct tree *, unsigned char *,
fs/unicode/mkutf8data.c
192
static utf8leaf_t *utf8lookup(struct tree *, unsigned char *, const char *);
fs/unicode/mkutf8data.c
2594
int utf8agemax(struct tree *, const char *);
fs/unicode/mkutf8data.c
2595
int utf8nagemax(struct tree *, const char *, size_t);
fs/unicode/mkutf8data.c
2596
int utf8agemin(struct tree *, const char *);
fs/unicode/mkutf8data.c
2597
int utf8nagemin(struct tree *, const char *, size_t);
fs/unicode/mkutf8data.c
2598
ssize_t utf8len(struct tree *, const char *);
fs/unicode/mkutf8data.c
2599
ssize_t utf8nlen(struct tree *, const char *, size_t);
fs/unicode/mkutf8data.c
2601
int utf8cursor(struct utf8cursor *, struct tree *, const char *);
fs/unicode/mkutf8data.c
2602
int utf8ncursor(struct utf8cursor *, struct tree *, const char *, size_t);
fs/unicode/mkutf8data.c
2706
static utf8leaf_t *utf8nlookup(struct tree *tree, unsigned char *hangul,
fs/unicode/mkutf8data.c
2715
if (!tree)
fs/unicode/mkutf8data.c
2720
trie = utf8data + tree->index;
fs/unicode/mkutf8data.c
2781
static utf8leaf_t *utf8lookup(struct tree *tree, unsigned char *hangul,
fs/unicode/mkutf8data.c
2784
return utf8nlookup(tree, hangul, s, (size_t)-1);
fs/unicode/mkutf8data.c
2803
int utf8agemax(struct tree *tree, const char *s)
fs/unicode/mkutf8data.c
2810
if (!tree)
fs/unicode/mkutf8data.c
2814
leaf = utf8lookup(tree, hangul, s);
fs/unicode/mkutf8data.c
2818
if (leaf_age <= tree->maxage && leaf_age > age)
fs/unicode/mkutf8data.c
2830
int utf8agemin(struct tree *tree, const char *s)
fs/unicode/mkutf8data.c
2837
if (!tree)
fs/unicode/mkutf8data.c
2839
age = tree->maxage;
fs/unicode/mkutf8data.c
2841
leaf = utf8lookup(tree, hangul, s);
fs/unicode/mkutf8data.c
2845
if (leaf_age <= tree->maxage && leaf_age < age)
fs/unicode/mkutf8data.c
2856
int utf8nagemax(struct tree *tree, const char *s, size_t len)
fs/unicode/mkutf8data.c
2863
if (!tree)
fs/unicode/mkutf8data.c
2867
leaf = utf8nlookup(tree, hangul, s, len);
fs/unicode/mkutf8data.c
2871
if (leaf_age <= tree->maxage && leaf_age > age)
fs/unicode/mkutf8data.c
2883
int utf8nagemin(struct tree *tree, const char *s, size_t len)
fs/unicode/mkutf8data.c
2890
if (!tree)
fs/unicode/mkutf8data.c
2892
age = tree->maxage;
fs/unicode/mkutf8data.c
2894
leaf = utf8nlookup(tree, hangul, s, len);
fs/unicode/mkutf8data.c
2898
if (leaf_age <= tree->maxage && leaf_age < age)
fs/unicode/mkutf8data.c
2912
ssize_t utf8len(struct tree *tree, const char *s)
fs/unicode/mkutf8data.c
2918
if (!tree)
fs/unicode/mkutf8data.c
2921
leaf = utf8lookup(tree, hangul, s);
fs/unicode/mkutf8data.c
2924
if (ages[LEAF_GEN(leaf)] > tree->maxage)
fs/unicode/mkutf8data.c
2939
ssize_t utf8nlen(struct tree *tree, const char *s, size_t len)
fs/unicode/mkutf8data.c
2945
if (!tree)
fs/unicode/mkutf8data.c
2948
leaf = utf8nlookup(tree, hangul, s, len);
fs/unicode/mkutf8data.c
2951
if (ages[LEAF_GEN(leaf)] > tree->maxage)
fs/unicode/mkutf8data.c
2967
struct tree *tree;
fs/unicode/mkutf8data.c
2990
int utf8ncursor(struct utf8cursor *u8c, struct tree *tree, const char *s,
fs/unicode/mkutf8data.c
2993
if (!tree)
fs/unicode/mkutf8data.c
2997
u8c->tree = tree;
fs/unicode/mkutf8data.c
3025
int utf8cursor(struct utf8cursor *u8c, struct tree *tree, const char *s)
fs/unicode/mkutf8data.c
3027
return utf8ncursor(u8c, tree, s, (unsigned int)-1);
fs/unicode/mkutf8data.c
3086
leaf = utf8lookup(u8c->tree, u8c->hangul, u8c->s);
fs/unicode/mkutf8data.c
3088
leaf = utf8nlookup(u8c->tree, u8c->hangul,
fs/unicode/mkutf8data.c
3097
if (ages[LEAF_GEN(leaf)] > u8c->tree->maxage) {
fs/unicode/mkutf8data.c
3110
leaf = utf8lookup(u8c->tree, u8c->hangul, u8c->s);
fs/unicode/mkutf8data.c
3174
static int normalize_line(struct tree *tree)
fs/unicode/mkutf8data.c
3184
if (utf8cursor(&u8c, tree, s))
fs/unicode/mkutf8data.c
3199
if (utf8cursor(&u8c, tree, s))
fs/unicode/mkutf8data.c
350
struct tree *next;
fs/unicode/mkutf8data.c
355
int *(*leaf_index)(struct tree *, void *);
fs/unicode/mkutf8data.c
380
static void *lookup(struct tree *tree, const char *key)
fs/unicode/mkutf8data.c
385
node = tree->root;
fs/unicode/mkutf8data.c
417
static void tree_walk(struct tree *tree)
fs/unicode/mkutf8data.c
428
printf("%s_%x root %p\n", tree->type, tree->maxage, tree->root);
fs/unicode/mkutf8data.c
429
if (tree->childnode == LEAF) {
fs/unicode/mkutf8data.c
430
assert(tree->root);
fs/unicode/mkutf8data.c
431
tree->leaf_print(tree->root, indent);
fs/unicode/mkutf8data.c
434
assert(tree->childnode == NODE);
fs/unicode/mkutf8data.c
435
node = tree->root;
fs/unicode/mkutf8data.c
454
tree->leaf_print(node->left,
fs/unicode/mkutf8data.c
468
tree->leaf_print(node->right,
fs/unicode/mkutf8data.c
533
static int insert(struct tree *tree, char *key, int keylen, void *leaf)
fs/unicode/mkutf8data.c
543
cursor = &tree->root;
fs/unicode/mkutf8data.c
574
if (! tree->leaf_equal(node->left, node->right))
fs/unicode/mkutf8data.c
582
tree->root = leaf;
fs/unicode/mkutf8data.c
583
tree->childnode = LEAF;
fs/unicode/mkutf8data.c
654
static void prune(struct tree *tree)
fs/unicode/mkutf8data.c
668
printf("Pruning %s_%x\n", tree->type, tree->maxage);
fs/unicode/mkutf8data.c
671
if (tree->childnode == LEAF)
fs/unicode/mkutf8data.c
673
if (!tree->root)
fs/unicode/mkutf8data.c
677
node = tree->root;
fs/unicode/mkutf8data.c
727
if (! tree->leaf_equal(leftleaf, rightleaf))
fs/unicode/mkutf8data.c
815
static void mark_nodes(struct tree *tree)
fs/unicode/mkutf8data.c
826
printf("Marking %s_%x\n", tree->type, tree->maxage);
fs/unicode/mkutf8data.c
827
if (tree->childnode == LEAF)
fs/unicode/mkutf8data.c
830
assert(tree->childnode == NODE);
fs/unicode/mkutf8data.c
831
node = tree->root;
fs/unicode/mkutf8data.c
839
if (tree->leaf_mark(node->left)) {
fs/unicode/mkutf8data.c
857
if (tree->leaf_mark(node->right)) {
fs/unicode/mkutf8data.c
878
assert(tree->childnode == NODE);
fs/unicode/mkutf8data.c
879
node = tree->root;
fs/unicode/mkutf8data.c
887
if (tree->leaf_mark(node->left)) {
fs/unicode/mkutf8data.c
909
if (tree->leaf_mark(node->right)) {
fs/unicode/mkutf8data.c
942
static int index_nodes(struct tree *tree, int index)
fs/unicode/mkutf8data.c
954
tree->index = index;
fs/unicode/mkutf8data.c
959
printf("Indexing %s_%x: %d\n", tree->type, tree->maxage, index);
fs/unicode/mkutf8data.c
960
if (tree->childnode == LEAF) {
fs/unicode/mkutf8data.c
961
index += tree->leaf_size(tree->root);
fs/unicode/mkutf8data.c
965
assert(tree->childnode == NODE);
fs/unicode/mkutf8data.c
966
node = tree->root;
fs/unicode/mkutf8data.c
982
*tree->leaf_index(tree, node->left) =
fs/unicode/mkutf8data.c
984
index += tree->leaf_size(node->left);
fs/unicode/mkutf8data.c
997
*tree->leaf_index(tree, node->right) = index;
fs/unicode/mkutf8data.c
998
index += tree->leaf_size(node->right);
fs/xfs/scrub/dqiterate.c
149
struct radix_tree_root *tree = xfs_dquot_tree(qi, cursor->dqtype);
fs/xfs/scrub/dqiterate.c
156
nr_found = radix_tree_gang_lookup(tree, (void **)&dq, cursor->id, 1);
fs/xfs/xfs_dquot.c
808
struct radix_tree_root *tree = xfs_dquot_tree(qi, type);
fs/xfs/xfs_dquot.c
813
dqp = radix_tree_lookup(tree, id);
fs/xfs/xfs_dquot.c
852
struct radix_tree_root *tree = xfs_dquot_tree(qi, type);
fs/xfs/xfs_dquot.c
858
error = radix_tree_insert(tree, id, dqp);
fs/xfs/xfs_qm.c
61
struct radix_tree_root *tree = xfs_dquot_tree(qi, type);
fs/xfs/xfs_qm.c
78
nr_found = radix_tree_gang_lookup(tree, (void **)batch,
include/drm/drm_gpuvm.h
262
struct rb_root_cached tree;
include/linux/audit.h
55
struct audit_tree *tree; /* associated watched tree */
include/linux/generic-radix-tree.h
115
.tree = { \
include/linux/generic-radix-tree.h
130
struct __genradix tree; \
include/linux/generic-radix-tree.h
156
#define genradix_free(_radix) __genradix_free(&(_radix)->tree)
include/linux/generic-radix-tree.h
206
__genradix_ptr_inlined(&(_radix)->tree, \
include/linux/generic-radix-tree.h
220
__genradix_ptr(&(_radix)->tree, \
include/linux/generic-radix-tree.h
228
(__genradix_ptr_inlined(&(_radix)->tree, \
include/linux/generic-radix-tree.h
230
__genradix_ptr_alloc(&(_radix)->tree, \
include/linux/generic-radix-tree.h
236
(__genradix_ptr_inlined(&(_radix)->tree, \
include/linux/generic-radix-tree.h
238
__genradix_ptr_alloc(&(_radix)->tree, \
include/linux/generic-radix-tree.h
253
__genradix_ptr_alloc(&(_radix)->tree, \
include/linux/generic-radix-tree.h
259
__genradix_ptr_alloc(&(_radix)->tree, \
include/linux/generic-radix-tree.h
291
__genradix_iter_peek(_iter, &(_radix)->tree, \
include/linux/generic-radix-tree.h
307
__genradix_iter_peek_prev(_iter, &(_radix)->tree, \
include/linux/generic-radix-tree.h
397
__genradix_prealloc(&(_radix)->tree, \
include/linux/maple_tree.h
431
struct maple_tree *tree; /* The tree we're operating in */
include/linux/maple_tree.h
464
#define mas_lock(mas) spin_lock(&((mas)->tree->ma_lock))
include/linux/maple_tree.h
466
spin_lock_nested(&((mas)->tree->ma_lock), subclass)
include/linux/maple_tree.h
467
#define mas_unlock(mas) spin_unlock(&((mas)->tree->ma_lock))
include/linux/maple_tree.h
483
.tree = mt, \
include/linux/maple_tree.h
506
#define MA_TOPIARY(name, tree) \
include/linux/maple_tree.h
510
.mtree = tree, \
include/linux/maple_tree.h
546
static inline void mas_init(struct ma_state *mas, struct maple_tree *tree,
include/linux/maple_tree.h
550
mas->tree = tree;
include/linux/maple_tree.h
646
mt_dump((__mas)->tree, mt_dump_hex); \
include/linux/maple_tree.h
663
mt_dump((__wrmas)->mas->tree, mt_dump_hex); \
include/linux/maple_tree.h
697
mt_dump((__mas)->tree, mt_dump_hex); \
include/linux/maple_tree.h
716
mt_dump((__wrmas)->mas->tree, mt_dump_hex); \
include/linux/mm_types.h
1504
.tree = &(__mm)->mm_mt, \
include/linux/netdevice.h
258
struct rb_root tree;
include/linux/perf_event.h
998
struct rb_root tree;
include/linux/rbtree.h
193
rb_add_cached(struct rb_node *node, struct rb_root_cached *tree,
include/linux/rbtree.h
196
struct rb_node **link = &tree->rb_root.rb_node;
include/linux/rbtree.h
211
rb_insert_color_cached(node, tree, leftmost);
include/linux/rbtree.h
223
rb_add(struct rb_node *node, struct rb_root *tree,
include/linux/rbtree.h
226
struct rb_node **link = &tree->rb_node;
include/linux/rbtree.h
238
rb_insert_color(node, tree);
include/linux/rbtree.h
251
rb_find_add_cached(struct rb_node *node, struct rb_root_cached *tree,
include/linux/rbtree.h
255
struct rb_node **link = &tree->rb_root.rb_node;
include/linux/rbtree.h
274
rb_insert_color_cached(node, tree, leftmost);
include/linux/rbtree.h
288
rb_find_add(struct rb_node *node, struct rb_root *tree,
include/linux/rbtree.h
291
struct rb_node **link = &tree->rb_node;
include/linux/rbtree.h
308
rb_insert_color(node, tree);
include/linux/rbtree.h
324
rb_find_add_rcu(struct rb_node *node, struct rb_root *tree,
include/linux/rbtree.h
327
struct rb_node **link = &tree->rb_node;
include/linux/rbtree.h
344
rb_insert_color(node, tree);
include/linux/rbtree.h
357
rb_find(const void *key, const struct rb_root *tree,
include/linux/rbtree.h
360
struct rb_node *node = tree->rb_node;
include/linux/rbtree.h
388
rb_find_rcu(const void *key, const struct rb_root *tree,
include/linux/rbtree.h
391
struct rb_node *node = tree->rb_node;
include/linux/rbtree.h
416
rb_find_first(const void *key, const struct rb_root *tree,
include/linux/rbtree.h
419
struct rb_node *node = tree->rb_node;
include/linux/rbtree.h
462
#define rb_for_each(node, key, tree, cmp) \
include/linux/rbtree.h
463
for ((node) = rb_find_first((key), (tree), (cmp)); \
include/linux/rbtree_augmented.h
64
rb_add_augmented_cached(struct rb_node *node, struct rb_root_cached *tree,
include/linux/rbtree_augmented.h
68
struct rb_node **link = &tree->rb_root.rb_node;
include/linux/rbtree_augmented.h
84
rb_insert_augmented_cached(node, tree, leftmost, augment);
include/linux/rbtree_latch.h
102
rb_erase(<n->node[idx], <r->tree[idx]);
include/linux/rbtree_latch.h
109
struct rb_node *node = rcu_dereference_raw(ltr->tree[idx].rb_node);
include/linux/rbtree_latch.h
46
struct rb_root tree[2];
include/linux/rbtree_latch.h
79
struct rb_root *root = <r->tree[idx];
include/linux/vmpressure.h
33
extern void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree,
include/linux/vmpressure.h
47
static inline void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree,
include/net/pkt_cls.h
508
struct tcf_ematch_tree *tree,
include/net/pkt_cls.h
511
if (tree->hdr.nmatches)
include/net/pkt_cls.h
512
return __tcf_em_tree_match(skb, tree, info);
include/trace/events/btrfs.h
2060
TP_PROTO(const struct extent_io_tree *tree,
include/trace/events/btrfs.h
2063
TP_ARGS(tree, start, len, set_bits),
include/trace/events/btrfs.h
2074
TP_fast_assign_btrfs(btrfs_extent_io_tree_to_fs_info(tree),
include/trace/events/btrfs.h
2075
const struct btrfs_inode *inode = btrfs_extent_io_tree_to_inode(tree);
include/trace/events/btrfs.h
2077
__entry->owner = tree->owner;
include/trace/events/btrfs.h
2093
TP_PROTO(const struct extent_io_tree *tree,
include/trace/events/btrfs.h
2096
TP_ARGS(tree, start, len, clear_bits),
include/trace/events/btrfs.h
2107
TP_fast_assign_btrfs(btrfs_extent_io_tree_to_fs_info(tree),
include/trace/events/btrfs.h
2108
const struct btrfs_inode *inode = btrfs_extent_io_tree_to_inode(tree);
include/trace/events/btrfs.h
2110
__entry->owner = tree->owner;
include/trace/events/btrfs.h
2126
TP_PROTO(const struct extent_io_tree *tree,
include/trace/events/btrfs.h
2129
TP_ARGS(tree, start, len, set_bits, clear_bits),
include/trace/events/btrfs.h
2141
TP_fast_assign_btrfs(btrfs_extent_io_tree_to_fs_info(tree),
include/trace/events/btrfs.h
2142
const struct btrfs_inode *inode = btrfs_extent_io_tree_to_inode(tree);
include/trace/events/btrfs.h
2144
__entry->owner = tree->owner;
kernel/audit.h
297
struct audit_tree *tree);
kernel/audit.h
303
extern const char *audit_tree_path(struct audit_tree *tree);
kernel/audit.h
304
extern void audit_put_tree(struct audit_tree *tree);
kernel/audit.h
333
#define audit_put_tree(tree) do { } while (0)
kernel/audit_tree.c
100
if (tree) {
kernel/audit_tree.c
101
refcount_set(&tree->count, 1);
kernel/audit_tree.c
102
tree->goner = 0;
kernel/audit_tree.c
103
INIT_LIST_HEAD(&tree->chunks);
kernel/audit_tree.c
104
INIT_LIST_HEAD(&tree->rules);
kernel/audit_tree.c
105
INIT_LIST_HEAD(&tree->list);
kernel/audit_tree.c
106
INIT_LIST_HEAD(&tree->same_root);
kernel/audit_tree.c
107
tree->root = NULL;
kernel/audit_tree.c
108
strscpy(tree->pathname, s, sz);
kernel/audit_tree.c
110
return tree;
kernel/audit_tree.c
113
static inline void get_tree(struct audit_tree *tree)
kernel/audit_tree.c
115
refcount_inc(&tree->count);
kernel/audit_tree.c
118
static inline void put_tree(struct audit_tree *tree)
kernel/audit_tree.c
120
if (refcount_dec_and_test(&tree->count))
kernel/audit_tree.c
121
kfree_rcu(tree, head);
kernel/audit_tree.c
125
const char *audit_tree_path(struct audit_tree *tree)
kernel/audit_tree.c
127
return tree->pathname;
kernel/audit_tree.c
263
bool audit_tree_match(struct audit_chunk *chunk, struct audit_tree *tree)
kernel/audit_tree.c
267
if (chunk->owners[n].owner == tree)
kernel/audit_tree.c
399
static int create_chunk(struct inode *inode, struct audit_tree *tree)
kernel/audit_tree.c
424
if (tree->goner) {
kernel/audit_tree.c
435
chunk->owners[0].owner = tree;
kernel/audit_tree.c
436
get_tree(tree);
kernel/audit_tree.c
437
list_add(&chunk->owners[0].list, &tree->chunks);
kernel/audit_tree.c
438
if (!tree->root) {
kernel/audit_tree.c
439
tree->root = chunk;
kernel/audit_tree.c
440
list_add(&tree->same_root, &chunk->trees);
kernel/audit_tree.c
460
static int tag_chunk(struct inode *inode, struct audit_tree *tree)
kernel/audit_tree.c
470
return create_chunk(inode, tree);
kernel/audit_tree.c
481
if (old->owners[n].owner == tree) {
kernel/audit_tree.c
498
if (tree->goner) {
kernel/audit_tree.c
507
p->owner = tree;
kernel/audit_tree.c
508
get_tree(tree);
kernel/audit_tree.c
509
list_add(&p->list, &tree->chunks);
kernel/audit_tree.c
510
if (!tree->root) {
kernel/audit_tree.c
511
tree->root = chunk;
kernel/audit_tree.c
512
list_add(&tree->same_root, &chunk->trees);
kernel/audit_tree.c
538
audit_log_untrustedstring(ab, rule->tree->pathname);
kernel/audit_tree.c
544
static void kill_rules(struct audit_context *context, struct audit_tree *tree)
kernel/audit_tree.c
549
list_for_each_entry_safe(rule, next, &tree->rules, rlist) {
kernel/audit_tree.c
553
if (rule->tree) {
kernel/audit_tree.c
558
rule->tree = NULL;
kernel/audit_tree.c
611
static void trim_marked(struct audit_tree *tree)
kernel/audit_tree.c
615
if (tree->goner) {
kernel/audit_tree.c
620
for (p = tree->chunks.next; p != &tree->chunks; p = q) {
kernel/audit_tree.c
625
list_add(p, &tree->chunks);
kernel/audit_tree.c
630
prune_tree_chunks(tree, true);
kernel/audit_tree.c
633
if (!tree->root && !tree->goner) {
kernel/audit_tree.c
634
tree->goner = 1;
kernel/audit_tree.c
637
kill_rules(audit_context(), tree);
kernel/audit_tree.c
638
list_del_init(&tree->list);
kernel/audit_tree.c
640
prune_one(tree);
kernel/audit_tree.c
651
struct audit_tree *tree;
kernel/audit_tree.c
652
tree = rule->tree;
kernel/audit_tree.c
653
if (tree) {
kernel/audit_tree.c
656
if (list_empty(&tree->rules) && !tree->goner) {
kernel/audit_tree.c
657
tree->root = NULL;
kernel/audit_tree.c
658
list_del_init(&tree->same_root);
kernel/audit_tree.c
659
tree->goner = 1;
kernel/audit_tree.c
660
list_move(&tree->list, &prune_list);
kernel/audit_tree.c
661
rule->tree = NULL;
kernel/audit_tree.c
666
rule->tree = NULL;
kernel/audit_tree.c
680
struct audit_tree *tree;
kernel/audit_tree.c
687
tree = container_of(cursor.next, struct audit_tree, list);
kernel/audit_tree.c
688
get_tree(tree);
kernel/audit_tree.c
689
list_move(&cursor, &tree->list);
kernel/audit_tree.c
692
err = kern_path(tree->pathname, 0, &path);
kernel/audit_tree.c
702
list_for_each_entry(node, &tree->chunks, list) {
kernel/audit_tree.c
715
trim_marked(tree);
kernel/audit_tree.c
718
put_tree(tree);
kernel/audit_tree.c
732
rule->inode_f || rule->watch || rule->tree)
kernel/audit_tree.c
734
rule->tree = alloc_tree(pathname);
kernel/audit_tree.c
735
if (!rule->tree)
kernel/audit_tree.c
740
void audit_put_tree(struct audit_tree *tree)
kernel/audit_tree.c
742
put_tree(tree);
kernel/audit_tree.c
745
static int tag_mounts(const struct path *paths, struct audit_tree *tree)
kernel/audit_tree.c
748
int err = tag_chunk(p->dentry->d_inode, tree);
kernel/audit_tree.c
807
struct audit_tree *seed = rule->tree, *tree;
kernel/audit_tree.c
813
rule->tree = NULL;
kernel/audit_tree.c
814
list_for_each_entry(tree, &tree_list, list) {
kernel/audit_tree.c
815
if (!strcmp(seed->pathname, tree->pathname)) {
kernel/audit_tree.c
817
rule->tree = tree;
kernel/audit_tree.c
818
list_add(&rule->rlist, &tree->rules);
kernel/audit_tree.c
822
tree = seed;
kernel/audit_tree.c
823
list_add(&tree->list, &tree_list);
kernel/audit_tree.c
824
list_add(&rule->rlist, &tree->rules);
kernel/audit_tree.c
834
err = kern_path(tree->pathname, 0, &path);
kernel/audit_tree.c
844
get_tree(tree);
kernel/audit_tree.c
845
err = tag_mounts(paths, tree);
kernel/audit_tree.c
851
list_for_each_entry(node, &tree->chunks, list)
kernel/audit_tree.c
855
trim_marked(tree);
kernel/audit_tree.c
861
put_tree(tree);
kernel/audit_tree.c
864
rule->tree = tree;
kernel/audit_tree.c
865
put_tree(tree);
kernel/audit_tree.c
870
list_del_init(&tree->list);
kernel/audit_tree.c
871
list_del_init(&tree->rules);
kernel/audit_tree.c
872
put_tree(tree);
kernel/audit_tree.c
904
struct audit_tree *tree;
kernel/audit_tree.c
907
tree = container_of(cursor.next, struct audit_tree, list);
kernel/audit_tree.c
908
get_tree(tree);
kernel/audit_tree.c
909
list_move(&cursor, &tree->list);
kernel/audit_tree.c
912
err = kern_path(tree->pathname, 0, &path2);
kernel/audit_tree.c
919
put_tree(tree);
kernel/audit_tree.c
924
failed = tag_mounts(paths, tree);
kernel/audit_tree.c
926
put_tree(tree);
kernel/audit_tree.c
933
if (!tree->goner) {
kernel/audit_tree.c
934
list_move(&tree->list, &tree_list);
kernel/audit_tree.c
937
put_tree(tree);
kernel/audit_tree.c
941
struct audit_tree *tree;
kernel/audit_tree.c
943
tree = container_of(barrier.prev, struct audit_tree, list);
kernel/audit_tree.c
944
get_tree(tree);
kernel/audit_tree.c
945
list_move(&tree->list, &barrier);
kernel/audit_tree.c
95
struct audit_tree *tree;
kernel/audit_tree.c
951
list_for_each_entry(node, &tree->chunks, list)
kernel/audit_tree.c
955
trim_marked(tree);
kernel/audit_tree.c
958
put_tree(tree);
kernel/audit_tree.c
99
tree = kmalloc_flex(*tree, pathname, sz);
kernel/audit_watch.c
189
krule->inode_f || krule->watch || krule->tree)
kernel/auditfilter.c
1026
struct audit_tree *tree = entry->rule.tree;
kernel/auditfilter.c
1051
if (e->rule.tree)
kernel/auditfilter.c
1072
if (tree)
kernel/auditfilter.c
1073
audit_put_tree(tree); /* that's the temporary one */
kernel/auditfilter.c
1431
if (r->watch || r->tree)
kernel/auditfilter.c
158
krule->inode_f || krule->watch || krule->tree ||
kernel/auditfilter.c
615
if (entry->rule.tree)
kernel/auditfilter.c
616
audit_put_tree(entry->rule.tree); /* that's the temporary one */
kernel/auditfilter.c
676
audit_tree_path(krule->tree));
kernel/auditfilter.c
741
if (strcmp(audit_tree_path(a->tree),
kernel/auditfilter.c
742
audit_tree_path(b->tree)))
kernel/auditfilter.c
851
new->tree = old->tree;
kernel/auditfilter.c
943
struct audit_tree *tree = entry->rule.tree;
kernel/auditfilter.c
964
if (tree)
kernel/auditfilter.c
965
audit_put_tree(tree);
kernel/auditfilter.c
978
if (tree)
kernel/auditfilter.c
979
audit_put_tree(tree);
kernel/auditfilter.c
983
if (tree) {
kernel/auditsc.c
310
static int match_tree_refs(struct audit_context *ctx, struct audit_tree *tree)
kernel/auditsc.c
315
if (!tree)
kernel/auditsc.c
320
if (audit_tree_match(p->c[n], tree))
kernel/auditsc.c
326
if (audit_tree_match(p->c[n], tree))
kernel/auditsc.c
643
result = match_tree_refs(ctx, rule->tree);
kernel/events/core.c
1749
groups->tree = RB_ROOT;
kernel/events/core.c
1871
rb_add(&event->group_node, &groups->tree, __group_less);
kernel/events/core.c
1894
RB_EMPTY_ROOT(&groups->tree));
kernel/events/core.c
1896
rb_erase(&event->group_node, &groups->tree);
kernel/events/core.c
1926
node = rb_find_first(&key, &groups->tree, __group_cmp);
kernel/events/core.c
1958
for (event = rb_entry_safe(rb_first(&((groups)->tree)), \
kernel/events/core.c
4285
if (!RB_EMPTY_ROOT(&ctx->pinned_groups.tree)) {
kernel/events/core.c
4294
if (!RB_EMPTY_ROOT(&ctx->pinned_groups.tree))
kernel/events/core.c
4542
struct rb_root *tree;
kernel/events/core.c
4554
tree = &pmu_ctx->ctx->flexible_groups.tree;
kernel/events/core.c
4559
node = rb_find_first(&key, tree, __group_cmp_ignore_cgroup);
kernel/events/core.c
4566
node = rb_find_first(&key, tree, __group_cmp_ignore_cgroup);
kernel/events/core.c
4573
node = rb_find_first(&key, tree, __group_cmp_ignore_cgroup);
kernel/locking/rtmutex.c
366
lockdep_assert(RB_EMPTY_NODE(&waiter->tree.entry));
kernel/locking/rtmutex.c
368
waiter->tree.prio = __waiter_prio(task);
kernel/locking/rtmutex.c
369
waiter->tree.deadline = task->dl.deadline;
kernel/locking/rtmutex.c
382
waiter->pi_tree.prio = waiter->tree.prio;
kernel/locking/rtmutex.c
383
waiter->pi_tree.deadline = waiter->tree.deadline;
kernel/locking/rtmutex.c
392
&(struct rt_mutex_waiter){ .tree = *task_to_waiter_node(p) }
kernel/locking/rtmutex.c
433
if (rt_waiter_node_less(&waiter->tree, &top_waiter->tree))
kernel/locking/rtmutex.c
441
if (rt_or_dl_prio(waiter->tree.prio))
kernel/locking/rtmutex.c
444
return rt_waiter_node_equal(&waiter->tree, &top_waiter->tree);
kernel/locking/rtmutex.c
451
rb_entry((node), struct rt_mutex_waiter, tree.entry)
kernel/locking/rtmutex.c
458
if (rt_waiter_node_less(&aw->tree, &bw->tree))
kernel/locking/rtmutex.c
464
if (rt_waiter_node_less(&bw->tree, &aw->tree))
kernel/locking/rtmutex.c
484
rb_add_cached(&waiter->tree.entry, &lock->waiters, __waiter_less);
kernel/locking/rtmutex.c
492
if (RB_EMPTY_NODE(&waiter->tree.entry))
kernel/locking/rtmutex.c
495
rb_erase_cached(&waiter->tree.entry, &lock->waiters);
kernel/locking/rtmutex.c
496
RB_CLEAR_NODE(&waiter->tree.entry);
kernel/locking/rtmutex.c
824
if (rt_waiter_node_equal(&waiter->tree, task_to_waiter_node(task))) {
kernel/locking/rtmutex_api.c
486
if (!waiter || rt_waiter_node_equal(&waiter->tree, task_to_waiter_node(task))) {
kernel/locking/rtmutex_common.h
126
return rb_entry(leftmost, struct rt_mutex_waiter, tree.entry) == waiter;
kernel/locking/rtmutex_common.h
137
w = rb_entry(leftmost, struct rt_mutex_waiter, tree.entry);
kernel/locking/rtmutex_common.h
207
RB_CLEAR_NODE(&waiter->tree.entry);
kernel/locking/rtmutex_common.h
53
struct rt_waiter_node tree;
kernel/locking/ww_mutex.h
105
struct rb_node *n = rb_next(&w->tree.entry);
kernel/locking/ww_mutex.h
108
return rb_entry(n, struct rt_mutex_waiter, tree.entry);
kernel/locking/ww_mutex.h
114
struct rb_node *n = rb_prev(&w->tree.entry);
kernel/locking/ww_mutex.h
117
return rb_entry(n, struct rt_mutex_waiter, tree.entry);
kernel/locking/ww_mutex.h
126
return rb_entry(n, struct rt_mutex_waiter, tree.entry);
kernel/locking/ww_mutex.h
99
return rb_entry(n, struct rt_mutex_waiter, tree.entry);
kernel/module/internal.h
311
struct module *mod_find(unsigned long addr, struct mod_tree_root *tree);
kernel/module/internal.h
317
static inline struct module *mod_find(unsigned long addr, struct mod_tree_root *tree)
kernel/module/main.c
105
if (min < tree->data_addr_min)
kernel/module/main.c
106
tree->data_addr_min = min;
kernel/module/main.c
107
if (max > tree->data_addr_max)
kernel/module/main.c
108
tree->data_addr_max = max;
kernel/module/main.c
112
if (min < tree->addr_min)
kernel/module/main.c
113
tree->addr_min = min;
kernel/module/main.c
114
if (max > tree->addr_max)
kernel/module/main.c
115
tree->addr_max = max;
kernel/module/main.c
98
unsigned int size, struct mod_tree_root *tree)
kernel/module/tree_lookup.c
103
struct module *mod_find(unsigned long addr, struct mod_tree_root *tree)
kernel/module/tree_lookup.c
107
ltn = latch_tree_find((void *)addr, &tree->root, &mod_tree_ops);
kernel/module/tree_lookup.c
64
static noinline void __mod_tree_insert(struct mod_tree_node *node, struct mod_tree_root *tree)
kernel/module/tree_lookup.c
66
latch_tree_insert(&node->node, &tree->root, &mod_tree_ops);
kernel/module/tree_lookup.c
69
static void __mod_tree_remove(struct mod_tree_node *node, struct mod_tree_root *tree)
kernel/module/tree_lookup.c
71
latch_tree_erase(&node->node, &tree->root, &mod_tree_ops);
lib/interval_tree_test.c
246
DEFINE_MTREE(tree);
lib/interval_tree_test.c
248
MA_STATE(mas, &tree, 0, 0);
lib/interval_tree_test.c
259
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/interval_tree_test.c
260
mt_set_in_rcu(&tree);
lib/interval_tree_test.c
263
WARN_ON_ONCE(mtree_store_range(&tree, nodes[j].start,
lib/interval_tree_test.c
301
mtree_destroy(&tree);
lib/maple_tree.c
1500
if (!mt_is_alloc(mas->tree))
lib/maple_tree.c
1548
__must_hold(mas->tree->ma_lock)
lib/maple_tree.c
1555
rcu_assign_pointer(mas->tree->ma_root, mte_mk_root(mas->node));
lib/maple_tree.c
1556
mt_set_height(mas->tree, new_height);
lib/maple_tree.c
1578
__must_hold(mas->tree->ma_lock)
lib/maple_tree.c
1590
__must_hold(mas->tree->ma_lock)
lib/maple_tree.c
1775
if (!ma_is_leaf(mt) && mt_is_alloc(mas->tree)) {
lib/maple_tree.c
1832
if (likely(!ma_is_leaf(mt) && mt_is_alloc(mas->tree))) {
lib/maple_tree.c
1951
MA_STATE(parent, mas->tree, mas->index, mas->last);
lib/maple_tree.c
2197
if (mt_is_alloc(mas->tree))
lib/maple_tree.c
223
return mt_height(mas->tree);
lib/maple_tree.c
2336
MA_TOPIARY(subtrees, mas->tree);
lib/maple_tree.c
2382
in_rcu = mt_in_rcu(mas->tree);
lib/maple_tree.c
2579
next = mt_slot(mas->tree, slots, offset);
lib/maple_tree.c
2623
MA_STATE(l_mas, mas->tree, mas->index, mas->index);
lib/maple_tree.c
2624
MA_STATE(r_mas, mas->tree, mas->index, mas->last);
lib/maple_tree.c
2625
MA_STATE(m_mas, mas->tree, mas->index, mas->index);
lib/maple_tree.c
2758
MA_STATE(l_mas, mas->tree, mas->index, mas->last);
lib/maple_tree.c
2759
MA_STATE(r_mas, mas->tree, mas->index, mas->last);
lib/maple_tree.c
2807
if (mt_is_alloc(mas->tree))
lib/maple_tree.c
2911
MA_STATE(tmp_mas, mas->tree, mas->index, mas->last);
lib/maple_tree.c
2997
MA_STATE(l_mas, mas->tree, mas->index, mas->last);
lib/maple_tree.c
2998
MA_STATE(r_mas, mas->tree, mas->index, mas->last);
lib/maple_tree.c
2999
MA_STATE(prev_l_mas, mas->tree, mas->index, mas->last);
lib/maple_tree.c
3000
MA_STATE(prev_r_mas, mas->tree, mas->index, mas->last);
lib/maple_tree.c
3111
mt_set_height(mas->tree, 1);
lib/maple_tree.c
3114
rcu_assign_pointer(mas->tree->ma_root, mte_mk_root(mas->node));
lib/maple_tree.c
3130
rcu_assign_pointer(mas->tree->ma_root, NULL);
lib/maple_tree.c
3136
rcu_assign_pointer(mas->tree->ma_root, entry);
lib/maple_tree.c
3333
next = mt_slot(mas->tree, slots, offset);
lib/maple_tree.c
3365
mt_set_height(mas->tree, 0);
lib/maple_tree.c
3366
rcu_assign_pointer(mas->tree->ma_root, entry);
lib/maple_tree.c
3379
mt_set_height(mas->tree, 1);
lib/maple_tree.c
3380
rcu_assign_pointer(mas->tree->ma_root, mte_mk_root(mas->node));
lib/maple_tree.c
3384
mte_destroy_walk(root, mas->tree);
lib/maple_tree.c
3497
bool in_rcu = mt_in_rcu(mas->tree);
lib/maple_tree.c
3574
gap |= !mt_slot_locked(mas->tree, slots, offset);
lib/maple_tree.c
3575
gap |= !mt_slot_locked(mas->tree, slots, offset + 1);
lib/maple_tree.c
3589
WARN_ON_ONCE(mt_in_rcu(mas->tree));
lib/maple_tree.c
3594
gap |= !mt_slot_locked(mas->tree, slots, offset + 2);
lib/maple_tree.c
3780
MT_BUG_ON(mas->tree, 1);
lib/maple_tree.c
3866
ret = mt_in_rcu(mas->tree) ? 1 : 0;
lib/maple_tree.c
3926
if (!mt_in_rcu(mas->tree) && (mas->offset == mas->end))
lib/maple_tree.c
3929
if ((new_end == mas->end) && (!mt_in_rcu(mas->tree) ||
lib/maple_tree.c
4034
if ((mas->tree->ma_flags & MT_FLAGS_ALLOC_WRAPPED) && ret == 0) {
lib/maple_tree.c
4035
mas->tree->ma_flags &= ~MT_FLAGS_ALLOC_WRAPPED;
lib/maple_tree.c
4056
mas->tree->ma_flags |= MT_FLAGS_ALLOC_WRAPPED;
lib/maple_tree.c
4406
entry = mt_slot(mas->tree, slots, mas->offset);
lib/maple_tree.c
478
if (mt_is_alloc(mas->tree))
lib/maple_tree.c
5843
__must_hold(mas->tree->ma_lock)
lib/maple_tree.c
5848
if (gfpflags_allow_blocking(gfp) && !mt_external_lock(mas->tree)) {
lib/maple_tree.c
5849
mtree_unlock(mas->tree);
lib/maple_tree.c
5851
mtree_lock(mas->tree);
lib/maple_tree.c
6261
val = (unsigned long)mt_slot_locked(mas->tree, slots, i);
lib/maple_tree.c
6289
if (unlikely(mt_attr(mas->tree) != mt_attr(new_mas->tree)) ||
lib/maple_tree.c
6290
unlikely(!mtree_empty(new_mas->tree))) {
lib/maple_tree.c
6349
new_mas->tree->ma_flags = mas->tree->ma_flags;
lib/maple_tree.c
6350
rcu_assign_pointer(new_mas->tree->ma_root, root);
lib/maple_tree.c
6906
MT_BUG_ON(mas->tree, !entry);
lib/maple_tree.c
6912
MT_BUG_ON(mas->tree, gap > p_end - p_start + 1);
lib/maple_tree.c
6926
MT_BUG_ON(mas->tree, !gaps);
lib/maple_tree.c
6930
MT_BUG_ON(mas->tree, 1);
lib/maple_tree.c
6936
MT_BUG_ON(mas->tree, 1);
lib/maple_tree.c
6943
MT_BUG_ON(mas->tree, 1);
lib/maple_tree.c
6953
MT_BUG_ON(mas->tree, max_gap > mas->max);
lib/maple_tree.c
6956
mt_dump(mas->tree, mt_dump_hex);
lib/maple_tree.c
6957
MT_BUG_ON(mas->tree, 1);
lib/maple_tree.c
6977
MT_BUG_ON(mas->tree, mas_mn(mas) == parent);
lib/maple_tree.c
6987
MT_BUG_ON(mas->tree, node != mas->node);
lib/maple_tree.c
6991
MT_BUG_ON(mas->tree, node == mas->node);
lib/maple_tree.c
7013
MT_BUG_ON(mas->tree, 1);
lib/maple_tree.c
7020
MT_BUG_ON(mas->tree, 1);
lib/maple_tree.c
7027
MT_BUG_ON(mas->tree, 1);
lib/maple_tree.c
7083
MT_BUG_ON(mas->tree, 1);
lib/maple_tree.c
7092
MT_BUG_ON(mas->tree, entry != NULL);
lib/maple_tree.c
7151
__must_hold(mas->tree->ma_lock)
lib/maple_tree.c
7186
mas->tree, mas->node);
lib/maple_tree.c
755
return mt_slot_locked(mas->tree, slots, offset);
lib/maple_tree.c
769
return mt_slot(mas->tree, slots, offset);
lib/maple_tree.c
780
return rcu_dereference_check(mas->tree->ma_root, mt_locked(mas->tree));
lib/maple_tree.c
796
return mt_root_locked(mas->tree);
lib/maple_tree.c
93
#define mas_tree_parent(x) ((unsigned long)(x->tree) | MA_ROOT_PARENT)
lib/maple_tree.c
935
bool in_rcu = mt_in_rcu(mas->tree);
lib/maple_tree.c
940
mt_destroy_walk(mat->head, mas->tree, !in_rcu);
lib/test_maple_tree.c
2092
newmas.tree = &newmt;
lib/test_maple_tree.c
3678
static DEFINE_MTREE(tree);
lib/test_maple_tree.c
3690
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3691
bench_slot_store(&tree);
lib/test_maple_tree.c
3692
mtree_destroy(&tree);
lib/test_maple_tree.c
3697
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3698
bench_node_store(&tree);
lib/test_maple_tree.c
3699
mtree_destroy(&tree);
lib/test_maple_tree.c
3704
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3705
bench_awalk(&tree);
lib/test_maple_tree.c
3706
mtree_destroy(&tree);
lib/test_maple_tree.c
3711
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3712
bench_walk(&tree);
lib/test_maple_tree.c
3713
mtree_destroy(&tree);
lib/test_maple_tree.c
3718
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3719
bench_load(&tree);
lib/test_maple_tree.c
3720
mtree_destroy(&tree);
lib/test_maple_tree.c
3730
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3731
bench_mt_for_each(&tree);
lib/test_maple_tree.c
3732
mtree_destroy(&tree);
lib/test_maple_tree.c
3737
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3738
bench_mas_for_each(&tree);
lib/test_maple_tree.c
3739
mtree_destroy(&tree);
lib/test_maple_tree.c
3744
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3745
bench_mas_prev(&tree);
lib/test_maple_tree.c
3746
mtree_destroy(&tree);
lib/test_maple_tree.c
3750
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3751
check_deficient_node(&tree);
lib/test_maple_tree.c
3752
mtree_destroy(&tree);
lib/test_maple_tree.c
3754
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3755
check_store_null(&tree);
lib/test_maple_tree.c
3756
mtree_destroy(&tree);
lib/test_maple_tree.c
3758
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3759
check_root_expand(&tree);
lib/test_maple_tree.c
3760
mtree_destroy(&tree);
lib/test_maple_tree.c
3762
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3763
check_iteration(&tree);
lib/test_maple_tree.c
3764
mtree_destroy(&tree);
lib/test_maple_tree.c
3768
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3769
check_mas_store_gfp(&tree);
lib/test_maple_tree.c
3770
mtree_destroy(&tree);
lib/test_maple_tree.c
3773
mt_init_flags(&tree, 0);
lib/test_maple_tree.c
3774
check_ranges(&tree);
lib/test_maple_tree.c
3775
mtree_destroy(&tree);
lib/test_maple_tree.c
3779
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3780
check_alloc_range(&tree);
lib/test_maple_tree.c
3781
mtree_destroy(&tree);
lib/test_maple_tree.c
3783
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3784
check_alloc_rev_range(&tree);
lib/test_maple_tree.c
3785
mtree_destroy(&tree);
lib/test_maple_tree.c
3788
mt_init_flags(&tree, 0);
lib/test_maple_tree.c
3790
check_load(&tree, set[0], NULL); /* See if 5015 -> NULL */
lib/test_maple_tree.c
3792
check_insert(&tree, set[9], &tree); /* Insert 0 */
lib/test_maple_tree.c
3793
check_load(&tree, set[9], &tree); /* See if 0 -> &tree */
lib/test_maple_tree.c
3794
check_load(&tree, set[0], NULL); /* See if 5015 -> NULL */
lib/test_maple_tree.c
3796
check_insert(&tree, set[10], ptr); /* Insert 5003 */
lib/test_maple_tree.c
3797
check_load(&tree, set[9], &tree); /* See if 0 -> &tree */
lib/test_maple_tree.c
3798
check_load(&tree, set[11], NULL); /* See if 5002 -> NULL */
lib/test_maple_tree.c
3799
check_load(&tree, set[10], ptr); /* See if 5003 -> ptr */
lib/test_maple_tree.c
3802
mtree_destroy(&tree);
lib/test_maple_tree.c
3805
mt_init_flags(&tree, 0);
lib/test_maple_tree.c
3806
check_insert(&tree, set[0], &tree); /* Insert 5015 */
lib/test_maple_tree.c
3807
check_dup_insert(&tree, set[0], &tree); /* Insert 5015 again */
lib/test_maple_tree.c
3808
check_load(&tree, set[0], &tree); /* See if 5015 -> &tree */
lib/test_maple_tree.c
3814
check_load(&tree, set[1], NULL); /* See if 5014 -> NULL */
lib/test_maple_tree.c
3815
check_insert(&tree, set[1], ptr); /* insert 5014 -> ptr */
lib/test_maple_tree.c
3816
check_load(&tree, set[1], ptr); /* See if 5014 -> ptr */
lib/test_maple_tree.c
3817
check_load(&tree, set[0], &tree); /* See if 5015 -> &tree */
lib/test_maple_tree.c
3822
check_insert(&tree, set[6], ptr); /* insert 1002 -> ptr */
lib/test_maple_tree.c
3823
check_insert(&tree, set[7], &tree); /* insert 1003 -> &tree */
lib/test_maple_tree.c
3825
check_load(&tree, set[0], &tree); /* See if 5015 -> &tree */
lib/test_maple_tree.c
3826
check_load(&tree, set[1], ptr); /* See if 5014 -> ptr */
lib/test_maple_tree.c
3827
check_load(&tree, set[6], ptr); /* See if 1002 -> ptr */
lib/test_maple_tree.c
3828
check_load(&tree, set[7], &tree); /* 1003 = &tree ? */
lib/test_maple_tree.c
3831
mtree_destroy(&tree);
lib/test_maple_tree.c
3833
mt_init_flags(&tree, 0);
lib/test_maple_tree.c
3835
check_insert(&tree, set[5], ptr); /* insert 1001 -> ptr */
lib/test_maple_tree.c
3836
check_insert(&tree, set[7], &tree); /* insert 1003 -> &tree */
lib/test_maple_tree.c
3837
check_insert(&tree, set[6], ptr); /* insert 1002 -> ptr */
lib/test_maple_tree.c
3838
check_load(&tree, set[5], ptr); /* See if 1001 -> ptr */
lib/test_maple_tree.c
3839
check_load(&tree, set[6], ptr); /* See if 1002 -> ptr */
lib/test_maple_tree.c
3840
check_load(&tree, set[7], &tree); /* See if 1003 -> &tree */
lib/test_maple_tree.c
3843
mtree_destroy(&tree);
lib/test_maple_tree.c
3845
mt_init_flags(&tree, 0);
lib/test_maple_tree.c
3852
check_insert(&tree, set[0], ptr); /* 5015 */
lib/test_maple_tree.c
3853
check_insert(&tree, set[1], &tree); /* 5014 */
lib/test_maple_tree.c
3854
check_insert(&tree, set[2], ptr); /* 5017 */
lib/test_maple_tree.c
3855
check_insert(&tree, set[3], &tree); /* 25 */
lib/test_maple_tree.c
3856
check_load(&tree, set[0], ptr);
lib/test_maple_tree.c
3857
check_load(&tree, set[1], &tree);
lib/test_maple_tree.c
3858
check_load(&tree, set[2], ptr);
lib/test_maple_tree.c
3859
check_load(&tree, set[3], &tree);
lib/test_maple_tree.c
3860
check_insert(&tree, set[4], ptr); /* 1000 < Should split. */
lib/test_maple_tree.c
3861
check_load(&tree, set[0], ptr);
lib/test_maple_tree.c
3862
check_load(&tree, set[1], &tree);
lib/test_maple_tree.c
3863
check_load(&tree, set[2], ptr);
lib/test_maple_tree.c
3864
check_load(&tree, set[3], &tree); /*25 */
lib/test_maple_tree.c
3865
check_load(&tree, set[4], ptr);
lib/test_maple_tree.c
3866
check_insert(&tree, set[5], &tree); /* 1001 */
lib/test_maple_tree.c
3867
check_load(&tree, set[0], ptr);
lib/test_maple_tree.c
3868
check_load(&tree, set[1], &tree);
lib/test_maple_tree.c
3869
check_load(&tree, set[2], ptr);
lib/test_maple_tree.c
3870
check_load(&tree, set[3], &tree);
lib/test_maple_tree.c
3871
check_load(&tree, set[4], ptr);
lib/test_maple_tree.c
3872
check_load(&tree, set[5], &tree);
lib/test_maple_tree.c
3873
check_insert(&tree, set[6], ptr);
lib/test_maple_tree.c
3874
check_load(&tree, set[0], ptr);
lib/test_maple_tree.c
3875
check_load(&tree, set[1], &tree);
lib/test_maple_tree.c
3876
check_load(&tree, set[2], ptr);
lib/test_maple_tree.c
3877
check_load(&tree, set[3], &tree);
lib/test_maple_tree.c
3878
check_load(&tree, set[4], ptr);
lib/test_maple_tree.c
3879
check_load(&tree, set[5], &tree);
lib/test_maple_tree.c
3880
check_load(&tree, set[6], ptr);
lib/test_maple_tree.c
3881
check_insert(&tree, set[7], &tree);
lib/test_maple_tree.c
3882
check_load(&tree, set[0], ptr);
lib/test_maple_tree.c
3883
check_insert(&tree, set[8], ptr);
lib/test_maple_tree.c
3885
check_insert(&tree, set[9], &tree);
lib/test_maple_tree.c
3887
check_load(&tree, set[0], ptr);
lib/test_maple_tree.c
3888
check_load(&tree, set[1], &tree);
lib/test_maple_tree.c
3889
check_load(&tree, set[2], ptr);
lib/test_maple_tree.c
3890
check_load(&tree, set[3], &tree);
lib/test_maple_tree.c
3891
check_load(&tree, set[4], ptr);
lib/test_maple_tree.c
3892
check_load(&tree, set[5], &tree);
lib/test_maple_tree.c
3893
check_load(&tree, set[6], ptr);
lib/test_maple_tree.c
3894
check_load(&tree, set[9], &tree);
lib/test_maple_tree.c
3895
mtree_destroy(&tree);
lib/test_maple_tree.c
3897
mt_init_flags(&tree, 0);
lib/test_maple_tree.c
3898
check_seq(&tree, 16, false);
lib/test_maple_tree.c
3899
mtree_destroy(&tree);
lib/test_maple_tree.c
3901
mt_init_flags(&tree, 0);
lib/test_maple_tree.c
3902
check_seq(&tree, 1000, true);
lib/test_maple_tree.c
3903
mtree_destroy(&tree);
lib/test_maple_tree.c
3905
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3906
check_rev_seq(&tree, 1000, true);
lib/test_maple_tree.c
3907
mtree_destroy(&tree);
lib/test_maple_tree.c
3909
check_lower_bound_split(&tree);
lib/test_maple_tree.c
3910
check_upper_bound_split(&tree);
lib/test_maple_tree.c
3911
check_mid_split(&tree);
lib/test_maple_tree.c
3913
mt_init_flags(&tree, 0);
lib/test_maple_tree.c
3914
check_next_entry(&tree);
lib/test_maple_tree.c
3915
check_find(&tree);
lib/test_maple_tree.c
3916
check_find_2(&tree);
lib/test_maple_tree.c
3917
mtree_destroy(&tree);
lib/test_maple_tree.c
3919
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3920
check_prev_entry(&tree);
lib/test_maple_tree.c
3921
mtree_destroy(&tree);
lib/test_maple_tree.c
3923
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3924
check_gap_combining(&tree);
lib/test_maple_tree.c
3925
mtree_destroy(&tree);
lib/test_maple_tree.c
3927
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3928
check_node_overwrite(&tree);
lib/test_maple_tree.c
3929
mtree_destroy(&tree);
lib/test_maple_tree.c
3931
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3932
next_prev_test(&tree);
lib/test_maple_tree.c
3933
mtree_destroy(&tree);
lib/test_maple_tree.c
3935
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3936
check_spanning_relatives(&tree);
lib/test_maple_tree.c
3937
mtree_destroy(&tree);
lib/test_maple_tree.c
3939
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3940
check_rev_find(&tree);
lib/test_maple_tree.c
3941
mtree_destroy(&tree);
lib/test_maple_tree.c
3943
mt_init_flags(&tree, 0);
lib/test_maple_tree.c
3944
check_fuzzer(&tree);
lib/test_maple_tree.c
3945
mtree_destroy(&tree);
lib/test_maple_tree.c
3947
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3948
check_bnode_min_spanning(&tree);
lib/test_maple_tree.c
3949
mtree_destroy(&tree);
lib/test_maple_tree.c
3951
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3952
check_empty_area_window(&tree);
lib/test_maple_tree.c
3953
mtree_destroy(&tree);
lib/test_maple_tree.c
3955
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3956
check_empty_area_fill(&tree);
lib/test_maple_tree.c
3957
mtree_destroy(&tree);
lib/test_maple_tree.c
3959
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3960
check_state_handling(&tree);
lib/test_maple_tree.c
3961
mtree_destroy(&tree);
lib/test_maple_tree.c
3963
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
lib/test_maple_tree.c
3964
alloc_cyclic_testing(&tree);
lib/test_maple_tree.c
3965
mtree_destroy(&tree);
lib/test_maple_tree.c
557
MT_BUG_ON(mas.tree, entry == NULL);
lib/zlib_deflate/deftree.c
133
static void pqdownheap (deflate_state *s, ct_data *tree, int k);
lib/zlib_deflate/deftree.c
135
static void gen_codes (ct_data *tree, int max_code, ush *bl_count);
lib/zlib_deflate/deftree.c
137
static void scan_tree (deflate_state *s, ct_data *tree, int max_code);
lib/zlib_deflate/deftree.c
138
static void send_tree (deflate_state *s, ct_data *tree, int max_code);
lib/zlib_deflate/deftree.c
150
# define send_code(s, c, tree) send_bits(s, tree[c].Code, tree[c].Len)
lib/zlib_deflate/deftree.c
154
# define send_code(s, c, tree) \
lib/zlib_deflate/deftree.c
156
send_bits(s, tree[c].Code, tree[c].Len); }
lib/zlib_deflate/deftree.c
296
#define pqremove(s, tree, top) \
lib/zlib_deflate/deftree.c
300
pqdownheap(s, tree, SMALLEST); \
lib/zlib_deflate/deftree.c
307
#define smaller(tree, n, m, depth) \
lib/zlib_deflate/deftree.c
308
(tree[n].Freq < tree[m].Freq || \
lib/zlib_deflate/deftree.c
309
(tree[n].Freq == tree[m].Freq && depth[n] <= depth[m]))
lib/zlib_deflate/deftree.c
319
ct_data *tree, /* the tree to restore */
lib/zlib_deflate/deftree.c
328
smaller(tree, s->heap[j+1], s->heap[j], s->depth)) {
lib/zlib_deflate/deftree.c
332
if (smaller(tree, v, s->heap[j], s->depth)) break;
lib/zlib_deflate/deftree.c
358
ct_data *tree = desc->dyn_tree;
lib/zlib_deflate/deftree.c
376
tree[s->heap[s->heap_max]].Len = 0; /* root of the heap */
lib/zlib_deflate/deftree.c
380
bits = tree[tree[n].Dad].Len + 1;
lib/zlib_deflate/deftree.c
382
tree[n].Len = (ush)bits;
lib/zlib_deflate/deftree.c
390
f = tree[n].Freq;
lib/zlib_deflate/deftree.c
422
if (tree[m].Len != (unsigned) bits) {
lib/zlib_deflate/deftree.c
423
Trace((stderr,"code %d bits %d->%d\n", m, tree[m].Len, bits));
lib/zlib_deflate/deftree.c
424
s->opt_len += ((long)bits - (long)tree[m].Len)
lib/zlib_deflate/deftree.c
425
*(long)tree[m].Freq;
lib/zlib_deflate/deftree.c
426
tree[m].Len = (ush)bits;
lib/zlib_deflate/deftree.c
442
ct_data *tree, /* the tree to decorate */
lib/zlib_deflate/deftree.c
466
int len = tree[n].Len;
lib/zlib_deflate/deftree.c
469
tree[n].Code = bitrev32((u32)(next_code[len]++)) >> (32 - len);
lib/zlib_deflate/deftree.c
471
Tracecv(tree != static_ltree, (stderr,"\nn %3d %c l %2d c %4x (%x) ",
lib/zlib_deflate/deftree.c
472
n, (isgraph(n) ? n : ' '), len, tree[n].Code, next_code[len]-1));
lib/zlib_deflate/deftree.c
489
ct_data *tree = desc->dyn_tree;
lib/zlib_deflate/deftree.c
503
if (tree[n].Freq != 0) {
lib/zlib_deflate/deftree.c
507
tree[n].Len = 0;
lib/zlib_deflate/deftree.c
518
tree[node].Freq = 1;
lib/zlib_deflate/deftree.c
528
for (n = s->heap_len/2; n >= 1; n--) pqdownheap(s, tree, n);
lib/zlib_deflate/deftree.c
535
pqremove(s, tree, n); /* n = node of least frequency */
lib/zlib_deflate/deftree.c
542
tree[node].Freq = tree[n].Freq + tree[m].Freq;
lib/zlib_deflate/deftree.c
544
tree[n].Dad = tree[m].Dad = (ush)node;
lib/zlib_deflate/deftree.c
546
if (tree == s->bl_tree) {
lib/zlib_deflate/deftree.c
548
node, tree[node].Freq, n, tree[n].Freq, m, tree[m].Freq);
lib/zlib_deflate/deftree.c
553
pqdownheap(s, tree, SMALLEST);
lib/zlib_deflate/deftree.c
565
gen_codes ((ct_data *)tree, max_code, s->bl_count);
lib/zlib_deflate/deftree.c
574
ct_data *tree, /* the tree to be scanned */
lib/zlib_deflate/deftree.c
581
int nextlen = tree[0].Len; /* length of next code */
lib/zlib_deflate/deftree.c
587
tree[max_code+1].Len = (ush)0xffff; /* guard */
lib/zlib_deflate/deftree.c
590
curlen = nextlen; nextlen = tree[n+1].Len;
lib/zlib_deflate/deftree.c
620
ct_data *tree, /* the tree to be scanned */
lib/zlib_deflate/deftree.c
627
int nextlen = tree[0].Len; /* length of next code */
lib/zlib_deflate/deftree.c
636
curlen = nextlen; nextlen = tree[n+1].Len;
lib/zstd/common/huf.h
164
size_t HUF_buildCTable_wksp (HUF_CElt* tree,
mm/debug.c
360
mt_dump(vmi->mas.tree, mt_dump_hex);
mm/memcontrol-v1.c
1763
int nid, unsigned int lru_mask, bool tree)
mm/memcontrol-v1.c
1774
if (tree)
mm/memcontrol-v1.c
1784
bool tree)
mm/memcontrol-v1.c
1792
if (tree)
mm/mmap.c
1762
mt_clear_in_rcu(vmi.mas.tree);
mm/mmap.c
1853
mt_set_in_rcu(vmi.mas.tree);
mm/vma.c
1349
__mt_destroy(mas_detach->tree);
mm/vma.c
1366
__mt_destroy(mas_detach->tree);
mm/vma.c
1484
MA_STATE(test, mas_detach->tree, 0, 0);
mm/vma.c
1570
mt_init_flags(&mt_detach, vmi->mas.tree->ma_flags & MT_FLAGS_LOCK_MASK);
mm/vma.c
2407
vmi->mas.tree->ma_flags & MT_FLAGS_LOCK_MASK);
mm/vmpressure.c
239
void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree,
mm/vmpressure.c
252
if (!cgroup_subsys_on_dfl(memory_cgrp_subsys) && !tree)
mm/vmpressure.c
282
if (tree) {
mm/zswap.c
1038
tree = swap_zswap_tree(swpentry);
mm/zswap.c
1039
if (entry != xa_load(tree, offset)) {
mm/zswap.c
1049
xa_erase(tree, offset);
mm/zswap.c
1561
struct xarray *tree;
mm/zswap.c
1564
tree = swap_zswap_tree(swp_entry(type, offset + index));
mm/zswap.c
1565
entry = xa_erase(tree, offset + index);
mm/zswap.c
1599
struct xarray *tree = swap_zswap_tree(swp);
mm/zswap.c
1617
entry = xa_load(tree, offset);
mm/zswap.c
1646
xa_erase(tree, offset);
mm/zswap.c
1657
struct xarray *tree = swap_zswap_tree(swp);
mm/zswap.c
1660
if (xa_empty(tree))
mm/zswap.c
1663
entry = xa_erase(tree, offset);
mm/zswap.c
1670
struct xarray *trees, *tree;
mm/zswap.c
1674
trees = kvzalloc_objs(*tree, nr);
mm/zswap.c
997
struct xarray *tree;
net/core/dev_addr_lists.c
123
rb_insert_color(&ha->node, &list->tree);
net/core/dev_addr_lists.c
158
rb_erase(&ha->node, &list->tree);
net/core/dev_addr_lists.c
172
node = list->tree.rb_node;
net/core/dev_addr_lists.c
24
struct rb_node **ins_point = &list->tree.rb_node, *parent = NULL;
net/core/dev_addr_lists.c
45
rb_insert_color(&new->node, &list->tree);
net/core/dev_addr_lists.c
488
list->tree = RB_ROOT;
net/core/dev_addr_lists.c
500
list->tree = RB_ROOT;
net/core/dev_addr_lists.c
581
rb_erase(&ha->node, &dev->dev_addrs.tree);
net/core/dev_addr_lists.c
78
struct rb_node **ins_point = &list->tree.rb_node, *parent = NULL;
net/netfilter/nf_conncount.c
388
static void schedule_gc_worker(struct nf_conncount_data *data, int tree)
net/netfilter/nf_conncount.c
390
set_bit(tree, data->pending_trees);
net/netfilter/nf_conncount.c
556
unsigned int tree, next_tree, gc_count = 0;
net/netfilter/nf_conncount.c
558
tree = data->gc_tree % CONNCOUNT_SLOTS;
net/netfilter/nf_conncount.c
559
root = &data->root[tree];
net/netfilter/nf_conncount.c
573
spin_lock_bh(&nf_conncount_locks[tree]);
net/netfilter/nf_conncount.c
595
clear_bit(tree, data->pending_trees);
net/netfilter/nf_conncount.c
597
next_tree = (tree + 1) % CONNCOUNT_SLOTS;
net/netfilter/nf_conncount.c
605
spin_unlock_bh(&nf_conncount_locks[tree]);
net/sched/ematch.c
162
static inline struct tcf_ematch *tcf_em_get_match(struct tcf_ematch_tree *tree,
net/sched/ematch.c
165
return &tree->matches[index];
net/sched/ematch.c
306
struct tcf_ematch_tree *tree)
net/sched/ematch.c
314
memset(tree, 0, sizeof(*tree));
net/sched/ematch.c
331
memcpy(&tree->hdr, tree_hdr, sizeof(*tree_hdr));
net/sched/ematch.c
337
tree->matches = kzalloc(matches_len, GFP_KERNEL);
net/sched/ematch.c
338
if (tree->matches == NULL)
net/sched/ematch.c
362
em = tcf_em_get_match(tree, idx);
net/sched/ematch.c
386
tcf_em_tree_destroy(tree);
net/sched/ematch.c
400
void tcf_em_tree_destroy(struct tcf_ematch_tree *tree)
net/sched/ematch.c
404
if (tree->matches == NULL)
net/sched/ematch.c
407
for (i = 0; i < tree->hdr.nmatches; i++) {
net/sched/ematch.c
408
struct tcf_ematch *em = tcf_em_get_match(tree, i);
net/sched/ematch.c
419
tree->hdr.nmatches = 0;
net/sched/ematch.c
420
kfree(tree->matches);
net/sched/ematch.c
421
tree->matches = NULL;
net/sched/ematch.c
437
int tcf_em_tree_dump(struct sk_buff *skb, struct tcf_ematch_tree *tree, int tlv)
net/sched/ematch.c
448
if (nla_put(skb, TCA_EMATCH_TREE_HDR, sizeof(tree->hdr), &tree->hdr))
net/sched/ematch.c
456
for (i = 0; i < tree->hdr.nmatches; i++) {
net/sched/ematch.c
458
struct tcf_ematch *em = tcf_em_get_match(tree, i);
net/sched/ematch.c
500
int __tcf_em_tree_match(struct sk_buff *skb, struct tcf_ematch_tree *tree,
net/sched/ematch.c
508
while (match_idx < tree->hdr.nmatches) {
net/sched/ematch.c
509
cur_match = tcf_em_get_match(tree, match_idx);
net/sched/ematch.c
531
cur_match = tcf_em_get_match(tree, match_idx);
net/sctp/stream.c
223
new->out.tree.root = NULL;
net/sctp/stream.c
224
new->in.tree.root = NULL;
net/tipc/group.c
199
struct rb_root *tree = &grp->members;
net/tipc/group.c
204
rbtree_postorder_for_each_entry_safe(m, tmp, tree, tree_node) {
net/tipc/group.c
214
struct rb_root *tree = &grp->members;
net/tipc/group.c
220
rbtree_postorder_for_each_entry_safe(m, tmp, tree, tree_node) {
scripts/dtc/dtc.c
31
static void fill_fullpaths(struct node *tree, const char *prefix)
scripts/dtc/dtc.c
36
tree->fullpath = join_path(prefix, tree->name);
scripts/dtc/dtc.c
38
unit = strchr(tree->name, '@');
scripts/dtc/dtc.c
40
tree->basenamelen = unit - tree->name;
scripts/dtc/dtc.c
42
tree->basenamelen = strlen(tree->name);
scripts/dtc/dtc.c
44
for_each_child(tree, child)
scripts/dtc/dtc.c
45
fill_fullpaths(child, tree->fullpath);
scripts/dtc/dtc.h
296
struct property *get_property_by_label(struct node *tree, const char *label,
scripts/dtc/dtc.h
298
struct marker *get_marker_label(struct node *tree, const char *label,
scripts/dtc/dtc.h
301
struct node *get_node_by_path(struct node *tree, const char *path);
scripts/dtc/dtc.h
302
struct node *get_node_by_label(struct node *tree, const char *label);
scripts/dtc/dtc.h
303
struct node *get_node_by_phandle(struct node *tree, cell_t phandle);
scripts/dtc/dtc.h
304
struct node *get_node_by_ref(struct node *tree, const char *ref);
scripts/dtc/dtc.h
307
uint32_t guess_boot_cpuid(struct node *tree);
scripts/dtc/dtc.h
340
struct node *tree, uint32_t boot_cpuid_phys);
scripts/dtc/flattree.c
236
static void flatten_tree(struct node *tree, struct emitter *emit,
scripts/dtc/flattree.c
244
if (tree->deleted)
scripts/dtc/flattree.c
247
emit->beginnode(etarget, tree->labels);
scripts/dtc/flattree.c
250
emit->string(etarget, tree->fullpath, 0);
scripts/dtc/flattree.c
252
emit->string(etarget, tree->name, 0);
scripts/dtc/flattree.c
256
for_each_property(tree, prop) {
scripts/dtc/flattree.c
277
emit->cell(etarget, tree->basenamelen+1);
scripts/dtc/flattree.c
280
if ((vi->flags & FTF_VARALIGN) && ((tree->basenamelen+1) >= 8))
scripts/dtc/flattree.c
283
emit->string(etarget, tree->name, tree->basenamelen);
scripts/dtc/flattree.c
287
for_each_child(tree, child) {
scripts/dtc/flattree.c
291
emit->endnode(etarget, tree->labels);
scripts/dtc/flattree.c
807
struct node *tree;
scripts/dtc/flattree.c
913
tree = unflatten_tree(&dtbuf, &strbuf, "", flags);
scripts/dtc/flattree.c
923
if (get_subnode(tree, "__fixups__") || get_subnode(tree, "__local_fixups__"))
scripts/dtc/flattree.c
926
return build_dt_info(dtsflags, reservelist, tree, boot_cpuid_phys);
scripts/dtc/fstree.c
16
struct node *tree;
scripts/dtc/fstree.c
22
tree = build_node(NULL, NULL, NULL);
scripts/dtc/fstree.c
50
add_property(tree, prop);
scripts/dtc/fstree.c
58
add_child(tree, newchild);
scripts/dtc/fstree.c
65
return tree;
scripts/dtc/fstree.c
70
struct node *tree;
scripts/dtc/fstree.c
72
tree = read_fstree(dirname);
scripts/dtc/fstree.c
73
tree = name_node(tree, "");
scripts/dtc/fstree.c
75
return build_dt_info(DTSF_V1, NULL, tree, guess_boot_cpuid(tree));
scripts/dtc/livetree.c
453
struct node *tree, uint32_t boot_cpuid_phys)
scripts/dtc/livetree.c
460
dti->dt = tree;
scripts/dtc/livetree.c
501
struct property *get_property_by_label(struct node *tree, const char *label,
scripts/dtc/livetree.c
507
*node = tree;
scripts/dtc/livetree.c
509
for_each_property(tree, prop) {
scripts/dtc/livetree.c
517
for_each_child(tree, c) {
scripts/dtc/livetree.c
527
struct marker *get_marker_label(struct node *tree, const char *label,
scripts/dtc/livetree.c
534
*node = tree;
scripts/dtc/livetree.c
536
for_each_property(tree, p) {
scripts/dtc/livetree.c
544
for_each_child(tree, c) {
scripts/dtc/livetree.c
566
struct node *get_node_by_path(struct node *tree, const char *path)
scripts/dtc/livetree.c
572
if (tree->deleted)
scripts/dtc/livetree.c
574
return tree;
scripts/dtc/livetree.c
582
for_each_child(tree, child) {
scripts/dtc/livetree.c
592
struct node *get_node_by_label(struct node *tree, const char *label)
scripts/dtc/livetree.c
599
for_each_label(tree->labels, l)
scripts/dtc/livetree.c
601
return tree;
scripts/dtc/livetree.c
603
for_each_child(tree, child) {
scripts/dtc/livetree.c
612
struct node *get_node_by_phandle(struct node *tree, cell_t phandle)
scripts/dtc/livetree.c
621
if (tree->phandle == phandle) {
scripts/dtc/livetree.c
622
if (tree->deleted)
scripts/dtc/livetree.c
624
return tree;
scripts/dtc/livetree.c
627
for_each_child(tree, child) {
scripts/dtc/livetree.c
636
struct node *get_node_by_ref(struct node *tree, const char *ref)
scripts/dtc/livetree.c
638
struct node *target = tree;
scripts/dtc/livetree.c
642
return tree;
scripts/dtc/livetree.c
659
target = get_node_by_label(tree, label);
scripts/dtc/livetree.c
711
uint32_t guess_boot_cpuid(struct node *tree)
scripts/dtc/livetree.c
716
cpus = get_node_by_path(tree, "/cpus");
scripts/dtc/treesource.c
370
static void write_tree_source_node(FILE *f, struct node *tree, int level)
scripts/dtc/treesource.c
378
for_each_label(tree->labels, l)
scripts/dtc/treesource.c
380
if (tree->name && (*tree->name))
scripts/dtc/treesource.c
381
fprintf(f, "%s {", tree->name);
scripts/dtc/treesource.c
386
srcstr = srcpos_string_first(tree->srcpos, annotate);
scripts/dtc/treesource.c
394
for_each_property(tree, prop) {
scripts/dtc/treesource.c
401
for_each_child(tree, child) {
scripts/dtc/treesource.c
408
srcstr = srcpos_string_last(tree->srcpos, annotate);
scripts/dtc/yamltree.c
178
static void yaml_tree(struct node *tree, yaml_emitter_t *emitter)
scripts/dtc/yamltree.c
184
if (tree->deleted)
scripts/dtc/yamltree.c
191
for_each_property(tree, prop)
scripts/dtc/yamltree.c
195
for_each_child(tree, child) {
scripts/gcc-plugins/gcc-common.h
103
static inline tree build_const_char_string(int len, const char *str)
scripts/gcc-plugins/gcc-common.h
105
tree cstr, elem, index, type;
scripts/gcc-plugins/gcc-common.h
118
static inline void __add_type_attr(tree type, const char *attr, tree args)
scripts/gcc-plugins/gcc-common.h
120
tree oldattr;
scripts/gcc-plugins/gcc-common.h
134
static inline void add_type_attr(tree type, const char *attr, tree args)
scripts/gcc-plugins/gcc-common.h
136
tree main_variant = TYPE_MAIN_VARIANT(type);
scripts/gcc-plugins/gcc-common.h
217
static inline void change_decl_assembler_name(tree decl, tree name)
scripts/gcc-plugins/gcc-common.h
222
static inline void varpool_finalize_decl(tree decl)
scripts/gcc-plugins/gcc-common.h
227
static inline void varpool_add_new_variable(tree decl)
scripts/gcc-plugins/gcc-common.h
315
static inline gimple gimple_build_assign_with_ops(enum tree_code subcode, tree lhs, tree op1, tree op2 MEM_STAT_DECL)
scripts/gcc-plugins/latent_entropy_plugin.c
116
static tree tree_get_random_const(tree type)
scripts/gcc-plugins/latent_entropy_plugin.c
128
static tree handle_latent_entropy_attribute(tree *node, tree name,
scripts/gcc-plugins/latent_entropy_plugin.c
129
tree args __unused,
scripts/gcc-plugins/latent_entropy_plugin.c
133
tree type;
scripts/gcc-plugins/latent_entropy_plugin.c
167
tree fld, lst = TYPE_FIELDS(type);
scripts/gcc-plugins/latent_entropy_plugin.c
171
tree fieldtype;
scripts/gcc-plugins/latent_entropy_plugin.c
189
tree random_const, fld_t = TREE_TYPE(fld);
scripts/gcc-plugins/latent_entropy_plugin.c
206
tree elt_type, array_size, elt_size;
scripts/gcc-plugins/latent_entropy_plugin.c
229
tree cst = size_int(i);
scripts/gcc-plugins/latent_entropy_plugin.c
230
tree rand_cst = tree_get_random_const(elt_type);
scripts/gcc-plugins/latent_entropy_plugin.c
265
tree list;
scripts/gcc-plugins/latent_entropy_plugin.c
279
static tree create_var(tree type, const char *name)
scripts/gcc-plugins/latent_entropy_plugin.c
281
tree var;
scripts/gcc-plugins/latent_entropy_plugin.c
301
static enum tree_code get_op(tree *rhs)
scripts/gcc-plugins/latent_entropy_plugin.c
334
static gimple create_assign(enum tree_code code, tree lhs, tree op1,
scripts/gcc-plugins/latent_entropy_plugin.c
335
tree op2)
scripts/gcc-plugins/latent_entropy_plugin.c
340
static void perturb_local_entropy(basic_block bb, tree local_entropy)
scripts/gcc-plugins/latent_entropy_plugin.c
344
tree rhs;
scripts/gcc-plugins/latent_entropy_plugin.c
355
tree local_entropy)
scripts/gcc-plugins/latent_entropy_plugin.c
358
tree temp;
scripts/gcc-plugins/latent_entropy_plugin.c
383
static bool handle_tail_calls(basic_block bb, tree local_entropy)
scripts/gcc-plugins/latent_entropy_plugin.c
405
static void perturb_latent_entropy(tree local_entropy)
scripts/gcc-plugins/latent_entropy_plugin.c
431
static void init_local_entropy(basic_block bb, tree local_entropy)
scripts/gcc-plugins/latent_entropy_plugin.c
434
tree frame_addr, rand_const, tmp, fndecl, udi_frame_addr;
scripts/gcc-plugins/latent_entropy_plugin.c
485
tree name, var = NODE_DECL(node);
scripts/gcc-plugins/latent_entropy_plugin.c
504
tree local_entropy;
scripts/gcc-plugins/latent_entropy_plugin.c
543
tree type, id;
scripts/gcc-plugins/latent_entropy_plugin.c
82
static GTY(()) tree latent_entropy_decl;
scripts/gcc-plugins/randomize_layout_plugin.c
100
static tree handle_randomize_considered_attr(tree *node, tree name, tree args, int flags, bool *no_add_attrs)
scripts/gcc-plugins/randomize_layout_plugin.c
110
static tree handle_randomize_performed_attr(tree *node, tree name, tree args, int flags, bool *no_add_attrs)
scripts/gcc-plugins/randomize_layout_plugin.c
150
tree tree_start;
scripts/gcc-plugins/randomize_layout_plugin.c
155
static void partition_struct(tree *fields, unsigned long length, struct partition_group *size_groups, unsigned long *num_groups)
scripts/gcc-plugins/randomize_layout_plugin.c
191
static void performance_shuffle(tree *newtree, unsigned long length, ranctx *prng_state)
scripts/gcc-plugins/randomize_layout_plugin.c
211
tree tmp;
scripts/gcc-plugins/randomize_layout_plugin.c
228
static void full_shuffle(tree *newtree, unsigned long length, ranctx *prng_state)
scripts/gcc-plugins/randomize_layout_plugin.c
233
tree tmp;
scripts/gcc-plugins/randomize_layout_plugin.c
242
static void shuffle(const_tree type, tree *newtree, unsigned long length)
scripts/gcc-plugins/randomize_layout_plugin.c
259
debug_tree((tree)type);
scripts/gcc-plugins/randomize_layout_plugin.c
296
static int relayout_struct(tree type)
scripts/gcc-plugins/randomize_layout_plugin.c
300
tree field;
scripts/gcc-plugins/randomize_layout_plugin.c
301
tree newtree[num_fields];
scripts/gcc-plugins/randomize_layout_plugin.c
303
tree list;
scripts/gcc-plugins/randomize_layout_plugin.c
304
tree variant;
scripts/gcc-plugins/randomize_layout_plugin.c
305
tree main_variant;
scripts/gcc-plugins/randomize_layout_plugin.c
348
shuffle(type, (tree *)newtree, shuffle_length);
scripts/gcc-plugins/randomize_layout_plugin.c
417
static void randomize_type(tree type)
scripts/gcc-plugins/randomize_layout_plugin.c
419
tree variant;
scripts/gcc-plugins/randomize_layout_plugin.c
439
static void update_decl_size(tree decl)
scripts/gcc-plugins/randomize_layout_plugin.c
441
tree lastval, lastidx, field, init, type, flexsize;
scripts/gcc-plugins/randomize_layout_plugin.c
486
tree decl = (tree)event_data;
scripts/gcc-plugins/randomize_layout_plugin.c
487
tree type;
scripts/gcc-plugins/randomize_layout_plugin.c
514
tree type = (tree)event_data;
scripts/gcc-plugins/randomize_layout_plugin.c
54
static tree handle_randomize_layout_attr(tree *node, tree name, tree args, int flags, bool *no_add_attrs)
scripts/gcc-plugins/randomize_layout_plugin.c
56
tree type;
scripts/gcc-plugins/randomize_layout_plugin.c
570
static void check_bad_casts_in_constructor(tree var, tree init)
scripts/gcc-plugins/randomize_layout_plugin.c
573
tree field, val;
scripts/gcc-plugins/randomize_layout_plugin.c
574
tree field_type, val_type;
scripts/gcc-plugins/randomize_layout_plugin.c
615
tree init;
scripts/gcc-plugins/randomize_layout_plugin.c
618
tree var = NODE_DECL(node);
scripts/gcc-plugins/randomize_layout_plugin.c
688
tree var;
scripts/gcc-plugins/randomize_layout_plugin.c
692
tree init = DECL_INITIAL(var);
scripts/gcc-plugins/stackleak_plugin.c
116
tree sp_decl, input;
scripts/gcc-plugins/stackleak_plugin.c
117
vec<tree, va_gc> *inputs = NULL;
scripts/gcc-plugins/stackleak_plugin.c
436
static inline bool string_equal(tree node, const char *string, int length)
scripts/gcc-plugins/stackleak_plugin.c
44
static GTY(()) tree track_function_decl;
scripts/gcc-plugins/stackleak_plugin.c
451
tree section;
scripts/gcc-plugins/stackleak_plugin.c
481
tree fntype;
scripts/gcc-plugins/stackleak_plugin.c
89
static tree get_current_stack_pointer_decl(void)
scripts/gcc-plugins/stackleak_plugin.c
94
tree var = NODE_DECL(node);
scripts/gcc-plugins/stackleak_plugin.c
95
tree name = DECL_NAME(var);
scripts/kconfig/gconf.c
151
static void set_node(GtkTreeStore *tree, GtkTreeIter *node, struct menu *menu)
scripts/kconfig/gconf.c
228
gtk_tree_store_set(tree, node,
scripts/kconfig/gconf.c
891
static void _display_tree(GtkTreeStore *tree, struct menu *menu,
scripts/kconfig/gconf.c
908
&& !(child->flags & MENU_ROOT) && (tree == tree1))
scripts/kconfig/gconf.c
912
&& (tree == tree2))
scripts/kconfig/gconf.c
915
gtk_tree_store_append(tree, &iter, parent);
scripts/kconfig/gconf.c
916
set_node(tree, &iter, child);
scripts/kconfig/gconf.c
919
_display_tree(tree, child, &iter);
sound/hda/core/sysfs.c
322
struct hdac_widget_tree *tree = codec->widgets;
sound/hda/core/sysfs.c
325
if (!tree)
sound/hda/core/sysfs.c
327
free_widget_node(tree->afg, &widget_afg_group);
sound/hda/core/sysfs.c
328
if (tree->nodes) {
sound/hda/core/sysfs.c
329
for (p = tree->nodes; *p; p++)
sound/hda/core/sysfs.c
331
kfree(tree->nodes);
sound/hda/core/sysfs.c
333
kobject_put(tree->root);
sound/hda/core/sysfs.c
334
kfree(tree);
sound/hda/core/sysfs.c
365
struct hdac_widget_tree *tree;
sound/hda/core/sysfs.c
369
tree = codec->widgets = kzalloc_obj(*tree);
sound/hda/core/sysfs.c
370
if (!tree)
sound/hda/core/sysfs.c
373
tree->root = kobject_create_and_add("widgets", &codec->dev.kobj);
sound/hda/core/sysfs.c
374
if (!tree->root)
sound/hda/core/sysfs.c
377
tree->nodes = kzalloc_objs(*tree->nodes, codec->num_nodes + 1);
sound/hda/core/sysfs.c
378
if (!tree->nodes)
sound/hda/core/sysfs.c
382
err = add_widget_node(tree->root, nid, &widget_node_group,
sound/hda/core/sysfs.c
383
&tree->nodes[i]);
sound/hda/core/sysfs.c
389
err = add_widget_node(tree->root, codec->afg,
sound/hda/core/sysfs.c
390
&widget_afg_group, &tree->afg);
sound/hda/core/sysfs.c
395
kobject_uevent(tree->root, KOBJ_CHANGE);
sound/hda/core/sysfs.c
426
struct hdac_widget_tree *tree;
sound/hda/core/sysfs.c
434
tree = kmemdup(codec->widgets, sizeof(*tree), GFP_KERNEL);
sound/hda/core/sysfs.c
435
if (!tree)
sound/hda/core/sysfs.c
438
tree->nodes = kzalloc_objs(*tree->nodes, num_nodes + 1);
sound/hda/core/sysfs.c
439
if (!tree->nodes) {
sound/hda/core/sysfs.c
440
kfree(tree);
sound/hda/core/sysfs.c
454
add_widget_node(tree->root, nid, &widget_node_group,
sound/hda/core/sysfs.c
455
&tree->nodes[i]);
sound/hda/core/sysfs.c
457
tree->nodes[i] =
sound/hda/core/sysfs.c
464
codec->widgets = tree;
sound/hda/core/sysfs.c
466
kobject_uevent(tree->root, KOBJ_CHANGE);
tools/include/linux/rbtree.h
178
rb_add_cached(struct rb_node *node, struct rb_root_cached *tree,
tools/include/linux/rbtree.h
181
struct rb_node **link = &tree->rb_root.rb_node;
tools/include/linux/rbtree.h
196
rb_insert_color_cached(node, tree, leftmost);
tools/include/linux/rbtree.h
206
rb_add(struct rb_node *node, struct rb_root *tree,
tools/include/linux/rbtree.h
209
struct rb_node **link = &tree->rb_node;
tools/include/linux/rbtree.h
221
rb_insert_color(node, tree);
tools/include/linux/rbtree.h
234
rb_find_add(struct rb_node *node, struct rb_root *tree,
tools/include/linux/rbtree.h
237
struct rb_node **link = &tree->rb_node;
tools/include/linux/rbtree.h
254
rb_insert_color(node, tree);
tools/include/linux/rbtree.h
267
rb_find(const void *key, const struct rb_root *tree,
tools/include/linux/rbtree.h
270
struct rb_node *node = tree->rb_node;
tools/include/linux/rbtree.h
295
rb_find_first(const void *key, const struct rb_root *tree,
tools/include/linux/rbtree.h
298
struct rb_node *node = tree->rb_node;
tools/include/linux/rbtree.h
341
#define rb_for_each(node, key, tree, cmp) \
tools/include/linux/rbtree.h
342
for ((node) = rb_find_first((key), (tree), (cmp)); \
tools/objtool/elf.c
172
struct rb_root_cached *tree = (struct rb_root_cached *)&sec->symbol_tree;
tools/objtool/elf.c
175
__sym_for_each(sym, tree, offset, offset) {
tools/objtool/elf.c
185
struct rb_root_cached *tree = (struct rb_root_cached *)&sec->symbol_tree;
tools/objtool/elf.c
188
__sym_for_each(func, tree, offset, offset) {
tools/objtool/elf.c
198
struct rb_root_cached *tree = (struct rb_root_cached *)&sec->symbol_tree;
tools/objtool/elf.c
201
__sym_for_each(tmp, tree, offset, offset) {
tools/objtool/elf.c
266
struct rb_root_cached *tree = (struct rb_root_cached *)&sec->symbol_tree;
tools/objtool/elf.c
269
__sym_for_each(func, tree, offset, offset) {
tools/perf/ui/stdio/hist.c
309
static size_t callchain__fprintf_flat(FILE *fp, struct rb_root *tree,
tools/perf/ui/stdio/hist.c
315
struct rb_node *rb_node = rb_first(tree);
tools/perf/ui/stdio/hist.c
360
static size_t callchain__fprintf_folded(FILE *fp, struct rb_root *tree,
tools/perf/ui/stdio/hist.c
366
struct rb_node *rb_node = rb_first(tree);
tools/perf/util/cgroup.c
564
cgrp = __cgroup__findnew(&env->cgroups.tree, id, true, path);
tools/perf/util/cgroup.c
579
cgrp = __cgroup__findnew(&env->cgroups.tree, id, false, NULL);
tools/perf/util/cgroup.c
590
while (!RB_EMPTY_ROOT(&env->cgroups.tree)) {
tools/perf/util/cgroup.c
591
node = rb_first(&env->cgroups.tree);
tools/perf/util/cgroup.c
594
rb_erase(node, &env->cgroups.tree);
tools/perf/util/env.h
141
struct rb_root tree;
tools/perf/util/srcline.c
363
void srcline__tree_insert(struct rb_root_cached *tree, u64 addr, char *srcline)
tools/perf/util/srcline.c
365
struct rb_node **p = &tree->rb_root.rb_node;
tools/perf/util/srcline.c
390
rb_insert_color_cached(&node->rb_node, tree, leftmost);
tools/perf/util/srcline.c
393
char *srcline__tree_find(struct rb_root_cached *tree, u64 addr)
tools/perf/util/srcline.c
395
struct rb_node *n = tree->rb_root.rb_node;
tools/perf/util/srcline.c
412
void srcline__tree_delete(struct rb_root_cached *tree)
tools/perf/util/srcline.c
415
struct rb_node *next = rb_first_cached(tree);
tools/perf/util/srcline.c
420
rb_erase_cached(&pos->rb_node, tree);
tools/perf/util/srcline.c
454
void inlines__tree_insert(struct rb_root_cached *tree,
tools/perf/util/srcline.c
457
struct rb_node **p = &tree->rb_root.rb_node;
tools/perf/util/srcline.c
474
rb_insert_color_cached(&inlines->rb_node, tree, leftmost);
tools/perf/util/srcline.c
477
struct inline_node *inlines__tree_find(struct rb_root_cached *tree, u64 addr)
tools/perf/util/srcline.c
479
struct rb_node *n = tree->rb_root.rb_node;
tools/perf/util/srcline.c
496
void inlines__tree_delete(struct rb_root_cached *tree)
tools/perf/util/srcline.c
499
struct rb_node *next = rb_first_cached(tree);
tools/perf/util/srcline.c
504
rb_erase_cached(&pos->rb_node, tree);
tools/perf/util/srcline.h
22
void srcline__tree_insert(struct rb_root_cached *tree, u64 addr, char *srcline);
tools/perf/util/srcline.h
24
char *srcline__tree_find(struct rb_root_cached *tree, u64 addr);
tools/perf/util/srcline.h
26
void srcline__tree_delete(struct rb_root_cached *tree);
tools/perf/util/srcline.h
52
void inlines__tree_insert(struct rb_root_cached *tree,
tools/perf/util/srcline.h
55
struct inline_node *inlines__tree_find(struct rb_root_cached *tree, u64 addr);
tools/perf/util/srcline.h
57
void inlines__tree_delete(struct rb_root_cached *tree);
tools/testing/radix-tree/benchmark.c
117
RADIX_TREE(tree, GFP_KERNEL);
tools/testing/radix-tree/benchmark.c
120
benchmark_insert(&tree, size, step);
tools/testing/radix-tree/benchmark.c
121
benchmark_tagging(&tree, size, step);
tools/testing/radix-tree/benchmark.c
123
tagged = benchmark_iter(&tree, true);
tools/testing/radix-tree/benchmark.c
124
normal = benchmark_iter(&tree, false);
tools/testing/radix-tree/benchmark.c
131
benchmark_delete(&tree, size, step);
tools/testing/radix-tree/benchmark.c
133
item_kill_tree(&tree);
tools/testing/radix-tree/main.c
101
assert(item_delete(&tree, 1000000));
tools/testing/radix-tree/main.c
102
tree_verify_min_height(&tree, 42);
tools/testing/radix-tree/main.c
104
assert(item_delete(&tree, 42));
tools/testing/radix-tree/main.c
105
tree_verify_min_height(&tree, 0);
tools/testing/radix-tree/main.c
108
item_insert(&tree, i);
tools/testing/radix-tree/main.c
109
tree_verify_min_height(&tree, i);
tools/testing/radix-tree/main.c
114
assert(item_delete(&tree, i));
tools/testing/radix-tree/main.c
116
tree_verify_min_height(&tree, 0);
tools/testing/radix-tree/main.c
120
tree_verify_min_height(&tree, i);
tools/testing/radix-tree/main.c
123
item_kill_tree(&tree);
tools/testing/radix-tree/main.c
126
void check_copied_tags(struct radix_tree_root *tree, unsigned long start, unsigned long end, unsigned long *idx, int count, int fromtag, int totag)
tools/testing/radix-tree/main.c
134
if (item_tag_get(tree, idx[i], totag)) {
tools/testing/radix-tree/main.c
136
end, idx[i], item_tag_get(tree, idx[i],
tools/testing/radix-tree/main.c
138
item_tag_get(tree, idx[i], totag));
tools/testing/radix-tree/main.c
140
assert(!item_tag_get(tree, idx[i], totag));
tools/testing/radix-tree/main.c
143
if (item_tag_get(tree, idx[i], fromtag) ^
tools/testing/radix-tree/main.c
144
item_tag_get(tree, idx[i], totag)) {
tools/testing/radix-tree/main.c
146
idx[i], item_tag_get(tree, idx[i], fromtag),
tools/testing/radix-tree/main.c
147
item_tag_get(tree, idx[i], totag));
tools/testing/radix-tree/main.c
149
assert(!(item_tag_get(tree, idx[i], fromtag) ^
tools/testing/radix-tree/main.c
150
item_tag_get(tree, idx[i], totag)));
tools/testing/radix-tree/main.c
158
RADIX_TREE(tree, GFP_KERNEL);
tools/testing/radix-tree/main.c
175
item_insert(&tree, start);
tools/testing/radix-tree/main.c
179
item_tag_set(&tree, start, 0);
tools/testing/radix-tree/main.c
18
RADIX_TREE(tree, GFP_KERNEL);
tools/testing/radix-tree/main.c
183
item_insert(&tree, start-1);
tools/testing/radix-tree/main.c
185
item_tag_set(&tree, start-1, 0);
tools/testing/radix-tree/main.c
188
item_insert(&tree, end);
tools/testing/radix-tree/main.c
192
item_tag_set(&tree, end, 0);
tools/testing/radix-tree/main.c
196
item_insert(&tree, end+1);
tools/testing/radix-tree/main.c
198
item_tag_set(&tree, end+1, 0);
tools/testing/radix-tree/main.c
204
} while (item_lookup(&tree, idx[i]));
tools/testing/radix-tree/main.c
206
item_insert(&tree, idx[i]);
tools/testing/radix-tree/main.c
208
item_tag_set(&tree, idx[i], 0);
tools/testing/radix-tree/main.c
217
tagged = tag_tagged_items(&tree, start, end, ITEMS, XA_MARK_0, XA_MARK_1);
tools/testing/radix-tree/main.c
221
check_copied_tags(&tree, start, end, idx, ITEMS, 0, 1);
tools/testing/radix-tree/main.c
226
tagged = tag_tagged_items(&tree, start, end, tmp, XA_MARK_0, XA_MARK_2);
tools/testing/radix-tree/main.c
23
item_insert(&tree, middle + idx);
tools/testing/radix-tree/main.c
231
check_copied_tags(&tree, start, end, idx, ITEMS, 0, 2);
tools/testing/radix-tree/main.c
232
verify_tag_consistency(&tree, 0);
tools/testing/radix-tree/main.c
233
verify_tag_consistency(&tree, 1);
tools/testing/radix-tree/main.c
234
verify_tag_consistency(&tree, 2);
tools/testing/radix-tree/main.c
236
item_kill_tree(&tree);
tools/testing/radix-tree/main.c
25
item_check_absent(&tree, middle - down - 1);
tools/testing/radix-tree/main.c
27
item_check_present(&tree, middle + idx);
tools/testing/radix-tree/main.c
28
item_check_absent(&tree, middle + up);
tools/testing/radix-tree/main.c
31
item_gang_check_present(&tree, middle - down, up + down,
tools/testing/radix-tree/main.c
33
item_full_scan(&tree, middle - down, down + up, chunk);
tools/testing/radix-tree/main.c
35
item_kill_tree(&tree);
tools/testing/radix-tree/main.c
81
RADIX_TREE(tree, GFP_KERNEL);
tools/testing/radix-tree/main.c
83
item_insert(&tree, 44);
tools/testing/radix-tree/main.c
84
item_check_present(&tree, 44);
tools/testing/radix-tree/main.c
85
item_check_absent(&tree, 43);
tools/testing/radix-tree/main.c
86
item_kill_tree(&tree);
tools/testing/radix-tree/main.c
92
RADIX_TREE(tree, GFP_KERNEL);
tools/testing/radix-tree/main.c
93
tree_verify_min_height(&tree, 0);
tools/testing/radix-tree/main.c
95
item_insert(&tree, 42);
tools/testing/radix-tree/main.c
96
tree_verify_min_height(&tree, 42);
tools/testing/radix-tree/main.c
98
item_insert(&tree, 1000000);
tools/testing/radix-tree/main.c
99
tree_verify_min_height(&tree, 1000000);
tools/testing/radix-tree/maple.c
33568
mas.tree = mt;
tools/testing/radix-tree/maple.c
35660
MA_STATE(newmas, mas->tree, 0, 0);
tools/testing/radix-tree/maple.c
35938
mtree_lock(mas.tree);
tools/testing/radix-tree/maple.c
35941
mtree_unlock(mas.tree);
tools/testing/radix-tree/maple.c
36036
DEFINE_MTREE(tree);
tools/testing/radix-tree/maple.c
36038
mt_dump(&tree, mt_dump_dec);
tools/testing/radix-tree/maple.c
36040
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE | MT_FLAGS_LOCK_EXTERN | MT_FLAGS_USE_RCU);
tools/testing/radix-tree/maple.c
36041
check_vma_modification(&tree);
tools/testing/radix-tree/maple.c
36042
mtree_destroy(&tree);
tools/testing/radix-tree/maple.c
36044
tree.ma_root = xa_mk_value(0);
tools/testing/radix-tree/maple.c
36045
mt_dump(&tree, mt_dump_dec);
tools/testing/radix-tree/maple.c
36048
node->parent = (void *)((unsigned long)(&tree) | 1);
tools/testing/radix-tree/maple.c
36054
tree.ma_root = mt_mk_node(node, maple_leaf_64);
tools/testing/radix-tree/maple.c
36055
mt_dump(&tree, mt_dump_dec);
tools/testing/radix-tree/maple.c
36061
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
tools/testing/radix-tree/maple.c
36062
check_locky(&tree);
tools/testing/radix-tree/maple.c
36063
mtree_destroy(&tree);
tools/testing/radix-tree/maple.c
36066
mt_init_flags(&tree, 0);
tools/testing/radix-tree/maple.c
36067
check_dfs_preorder(&tree);
tools/testing/radix-tree/maple.c
36068
mtree_destroy(&tree);
tools/testing/radix-tree/maple.c
36070
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE | MT_FLAGS_USE_RCU);
tools/testing/radix-tree/maple.c
36071
check_nomem_writer_race(&tree);
tools/testing/radix-tree/maple.c
36072
mtree_destroy(&tree);
tools/testing/radix-tree/maple.c
36074
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
tools/testing/radix-tree/maple.c
36075
check_prealloc(&tree);
tools/testing/radix-tree/maple.c
36076
mtree_destroy(&tree);
tools/testing/radix-tree/maple.c
36078
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
tools/testing/radix-tree/maple.c
36079
check_spanning_write(&tree);
tools/testing/radix-tree/maple.c
36080
mtree_destroy(&tree);
tools/testing/radix-tree/maple.c
36082
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
tools/testing/radix-tree/maple.c
36083
check_spanning_store_height(&tree);
tools/testing/radix-tree/maple.c
36084
mtree_destroy(&tree);
tools/testing/radix-tree/maple.c
36086
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
tools/testing/radix-tree/maple.c
36087
check_collapsing_rebalance(&tree);
tools/testing/radix-tree/maple.c
36088
mtree_destroy(&tree);
tools/testing/radix-tree/maple.c
36090
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
tools/testing/radix-tree/maple.c
36091
check_null_expand(&tree);
tools/testing/radix-tree/maple.c
36092
mtree_destroy(&tree);
tools/testing/radix-tree/maple.c
36094
mt_init_flags(&tree, 0);
tools/testing/radix-tree/maple.c
36095
check_mtree_dup(&tree);
tools/testing/radix-tree/maple.c
36096
mtree_destroy(&tree);
tools/testing/radix-tree/maple.c
36099
mt_init_flags(&tree, 0);
tools/testing/radix-tree/maple.c
36100
check_erase_testset(&tree);
tools/testing/radix-tree/maple.c
36101
mtree_destroy(&tree);
tools/testing/radix-tree/maple.c
36104
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
tools/testing/radix-tree/maple.c
36105
check_rcu_simulated(&tree);
tools/testing/radix-tree/maple.c
36106
mtree_destroy(&tree);
tools/testing/radix-tree/maple.c
36108
mt_init_flags(&tree, MT_FLAGS_ALLOC_RANGE);
tools/testing/radix-tree/maple.c
36109
check_rcu_threaded(&tree);
tools/testing/radix-tree/maple.c
36110
mtree_destroy(&tree);
tools/testing/radix-tree/maple.c
36116
mt_init_flags(&tree, 0);
tools/testing/radix-tree/maple.c
36117
check_erase2_sets(&tree);
tools/testing/radix-tree/maple.c
36118
mtree_destroy(&tree);
tools/testing/radix-tree/maple.c
36123
check_nomem(&tree);
tools/testing/radix-tree/maple.c
36145
DEFINE_MTREE(tree);
tools/testing/radix-tree/maple.c
36146
MA_STATE(mas, &tree, 0, 0);
tools/testing/radix-tree/maple.c
36153
build_full_tree(&tree, 0, 3);
tools/testing/radix-tree/maple.c
36198
mt_validate(&tree);
tools/testing/radix-tree/maple.c
36201
__mt_destroy(&tree);
tools/testing/radix-tree/maple.c
696
mt_dump(mas.tree, mt_dump_hex);
tools/testing/radix-tree/maple.c
719
MT_BUG_ON(mt, mtree_load(mas.tree, 0) != NULL);
tools/testing/radix-tree/multiorder.c
168
struct radix_tree_root *tree = ptr;
tools/testing/radix-tree/multiorder.c
172
item_insert_order(tree, 0, order);
tools/testing/radix-tree/multiorder.c
173
item_delete_rcu(tree, 0);
tools/testing/radix-tree/multiorder.c
219
struct radix_tree_root *tree = ptr;
tools/testing/radix-tree/multiorder.c
223
item_insert_order(tree, 3 << RADIX_TREE_MAP_SHIFT, 0);
tools/testing/radix-tree/multiorder.c
224
item_insert_order(tree, 2 << RADIX_TREE_MAP_SHIFT, 0);
tools/testing/radix-tree/multiorder.c
229
item_insert_order(tree, index, order);
tools/testing/radix-tree/multiorder.c
230
xa_set_mark(tree, index, XA_MARK_1);
tools/testing/radix-tree/multiorder.c
231
item_delete_rcu(tree, index);
tools/testing/radix-tree/tag_check.c
100
item_tag_set(&tree, tmp+1, 0);
tools/testing/radix-tree/tag_check.c
101
item_delete(&tree, tmp+1);
tools/testing/radix-tree/tag_check.c
102
item_tag_clear(&tree, tmp, 1);
tools/testing/radix-tree/tag_check.c
104
assert(radix_tree_gang_lookup_tag(&tree, (void **)&item, 0, 1, 0) == 1);
tools/testing/radix-tree/tag_check.c
105
assert(radix_tree_gang_lookup_tag(&tree, (void **)&item, 0, 1, 1) == 0);
tools/testing/radix-tree/tag_check.c
107
assert(item_tag_get(&tree, tmp, 0) == 1);
tools/testing/radix-tree/tag_check.c
108
assert(item_tag_get(&tree, tmp, 1) == 0);
tools/testing/radix-tree/tag_check.c
110
verify_tag_consistency(&tree, 0);
tools/testing/radix-tree/tag_check.c
111
item_kill_tree(&tree);
tools/testing/radix-tree/tag_check.c
131
static void gang_check(struct radix_tree_root *tree,
tools/testing/radix-tree/tag_check.c
139
while ((nr_found = radix_tree_gang_lookup_tag(tree, (void **)items,
tools/testing/radix-tree/tag_check.c
14
__simple_checks(struct radix_tree_root *tree, unsigned long index, int tag)
tools/testing/radix-tree/tag_check.c
157
static void do_thrash(struct radix_tree_root *tree, char *thrash_state, int tag)
tools/testing/radix-tree/tag_check.c
183
item_check_absent(tree, index);
tools/testing/radix-tree/tag_check.c
184
item_insert(tree, index);
tools/testing/radix-tree/tag_check.c
19
item_check_absent(tree, index);
tools/testing/radix-tree/tag_check.c
195
item_check_present(tree, index);
tools/testing/radix-tree/tag_check.c
196
if (item_tag_get(tree, index, tag)) {
tools/testing/radix-tree/tag_check.c
20
assert(item_tag_get(tree, index, tag) == 0);
tools/testing/radix-tree/tag_check.c
202
item_delete(tree, index);
tools/testing/radix-tree/tag_check.c
212
if (item_lookup(tree, index))
tools/testing/radix-tree/tag_check.c
213
assert(item_tag_get(tree, index, tag));
tools/testing/radix-tree/tag_check.c
216
item_tag_set(tree, index, tag);
tools/testing/radix-tree/tag_check.c
217
item_tag_set(tree, index, tag);
tools/testing/radix-tree/tag_check.c
22
item_insert(tree, index);
tools/testing/radix-tree/tag_check.c
228
item_check_present(tree, index);
tools/testing/radix-tree/tag_check.c
229
assert(item_tag_get(tree, index, tag));
tools/testing/radix-tree/tag_check.c
23
assert(item_tag_get(tree, index, tag) == 0);
tools/testing/radix-tree/tag_check.c
230
item_tag_clear(tree, index, tag);
tools/testing/radix-tree/tag_check.c
231
item_tag_clear(tree, index, tag);
tools/testing/radix-tree/tag_check.c
24
item_tag_set(tree, index, tag);
tools/testing/radix-tree/tag_check.c
243
item_check_absent(tree, index);
tools/testing/radix-tree/tag_check.c
246
item_check_present(tree, index);
tools/testing/radix-tree/tag_check.c
247
assert(!item_tag_get(tree, index, tag));
tools/testing/radix-tree/tag_check.c
25
ret = item_tag_get(tree, index, tag);
tools/testing/radix-tree/tag_check.c
251
item_check_present(tree, index);
tools/testing/radix-tree/tag_check.c
252
assert(item_tag_get(tree, index, tag));
tools/testing/radix-tree/tag_check.c
259
gang_check(tree, thrash_state, tag);
tools/testing/radix-tree/tag_check.c
27
ret = tag_tagged_items(tree, first, ~0UL, 10, tag, !tag);
tools/testing/radix-tree/tag_check.c
274
RADIX_TREE(tree, GFP_KERNEL);
tools/testing/radix-tree/tag_check.c
280
do_thrash(&tree, thrash_state, 0);
tools/testing/radix-tree/tag_check.c
282
verify_tag_consistency(&tree, 0);
tools/testing/radix-tree/tag_check.c
283
item_kill_tree(&tree);
tools/testing/radix-tree/tag_check.c
289
RADIX_TREE(tree, GFP_KERNEL);
tools/testing/radix-tree/tag_check.c
29
ret = item_tag_get(tree, index, !tag);
tools/testing/radix-tree/tag_check.c
291
item_insert(&tree, 1000000);
tools/testing/radix-tree/tag_check.c
292
item_delete(&tree, 1000000);
tools/testing/radix-tree/tag_check.c
293
item_kill_tree(&tree);
tools/testing/radix-tree/tag_check.c
298
RADIX_TREE(tree, GFP_KERNEL);
tools/testing/radix-tree/tag_check.c
301
item_insert(&tree, 1000000);
tools/testing/radix-tree/tag_check.c
303
item_delete(&tree, 1000000);
tools/testing/radix-tree/tag_check.c
305
item_kill_tree(&tree);
tools/testing/radix-tree/tag_check.c
31
ret = item_delete(tree, index);
tools/testing/radix-tree/tag_check.c
312
RADIX_TREE(tree, GFP_KERNEL);
tools/testing/radix-tree/tag_check.c
316
item_insert(&tree, 0);
tools/testing/radix-tree/tag_check.c
317
item_tag_set(&tree, 0, 0);
tools/testing/radix-tree/tag_check.c
318
ret = radix_tree_gang_lookup_tag(&tree, (void **)items, 0, BATCH, 0);
tools/testing/radix-tree/tag_check.c
320
ret = radix_tree_gang_lookup_tag(&tree, (void **)items, 1, BATCH, 0);
tools/testing/radix-tree/tag_check.c
322
verify_tag_consistency(&tree, 0);
tools/testing/radix-tree/tag_check.c
323
verify_tag_consistency(&tree, 1);
tools/testing/radix-tree/tag_check.c
324
ret = tag_tagged_items(&tree, first, 10, 10, XA_MARK_0, XA_MARK_1);
tools/testing/radix-tree/tag_check.c
326
ret = radix_tree_gang_lookup_tag(&tree, (void **)items, 0, BATCH, 1);
tools/testing/radix-tree/tag_check.c
328
item_tag_clear(&tree, 0, 0);
tools/testing/radix-tree/tag_check.c
329
ret = radix_tree_gang_lookup_tag(&tree, (void **)items, 0, BATCH, 0);
tools/testing/radix-tree/tag_check.c
33
item_insert(tree, index);
tools/testing/radix-tree/tag_check.c
331
item_kill_tree(&tree);
tools/testing/radix-tree/tag_check.c
34
ret = item_tag_get(tree, index, tag);
tools/testing/radix-tree/tag_check.c
36
ret = item_delete(tree, index);
tools/testing/radix-tree/tag_check.c
38
ret = item_delete(tree, index);
tools/testing/radix-tree/tag_check.c
45
RADIX_TREE(tree, GFP_KERNEL);
tools/testing/radix-tree/tag_check.c
48
__simple_checks(&tree, index, 0);
tools/testing/radix-tree/tag_check.c
49
__simple_checks(&tree, index, 1);
tools/testing/radix-tree/tag_check.c
51
verify_tag_consistency(&tree, 0);
tools/testing/radix-tree/tag_check.c
52
verify_tag_consistency(&tree, 1);
tools/testing/radix-tree/tag_check.c
54
item_kill_tree(&tree);
tools/testing/radix-tree/tag_check.c
64
RADIX_TREE(tree, GFP_KERNEL);
tools/testing/radix-tree/tag_check.c
66
item_insert(&tree, 43);
tools/testing/radix-tree/tag_check.c
67
assert(item_tag_get(&tree, 43, 0) == 0);
tools/testing/radix-tree/tag_check.c
68
item_tag_set(&tree, 43, 0);
tools/testing/radix-tree/tag_check.c
69
assert(item_tag_get(&tree, 43, 0) == 1);
tools/testing/radix-tree/tag_check.c
70
item_insert(&tree, 1000000);
tools/testing/radix-tree/tag_check.c
71
assert(item_tag_get(&tree, 43, 0) == 1);
tools/testing/radix-tree/tag_check.c
73
item_insert(&tree, 0);
tools/testing/radix-tree/tag_check.c
74
item_tag_set(&tree, 0, 0);
tools/testing/radix-tree/tag_check.c
75
item_delete(&tree, 1000000);
tools/testing/radix-tree/tag_check.c
76
assert(item_tag_get(&tree, 43, 0) != 0);
tools/testing/radix-tree/tag_check.c
77
item_delete(&tree, 43);
tools/testing/radix-tree/tag_check.c
78
assert(item_tag_get(&tree, 43, 0) == 0); /* crash */
tools/testing/radix-tree/tag_check.c
79
assert(item_tag_get(&tree, 0, 0) == 1);
tools/testing/radix-tree/tag_check.c
81
verify_tag_consistency(&tree, 0);
tools/testing/radix-tree/tag_check.c
83
item_kill_tree(&tree);
tools/testing/radix-tree/tag_check.c
93
RADIX_TREE(tree, GFP_KERNEL);
tools/testing/radix-tree/tag_check.c
96
item_insert(&tree, tmp);
tools/testing/radix-tree/tag_check.c
97
item_insert(&tree, tmp+1);
tools/testing/radix-tree/tag_check.c
98
item_tag_set(&tree, tmp, 0);
tools/testing/radix-tree/tag_check.c
99
item_tag_set(&tree, tmp, 1);
tools/testing/vma/include/dup.h
407
.tree = &(__mm)->mm_mt, \