tnode
phandle tnode;
tnode = prom_getchild(prom_root_node);
fd_node = prom_searchsiblings(tnode, "obio");
tnode = prom_getchild(fd_node);
fd_node = prom_searchsiblings(tnode, "SUNW,fdtwo");
fd_node = prom_searchsiblings(tnode, "fd");
int tnode;
tnode = find_best_numa_node_for_mlgroup(m);
if (tnode == MAX_NUMNODES)
numa_latency[index][tnode] = m->latency;
struct device_node *tnode = NULL, *pnode = parent->of_node;
for (i = 0; i < ARRAY_SIZE(stm32_timer_trigger_compat) && !tnode; i++)
tnode = of_get_compatible_child(pnode, stm32_timer_trigger_compat[i]);
if (!tnode) {
ret = of_property_read_u32(tnode, "reg", &idx);
of_node_put(tnode);
struct kern_tid_node *node = &flow->tnode[flow->tnode_cnt++];
struct kern_tid_node *node = &flow->tnode[grp_num];
struct kern_tid_node *node = &flow->tnode[grp_num];
struct kern_tid_node tnode[TID_RDMA_MAX_PAGES];
struct tid_rb_node *tnode);
struct tid_rb_node *tnode)
fdata->entry_to_rb[tnode->rcventry - base] = NULL;
clear_tid_node(fdata, tnode);
struct device_node *ipar, *tnode, *old = NULL;
tnode = ipar;
of_node_put(tnode);
tnode = of_get_parent(old);
old = tnode;
struct latch_tree_node tnode;
return container_of(n, struct bpf_ksym, tnode)->start;
ksym = container_of(n, struct bpf_ksym, tnode);
latch_tree_insert(&ksym->tnode, &bpf_tree, &bpf_tree_ops);
latch_tree_erase(&ksym->tnode, &bpf_tree, &bpf_tree_ops);
return n ? container_of(n, struct bpf_ksym, tnode) : NULL;
struct tnode *t;
t = lwq_dequeue(&q, struct tnode, n);
while ((t = lwq_dequeue(&q, struct tnode, n)) != NULL) {
struct tnode *t;
wait_var_event(q, (t = lwq_dequeue(q, struct tnode, n)) != NULL);
struct rhash_head tnode;
.head_offset = offsetof(struct net_bridge_vlan, tnode),
rhashtable_remove_fast(&vg->tunnel_hash, &vlan->tnode,
err = rhashtable_lookup_insert_fast(&vg->tunnel_hash, &vlan->tnode,
DECLARE_FLEX_ARRAY(struct key_vector __rcu *, tnode);
#define TNODE_SIZE(n) offsetof(struct tnode, kv[0].tnode[n])
struct key_vector __rcu **cptr = n->tnode;
cptr = &pn->tnode[cindex];
static inline struct tnode *tn_info(struct key_vector *kv)
return container_of(kv, struct tnode, kv[0]);
#define get_child(tn, i) rtnl_dereference((tn)->tnode[i])
#define get_child_rcu(tn, i) rcu_dereference_rtnl((tn)->tnode[i])
struct key_vector *tnode;
struct key_vector *pn = iter->tnode;
iter->tnode, iter->index, iter->depth);
iter->tnode = pn;
iter->tnode = n;
iter->tnode = pn;
n = rcu_dereference(pn->tnode[0]);
iter->tnode = n;
iter->tnode = pn;
struct key_vector *tnode;
struct key_vector *l, **tp = &iter->tnode;
iter->tnode = t->kv;
l = leaf_walk_rcu(&iter->tnode, key);
struct tnode *n = container_of(head, struct tnode, rcu);
static struct tnode *tnode_alloc(int bits)
struct tnode *kv;
struct tnode *tnode;
tnode = tnode_alloc(bits);
if (!tnode)
pr_debug("AT %p s=%zu %zu\n", tnode, TNODE_SIZE(0),
tnode->full_children = 1;
tnode->empty_children = 1ul << bits;
tn = tnode->kv;
rcu_assign_pointer(tn->tnode[i], n);
rcu_assign_pointer(tp->tnode[0], n);
tn = container_of(head, struct tnode, rcu)->kv;
struct ebitmap_node *snode, *tnode;
ebitmap_for_each_positive_bit(tattr, tnode, j) {
struct ebitmap_node *rnode, *tnode;
ebitmap_for_each_positive_bit(&role->types, tnode, j) {
struct ebitmap_node *snode, *tnode;
ebitmap_for_each_positive_bit(tattr, tnode, j) {