Home
last modified time | relevance | path

Searched refs:rbnode (Results 1 – 15 of 15) sorted by relevance

/Linux-v5.4/drivers/base/regmap/
Dregcache-rbtree.c41 struct regcache_rbtree_node *rbnode, in regcache_rbtree_get_base_top_reg() argument
44 *base = rbnode->base_reg; in regcache_rbtree_get_base_top_reg()
45 *top = rbnode->base_reg + ((rbnode->blklen - 1) * map->reg_stride); in regcache_rbtree_get_base_top_reg()
49 struct regcache_rbtree_node *rbnode, unsigned int idx) in regcache_rbtree_get_register() argument
51 return regcache_get_val(map, rbnode->block, idx); in regcache_rbtree_get_register()
55 struct regcache_rbtree_node *rbnode, in regcache_rbtree_set_register() argument
58 set_bit(idx, rbnode->cache_present); in regcache_rbtree_set_register()
59 regcache_set_val(map, rbnode->block, idx, val); in regcache_rbtree_set_register()
67 struct regcache_rbtree_node *rbnode; in regcache_rbtree_lookup() local
70 rbnode = rbtree_ctx->cached_rbnode; in regcache_rbtree_lookup()
[all …]
/Linux-v5.4/net/ipv4/
Dinet_fragment.c48 RB_CLEAR_NODE(&skb->rbnode); in fragcb_clear()
71 rb_link_node(&skb->rbnode, &q->last_run_head->rbnode, in fragrun_create()
72 &q->last_run_head->rbnode.rb_right); in fragrun_create()
74 rb_link_node(&skb->rbnode, NULL, &q->rb_fragments.rb_node); in fragrun_create()
75 rb_insert_color(&skb->rbnode, &q->rb_fragments); in fragrun_create()
238 struct sk_buff *skb = rb_entry(p, struct sk_buff, rbnode); in inet_frag_rbtree_purge()
241 rb_erase(&skb->rbnode, root); in inet_frag_rbtree_purge()
398 rb_link_node(&skb->rbnode, parent, rbn); in inet_frag_queue_insert()
399 rb_insert_color(&skb->rbnode, &q->rb_fragments); in inet_frag_queue_insert()
420 if (RB_EMPTY_NODE(&skb->rbnode)) in inet_frag_reasm_prepare()
[all …]
Dtcp_input.c4494 rb_erase(&skb->rbnode, &tp->out_of_order_queue); in tcp_ofo_queue()
4573 rb_link_node(&skb->rbnode, NULL, p); in tcp_data_queue_ofo()
4574 rb_insert_color(&skb->rbnode, &tp->out_of_order_queue); in tcp_data_queue_ofo()
4592 parent = &tp->ooo_last_skb->rbnode; in tcp_data_queue_ofo()
4623 rb_replace_node(&skb1->rbnode, &skb->rbnode, in tcp_data_queue_ofo()
4641 rb_link_node(&skb->rbnode, parent, p); in tcp_data_queue_ofo()
4642 rb_insert_color(&skb->rbnode, &tp->out_of_order_queue); in tcp_data_queue_ofo()
4654 rb_erase(&skb1->rbnode, &tp->out_of_order_queue); in tcp_data_queue_ofo()
4871 rb_erase(&skb->rbnode, root); in tcp_collapse_one()
4894 rb_link_node(&skb->rbnode, parent, p); in tcp_rbtree_insert()
[all …]
/Linux-v5.4/drivers/infiniband/hw/hfi1/
Dmmu_rb.c131 struct mmu_rb_node *rbnode; in hfi1_mmu_rb_unregister() local
149 rbnode = rb_entry(node, struct mmu_rb_node, node); in hfi1_mmu_rb_unregister()
152 list_move(&rbnode->list, &del_list); in hfi1_mmu_rb_unregister()
237 struct mmu_rb_node *rbnode, *ptr; in hfi1_mmu_rb_evict() local
245 list_for_each_entry_safe_reverse(rbnode, ptr, &handler->lru_list, in hfi1_mmu_rb_evict()
247 if (handler->ops->evict(handler->ops_arg, rbnode, evict_arg, in hfi1_mmu_rb_evict()
249 __mmu_int_rb_remove(rbnode, &handler->root); in hfi1_mmu_rb_evict()
251 list_move(&rbnode->list, &del_list); in hfi1_mmu_rb_evict()
259 rbnode = list_first_entry(&del_list, struct mmu_rb_node, list); in hfi1_mmu_rb_evict()
260 list_del(&rbnode->list); in hfi1_mmu_rb_evict()
[all …]
/Linux-v5.4/net/netfilter/
Dnf_conncount.c309 struct rb_node **rbnode, *parent; in insert_tree() local
319 rbnode = &(root->rb_node); in insert_tree()
320 while (*rbnode) { in insert_tree()
322 rbconn = rb_entry(*rbnode, struct nf_conncount_rb, node); in insert_tree()
324 parent = *rbnode; in insert_tree()
327 rbnode = &((*rbnode)->rb_left); in insert_tree()
329 rbnode = &((*rbnode)->rb_right); in insert_tree()
377 rb_link_node_rcu(&rbconn->node, parent, rbnode); in insert_tree()
/Linux-v5.4/mm/
Dzswap.c152 struct rb_node rbnode; member
255 RB_CLEAR_NODE(&entry->rbnode); in zswap_entry_cache_alloc()
273 entry = rb_entry(node, struct zswap_entry, rbnode); in zswap_rb_search()
296 myentry = rb_entry(parent, struct zswap_entry, rbnode); in zswap_rb_insert()
306 rb_link_node(&entry->rbnode, parent, link); in zswap_rb_insert()
307 rb_insert_color(&entry->rbnode, root); in zswap_rb_insert()
313 if (!RB_EMPTY_NODE(&entry->rbnode)) { in zswap_rb_erase()
314 rb_erase(&entry->rbnode, root); in zswap_rb_erase()
315 RB_CLEAR_NODE(&entry->rbnode); in zswap_rb_erase()
1212 rbtree_postorder_for_each_entry_safe(entry, n, &tree->rbroot, rbnode) in zswap_frontswap_invalidate_area()
/Linux-v5.4/net/sched/
Dsch_etf.c187 rb_link_node(&nskb->rbnode, parent, p); in etf_enqueue_timesortedlist()
188 rb_insert_color_cached(&nskb->rbnode, &q->head, leftmost); in etf_enqueue_timesortedlist()
210 rb_erase_cached(&skb->rbnode, &q->head); in timesortedlist_drop()
234 rb_erase_cached(&skb->rbnode, &q->head); in timesortedlist_remove()
429 rb_erase_cached(&skb->rbnode, &q->head); in timesortedlist_clear()
Dsch_fq.c359 rb_erase(&skb->rbnode, &flow->t_root); in fq_erase_head()
409 rb_link_node(&skb->rbnode, parent, p); in flow_queue_add()
410 rb_insert_color(&skb->rbnode, &flow->t_root); in flow_queue_add()
611 rb_erase(&skb->rbnode, &flow->t_root); in fq_flow_purge()
Dsch_netem.c369 rb_erase(&skb->rbnode, &q->t_root); in tfifo_reset()
402 rb_link_node(&nskb->rbnode, parent, p); in tfifo_enqueue()
403 rb_insert_color(&nskb->rbnode, &q->t_root); in tfifo_enqueue()
672 rb_erase(&skb->rbnode, &q->t_root); in netem_erase_head()
/Linux-v5.4/fs/btrfs/
Dbackref.h67 struct rb_node rbnode; member
Dbackref.c240 ref = rb_entry(parent, struct prelim_ref, rbnode); in prelim_ref_insert()
276 rb_link_node(&newref->rbnode, parent, p); in prelim_ref_insert()
277 rb_insert_color_cached(&newref->rbnode, root, leftmost); in prelim_ref_insert()
289 &preftree->root.rb_root, rbnode) in prelim_release()
635 ref = rb_entry(rnode, struct prelim_ref, rbnode); in resolve_indirect_refs()
642 rb_erase_cached(&ref->rbnode, &preftrees->indirect.root); in resolve_indirect_refs()
723 ref = rb_entry(node, struct prelim_ref, rbnode); in add_missing_keys()
1254 ref = rb_entry(node, struct prelim_ref, rbnode);
1255 node = rb_next(&ref->rbnode);
/Linux-v5.4/include/linux/
Dskbuff.h701 struct rb_node rbnode; /* used in netem, ip4 defrag, and tcp stack */ member
3394 #define rb_to_skb(rb) rb_entry_safe(rb, struct sk_buff, rbnode)
3397 #define skb_rb_next(skb) rb_to_skb(rb_next(&(skb)->rbnode))
3398 #define skb_rb_prev(skb) rb_to_skb(rb_prev(&(skb)->rbnode))
/Linux-v5.4/Documentation/networking/
Drds.txt298 wraps the raw congestion bitmap, contains rbnode, waitq, etc.
/Linux-v5.4/include/net/
Dtcp.h1795 rb_erase(&skb->rbnode, &sk->tcp_rtx_queue); in tcp_rtx_queue_unlink()
/Linux-v5.4/net/core/
Dskbuff.c3098 struct sk_buff *skb = rb_entry(p, struct sk_buff, rbnode); in skb_rbtree_purge()
3101 rb_erase(&skb->rbnode, root); in skb_rbtree_purge()