/Linux-v5.4/tools/perf/tests/ |
D | hists_output.c | 109 he = rb_entry(node, struct hist_entry, rb_node); in del_hist_entries() 167 he = rb_entry(node, struct hist_entry, rb_node); in test1() 173 he = rb_entry(node, struct hist_entry, rb_node); in test1() 179 he = rb_entry(node, struct hist_entry, rb_node); in test1() 185 he = rb_entry(node, struct hist_entry, rb_node); in test1() 191 he = rb_entry(node, struct hist_entry, rb_node); in test1() 197 he = rb_entry(node, struct hist_entry, rb_node); in test1() 203 he = rb_entry(node, struct hist_entry, rb_node); in test1() 209 he = rb_entry(node, struct hist_entry, rb_node); in test1() 215 he = rb_entry(node, struct hist_entry, rb_node); in test1() [all …]
|
/Linux-v5.4/fs/f2fs/ |
D | extent_cache.c | 18 static struct rb_entry *__lookup_rb_tree_fast(struct rb_entry *cached_re, in __lookup_rb_tree_fast() 30 static struct rb_entry *__lookup_rb_tree_slow(struct rb_root_cached *root, in __lookup_rb_tree_slow() 34 struct rb_entry *re; in __lookup_rb_tree_slow() 37 re = rb_entry(node, struct rb_entry, rb_node); in __lookup_rb_tree_slow() 49 struct rb_entry *f2fs_lookup_rb_tree(struct rb_root_cached *root, in f2fs_lookup_rb_tree() 50 struct rb_entry *cached_re, unsigned int ofs) in f2fs_lookup_rb_tree() 52 struct rb_entry *re; in f2fs_lookup_rb_tree() 67 struct rb_entry *re; in f2fs_lookup_rb_tree_for_insert() 71 re = rb_entry(*parent, struct rb_entry, rb_node); in f2fs_lookup_rb_tree_for_insert() 95 struct rb_entry *f2fs_lookup_rb_tree_ret(struct rb_root_cached *root, in f2fs_lookup_rb_tree_ret() [all …]
|
/Linux-v5.4/include/linux/ |
D | interval_tree_generic.h | 48 parent = rb_entry(rb_parent, ITSTRUCT, ITRB); \ 89 ITSTRUCT *left = rb_entry(node->ITRB.rb_left, \ 108 node = rb_entry(node->ITRB.rb_right, \ 140 node = rb_entry(root->rb_root.rb_node, ITSTRUCT, ITRB); \ 144 leftmost = rb_entry(root->rb_leftmost, ITSTRUCT, ITRB); \ 165 ITSTRUCT *right = rb_entry(rb, ITSTRUCT, ITRB); \ 177 node = rb_entry(rb, ITSTRUCT, ITRB); \
|
D | rbtree_augmented.h | 80 RBSTRUCT *node = rb_entry(rb, RBSTRUCT, RBFIELD); \ 89 RBSTRUCT *old = rb_entry(rb_old, RBSTRUCT, RBFIELD); \ 90 RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \ 96 RBSTRUCT *old = rb_entry(rb_old, RBSTRUCT, RBFIELD); \ 97 RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \ 127 child = rb_entry(node->RBFIELD.rb_left, RBSTRUCT, RBFIELD); \ 132 child = rb_entry(node->RBFIELD.rb_right, RBSTRUCT, RBFIELD); \
|
/Linux-v5.4/fs/jffs2/ |
D | nodelist.h | 334 return rb_entry(node, struct jffs2_node_frag, rb); in frag_first() 344 return rb_entry(node, struct jffs2_node_frag, rb); in frag_last() 347 #define frag_next(frag) rb_entry(rb_next(&(frag)->rb), struct jffs2_node_frag, rb) 348 #define frag_prev(frag) rb_entry(rb_prev(&(frag)->rb), struct jffs2_node_frag, rb) 349 #define frag_parent(frag) rb_entry(rb_parent(&(frag)->rb), struct jffs2_node_frag, rb) 350 #define frag_left(frag) rb_entry((frag)->rb.rb_left, struct jffs2_node_frag, rb) 351 #define frag_right(frag) rb_entry((frag)->rb.rb_right, struct jffs2_node_frag, rb) 354 #define tn_next(tn) rb_entry(rb_next(&(tn)->rb), struct jffs2_tmp_dnode_info, rb) 355 #define tn_prev(tn) rb_entry(rb_prev(&(tn)->rb), struct jffs2_tmp_dnode_info, rb) 356 #define tn_parent(tn) rb_entry(rb_parent(&(tn)->rb), struct jffs2_tmp_dnode_info, rb) [all …]
|
/Linux-v5.4/drivers/block/drbd/ |
D | drbd_interval.c | 12 struct drbd_interval *this = rb_entry(node, struct drbd_interval, rb); in interval_end() 34 rb_entry(*new, struct drbd_interval, rb); in drbd_insert_interval() 75 rb_entry(node, struct drbd_interval, rb); in drbd_contains_interval() 122 rb_entry(node, struct drbd_interval, rb); in drbd_find_overlap() 151 i = rb_entry(node, struct drbd_interval, rb); in drbd_next_overlap()
|
/Linux-v5.4/tools/include/linux/ |
D | rbtree_augmented.h | 82 RBSTRUCT *node = rb_entry(rb, RBSTRUCT, RBFIELD); \ 91 RBSTRUCT *old = rb_entry(rb_old, RBSTRUCT, RBFIELD); \ 92 RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \ 98 RBSTRUCT *old = rb_entry(rb_old, RBSTRUCT, RBFIELD); \ 99 RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \ 129 child = rb_entry(node->RBFIELD.rb_left, RBSTRUCT, RBFIELD); \ 134 child = rb_entry(node->RBFIELD.rb_right, RBSTRUCT, RBFIELD); \
|
/Linux-v5.4/arch/x86/mm/ |
D | pat_rbtree.c | 51 struct memtype *data = rb_entry(node, struct memtype, rb); in get_subtree_max_end() 70 struct memtype *data = rb_entry(node, struct memtype, rb); in RB_DECLARE_CALLBACKS_MAX() 112 match = rb_entry(node, struct memtype, rb); in memtype_rb_match() 141 match = rb_entry(node, struct memtype, rb); in memtype_rb_check_conflict() 172 struct memtype *data = rb_entry(*node, struct memtype, rb); in memtype_rb_insert() 261 struct memtype *this = rb_entry(node, struct memtype, rb); in rbt_memtype_copy_nth_element()
|
/Linux-v5.4/net/ceph/ |
D | debugfs.c | 71 rb_entry(n, struct ceph_pg_pool_info, node); in osdmap_show() 92 rb_entry(n, struct ceph_pg_mapping, node); in osdmap_show() 103 rb_entry(n, struct ceph_pg_mapping, node); in osdmap_show() 110 rb_entry(n, struct ceph_pg_mapping, node); in osdmap_show() 121 rb_entry(n, struct ceph_pg_mapping, node); in osdmap_show() 160 req = rb_entry(rp, struct ceph_mon_generic_request, node); in monc_show() 236 rb_entry(n, struct ceph_osd_request, r_node); in dump_requests() 262 rb_entry(n, struct ceph_osd_linger_request, node); in dump_linger_requests() 323 rb_entry(n, struct ceph_osd_backoff, id_node); in dump_backoffs() 348 struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node); in osdc_show() [all …]
|
/Linux-v5.4/security/keys/ |
D | proc.c | 70 struct key *key = rb_entry(n, struct key, serial_node); in key_serial_next() 85 struct key *key = rb_entry(n, struct key, serial_node); in find_ge_key() 108 minkey = rb_entry(n, struct key, serial_node); in find_ge_key() 131 struct key *key = rb_entry(n, struct key, serial_node); in key_node_serial() 154 struct key *key = rb_entry(_p, struct key, serial_node); in proc_keys_show() 253 struct key_user *user = rb_entry(n, struct key_user, node); in __key_user_next() 304 struct key_user *user = rb_entry(_p, struct key_user, node); in proc_key_users_show()
|
/Linux-v5.4/arch/powerpc/kernel/ |
D | eeh_cache.c | 60 piar = rb_entry(n, struct pci_io_addr_range, rb_node); in __eeh_addr_cache_get_device() 106 piar = rb_entry(n, struct pci_io_addr_range, rb_node); in eeh_addr_cache_print() 128 piar = rb_entry(parent, struct pci_io_addr_range, rb_node); in eeh_addr_cache_insert() 229 piar = rb_entry(n, struct pci_io_addr_range, rb_node); in __eeh_addr_cache_rmv_dev() 278 piar = rb_entry(n, struct pci_io_addr_range, rb_node); in eeh_addr_cache_show()
|
/Linux-v5.4/fs/btrfs/ |
D | extent_map.c | 105 entry = rb_entry(parent, struct extent_map, rb_node); in tree_insert() 120 entry = rb_entry(parent, struct extent_map, rb_node); in tree_insert() 127 entry = rb_entry(parent, struct extent_map, rb_node); in tree_insert() 130 entry = rb_entry(parent, struct extent_map, rb_node); in tree_insert() 156 entry = rb_entry(n, struct extent_map, rb_node); in __tree_search() 172 prev_entry = rb_entry(prev, struct extent_map, rb_node); in __tree_search() 179 prev_entry = rb_entry(prev, struct extent_map, rb_node); in __tree_search() 182 prev_entry = rb_entry(prev, struct extent_map, rb_node); in __tree_search() 239 merge = rb_entry(rb, struct extent_map, rb_node); in try_merge_map() 258 merge = rb_entry(rb, struct extent_map, rb_node); in try_merge_map() [all …]
|
D | ordered-data.c | 40 entry = rb_entry(parent, struct btrfs_ordered_extent, rb_node); in tree_insert() 77 entry = rb_entry(n, struct btrfs_ordered_extent, rb_node); in __tree_search() 95 prev_entry = rb_entry(test, struct btrfs_ordered_extent, in __tree_search() 103 prev_entry = rb_entry(prev, struct btrfs_ordered_extent, in __tree_search() 109 prev_entry = rb_entry(test, struct btrfs_ordered_extent, in __tree_search() 150 entry = rb_entry(tree->last, struct btrfs_ordered_extent, in tree_search() 324 entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); in btrfs_dec_test_first_ordered_pending() 396 entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); in btrfs_dec_test_ordered_pending() 721 entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); in btrfs_lookup_ordered_extent() 751 entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); in btrfs_lookup_ordered_range() [all …]
|
/Linux-v5.4/drivers/gpu/drm/ |
D | drm_vma_manager.c | 152 node = rb_entry(iter, struct drm_mm_node, rb); in drm_vma_offset_lookup_locked() 282 entry = rb_entry(*iter, struct drm_vma_offset_file, vm_rb); in drm_vma_node_allow() 335 entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb); in drm_vma_node_revoke() 376 entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb); in drm_vma_node_is_allowed()
|
/Linux-v5.4/fs/ext4/ |
D | block_validity.c | 77 entry = rb_entry(parent, struct ext4_system_zone, node); in add_system_zone() 88 new_entry = rb_entry(new_node, struct ext4_system_zone, in add_system_zone() 110 entry = rb_entry(node, struct ext4_system_zone, node); in add_system_zone() 122 entry = rb_entry(node, struct ext4_system_zone, node); in add_system_zone() 141 entry = rb_entry(node, struct ext4_system_zone, node); in debug_print_tree() 175 entry = rb_entry(n, struct ext4_system_zone, node); in ext4_data_block_valid_rcu()
|
D | extents_status.c | 188 es = rb_entry(node, struct extent_status, rb_node); in ext4_es_print_tree() 217 es = rb_entry(node, struct extent_status, rb_node); in __es_tree_search() 231 return node ? rb_entry(node, struct extent_status, rb_node) : in __es_tree_search() 287 es1 = rb_entry(node, struct extent_status, rb_node); in __es_find_extent_range() 539 es1 = rb_entry(node, struct extent_status, rb_node); in ext4_es_try_to_merge_left() 563 es1 = rb_entry(node, struct extent_status, rb_node); in ext4_es_try_to_merge_right() 757 es = rb_entry(parent, struct extent_status, rb_node); in __es_insert_extent() 931 es1 = rb_entry(node, struct extent_status, rb_node); in ext4_es_lookup_extent() 955 es1 = rb_entry(node, struct extent_status, in ext4_es_lookup_extent() 1012 rc->left_es = node ? rb_entry(node, in init_rsvd() [all …]
|
/Linux-v5.4/lib/ |
D | rbtree_test.c | 39 if (key < rb_entry(parent, struct test_node, rb)->key) in insert() 57 if (key < rb_entry(parent, struct test_node, rb)->key) in insert_cached() 95 parent = rb_entry(rb_parent, struct test_node, rb); in RB_DECLARE_CALLBACKS_MAX() 120 parent = rb_entry(rb_parent, struct test_node, rb); in insert_augmented_cached() 198 struct test_node *node = rb_entry(rb, struct test_node, rb); in check() 224 struct test_node *node = rb_entry(rb, struct test_node, rb); in check_augmented() 227 subtree = rb_entry(node->rb.rb_left, struct test_node, in check_augmented() 233 subtree = rb_entry(node->rb.rb_right, struct test_node, in check_augmented()
|
/Linux-v5.4/tools/perf/util/ |
D | env.c | 29 node = rb_entry(parent, struct bpf_prog_info_node, rb_node); in perf_env__insert_bpf_prog_info() 57 node = rb_entry(n, struct bpf_prog_info_node, rb_node); in perf_env__find_bpf_prog_info() 84 node = rb_entry(parent, struct btf_node, rb_node); in perf_env__insert_btf() 111 node = rb_entry(n, struct btf_node, rb_node); in perf_env__find_btf() 140 node = rb_entry(next, struct bpf_prog_info_node, rb_node); in perf_env__purge_bpf() 154 node = rb_entry(next, struct btf_node, rb_node); in perf_env__purge_bpf()
|
D | hist.c | 233 n = rb_entry(next, struct hist_entry, rb_node); in hists__output_recalc_col_len() 323 child = rb_entry(node, struct hist_entry, rb_node); in hists__decay_entry() 366 n = rb_entry(next, struct hist_entry, rb_node); in hists__decay_entries() 382 n = rb_entry(next, struct hist_entry, rb_node); in hists__delete_entries() 396 n = rb_entry(next, struct hist_entry, rb_node); in hists__get_entry() 581 he = rb_entry(parent, struct hist_entry, rb_node_in); in hists__findnew_entry() 1409 iter = rb_entry(parent, struct hist_entry, rb_node_in); in hierarchy_insert_entry() 1528 iter = rb_entry(parent, struct hist_entry, rb_node_in); in hists__collapse_insert_entry() 1606 n = rb_entry(next, struct hist_entry, rb_node_in); in hists__collapse_resort() 1691 he = rb_entry(node, struct hist_entry, rb_node); in hierarchy_recalc_total_periods() [all …]
|
D | rb_resort.h | 66 a = rb_entry(nda, struct __name##_sorted_entry, rb_node); \ 67 b = rb_entry(ndb, struct __name##_sorted_entry, rb_node); \ 129 __name##_entry = rb_entry(__nd, struct __name##_sorted_entry, \
|
D | map.c | 312 struct symbol *sym = rb_entry(nd, struct symbol, rb_node); in map__fixup_start() 322 struct symbol *sym = rb_entry(nd, struct symbol, rb_node); in map__fixup_end() 601 struct map *pos = rb_entry(next, struct map, rb_node); in __maps__purge() 615 struct map *pos = rb_entry(next, struct map, rb_node_name); in __maps__purge_names() 695 struct map *pos = rb_entry(nd, struct map, rb_node); in maps__find_symbol_by_name() 747 struct map *pos = rb_entry(nd, struct map, rb_node); in maps__fprintf() 790 struct map *pos = rb_entry(next, struct map, rb_node); in maps__fixup_overlappings() 803 struct map *pos = rb_entry(next, struct map, rb_node); in maps__fixup_overlappings() 920 m = rb_entry(parent, struct map, rb_node); in __maps__insert() 941 m = rb_entry(parent, struct map, rb_node_name); in __maps__insert_name() [all …]
|
D | symbol.c | 196 curr = rb_entry(nd, struct symbol, rb_node); in symbols__fixup_duplicate() 199 next = rb_entry(nd, struct symbol, rb_node); in symbols__fixup_duplicate() 227 curr = rb_entry(prevnd, struct symbol, rb_node); in symbols__fixup_end() 231 curr = rb_entry(nd, struct symbol, rb_node); in symbols__fixup_end() 310 pos = rb_entry(next, struct symbol, rb_node); in symbols__delete() 339 s = rb_entry(parent, struct symbol, rb_node); in __symbols__insert() 366 struct symbol *s = rb_entry(n, struct symbol, rb_node); in symbols__find() 384 return rb_entry(n, struct symbol, rb_node); in symbols__first() 394 return rb_entry(n, struct symbol, rb_node); in symbols__last() 404 return rb_entry(n, struct symbol, rb_node); in symbols__next() [all …]
|
/Linux-v5.4/fs/ocfs2/ |
D | reservations.c | 90 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); in ocfs2_dump_resv() 145 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); in ocfs2_check_resmap() 280 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); in ocfs2_resmap_clear_all_resv() 321 tmp = rb_entry(parent, struct ocfs2_alloc_reservation, r_node); in ocfs2_resv_insert() 372 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); in ocfs2_find_resv_lhs() 530 next_resv = rb_entry(next, struct ocfs2_alloc_reservation, in __ocfs2_resv_find_window() 567 next_resv = rb_entry(next, in __ocfs2_resv_find_window() 610 prev_resv = rb_entry(prev, struct ocfs2_alloc_reservation, in __ocfs2_resv_find_window()
|
/Linux-v5.4/drivers/base/regmap/ |
D | regcache-rbtree.c | 80 rbnode = rb_entry(node, struct regcache_rbtree_node, node); in regcache_rbtree_lookup() 107 rbnode_tmp = rb_entry(*new, struct regcache_rbtree_node, node); in regcache_rbtree_insert() 150 n = rb_entry(node, struct regcache_rbtree_node, node); in rbtree_show() 227 rbtree_node = rb_entry(next, struct regcache_rbtree_node, node); in regcache_rbtree_exit() 405 rbnode_tmp = rb_entry(node, struct regcache_rbtree_node, in regcache_rbtree_write() 477 rbnode = rb_entry(node, struct regcache_rbtree_node, node); in regcache_rbtree_sync() 517 rbnode = rb_entry(node, struct regcache_rbtree_node, node); in regcache_rbtree_drop()
|
/Linux-v5.4/Documentation/ |
D | rbtree.txt | 71 individual members may be accessed directly via rb_entry(node, type, member). 188 rb_entry(node, type, member). 194 printk("key=%s\n", rb_entry(node, struct mytype, node)->keystring); 309 node = rb_entry(root->rb_node, struct interval_tree_node, rb); 314 rb_entry(node->rb.rb_left, 333 node = rb_entry(node->rb.rb_right, 350 subtree_last = rb_entry(node->rb.rb_left, 356 subtree_last = rb_entry(node->rb.rb_right, 368 rb_entry(rb, struct interval_tree_node, rb); 380 rb_entry(rb_old, struct interval_tree_node, rb); [all …]
|