| /linux/tools/perf/tests/ |
| H A D | hists_output.c | 112 he = rb_entry(node, struct hist_entry, rb_node); in del_hist_entries() 180 he = rb_entry(node, struct hist_entry, rb_node); in test1() 186 he = rb_entry(node, struct hist_entry, rb_node); in test1() 192 he = rb_entry(node, struct hist_entry, rb_node); in test1() 198 he = rb_entry(node, struct hist_entry, rb_node); in test1() 204 he = rb_entry(node, struct hist_entry, rb_node); in test1() 210 he = rb_entry(node, struct hist_entry, rb_node); in test1() 216 he = rb_entry(node, struct hist_entry, rb_node); in test1() 222 he = rb_entry(node, struct hist_entry, rb_node); in test1() 228 he = rb_entry(node, struct hist_entry, rb_node); in test1() [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | rbtree_search.c | 24 #define rb_entry(ptr, type, member) container_of(ptr, type, member) macro 34 node_a = rb_entry(a, struct node_data, r0); in less0() 35 node_b = rb_entry(b, struct node_data, r0); in less0() 45 node_a = rb_entry(a, struct node_data, r1); in less1() 46 node_b = rb_entry(b, struct node_data, r1); in less1() 88 n = rb_entry(rb_n, struct node_data, r0); in rbtree_search() 115 n = rb_entry(rb_n, struct node_data, r0); in rbtree_search() 129 bpf_obj_drop(rb_entry(rb_m, struct node_data, r1)); in rbtree_search() 168 n = rb_entry(rb_n, struct node_data, r0); \
|
| /linux/Documentation/translations/zh_CN/core-api/ |
| H A D | rbtree.rst | 65 宏访问。此外,个体成员可直接用rb_entry(node, type, member)访问。 175 标准的container_of()宏访问。此外,个体成员可直接用rb_entry(node, type, member) 182 printk("key=%s\n", rb_entry(node, struct mytype, node)->keystring); 271 node = rb_entry(root->rb_node, struct interval_tree_node, rb); 276 rb_entry(node->rb.rb_left, 295 node = rb_entry(node->rb.rb_right, 312 subtree_last = rb_entry(node->rb.rb_left, 318 subtree_last = rb_entry(node->rb.rb_right, 330 rb_entry(rb, struct interval_tree_node, rb); 342 rb_entry(rb_old, struct interval_tree_node, rb); [all …]
|
| /linux/include/linux/ |
| H A D | interval_tree_generic.h | 48 parent = rb_entry(rb_parent, ITSTRUCT, ITRB); \ 89 ITSTRUCT *left = rb_entry(node->ITRB.rb_left, \ 107 node = rb_entry(node->ITRB.rb_right, ITSTRUCT, ITRB); \ 136 node = rb_entry(root->rb_root.rb_node, ITSTRUCT, ITRB); \ 140 leftmost = rb_entry(root->rb_leftmost, ITSTRUCT, ITRB); \ 161 ITSTRUCT *right = rb_entry(rb, ITSTRUCT, ITRB); \ 173 node = rb_entry(rb, ITSTRUCT, ITRB); \
|
| H A D | rbtree_augmented.h | 106 RBSTRUCT *node = rb_entry(rb, RBSTRUCT, RBFIELD); \ 115 RBSTRUCT *old = rb_entry(rb_old, RBSTRUCT, RBFIELD); \ 116 RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \ 122 RBSTRUCT *old = rb_entry(rb_old, RBSTRUCT, RBFIELD); \ 123 RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \ 153 child = rb_entry(node->RBFIELD.rb_left, RBSTRUCT, RBFIELD); \ 158 child = rb_entry(node->RBFIELD.rb_right, RBSTRUCT, RBFIELD); \
|
| /linux/tools/include/linux/ |
| H A D | interval_tree_generic.h | 48 parent = rb_entry(rb_parent, ITSTRUCT, ITRB); \ 89 ITSTRUCT *left = rb_entry(node->ITRB.rb_left, \ 107 node = rb_entry(node->ITRB.rb_right, ITSTRUCT, ITRB); \ 136 node = rb_entry(root->rb_root.rb_node, ITSTRUCT, ITRB); \ 140 leftmost = rb_entry(root->rb_leftmost, ITSTRUCT, ITRB); \ 161 ITSTRUCT *right = rb_entry(rb, ITSTRUCT, ITRB); \ 173 node = rb_entry(rb, ITSTRUCT, ITRB); \
|
| H A D | rbtree_augmented.h | 82 RBSTRUCT *node = rb_entry(rb, RBSTRUCT, RBFIELD); \ 91 RBSTRUCT *old = rb_entry(rb_old, RBSTRUCT, RBFIELD); \ 92 RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \ 98 RBSTRUCT *old = rb_entry(rb_old, RBSTRUCT, RBFIELD); \ 99 RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \ 129 child = rb_entry(node->RBFIELD.rb_left, RBSTRUCT, RBFIELD); \ 134 child = rb_entry(node->RBFIELD.rb_right, RBSTRUCT, RBFIELD); \
|
| /linux/fs/jffs2/ |
| H A D | nodelist.h | 334 return rb_entry(node, struct jffs2_node_frag, rb); in frag_first() 344 return rb_entry(node, struct jffs2_node_frag, rb); in frag_last() 347 #define frag_next(frag) rb_entry(rb_next(&(frag)->rb), struct jffs2_node_frag, rb) 348 #define frag_prev(frag) rb_entry(rb_prev(&(frag)->rb), struct jffs2_node_frag, rb) 349 #define frag_parent(frag) rb_entry(rb_parent(&(frag)->rb), struct jffs2_node_frag, rb) 350 #define frag_left(frag) rb_entry((frag)->rb.rb_left, struct jffs2_node_frag, rb) 351 #define frag_right(frag) rb_entry((frag)->rb.rb_right, struct jffs2_node_frag, rb) 354 #define tn_next(tn) rb_entry(rb_next(&(tn)->rb), struct jffs2_tmp_dnode_info, rb) 355 #define tn_prev(tn) rb_entry(rb_prev(&(tn)->rb), struct jffs2_tmp_dnode_info, rb) 356 #define tn_parent(tn) rb_entry(rb_parent(&(tn)->rb), struct jffs2_tmp_dnode_info, rb) [all …]
|
| /linux/drivers/block/drbd/ |
| H A D | drbd_interval.c | 12 struct drbd_interval *this = rb_entry(node, struct drbd_interval, rb); in interval_end() 34 rb_entry(*new, struct drbd_interval, rb); in drbd_insert_interval() 76 rb_entry(node, struct drbd_interval, rb); in drbd_contains_interval() 128 rb_entry(node, struct drbd_interval, rb); in drbd_find_overlap() 157 i = rb_entry(node, struct drbd_interval, rb); in drbd_next_overlap()
|
| /linux/net/ceph/ |
| H A D | debugfs.c | 76 rb_entry(n, struct ceph_pg_pool_info, node); in osdmap_show() 99 rb_entry(n, struct ceph_pg_mapping, node); in osdmap_show() 110 rb_entry(n, struct ceph_pg_mapping, node); in osdmap_show() 117 rb_entry(n, struct ceph_pg_mapping, node); in osdmap_show() 128 rb_entry(n, struct ceph_pg_mapping, node); in osdmap_show() 168 req = rb_entry(rp, struct ceph_mon_generic_request, node); in monc_show() 247 rb_entry(n, struct ceph_osd_request, r_node); in dump_requests() 273 rb_entry(n, struct ceph_osd_linger_request, node); in dump_linger_requests() 334 rb_entry(n, struct ceph_osd_backoff, id_node); in dump_backoffs() 359 struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node); in osdc_show() [all …]
|
| /linux/security/keys/ |
| H A D | proc.c | 70 struct key *key = rb_entry(n, struct key, serial_node); in key_serial_next() 85 struct key *key = rb_entry(n, struct key, serial_node); in find_ge_key() 108 minkey = rb_entry(n, struct key, serial_node); in find_ge_key() 131 struct key *key = rb_entry(n, struct key, serial_node); in key_node_serial() 156 struct key *key = rb_entry(_p, struct key, serial_node); in proc_keys_show() 255 struct key_user *user = rb_entry(n, struct key_user, node); in __key_user_next() 306 struct key_user *user = rb_entry(_p, struct key_user, node); in proc_key_users_show()
|
| /linux/arch/powerpc/kernel/ |
| H A D | eeh_cache.c | 60 piar = rb_entry(n, struct pci_io_addr_range, rb_node); in __eeh_addr_cache_get_device() 106 piar = rb_entry(n, struct pci_io_addr_range, rb_node); in eeh_addr_cache_print() 128 piar = rb_entry(parent, struct pci_io_addr_range, rb_node); in eeh_addr_cache_insert() 221 piar = rb_entry(n, struct pci_io_addr_range, rb_node); in __eeh_addr_cache_rmv_dev() 271 piar = rb_entry(n, struct pci_io_addr_range, rb_node); in eeh_addr_cache_show()
|
| /linux/net/netfilter/ |
| H A D | nft_set_rbtree.c | 221 rbe_prev = rb_entry(prev, struct nft_rbtree_elem, node); in nft_rbtree_gc_elem() 231 rbe_prev = rb_entry(prev, struct nft_rbtree_elem, node); in nft_rbtree_gc_elem() 246 first_elem = rb_entry(first, struct nft_rbtree_elem, node); in nft_rbtree_update_first() 265 return rb_entry(node, struct nft_rbtree_elem, node); in nft_rbtree_prev_active() 274 next_rbe = rb_entry(node, struct nft_rbtree_elem, node); in __nft_rbtree_next_active() 357 rbe = rb_entry(parent, struct nft_rbtree_elem, node); in __nft_rbtree_insert() 386 rbe = rb_entry(node, struct nft_rbtree_elem, node); in __nft_rbtree_insert() 523 rbe = rb_entry(parent, struct nft_rbtree_elem, node); in __nft_rbtree_insert() 742 next_rbe = rb_entry(node, struct nft_rbtree_elem, node); in nft_rbtree_next_inactive() 806 rbe = rb_entry(parent, struct nft_rbtree_elem, node); in nft_rbtree_deactivate() [all …]
|
| /linux/lib/ |
| H A D | rbtree_test.c | 40 if (key < rb_entry(parent, struct test_node, rb)->key) in insert() 58 if (key < rb_entry(parent, struct test_node, rb)->key) in insert_cached() 96 parent = rb_entry(rb_parent, struct test_node, rb); in RB_DECLARE_CALLBACKS_MAX() 121 parent = rb_entry(rb_parent, struct test_node, rb); in insert_augmented_cached() 199 struct test_node *node = rb_entry(rb, struct test_node, rb); in check() 225 struct test_node *node = rb_entry(rb, struct test_node, rb); in check_augmented() 228 subtree = rb_entry(node->rb.rb_left, struct test_node, in check_augmented() 234 subtree = rb_entry(node->rb.rb_right, struct test_node, in check_augmented()
|
| /linux/fs/btrfs/ |
| H A D | misc.h | 133 entry = rb_entry(node, struct rb_simple_node, rb_node); in rb_simple_search() 161 entry = rb_entry(node, struct rb_simple_node, rb_node); in rb_simple_search_first() 182 struct rb_simple_node *new_entry = rb_entry(new, struct rb_simple_node, rb_node); in rb_simple_node_bytenr_cmp() 183 struct rb_simple_node *existing_entry = rb_entry(existing, struct rb_simple_node, rb_node); in rb_simple_node_bytenr_cmp()
|
| H A D | ordered-data.c | 45 entry = rb_entry(parent, struct btrfs_ordered_extent, rb_node); in tree_insert() 74 entry = rb_entry(n, struct btrfs_ordered_extent, rb_node); in __tree_search() 92 prev_entry = rb_entry(test, struct btrfs_ordered_extent, in __tree_search() 100 prev_entry = rb_entry(prev, struct btrfs_ordered_extent, in __tree_search() 106 prev_entry = rb_entry(test, struct btrfs_ordered_extent, in __tree_search() 135 entry = rb_entry(inode->ordered_tree_last, struct btrfs_ordered_extent, in ordered_tree_search() 499 entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); in btrfs_mark_ordered_io_finished() 511 entry = rb_entry(node, struct btrfs_ordered_extent, in btrfs_mark_ordered_io_finished() 584 entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); in btrfs_dec_test_ordered_pending() 975 entry = rb_entry(node, struct btrfs_ordered_extent, rb_node); in btrfs_lookup_ordered_extent() [all …]
|
| H A D | ref-verify.c | 81 const struct block_entry *entry = rb_entry(node, struct block_entry, node); in block_entry_bytenr_key_cmp() 93 const struct block_entry *new_entry = rb_entry(new, struct block_entry, node); in block_entry_bytenr_cmp() 118 const struct root_entry *entry = rb_entry(node, struct root_entry, node); in root_entry_root_objectid_key_cmp() 130 const struct root_entry *new_entry = rb_entry(new, struct root_entry, node); in root_entry_root_objectid_cmp() 167 struct ref_entry *new_entry = rb_entry(new, struct ref_entry, node); in ref_entry_cmp() 168 struct ref_entry *existing_entry = rb_entry(existing, struct ref_entry, node); in ref_entry_cmp() 225 re = rb_entry(n, struct root_entry, node); in free_block_entry() 231 ref = rb_entry(n, struct ref_entry, node); in free_block_entry() 627 ref = rb_entry(n, struct ref_entry, node); in dump_block_entry() 635 re = rb_entry(n, struct root_entry, node); in dump_block_entry() [all …]
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_vma_manager.c | 153 node = rb_entry(iter, struct drm_mm_node, rb); in drm_vma_offset_lookup_locked() 264 entry = rb_entry(*iter, struct drm_vma_offset_file, vm_rb); in vma_node_allow() 368 entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb); in drm_vma_node_revoke() 409 entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb); in drm_vma_node_is_allowed()
|
| /linux/tools/perf/util/ |
| H A D | hist.c | 272 n = rb_entry(next, struct hist_entry, rb_node); in hists__output_recalc_col_len() 434 child = rb_entry(node, struct hist_entry, rb_node); in hists__decay_entry() 477 n = rb_entry(next, struct hist_entry, rb_node); in hists__decay_entries() 493 n = rb_entry(next, struct hist_entry, rb_node); in hists__delete_entries() 507 n = rb_entry(next, struct hist_entry, rb_node); in hists__get_entry() 696 he = rb_entry(parent, struct hist_entry, rb_node_in); in hists__findnew_entry() 1638 iter = rb_entry(parent, struct hist_entry, rb_node_in); in hierarchy_insert_entry() 1761 iter = rb_entry(parent, struct hist_entry, rb_node_in); in hists__collapse_insert_entry() 1845 n = rb_entry(next, struct hist_entry, rb_node_in); in hists__collapse_resort() 1918 he = rb_entry(node, struct hist_entry, rb_node); in hierarchy_recalc_total_periods() [all …]
|
| /linux/fs/ext4/ |
| H A D | extents_status.c | 222 es = rb_entry(node, struct extent_status, rb_node); in ext4_es_print_tree() 273 es = rb_entry(node, struct extent_status, rb_node); in __es_tree_search() 287 return node ? rb_entry(node, struct extent_status, rb_node) : in __es_tree_search() 341 es1 = rb_entry(node, struct extent_status, rb_node); in __es_find_extent_range() 634 es1 = rb_entry(node, struct extent_status, rb_node); in ext4_es_try_to_merge_left() 658 es1 = rb_entry(node, struct extent_status, rb_node); in ext4_es_try_to_merge_right() 852 es = rb_entry(parent, struct extent_status, rb_node); in __es_insert_extent() 1123 es1 = rb_entry(node, struct extent_status, rb_node); in ext4_es_lookup_extent() 1147 es1 = rb_entry(node, struct extent_status, in ext4_es_lookup_extent() 1206 rc->left_es = node ? rb_entry(node, in init_rsvd() [all …]
|
| /linux/fs/ocfs2/ |
| H A D | reservations.c | 85 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); in ocfs2_dump_resv() 140 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); in ocfs2_check_resmap() 273 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); in ocfs2_resmap_clear_all_resv() 314 tmp = rb_entry(parent, struct ocfs2_alloc_reservation, r_node); in ocfs2_resv_insert() 365 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); in ocfs2_find_resv_lhs() 523 next_resv = rb_entry(next, struct ocfs2_alloc_reservation, in __ocfs2_resv_find_window() 560 next_resv = rb_entry(next, in __ocfs2_resv_find_window() 603 prev_resv = rb_entry(prev, struct ocfs2_alloc_reservation, in __ocfs2_resv_find_window()
|
| /linux/Documentation/core-api/ |
| H A D | rbtree.rst | 71 individual members may be accessed directly via rb_entry(node, type, member). 188 rb_entry(node, type, member). 194 printk("key=%s\n", rb_entry(node, struct mytype, node)->keystring); 309 node = rb_entry(root->rb_node, struct interval_tree_node, rb); 314 rb_entry(node->rb.rb_left, 333 node = rb_entry(node->rb.rb_right, 350 subtree_last = rb_entry(node->rb.rb_left, 356 subtree_last = rb_entry(node->rb.rb_right, 368 rb_entry(rb, struct interval_tree_node, rb); 380 rb_entry(rb_old, struct interval_tree_node, rb); [all …]
|
| /linux/fs/ntfs3/ |
| H A D | bitmap.c | 137 rb_entry(node, struct e_node, start.node)); in wnd_close() 150 k = rb_entry(*p, struct rb_node_key, node); in rb_lookup() 176 rb_entry(parent = *p, struct e_node, count.node); in rb_insert_count() 211 k = rb_entry(parent, struct e_node, start.node); in rb_insert_start() 252 e = rb_entry(n, struct e_node, start.node); in wnd_add_free_ext() 268 e = rb_entry(n, struct e_node, start.node); in wnd_add_free_ext() 320 e = rb_entry(n, struct e_node, count.node); in wnd_add_free_ext() 328 e2 = rb_entry(n, struct e_node, count.node); in wnd_add_free_ext() 375 e = rb_entry(n, struct e_node, start.node); in wnd_remove_free_ext() 395 e3 = rb_entry(n3, struct e_node, start.node); in wnd_remove_free_ext() [all …]
|
| /linux/arch/x86/events/intel/ |
| H A D | uncore_discovery.c | 33 rb_entry((cur), struct intel_uncore_discovery_type, node) 103 unit = rb_entry(b, struct intel_uncore_discovery_unit, node); in pmu_idx_cmp() 126 unit = rb_entry(pos, struct intel_uncore_discovery_unit, node); in intel_uncore_find_discovery_unit() 132 unit = rb_entry(pos, struct intel_uncore_discovery_unit, node); in intel_uncore_find_discovery_unit() 160 a_node = rb_entry(a, struct intel_uncore_discovery_unit, node); in unit_less() 161 b_node = rb_entry(b, struct intel_uncore_discovery_unit, node); in unit_less() 183 unit = rb_entry(node, struct intel_uncore_discovery_unit, node); in uncore_find_unit() 446 pos = rb_entry(node, struct intel_uncore_discovery_unit, node); in intel_uncore_clear_discovery_tables() 753 type = rb_entry(node, struct intel_uncore_discovery_type, node); in intel_uncore_generic_init_uncores()
|
| /linux/drivers/base/regmap/ |
| H A D | regcache-rbtree.c | 80 rbnode = rb_entry(node, struct regcache_rbtree_node, node); in regcache_rbtree_lookup() 107 rbnode_tmp = rb_entry(*new, struct regcache_rbtree_node, node); in regcache_rbtree_insert() 150 n = rb_entry(node, struct regcache_rbtree_node, node); in rbtree_show() 213 rbtree_node = rb_entry(next, struct regcache_rbtree_node, node); in regcache_rbtree_exit() 404 rbnode_tmp = rb_entry(node, struct regcache_rbtree_node, in regcache_rbtree_write() 479 rbnode = rb_entry(node, struct regcache_rbtree_node, node); in regcache_rbtree_sync() 521 rbnode = rb_entry(node, struct regcache_rbtree_node, node); in regcache_rbtree_drop()
|