Home
last modified time | relevance | path

Searched refs:rb_entry_safe (Results 1 – 14 of 14) sorted by relevance

/linux/tools/include/linux/
A Drbtree.h75 #define rb_entry_safe(ptr, type, member) \ macro
98 for (pos = rb_entry_safe(rb_first_postorder(root), typeof(*pos), field); \
99 pos && ({ n = rb_entry_safe(rb_next_postorder(&pos->field), \
/linux/include/linux/
A Drbtree.h77 #define rb_entry_safe(ptr, type, member) \ macro
100 for (pos = rb_entry_safe(rb_first_postorder(root), typeof(*pos), field); \
101 pos && ({ n = rb_entry_safe(rb_next_postorder(&pos->field), \
A Dtimerqueue.h27 return rb_entry_safe(leftmost, struct timerqueue_node, node); in timerqueue_getnext()
A Dskbuff.h3987 #define rb_to_skb(rb) rb_entry_safe(rb, struct sk_buff, rbnode)
/linux/fs/xfs/scrub/
A Dbitmap.c63 for ((bn) = rb_entry_safe(rb_first(&(bitmap)->xb_root.rb_root), \ in INTERVAL_TREE_DEFINE()
66 (bn) = rb_entry_safe(rb_next(&(bn)->bn_rbnode), \
338 for ((bn) = rb_entry_safe(rb_first(&(bitmap)->xb_root.rb_root), \ in INTERVAL_TREE_DEFINE()
341 (bn) = rb_entry_safe(rb_next(&(bn)->bn_rbnode), \
/linux/fs/f2fs/
A Dextent_cache.c230 *next_entry = rb_entry_safe(tmp_node, struct extent_node, rb_node); in __lookup_extent_node_ret()
235 *prev_entry = rb_entry_safe(tmp_node, struct extent_node, rb_node); in __lookup_extent_node_ret()
242 *prev_entry = rb_entry_safe(tmp_node, in __lookup_extent_node_ret()
248 *next_entry = rb_entry_safe(tmp_node, in __lookup_extent_node_ret()
676 next_en = rb_entry_safe(node, struct extent_node, in __update_extent_tree_range()
A Dsegment.c1067 *next_entry = rb_entry_safe(tmp_node, struct discard_cmd, rb_node); in __lookup_discard_cmd_ret()
1072 *prev_entry = rb_entry_safe(tmp_node, struct discard_cmd, rb_node); in __lookup_discard_cmd_ret()
1078 *prev_entry = rb_entry_safe(tmp_node, struct discard_cmd, rb_node); in __lookup_discard_cmd_ret()
1082 *next_entry = rb_entry_safe(tmp_node, struct discard_cmd, rb_node); in __lookup_discard_cmd_ret()
1544 next_dc = rb_entry_safe(node, struct discard_cmd, rb_node); in __update_discard_tree_range()
1620 dc = rb_entry_safe(node, struct discard_cmd, rb_node); in __issue_discard_cmd_orderly()
3341 dc = rb_entry_safe(node, struct discard_cmd, rb_node); in __issue_discard_cmd_range()
A Dgc.c587 ve = rb_entry_safe(node, struct victim_entry, rb_node); in atgc_lookup_victim()
/linux/drivers/gpu/drm/
A Ddrm_mm.c297 return rb_entry_safe(rb, struct drm_mm_node, rb_hole_size); in rb_hole_size_to_node()
302 return rb_entry_safe(rb, struct drm_mm_node, rb_hole_addr); in rb_hole_addr_to_node()
/linux/fs/proc/
A Dgeneric.c58 return rb_entry_safe(rb_first(&dir->subdir), struct proc_dir_entry, in pde_subdir_first()
64 return rb_entry_safe(rb_next(&dir->subdir_node), struct proc_dir_entry, in pde_subdir_next()
/linux/drivers/gpu/drm/i915/
A Di915_drv.h386 rb_entry_safe(rb, struct intel_engine_cs, uabi_node)
/linux/drivers/gpu/drm/i915/selftests/
A Dintel_memory_region.c1010 engine = rb_entry_safe(rb_next(&engine->uabi_node), in random_engine_class()
/linux/mm/
A Dvmalloc.c981 va = rb_entry_safe(node, struct vmap_area, rb_node); in get_subtree_max_size()
4519 return rb_entry_safe(n, struct vmap_area, rb_node); in node_to_va()
/linux/kernel/events/
A Dcore.c1802 for (event = rb_entry_safe(rb_first(&((groups)->tree)), \
1804 event = rb_entry_safe(rb_next(&event->group_node), \

Completed in 89 milliseconds