Home
last modified time | relevance | path

Searched refs:nr_entries (Results 1 – 25 of 174) sorted by relevance

1234567

/linux/fs/xfs/libxfs/
A Dxfs_iext_tree.c461 int *nr_entries) in xfs_iext_split_node() argument
473 *nr_entries = 0; in xfs_iext_split_node()
489 *nr_entries = nr_move; in xfs_iext_split_node()
549 int *nr_entries) in xfs_iext_split_leaf() argument
561 *nr_entries = 0; in xfs_iext_split_leaf()
640 int nr_entries, i; in xfs_iext_insert_raw() local
693 int nr_entries) in xfs_iext_rebalance_node() argument
700 if (nr_entries == 0) in xfs_iext_rebalance_node()
810 int nr_entries) in xfs_iext_rebalance_leaf() argument
817 if (nr_entries == 0) in xfs_iext_rebalance_leaf()
[all …]
/linux/lib/
A Dstackdepot.c389 if (WARN_ON_ONCE(!nr_entries)) in depot_alloc_stack()
394 nr_entries = CONFIG_STACKDEPOT_MAX_FRAMES; in depot_alloc_stack()
415 stack->size = nr_entries; in depot_alloc_stack()
585 unsigned int nr_entries, in stack_depot_save_flags() argument
609 nr_entries = filter_irq_stacks(entries, nr_entries); in stack_depot_save_flags()
614 hash = hash_stack(entries, nr_entries); in stack_depot_save_flags()
675 unsigned int nr_entries, in stack_depot_save() argument
742 unsigned int nr_entries; in stack_depot_print() local
745 if (nr_entries > 0) in stack_depot_print()
754 unsigned int nr_entries; in stack_depot_snprint() local
[all …]
/linux/drivers/dma/dw-edma/
A Ddw-edma-v0-debugfs.c101 int nr_entries, struct dentry *dent) in dw_edma_debugfs_create_x32() argument
111 for (i = 0; i < nr_entries; i++) { in dw_edma_debugfs_create_x32()
133 int nr_entries; in dw_edma_debugfs_regs_ch() local
135 nr_entries = ARRAY_SIZE(debugfs_regs); in dw_edma_debugfs_regs_ch()
178 int nr_entries, i; in dw_edma_debugfs_regs_wr() local
183 nr_entries = ARRAY_SIZE(debugfs_regs); in dw_edma_debugfs_regs_wr()
187 nr_entries = ARRAY_SIZE(debugfs_unroll_regs); in dw_edma_debugfs_regs_wr()
241 int nr_entries, i; in dw_edma_debugfs_regs_rd() local
246 nr_entries = ARRAY_SIZE(debugfs_regs); in dw_edma_debugfs_regs_rd()
271 int nr_entries; in dw_edma_debugfs_regs() local
[all …]
/linux/kernel/
A Dstacktrace.c24 void stack_trace_print(const unsigned long *entries, unsigned int nr_entries, in stack_trace_print() argument
32 for (i = 0; i < nr_entries; i++) in stack_trace_print()
48 unsigned int nr_entries, int spaces) in stack_trace_snprint() argument
55 for (i = 0; i < nr_entries && size; i++) { in stack_trace_snprint()
278 return trace.nr_entries; in stack_trace_save()
303 return trace.nr_entries; in stack_trace_save_tsk()
326 return trace.nr_entries; in stack_trace_save_regs()
351 return ret ? ret : trace.nr_entries; in stack_trace_save_tsk_reliable()
371 return trace.nr_entries; in stack_trace_save_user()
396 for (i = 0; i < nr_entries; i++) { in filter_irq_stacks()
[all …]
A Dbacktracetest.c44 unsigned int nr_entries; in backtrace_test_saved() local
49 nr_entries = stack_trace_save(entries, ARRAY_SIZE(entries), 0); in backtrace_test_saved()
50 stack_trace_print(entries, nr_entries, 0); in backtrace_test_saved()
/linux/drivers/md/persistent-data/
A Ddm-btree-remove.c62 uint32_t nr_entries = le32_to_cpu(n->header.nr_entries); in node_shift() local
67 BUG_ON(shift > nr_entries); in node_shift()
71 (nr_entries - shift) * sizeof(__le64)); in node_shift()
74 (nr_entries - shift) * value_size); in node_shift()
79 nr_entries * sizeof(__le64)); in node_shift()
82 nr_entries * value_size); in node_shift()
131 unsigned int nr_entries = le32_to_cpu(n->header.nr_entries); in delete_at() local
135 BUG_ON(index >= nr_entries); in delete_at()
147 n->header.nr_entries = cpu_to_le32(nr_entries - 1); in delete_at()
491 if (le32_to_cpu(n->header.nr_entries) == 1) { in rebalance_children()
[all …]
A Ddm-btree.c77 uint32_t nr_entries = le32_to_cpu(n->header.nr_entries); in inc_children() local
90 uint32_t nr_entries = le32_to_cpu(node->header.nr_entries); in insert_at() local
94 if (index > nr_entries || in insert_at()
106 node->header.nr_entries = cpu_to_le32(nr_entries + 1); in insert_at()
348 uint32_t flags, nr_entries; in btree_lookup_raw() local
358 nr_entries = le32_to_cpu(ro_node(s)->header.nr_entries); in btree_lookup_raw()
423 uint32_t flags, nr_entries; in dm_btree_lookup_next_single() local
433 nr_entries = le32_to_cpu(n->header.nr_entries); in dm_btree_lookup_next_single()
444 if (i >= nr_entries) { in dm_btree_lookup_next_single()
976 unsigned int nr_entries; in get_node_free_space() local
[all …]
A Ddm-array.c27 __le32 nr_entries; member
114 unsigned int nr_entries = le32_to_cpu(ab->nr_entries); in on_entries() local
164 (*ab)->nr_entries = cpu_to_le32(0); in alloc_ablock()
178 uint32_t nr_entries, delta, i; in fill_ablock() local
184 nr_entries = le32_to_cpu(ab->nr_entries); in fill_ablock()
185 delta = new_nr - nr_entries; in fill_ablock()
201 uint32_t nr_entries, delta; in trim_ablock() local
207 nr_entries = le32_to_cpu(ab->nr_entries); in trim_ablock()
208 delta = nr_entries - new_nr; in trim_ablock()
882 nr_entries = le32_to_cpu(ab->nr_entries); in walk_ablock()
[all …]
/linux/arch/x86/kernel/
A De820.c82 for (i = 0; i < table->nr_entries; i++) { in _e820__mapped_any()
168 int x = table->nr_entries; in __e820__range_add()
179 table->nr_entries++; in __e820__range_add()
332 if (table->nr_entries < 2) in e820__update_table()
417 table->nr_entries = new_nr_entries; in e820__update_table()
426 while (nr_entries) { in __append_e820_table()
439 nr_entries--; in __append_e820_table()
456 if (nr_entries < 2) in append_e820_table()
625 int i = e820_table->nr_entries; in e820_search_gap()
929 e820_table->nr_entries = 0; in parse_memmap_one()
[all …]
A Dldt.c81 set_ldt(ldt_slot_va(ldt->slot), ldt->nr_entries); in load_mm_ldt()
83 set_ldt(ldt->entries, ldt->nr_entries); in load_mm_ldt()
183 new_ldt->nr_entries = num_entries; in alloc_ldt_struct()
309 nr_pages = DIV_ROUND_UP(ldt->nr_entries * LDT_ENTRY_SIZE, PAGE_SIZE); in map_ldt_struct()
418 paravirt_alloc_ldt(ldt->entries, ldt->nr_entries); in finalize_ldt_struct()
439 paravirt_free_ldt(ldt->entries, ldt->nr_entries); in free_ldt_struct()
440 if (ldt->nr_entries * LDT_ENTRY_SIZE > PAGE_SIZE) in free_ldt_struct()
463 new_ldt = alloc_ldt_struct(old_mm->context.ldt->nr_entries); in ldt_dup_context()
470 new_ldt->nr_entries * LDT_ENTRY_SIZE); in ldt_dup_context()
518 entries_size = mm->context.ldt->nr_entries * LDT_ENTRY_SIZE; in read_ldt()
[all …]
/linux/tools/perf/ui/browsers/
A Dannotate-data.c130 bracket->nr_entries = 1; in add_child_entries()
138 entry->nr_entries = 1; in add_child_entries()
148 nr += entry->nr_entries; in count_visible_entries()
461 entry->nr_entries = 1; in annotated_data_browser__fold()
470 int nr_entries; in annotated_data_browser__unfold() local
482 nr_entries += child->nr_entries; in annotated_data_browser__unfold()
485 entry->nr_entries = nr_entries; in annotated_data_browser__unfold()
497 parent->nr_entries -= curr->nr_entries; in annotated_data_browser__toggle_fold()
500 browser->b.nr_entries -= curr->nr_entries; in annotated_data_browser__toggle_fold()
509 parent->nr_entries += curr->nr_entries; in annotated_data_browser__toggle_fold()
[all …]
/linux/tools/perf/util/
A Dsyscalltbl.c78 int nr_entries = 0, i, j; in syscalltbl__init_native() local
83 ++nr_entries; in syscalltbl__init_native()
85 entries = tbl->syscalls.entries = malloc(sizeof(struct syscall) * nr_entries); in syscalltbl__init_native()
97 qsort(tbl->syscalls.entries, nr_entries, sizeof(struct syscall), syscallcmp); in syscalltbl__init_native()
98 tbl->syscalls.nr_entries = nr_entries; in syscalltbl__init_native()
129 tbl->syscalls.nr_entries, sizeof(*sc), in syscalltbl__id()
139 return idx < tbl->syscalls.nr_entries ? syscalls[idx].id : -1; in syscalltbl__id_at_idx()
147 for (i = *idx + 1; i < tbl->syscalls.nr_entries; ++i) { in syscalltbl__strglobmatch_next()
/linux/mm/kmsan/
A Dreport.c89 unsigned int nr_entries, chained_nr_entries, skipnr; in kmsan_print_origin() local
100 nr_entries = stack_depot_fetch(origin, &entries); in kmsan_print_origin()
102 magic = nr_entries ? entries[0] : 0; in kmsan_print_origin()
103 if ((nr_entries == 4) && (magic == KMSAN_ALLOCA_MAGIC_ORIGIN)) { in kmsan_print_origin()
115 if ((nr_entries == 3) && (magic == KMSAN_CHAIN_MAGIC_ORIGIN)) { in kmsan_print_origin()
139 if (nr_entries) { in kmsan_print_origin()
140 skipnr = get_stack_skipnr(entries, nr_entries); in kmsan_print_origin()
141 stack_trace_print(entries + skipnr, nr_entries - skipnr, in kmsan_print_origin()
/linux/tools/testing/selftests/bpf/benchs/
A Dbench_bpf_hashmap_lookup.c24 __u32 nr_entries; member
30 .nr_entries = 500,
91 args.nr_entries = ret; in parse_arg()
121 if (args.nr_entries > args.max_entries) { in validate()
123 args.max_entries, args.nr_entries); in validate()
171 ctx.skel->bss->nr_entries = args.nr_entries; in setup()
172 ctx.skel->bss->nr_loops = args.nr_loops / args.nr_entries; in setup()
188 for (u64 i = 0; i < args.nr_entries; i++) { in setup()
/linux/tools/perf/ui/
A Dbrowser.c108 if (browser->nr_entries == 0) in ui_browser__list_head_seek()
318 if (browser->nr_entries > 1) { in ui_browser__scrollbar_set()
320 (browser->nr_entries - 1)); in ui_browser__scrollbar_set()
372 off_t offset = nr_entries - browser->nr_entries; in ui_browser__update_nr_entries()
374 browser->nr_entries = nr_entries; in ui_browser__update_nr_entries()
426 if (browser->index == browser->nr_entries - 1) in ui_browser__run()
462 offset = browser->nr_entries - 1 - browser->index; in ui_browser__run()
485 if (offset >= browser->nr_entries) in ui_browser__run()
486 offset = browser->nr_entries - 1; in ui_browser__run()
488 browser->index = browser->nr_entries - 1; in ui_browser__run()
[all …]
/linux/arch/mips/kernel/
A Dstacktrace.c30 trace->entries[trace->nr_entries++] = addr; in save_raw_context_stack()
31 if (trace->nr_entries >= trace->max_entries) in save_raw_context_stack()
58 trace->entries[trace->nr_entries++] = pc; in save_context_stack()
59 if (trace->nr_entries >= trace->max_entries) in save_context_stack()
83 WARN_ON(trace->nr_entries || !trace->max_entries); in save_stack_trace_tsk()
/linux/arch/x86/xen/
A Dsetup.c169 for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) { in xen_find_pfn_range()
431 for (i = 0; i < xen_e820_table.nr_entries; i++, entry++) { in xen_foreach_remap_area()
655 xen_e820_table.nr_entries; in xen_e820_swap_entry_with_ram()
656 xen_e820_table.nr_entries++; in xen_e820_swap_entry_with_ram()
809 memmap.nr_entries = ARRAY_SIZE(xen_e820_table.entries); in xen_memory_setup()
822 memmap.nr_entries = 1; in xen_memory_setup()
831 BUG_ON(memmap.nr_entries == 0); in xen_memory_setup()
832 xen_e820_table.nr_entries = memmap.nr_entries; in xen_memory_setup()
850 xen_e820_table.nr_entries++; in xen_memory_setup()
899 while (i < xen_e820_table.nr_entries) { in xen_memory_setup()
[all …]
/linux/tools/perf/trace/beauty/
A Dioctl.c41 if (nr < strarray__ioctl_tty_cmd.nr_entries && strarray__ioctl_tty_cmd.entries[nr] != NULL) in ioctl__scnprintf_tty_cmd()
52 if (nr < strarray__drm_ioctl_cmds.nr_entries && strarray__drm_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_drm_cmd()
63 …if (nr < strarray__sndrv_pcm_ioctl_cmds.nr_entries && strarray__sndrv_pcm_ioctl_cmds.entries[nr] !… in ioctl__scnprintf_sndrv_pcm_cmd()
74 …if (nr < strarray__sndrv_ctl_ioctl_cmds.nr_entries && strarray__sndrv_ctl_ioctl_cmds.entries[nr] !… in ioctl__scnprintf_sndrv_ctl_cmd()
85 if (nr < strarray__kvm_ioctl_cmds.nr_entries && strarray__kvm_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_kvm_cmd()
98 if (nr < s->nr_entries && s->entries[nr] != NULL) in ioctl__scnprintf_vhost_virtio_cmd()
109 if (nr < strarray__perf_ioctl_cmds.nr_entries && strarray__perf_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_perf_cmd()
120 …if (nr < strarray__usbdevfs_ioctl_cmds.nr_entries && strarray__usbdevfs_ioctl_cmds.entries[nr] != … in ioctl__scnprintf_usbdevfs_cmd()
/linux/drivers/md/
A Ddm-cache-policy-internal.h117 static inline size_t bitset_size_in_bytes(unsigned int nr_entries) in bitset_size_in_bytes() argument
119 return sizeof(unsigned long) * dm_div_up(nr_entries, BITS_PER_LONG); in bitset_size_in_bytes()
122 static inline unsigned long *alloc_bitset(unsigned int nr_entries) in alloc_bitset() argument
124 size_t s = bitset_size_in_bytes(nr_entries); in alloc_bitset()
129 static inline void clear_bitset(void *bitset, unsigned int nr_entries) in clear_bitset() argument
131 size_t s = bitset_size_in_bytes(nr_entries); in clear_bitset()
/linux/arch/sh/kernel/
A Dstacktrace.c33 if (trace->nr_entries < trace->max_entries) in save_stack_address()
34 trace->entries[trace->nr_entries++] = addr; in save_stack_address()
65 if (trace->nr_entries < trace->max_entries) in save_stack_address_nosched()
66 trace->entries[trace->nr_entries++] = addr; in save_stack_address_nosched()
/linux/include/linux/
A Dstacktrace.h66 void stack_trace_print(const unsigned long *trace, unsigned int nr_entries,
69 unsigned int nr_entries, int spaces);
78 unsigned int filter_irq_stacks(unsigned long *entries, unsigned int nr_entries);
83 unsigned int nr_entries, max_entries; member
/linux/arch/openrisc/kernel/
A Dstacktrace.c38 if (trace->nr_entries < trace->max_entries) in save_stack_address()
39 trace->entries[trace->nr_entries++] = addr; in save_stack_address()
64 if (trace->nr_entries < trace->max_entries) in save_stack_address_nosched()
65 trace->entries[trace->nr_entries++] = addr; in save_stack_address_nosched()
/linux/drivers/net/ethernet/mellanox/mlxsw/
A Dspectrum1_kvdl.c122 unsigned int entry_index, nr_entries; in mlxsw_sp1_kvdl_part_alloc() local
124 nr_entries = (info->end_index - info->start_index + 1) / in mlxsw_sp1_kvdl_part_alloc()
126 entry_index = find_first_zero_bit(part->usage, nr_entries); in mlxsw_sp1_kvdl_part_alloc()
127 if (entry_index == nr_entries) in mlxsw_sp1_kvdl_part_alloc()
215 unsigned int nr_entries; in mlxsw_sp1_kvdl_part_init() local
226 nr_entries = div_u64(resource_size, info->alloc_size); in mlxsw_sp1_kvdl_part_init()
227 part = kzalloc(struct_size(part, usage, BITS_TO_LONGS(nr_entries)), in mlxsw_sp1_kvdl_part_init()
280 unsigned int nr_entries; in mlxsw_sp1_kvdl_part_occ() local
284 nr_entries = (info->end_index - in mlxsw_sp1_kvdl_part_occ()
287 while ((bit = find_next_bit(part->usage, nr_entries, bit + 1)) in mlxsw_sp1_kvdl_part_occ()
[all …]
/linux/tools/perf/arch/x86/tests/
A Dhybrid.c29 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__hybrid_hw_event_with_pmu()
41 TEST_ASSERT_VAL("wrong number of entries", 2 == evlist->core.nr_entries); in test__hybrid_hw_group_event()
60 TEST_ASSERT_VAL("wrong number of entries", 2 == evlist->core.nr_entries); in test__hybrid_sw_hw_group_event()
77 TEST_ASSERT_VAL("wrong number of entries", 2 == evlist->core.nr_entries); in test__hybrid_hw_sw_group_event()
94 TEST_ASSERT_VAL("wrong number of entries", 2 == evlist->core.nr_entries); in test__hybrid_group_modifier1()
130 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__hybrid_raw2()
140 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__hybrid_cache_event()
151 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__checkevent_pmu()
171 TEST_ASSERT_VAL("wrong number of entries", 2 == evlist->core.nr_entries); in test__hybrid_hw_group_event_2()
/linux/arch/sparc/kernel/
A Dstacktrace.c58 trace->entries[trace->nr_entries++] = pc; in __save_stack_trace()
66 if (trace->nr_entries < in __save_stack_trace()
68 trace->entries[trace->nr_entries++] = pc; in __save_stack_trace()
74 } while (trace->nr_entries < trace->max_entries); in __save_stack_trace()

Completed in 89 milliseconds

1234567