| /kernel/printk/ |
| A D | printk_ringbuffer.c | 351 unsigned long id; member 580 unsigned long id; in data_make_reusable() local 594 id = blk->id; /* LMM(data_make_reusable:A) */ in data_make_reusable() 886 unsigned long id; in desc_reserve() local 1001 *id_out = id; in desc_reserve() 1082 blk->id = id; /* LMM(data_alloc:B) */ in data_alloc() 1092 blk->id = id; in data_alloc() 1164 blk->id = id; in data_realloc() 1320 *id_out = id; in desc_reopen_last() 1393 e->id = id; in prb_reserve_in_last() [all …]
|
| /kernel/livepatch/ |
| A D | shadow.c | 58 unsigned long id; member 71 unsigned long id) in klp_shadow_match() argument 73 return shadow->obj == obj && shadow->id == id; in klp_shadow_match() 83 void *klp_shadow_get(void *obj, unsigned long id) in klp_shadow_get() argument 92 if (klp_shadow_match(shadow, obj, id)) { in klp_shadow_get() 114 shadow_data = klp_shadow_get(obj, id); in __klp_shadow_get_or_alloc() 129 shadow_data = klp_shadow_get(obj, id); in __klp_shadow_get_or_alloc() 141 new_shadow->id = id; in __klp_shadow_get_or_alloc() 151 obj, id, err); in __klp_shadow_get_or_alloc() 264 if (klp_shadow_match(shadow, obj, id)) { in klp_shadow_free() [all …]
|
| A D | state.c | 16 for (state = patch->states; state && state->id; state++) 31 struct klp_state *klp_get_state(struct klp_patch *patch, unsigned long id) in klp_get_state() argument 36 if (state->id == id) in klp_get_state() 64 struct klp_state *klp_get_prev_state(unsigned long id) in klp_get_prev_state() argument 76 state = klp_get_state(patch, id); in klp_get_prev_state() 92 state = klp_get_state(patch, old_state->id); in klp_is_state_compatible()
|
| /kernel/trace/rv/ |
| A D | rv_trace.h | 73 TP_PROTO(int id, char *state, char *event, char *next_state, bool final_state), 75 TP_ARGS(id, state, event, next_state, final_state), 78 __field( int, id ) 89 __entry->id = id; 94 __entry->id, 103 TP_PROTO(int id, char *state, char *event), 105 TP_ARGS(id, state, event), 108 __field( int, id ) 116 __entry->id = id; 120 __entry->id,
|
| /kernel/ |
| A D | user_namespace.c | 264 if (key->id >= first && key->id <= last && in cmp_map_id() 285 key.id = id; in map_id_range_down_max() 308 if (id >= first && id <= last && in map_id_range_down_base() 328 id = (id - extent->first) + extent->lower_first; in map_id_range_down() 330 id = (u32) -1; in map_id_range_down() 332 return id; in map_id_range_down() 357 if (id >= first && id <= last && in map_id_range_up_base() 375 key.id = id; in map_id_range_up_max() 394 id = (id - extent->lower_first) + extent->first; in map_id_range_up() 396 id = (u32) -1; in map_id_range_up() [all …]
|
| A D | watch_queue.c | 196 u64 id) in __post_watch_notification() argument 210 if (watch->id != id) in __post_watch_notification() 469 if (wqueue == wq && watch->id == w->id) in add_one_watch() 532 u64 id, bool all) in remove_watch_from_object() argument 545 (watch->id == id && rcu_access_pointer(watch->queue) == wq)) in remove_watch_from_object() 562 n.id = id; in remove_watch_from_object() 563 if (id != 0) in remove_watch_from_object()
|
| /kernel/locking/ |
| A D | qspinlock_stat.h | 40 int cpu, id, len; in lockevent_read() local 46 id = (long)file_inode(file)->i_private; in lockevent_read() 48 if (id >= lockevent_num) in lockevent_read() 52 sum += per_cpu(lockevents[id], cpu); in lockevent_read() 56 switch (id) { in lockevent_read() 69 if (id == LOCKEVENT_pv_hash_hops) { in lockevent_read() 86 if ((id == LOCKEVENT_pv_latency_kick) || in lockevent_read() 87 (id == LOCKEVENT_pv_latency_wake)) { in lockevent_read()
|
| A D | lock_events.c | 65 int cpu, id, len; in lockevent_read() local 71 id = (long)file_inode(file)->i_private; in lockevent_read() 73 if (id >= lockevent_num) in lockevent_read() 77 sum += per_cpu(lockevents[id], cpu); in lockevent_read()
|
| /kernel/bpf/ |
| A D | stackmap.c | 252 return id; in __bpf_get_stackid() 271 return id; in __bpf_get_stackid() 280 return id; in __bpf_get_stackid() 297 return id; in __bpf_get_stackid() 677 u32 id; in stack_map_get_next_key() local 682 id = 0; in stack_map_get_next_key() 685 if (id >= smap->n_buckets || !smap->buckets[id]) in stack_map_get_next_key() 686 id = 0; in stack_map_get_next_key() 688 id++; in stack_map_get_next_key() 691 while (id < smap->n_buckets && !smap->buckets[id]) in stack_map_get_next_key() [all …]
|
| A D | mprog.c | 12 bool id = flags & BPF_F_ID; in bpf_mprog_link() local 14 if (id) in bpf_mprog_link() 35 bool id = flags & BPF_F_ID; in bpf_mprog_prog() local 37 if (id) in bpf_mprog_prog() 58 bool id = flags & BPF_F_ID; in bpf_mprog_tuple_relative() local 67 if (!id && !id_or_fd) in bpf_mprog_tuple_relative() 404 u32 id, count = 0; in bpf_mprog_query() local 435 id = prog->aux->id; in bpf_mprog_query() 436 if (copy_to_user(uprog_id + i, &id, sizeof(id))) in bpf_mprog_query() 441 id = cp->link ? cp->link->id : 0; in bpf_mprog_query() [all …]
|
| A D | syscall.c | 444 int id; in bpf_map_alloc_id() local 450 map->id = id; in bpf_map_alloc_id() 457 return id > 0 ? 0 : id; in bpf_map_alloc_id() 2289 int id; in bpf_prog_alloc_id() local 2295 prog->aux->id = id; in bpf_prog_alloc_id() 2303 return id > 0 ? 0 : id; in bpf_prog_alloc_id() 3347 primer->id = id; in bpf_link_prime() 3355 primer->link->id = primer->id; in bpf_link_settle() 4883 info.id = prog->aux->id; in bpf_prog_get_info_by_fd() 5165 info.id = map->id; in bpf_map_get_info_by_fd() [all …]
|
| A D | btf.c | 265 u32 id; member 571 int id; in bpf_find_btf_id() local 1060 id = 0; in btf_show_name() 1063 if (!id) in btf_show_name() 1649 int id; in btf_alloc_id() local 1655 btf->id = id; in btf_alloc_id() 1662 return id > 0 ? 0 : id; in btf_alloc_id() 3451 s32 id; in btf_find_graph_root() local 3811 s32 id; in btf_parse_kptr() local 8091 info.id = btf->id; in btf_get_info_by_fd() [all …]
|
| /kernel/time/ |
| A D | timekeeping.c | 72 return tk->id >= TIMEKEEPER_AUX_FIRST && tk->id <= TIMEKEEPER_AUX_LAST; in tk_is_aux() 719 ntp_clear(tk->id); in timekeeping_update_from_shadow() 1204 if (cs->id == scv->cs_id) in convert_base_to_cs() 1418 return base ? base->id == id : false; in timekeeping_clocksource_has_base() 1671 switch (id) { in ktime_get_clock_ts64() 1770 tkd->timekeeper.id = tkd->shadow_timekeeper.id = tk_id; in tkd_basic_setup() 2795 if (!clockid_aux_valid(id)) in aux_get_tk_data() 2804 unsigned int id; in tk_aux_update_clocksource() local 2823 unsigned int id; in tk_aux_advance() local 2890 if (!clockid_aux_valid(id)) in aux_get_res() [all …]
|
| A D | posix-clock.c | 197 static int get_clock_desc(const clockid_t id, struct posix_clock_desc *cd) in get_clock_desc() argument 199 struct file *fp = fget(clockid_to_fd(id)); in get_clock_desc() 224 static int pc_clock_adjtime(clockid_t id, struct __kernel_timex *tx) in pc_clock_adjtime() argument 229 err = get_clock_desc(id, &cd); in pc_clock_adjtime() 248 static int pc_clock_gettime(clockid_t id, struct timespec64 *ts) in pc_clock_gettime() argument 253 err = get_clock_desc(id, &cd); in pc_clock_gettime() 267 static int pc_clock_getres(clockid_t id, struct timespec64 *ts) in pc_clock_getres() argument 272 err = get_clock_desc(id, &cd); in pc_clock_getres() 286 static int pc_clock_settime(clockid_t id, const struct timespec64 *ts) in pc_clock_settime() argument 294 err = get_clock_desc(id, &cd); in pc_clock_settime()
|
| A D | posix-timers.c | 88 DEFINE_CLASS(lock_timer, struct k_itimer *, unlock_timer(_T), __lock_timer(id), timer_t id); 96 static struct k_itimer *posix_timer_by_id(timer_t id) in posix_timer_by_id() argument 122 timer_t id) in posix_timer_hashed() argument 144 if (!posix_timer_hashed(bucket, sig, id)) { in posix_timer_add_at() 155 timer->it_id = (timer_t)id; in posix_timer_add_at() 186 if (posix_timer_add_at(timer, sig, id)) in posix_timer_add() 187 return id; in posix_timer_add() 1536 clockid_t idx = id; in clockid_to_kclock() 1538 if (id < 0) { in clockid_to_kclock() 1539 return (id & CLOCKFD_MASK) == CLOCKFD ? in clockid_to_kclock() [all …]
|
| /kernel/unwind/ |
| A D | deferred.c | 28 return try_cmpxchg(&info->id.cnt, &old, cnt); in try_assign_cnt() 35 info->id.cnt = cnt; in try_assign_cnt() 82 if (info->id.cpu) in get_cookie() 83 return info->id.id; in get_cookie() 92 info->id.cpu = smp_processor_id() + 1; /* Must be non zero */ in get_cookie() 94 return info->id.id; in get_cookie() 176 cookie = info->id.id; in process_unwind_deferred()
|
| /kernel/trace/rv/monitors/snroc/ |
| A D | snroc_trace.h | 9 TP_PROTO(int id, char *state, char *event, char *next_state, bool final_state), 10 TP_ARGS(id, state, event, next_state, final_state)); 13 TP_PROTO(int id, char *state, char *event), 14 TP_ARGS(id, state, event));
|
| /kernel/trace/rv/monitors/sssw/ |
| A D | sssw_trace.h | 9 TP_PROTO(int id, char *state, char *event, char *next_state, bool final_state), 10 TP_ARGS(id, state, event, next_state, final_state)); 13 TP_PROTO(int id, char *state, char *event), 14 TP_ARGS(id, state, event));
|
| /kernel/trace/rv/monitors/wwnr/ |
| A D | wwnr_trace.h | 10 TP_PROTO(int id, char *state, char *event, char *next_state, bool final_state), 11 TP_ARGS(id, state, event, next_state, final_state)); 14 TP_PROTO(int id, char *state, char *event), 15 TP_ARGS(id, state, event));
|
| /kernel/trace/rv/monitors/nrp/ |
| A D | nrp_trace.h | 9 TP_PROTO(int id, char *state, char *event, char *next_state, bool final_state), 10 TP_ARGS(id, state, event, next_state, final_state)); 13 TP_PROTO(int id, char *state, char *event), 14 TP_ARGS(id, state, event));
|
| /kernel/cgroup/ |
| A D | debug.c | 64 css = cset->subsys[ss->id]; in current_css_set_read() 67 seq_printf(seq, "%2d: %-4s\t- %p[%d]\n", ss->id, ss->name, in current_css_set_read() 68 css, css->id); in current_css_set_read() 217 css = rcu_dereference_check(cgrp->subsys[ss->id], true); in cgroup_subsys_states_read() 226 css->parent->id); in cgroup_subsys_states_read() 227 seq_printf(seq, "%2d: %-4s\t- %p[%d] %d%s\n", ss->id, ss->name, in cgroup_subsys_states_read() 228 css, css->id, in cgroup_subsys_states_read()
|
| /kernel/trace/ |
| A D | trace_btf.c | 16 s32 id; in btf_find_func_proto() local 18 id = bpf_find_btf_id(func_name, BTF_KIND_FUNC, btf_p); in btf_find_func_proto() 19 if (id < 0) in btf_find_func_proto() 23 t = btf_type_by_id(*btf_p, id); in btf_find_func_proto()
|
| A D | trace_export.c | 32 #define FTRACE_ENTRY_REG(name, struct_name, id, tstruct, print, regfn) \ argument 33 FTRACE_ENTRY(name, struct_name, id, PARAMS(tstruct), PARAMS(print)) 70 #define FTRACE_ENTRY(name, struct_name, id, tstruct, print) \ argument 83 #define FTRACE_ENTRY_DUP(name, struct_name, id, tstruct, print) \ argument 84 FTRACE_ENTRY(name, struct_name, id, PARAMS(tstruct), PARAMS(print)) 133 #define FTRACE_ENTRY(name, struct_name, id, tstruct, print) \ argument
|
| /kernel/events/ |
| A D | hw_breakpoint_test.c | 65 static void fill_one_bp_slot(struct kunit *test, int *id, int cpu, struct task_struct *tsk) in fill_one_bp_slot() argument 67 struct perf_event *bp = register_test_bp(cpu, tsk, *id); in fill_one_bp_slot() 71 KUNIT_ASSERT_NULL(test, test_bps[*id]); in fill_one_bp_slot() 72 test_bps[(*id)++] = bp; in fill_one_bp_slot() 80 static bool fill_bp_slots(struct kunit *test, int *id, int cpu, struct task_struct *tsk, int skip) in fill_bp_slots() argument 83 fill_one_bp_slot(test, id, cpu, tsk); in fill_bp_slots() 85 return *id + get_test_bp_slots() <= MAX_TEST_BREAKPOINTS; in fill_bp_slots()
|
| /kernel/bpf/preload/iterators/ |
| A D | iterators.bpf.c | 16 __u32 id; member 41 __u32 id; member 92 map->id, map->name, map->max_entries, in dump_bpf_map() 113 BPF_SEQ_PRINTF(seq, "%4u %-16s %s %s\n", aux->id, in dump_bpf_prog()
|