| /kernel/bpf/ |
| A D | bpf_lru_list.c | 56 l->counts[type]++; in bpf_lru_list_count_inc() 63 l->counts[type]--; in bpf_lru_list_count_dec() 78 l->next_inactive_rotation = l->next_inactive_rotation->prev; in __bpf_lru_node_move_to_free() 124 l->next_inactive_rotation = l->next_inactive_rotation->prev; in __bpf_lru_node_move() 330 raw_spin_lock(&l->lock); in bpf_lru_list_pop_free_to_local() 409 struct bpf_lru_list *l; in bpf_percpu_lru_pop_free() local 544 struct bpf_lru_list *l; in bpf_percpu_lru_push_free() local 591 struct bpf_lru_list *l; in bpf_percpu_lru_populate() local 647 l->counts[i] = 0; in bpf_lru_list_init() 649 l->next_inactive_rotation = &l->lists[BPF_LRU_LIST_T_INACTIVE]; in bpf_lru_list_init() [all …]
|
| A D | hashtab.c | 682 if (l) in htab_map_lookup_elem() 719 if (l) { in __htab_lru_map_lookup_elem() 840 if (!l) in htab_map_get_next_key() 1425 if (l) in htab_map_delete_elem() 1432 if (l) in htab_map_delete_elem() 1462 if (l) in htab_lru_map_delete_elem() 1468 if (l) in htab_lru_map_delete_elem() 1636 if (l) { in __htab_map_lookup_and_delete_elem() 2300 if (l) in htab_percpu_map_lookup_elem() 2334 if (l) in htab_percpu_map_lookup_percpu_elem() [all …]
|
| A D | log.c | 338 int l, r, m; in find_linfo() local 353 l = 0; in find_linfo() 355 while (l < r) { in find_linfo() 363 m = l + (r - l + 1) / 2; in find_linfo() 365 l = m; in find_linfo() 370 return &linfo[l]; in find_linfo()
|
| A D | offload.c | 44 struct rhash_head l; member 56 .head_offset = offsetof(struct bpf_offload_netdev, l), 94 err = rhashtable_insert_fast(&offdevs, &ondev->l, offdevs_params); in __bpf_offload_dev_netdev_register() 159 WARN_ON(rhashtable_remove_fast(&offdevs, &ondev->l, offdevs_params)); in __bpf_offload_dev_netdev_unregister()
|
| A D | helpers.c | 288 arch_spinlock_t *l = (void *)lock; in __bpf_spin_lock() local 295 BUILD_BUG_ON(sizeof(*l) != sizeof(__u32)); in __bpf_spin_lock() 298 arch_spin_lock(l); in __bpf_spin_lock() 303 arch_spinlock_t *l = (void *)lock; in __bpf_spin_unlock() local 305 arch_spin_unlock(l); in __bpf_spin_unlock() 313 atomic_t *l = (void *)lock; in __bpf_spin_lock() local 315 BUILD_BUG_ON(sizeof(*l) != sizeof(*lock)); in __bpf_spin_lock() 317 atomic_cond_read_relaxed(l, !VAL); in __bpf_spin_lock() 318 } while (atomic_xchg(l, 1)); in __bpf_spin_lock() 323 atomic_t *l = (void *)lock; in __bpf_spin_unlock() local [all …]
|
| /kernel/cgroup/ |
| A D | cgroup-v1.c | 296 if (l->key.type == type && l->key.ns == ns) in cgroup_pidlist_find() 297 return l; in cgroup_pidlist_find() 315 if (l) in cgroup_pidlist_find_create() 316 return l; in cgroup_pidlist_find_create() 320 if (!l) in cgroup_pidlist_find_create() 321 return l; in cgroup_pidlist_find_create() 329 return l; in cgroup_pidlist_find_create() 377 if (!l) { in pidlist_array_load() 386 *lp = l; in pidlist_array_load() 463 if (l) in cgroup_pidlist_stop() [all …]
|
| A D | cgroup.c | 4837 struct list_head *l; in css_task_iter_next_css_set() local 4845 l = it->tcset_pos->next; in css_task_iter_next_css_set() 4847 if (l != it->tcset_head) { in css_task_iter_next_css_set() 4848 it->tcset_pos = l; in css_task_iter_next_css_set() 4849 return container_of(l, struct css_set, in css_task_iter_next_css_set() 4857 l = it->cset_pos; in css_task_iter_next_css_set() 4858 l = l->next; in css_task_iter_next_css_set() 4859 if (l == it->cset_head) { in css_task_iter_next_css_set() 4865 cset = container_of(l, struct css_set, e_cset_node[it->ss->id]); in css_task_iter_next_css_set() 4867 link = list_entry(l, struct cgrp_cset_link, cset_link); in css_task_iter_next_css_set() [all …]
|
| /kernel/module/ |
| A D | tracking.c | 56 size_t l; in print_unloaded_tainted_modules() local 58 l = module_flags_taint(mod_taint->taints, buf); in print_unloaded_tainted_modules() 59 buf[l++] = '\0'; in print_unloaded_tainted_modules() 89 size_t l; in unloaded_tainted_modules_seq_show() local 92 l = module_flags_taint(mod_taint->taints, buf); in unloaded_tainted_modules_seq_show() 93 buf[l++] = '\0'; in unloaded_tainted_modules_seq_show()
|
| A D | main.c | 953 size_t l = 0; in module_flags_taint() local 958 buf[l++] = taint_flags[i].c_true; in module_flags_taint() 961 return l; in module_flags_taint() 1047 size_t l; in show_taint() local 1049 l = module_flags_taint(mk->mod->taints, buffer); in show_taint() 1050 buffer[l++] = '\n'; in show_taint() 1051 return l; in show_taint()
|
| /kernel/locking/ |
| A D | mcs_spinlock.h | 26 #define arch_mcs_spin_lock_contended(l) \ argument 27 smp_cond_load_acquire(l, VAL) 36 #define arch_mcs_spin_unlock_contended(l) \ argument 37 smp_store_release((l), 1)
|
| A D | spinlock.c | 49 # define arch_read_relax(l) cpu_relax() argument 52 # define arch_write_relax(l) cpu_relax() argument 55 # define arch_spin_relax(l) cpu_relax() argument
|
| A D | rtmutex_common.h | 97 extern int rt_mutex_futex_trylock(struct rt_mutex_base *l); 98 extern int __rt_mutex_futex_trylock(struct rt_mutex_base *l);
|
| A D | qspinlock_paravirt.h | 80 #define queued_spin_trylock(l) pv_hybrid_queued_unfair_trylock(l) argument
|
| A D | lockdep_proc.c | 411 static int lock_stat_cmp(const void *l, const void *r) in lock_stat_cmp() argument 413 const struct lock_stat_data *dl = l, *dr = r; in lock_stat_cmp()
|
| /kernel/sched/ |
| A D | cpudeadline.c | 28 int l, r, largest; in cpudl_heapify_down() local 40 l = left_child(idx); in cpudl_heapify_down() 45 if ((l < cp->size) && dl_time_before(orig_dl, in cpudl_heapify_down() 46 cp->elements[l].dl)) { in cpudl_heapify_down() 47 largest = l; in cpudl_heapify_down() 48 largest_dl = cp->elements[l].dl; in cpudl_heapify_down()
|
| /kernel/trace/ |
| A D | trace_branch.c | 295 int l; in annotate_branch_stat_show() local 302 l = snprintf(NULL, 0, "/%lu", p->constant); in annotate_branch_stat_show() 303 l = l > 8 ? 0 : 8 - l; in annotate_branch_stat_show() 306 p->data.correct, p->constant, l, p->data.incorrect); in annotate_branch_stat_show()
|
| A D | trace_sched_switch.c | 541 loff_t l = 0; in saved_cmdlines_start() local 547 while (l <= *pos) { in saved_cmdlines_start() 548 v = saved_cmdlines_next(m, v, &l); in saved_cmdlines_start()
|
| A D | trace_stat.c | 214 struct stat_node *l = container_of(v, struct stat_node, node); in stat_seq_show() local 219 return session->ts->stat_show(s, l->stat); in stat_seq_show()
|
| A D | trace_events.c | 1561 loff_t l; in t_start() local 1566 for (l = 0; l <= *pos; ) { in t_start() 1567 file = t_next(m, file, &l); in t_start() 1629 loff_t l; in s_start() local 1640 for (l = 0; l <= *pos; ) { in s_start() 1641 iter = s_next(m, iter, &l); in s_start() 2043 loff_t l = 0; in f_start() local 2051 while (l < *pos && p) in f_start() 2052 p = f_next(m, p, &l); in f_start()
|
| A D | ftrace.c | 3995 loff_t l; in t_probe_start() local 4006 for (l = 0; l <= (*pos - iter->mod_pos); ) { in t_probe_start() 4007 p = t_probe_next(m, &l); in t_probe_start() 4070 loff_t l; in t_mod_start() local 4081 for (l = 0; l <= (*pos - iter->func_pos); ) { in t_mod_start() 4082 p = t_mod_next(m, &l); in t_mod_start() 4179 return t_mod_start(m, &l); in t_next() 4185 return t_mod_start(m, &l); in t_next() 4201 loff_t l; in t_start() local 4240 for (l = 0; l <= *pos; ) { in t_start() [all …]
|
| A D | trace.c | 772 loff_t l = 0; in trace_pid_start() local 780 for (pid++; pid && l < *pos; in trace_pid_start() 3950 loff_t l = 0; in s_start() local 3982 for (p = iter; p && l < *pos; p = s_next(m, p, &l)) in s_start() 3993 l = *pos - 1; in s_start() 3994 p = s_next(m, p, &l); in s_start() 4929 loff_t l = 0; in t_start() local 4934 for (; t && l < *pos; t = t_next(m, t, &l)) in t_start() 5744 loff_t l = 0; in eval_map_start() local 5752 while (v && l < *pos) { in eval_map_start() [all …]
|
| /kernel/ |
| A D | audit.h | 310 #define audit_to_watch(k, p, l, o) (-EINVAL) argument 311 #define audit_add_watch(k, l) (-EINVAL) argument 316 #define audit_alloc_mark(k, p, l) (ERR_PTR(-EINVAL)) argument
|
| A D | resource.c | 92 loff_t l = *pos; in r_start() local 96 if (l-- == 0) in r_start()
|
| /kernel/trace/rv/ |
| A D | rv.c | 513 loff_t l; in enabled_monitors_start() local 522 for (l = 0; l <= *pos; ) { in enabled_monitors_start() 523 mon = enabled_monitors_next(m, mon, &l); in enabled_monitors_start()
|
| /kernel/printk/ |
| A D | printk.c | 4795 size_t l = 0; in kmsg_dump_get_line() local 4807 l = record_print_text(&r, syslog, printk_time); in kmsg_dump_get_line() 4813 l = get_record_print_text_size(&info, line_count, syslog, in kmsg_dump_get_line() 4822 *len = l; in kmsg_dump_get_line()
|