| /kernel/ |
| A D | cpu.c | 719 st->target = target; in cpuhp_set_state() 1274 enum cpuhp_state target = max((int)st->target, CPUHP_AP_OFFLINE); in take_cpu_down() local 1421 st->target = max((int)target, CPUHP_TEARDOWN_CPU); in _cpu_down() 1437 st->target = target; in _cpu_down() 1473 struct cpu_down_work work = { .cpu = cpu, .target = target, }; in cpu_down_maps_locked() 1590 enum cpuhp_state target = min((int)st->target, CPUHP_AP_ONLINE); in notify_cpu_starting() local 1684 target = min((int)target, CPUHP_BRINGUP_CPU); in _cpu_up() 2751 if (target < CPUHP_OFFLINE || target > CPUHP_ONLINE) in target_store() 2754 if (target != CPUHP_OFFLINE && target != CPUHP_ONLINE) in target_store() 2773 else if (WARN_ON(st->target != target)) in target_store() [all …]
|
| A D | regset.c | 6 static int __regset_get(struct task_struct *target, in __regset_get() argument 23 res = regset->regset_get(target, regset, in __regset_get() 33 int regset_get(struct task_struct *target, in regset_get() argument 38 return __regset_get(target, regset, size, &data); in regset_get() 42 int regset_get_alloc(struct task_struct *target, in regset_get_alloc() argument 48 return __regset_get(target, regset, size, data); in regset_get_alloc() 61 int copy_regset_to_user(struct task_struct *target, in copy_regset_to_user() argument 71 ret = regset_get_alloc(target, regset, size, &buf); in copy_regset_to_user()
|
| A D | cfi.c | 15 unsigned long *target, u32 type) in report_cfi_failure() argument 17 if (target) in report_cfi_failure() 19 (void *)addr, (void *)*target, type); in report_cfi_failure()
|
| A D | capability.c | 111 const struct task_struct *target; in cap_get_target_pid() local 115 target = find_task_by_vpid(pid); in cap_get_target_pid() 116 if (!target) in cap_get_target_pid() 119 ret = security_capget(target, pEp, pIp, pPp); in cap_get_target_pid()
|
| A D | exit.c | 1608 struct task_struct *target) in is_effectively_child() argument 1611 !ptrace ? target->real_parent : target->parent; in is_effectively_child() 1624 struct task_struct *target; in do_wait_pid() local 1628 target = pid_task(wo->wo_pid, PIDTYPE_TGID); in do_wait_pid() 1629 if (target && is_effectively_child(wo, ptrace, target)) { in do_wait_pid() 1630 retval = wait_consider_task(wo, ptrace, target); in do_wait_pid() 1636 target = pid_task(wo->wo_pid, PIDTYPE_PID); in do_wait_pid() 1637 if (target && target->ptrace && in do_wait_pid() 1638 is_effectively_child(wo, ptrace, target)) { in do_wait_pid() 1639 retval = wait_consider_task(wo, ptrace, target); in do_wait_pid()
|
| A D | jump_label.c | 71 jea->target = jeb->target - delta; in jump_label_swap() 75 jeb->target = tmp.target + delta; in jump_label_swap()
|
| A D | Kconfig.kexec | 107 passing additional metadata to the target kernel. This is useful 109 both source and target kernels need to have this option enabled.
|
| /kernel/gcov/ |
| A D | fs.c | 415 char *target; in link_target() local 428 target = kasprintf(GFP_KERNEL, "%s.%s", copy, ext); in link_target() 431 return target; in link_target() 479 char *target; in add_links() local 489 target = get_link_target( in add_links() 492 if (!target) in add_links() 494 basename = kbasename(target); in add_links() 495 if (basename == target) in add_links() 498 parent, target); in add_links() 499 kfree(target); in add_links() [all …]
|
| /kernel/trace/ |
| A D | trace_events_filter.c | 151 int target; member 172 t = prog[N].target; in update_preds() 173 s = prog[t].target; in update_preds() 175 prog[t].target = N; in update_preds() 176 prog[N].target = s; in update_preds() 617 prog[N-1].target = N; in predicate_parse() 622 int target = prog[i].target; in predicate_parse() local 624 prog[i].target = prog[target].target; in predicate_parse() 2659 int target; in is_or() local 2666 target = prog[i].target + 1; in is_or() [all …]
|
| A D | trace_uprobe.c | 1268 if (event->hw.target->mm == mm) in __uprobe_perf_filter() 1279 return __uprobe_perf_filter(filter, event->hw.target->mm); in trace_uprobe_filter_event() 1288 if (event->hw.target) { in trace_uprobe_filter_remove() 1291 (event->hw.target->flags & PF_EXITING) || in trace_uprobe_filter_remove() 1309 if (event->hw.target) { in trace_uprobe_filter_add()
|
| /kernel/events/ |
| A D | hw_breakpoint.c | 76 .key_offset = offsetof(struct hw_perf_event, target), 77 .key_len = sizeof_field(struct hw_perf_event, target), 110 struct task_struct *tsk = bp->hw.target; in get_task_bps_mutex() 334 head = rhltable_lookup(&task_bps_ht, &bp->hw.target, task_bps_ht_params); in task_bp_pinned() 376 if (bp->hw.target && bp->cpu < 0) { in max_bp_pinned_slots() 394 if (!bp->hw.target) in max_bp_pinned_slots() 416 if (!bp->hw.target) { in toggle_bp_slot()
|
| /kernel/bpf/ |
| A D | map_iter.c | 91 .target = "bpf_map", 168 .target = "bpf_map_elem",
|
| A D | link_iter.c | 91 .target = "bpf_link",
|
| A D | prog_iter.c | 91 .target = "bpf_prog",
|
| A D | dmabuf_iter.c | 82 .target = "dmabuf",
|
| A D | bpf_iter.c | 354 if (!strcmp(attach_fname + prefix_len, iter->reg_info->target)) { in bpf_iter_prog_supported() 444 iter_link->tinfo->reg_info->target); in bpf_iter_link_show_fdinfo() 465 target_name = iter_link->tinfo->reg_info->target; in bpf_iter_link_fill_link_info()
|
| A D | kmem_cache_iter.c | 221 .target = "kmem_cache",
|
| A D | task_iter.c | 692 .target = "task", 713 .target = "task_file", 736 .target = "task_vma",
|
| A D | cgroup_iter.c | 276 .target = "cgroup",
|
| /kernel/locking/ |
| A D | lockdep.c | 1881 print_lock_name(NULL, target->class); in print_circular_bug_entry() 1883 print_lock_trace(target->trace, 6); in print_circular_bug_entry() 1892 struct lock_class *target = hlock_class(tgt); in print_circular_lock_scenario() local 1916 __print_lock_name(tgt, target); in print_circular_lock_scenario() 1927 __print_lock_name(tgt, target); in print_circular_lock_scenario() 1933 __print_lock_name(tgt, target); in print_circular_lock_scenario() 2006 struct lock_list *target, in print_circular_bug() argument 2022 depth = get_lock_depth(target); in print_circular_bug() 2028 parent = get_lock_parent(target); in print_circular_bug() 2172 if (src->class_idx == target->class_idx) in check_noncircular() [all …]
|
| /kernel/sched/ |
| A D | fair.c | 7323 if (target != this_cpu) in wake_affine() 7328 return target; in wake_affine() 7550 if (cpu == target) in select_idle_smt() 7751 if ((available_idle_cpu(target) || sched_idle_cpu(target)) && in select_idle_sibling() 7753 return target; in select_idle_sibling() 7758 if (prev != target && cpus_share_cache(prev, target) && in select_idle_sibling() 7825 return target; in select_idle_sibling() 7852 return target; in select_idle_sibling() 8306 target = prev_cpu; in find_energy_efficient_cpu() 8455 return target; in find_energy_efficient_cpu() [all …]
|
| A D | rt.c | 1546 int target = find_lowest_rq(p); in select_task_rq_rt() local 1552 if (!test && target != -1 && !rt_task_fits_capacity(p, target)) in select_task_rq_rt() 1559 if (target != -1 && in select_task_rq_rt() 1560 p->prio < cpu_rq(target)->rt.highest_prio.curr) in select_task_rq_rt() 1561 cpu = target; in select_task_rq_rt()
|
| /kernel/module/ |
| A D | kdb.c | 56 kdb_printf("%s ", use->target->name); in kdb_lsmod()
|
| A D | sysfs.c | 245 sysfs_remove_link(use->target->holders_dir, mod->name); in del_usage_links() 258 ret = sysfs_create_link(use->target->holders_dir, in add_usage_links()
|
| /kernel/bpf/preload/iterators/ |
| A D | Makefile | 53 $(Q)$(CLANG) -g -O2 --target=bpf -m$* $(INCLUDES) \
|