| /kernel/trace/ |
| A D | trace_probe.c | 921 switch (code->op) { in store_trace_entry_data() 1213 code->op = deref; in parse_probe_arg() 1366 if (code->op != FETCH_OP_REG && code->op != FETCH_OP_STACK && in finalize_fetch_insn() 1367 code->op != FETCH_OP_RETVAL && code->op != FETCH_OP_ARG && in finalize_fetch_insn() 1368 code->op != FETCH_OP_DEREF && code->op != FETCH_OP_TP_ARG) { in finalize_fetch_insn() 1374 if (code->op != FETCH_OP_DEREF && code->op != FETCH_OP_UDEREF && in finalize_fetch_insn() 1375 code->op != FETCH_OP_IMM && code->op != FETCH_OP_COMM && in finalize_fetch_insn() 1376 code->op != FETCH_OP_DATA && code->op != FETCH_OP_TP_ARG) { in finalize_fetch_insn() 1384 (code->op == FETCH_OP_IMM || code->op == FETCH_OP_COMM || in finalize_fetch_insn() 1385 code->op == FETCH_OP_DATA) || code->op == FETCH_OP_TP_ARG || in finalize_fetch_insn() [all …]
|
| A D | trace_probe_tmpl.h | 105 switch (code->op) { in process_common_fetch_insn() 134 if (code->op == FETCH_OP_DEREF) { in process_fetch_insn_bottom() 138 } else if (code->op == FETCH_OP_UDEREF) { in process_fetch_insn_bottom() 153 switch (code->op) { in process_fetch_insn_bottom() 171 switch (code->op) { in process_fetch_insn_bottom() 199 if (code->op == FETCH_OP_MOD_BF) { in process_fetch_insn_bottom() 206 if (code->op == FETCH_OP_LP_ARRAY) { in process_fetch_insn_bottom() 212 if (s3->op != FETCH_OP_ST_STRING && in process_fetch_insn_bottom() 213 s3->op != FETCH_OP_ST_USTRING) { in process_fetch_insn_bottom() 230 return code->op == FETCH_OP_END ? ret : -EILSEQ; in process_fetch_insn_bottom()
|
| A D | ftrace.c | 188 op->saved_func(ip, parent_ip, op, fregs); in ftrace_pid_func() 404 op->func = ftrace_pids_enabled(op) ? in ftrace_update_pid_func() 1094 if (op->trampoline && op->trampoline_size) in ftrace_ops_trampoline() 1096 addr < op->trampoline + op->trampoline_size) { in ftrace_ops_trampoline() 2420 return op; in ftrace_find_tramp_ops_any() 2434 if (op == op_exclude || !op->trampoline) in ftrace_find_tramp_ops_any_other() 2438 return op; in ftrace_find_tramp_ops_any_other() 2456 return op; in ftrace_find_tramp_ops_next() 8058 pr_warn("op=%p %pS\n", op, op); in __ftrace_ops_list_func() 8061 op->func(ip, parent_ip, op, fregs); in __ftrace_ops_list_func() [all …]
|
| A D | trace_events_filter.c | 89 int op; member 655 switch (op) { in do_filter_cpumask() 675 switch (op) { in do_filter_scalar_cpumask() 690 switch (op) { in do_filter_cpumask_scalar() 1512 switch (op) { in select_comparison_fn() 1647 int op; in parse_pred() local 1693 for (op = 0; ops[op]; op++) { in parse_pred() 1695 if (strncmp(str + i, ops[op], strlen(ops[op])) == 0) in parse_pred() 1717 pred->op = op; in parse_pred() 1727 switch (op) { in parse_pred() [all …]
|
| A D | trace_functions.c | 29 struct ftrace_ops *op, struct ftrace_regs *fregs); 32 struct ftrace_ops *op, struct ftrace_regs *fregs); 38 struct ftrace_ops *op, 211 struct trace_array *tr = op->private; in function_trace_call() 235 struct trace_array *tr = op->private; in function_args_trace_call() 275 struct trace_array *tr = op->private; in function_stack_trace_call() 300 if (ftrace_pids_enabled(op)) in function_stack_trace_call() 342 struct ftrace_ops *op, in function_no_repeats_trace_call() argument 346 struct trace_array *tr = op->private; in function_no_repeats_trace_call() 383 struct ftrace_ops *op, in function_stack_no_repeats_trace_call() argument [all …]
|
| A D | trace_selftest.c | 110 struct ftrace_ops *op, in trace_selftest_test_probe1_func() argument 119 struct ftrace_ops *op, in trace_selftest_test_probe2_func() argument 128 struct ftrace_ops *op, in trace_selftest_test_probe3_func() argument 137 struct ftrace_ops *op, in trace_selftest_test_global_func() argument 146 struct ftrace_ops *op, in trace_selftest_test_dyn_func() argument 451 struct ftrace_ops *op, in trace_selftest_test_recursion_func() argument 466 struct ftrace_ops *op, in trace_selftest_test_recursion_safe_func() argument 585 struct ftrace_ops *op, in trace_selftest_test_regs_func() argument
|
| A D | fgraph.c | 1209 struct ftrace_ops *op; in fgraph_update_pid_func() local 1214 list_for_each_entry(op, &graph_ops.subop_list, list) { in fgraph_update_pid_func() 1215 if (op->flags & FTRACE_OPS_FL_PID) { in fgraph_update_pid_func() 1216 gops = container_of(op, struct fgraph_ops, ops); in fgraph_update_pid_func() 1217 gops->entryfunc = ftrace_pids_enabled(op) ? in fgraph_update_pid_func()
|
| /kernel/ |
| A D | kprobes.c | 446 kfree(op); in free_aggr_kprobe() 568 if (kprobe_disabled(&op->kp) && !kprobe_gone(&op->kp)) in do_unoptimize_kprobes() 578 list_del_init(&op->list); in do_unoptimize_kprobes() 588 list_del_init(&op->list); in do_free_cleaned_kprobes() 672 if (op == _op) in optprobe_queued_unopt() 798 list_del_init(&op->list); in kill_optimized_kprobe() 837 if (!op) in alloc_aggr_kprobe() 840 INIT_LIST_HEAD(&op->list); in alloc_aggr_kprobe() 841 op->kp.addr = p->addr; in alloc_aggr_kprobe() 844 return &op->kp; in alloc_aggr_kprobe() [all …]
|
| A D | auditsc.c | 122 enum audit_nfcfgop op; member 650 f->op, f->uid); in audit_filter_rules() 658 f->op, f->val); in audit_filter_rules() 687 f->op, in audit_filter_rules() 704 f->op, in audit_filter_rules() 711 f->op, in audit_filter_rules() 722 f->type, f->op, in audit_filter_rules() 831 unsigned long op) in __audit_filter_op() argument 1910 ctx->uring_op = op; in __audit_uring_entry() 2310 f->op, f->val) in __audit_inode() [all …]
|
| A D | auditfilter.c | 159 (f->op != Audit_equal && f->op != Audit_not_equal)) in audit_to_inode() 318 static u32 audit_to_op(u32 op) in audit_to_op() argument 391 if (f->op == Audit_bitmask || f->op == Audit_bittest) in audit_field_valid() 411 if (f->op != Audit_not_equal && f->op != Audit_equal) in audit_field_valid() 472 if (f->op == Audit_bad) in audit_data_to_entry() 719 a->fields[i].op != b->fields[i].op) in audit_compare_rule() 1208 switch (op) { in audit_comparator() 1232 switch (op) { in audit_uid_comparator() 1254 switch (op) { in audit_gid_comparator() 1365 f->op, f->uid); in audit_filter() [all …]
|
| A D | audit.h | 232 extern int audit_comparator(const u32 left, const u32 op, const u32 right); 233 extern int audit_uid_comparator(kuid_t left, u32 op, kuid_t right); 234 extern int audit_gid_comparator(kgid_t left, u32 op, kgid_t right); 271 u32 op); 293 extern int audit_make_tree(struct audit_krule *rule, char *pathname, u32 op); 326 #define audit_make_tree(rule, str, op) -EINVAL argument
|
| A D | audit_watch.c | 178 int audit_to_watch(struct audit_krule *krule, char *path, int len, u32 op) in audit_to_watch() argument 188 op != Audit_equal || in audit_to_watch() 227 static void audit_watch_log_rule_change(struct audit_krule *r, struct audit_watch *w, char *op) in audit_watch_log_rule_change() argument 237 audit_log_format(ab, "op=%s path=", op); in audit_watch_log_rule_change()
|
| A D | audit_fsnotify.c | 115 static void audit_mark_log_rule_change(struct audit_fsnotify_mark *audit_mark, char *op) in audit_mark_log_rule_change() argument 126 audit_log_format(ab, " op=%s path=", op); in audit_mark_log_rule_change()
|
| A D | seccomp.c | 2085 static long do_seccomp(unsigned int op, unsigned int flags, in do_seccomp() argument 2088 switch (op) { in do_seccomp() 2110 SYSCALL_DEFINE3(seccomp, unsigned int, op, unsigned int, flags, in SYSCALL_DEFINE3() argument 2113 return do_seccomp(op, flags, uargs); in SYSCALL_DEFINE3() 2125 unsigned int op; in prctl_set_seccomp() local 2130 op = SECCOMP_SET_MODE_STRICT; in prctl_set_seccomp() 2139 op = SECCOMP_SET_MODE_FILTER; in prctl_set_seccomp() 2147 return do_seccomp(op, 0, uargs); in prctl_set_seccomp()
|
| A D | ptrace.c | 999 .op = ptrace_get_syscall_info_op(child), in ptrace_get_syscall_info() 1007 switch (info.op) { in ptrace_get_syscall_info() 1119 if (ptrace_get_syscall_info_op(child) != info.op) in ptrace_set_syscall_info() 1122 switch (info.op) { in ptrace_set_syscall_info()
|
| /kernel/futex/ |
| A D | syscalls.c | 84 long do_futex(u32 __user *uaddr, int op, u32 val, ktime_t *timeout, in do_futex() argument 87 unsigned int flags = futex_to_flags(op); in do_futex() 88 int cmd = op & FUTEX_CMD_MASK; in do_futex() 147 futex_init_timeout(u32 cmd, u32 op, struct timespec64 *ts, ktime_t *t) in futex_init_timeout() argument 155 else if (cmd != FUTEX_LOCK_PI && !(op & FUTEX_CLOCK_REALTIME)) in futex_init_timeout() 160 SYSCALL_DEFINE6(futex, u32 __user *, uaddr, int, op, u32, val, in SYSCALL_DEFINE6() argument 164 int ret, cmd = op & FUTEX_CMD_MASK; in SYSCALL_DEFINE6() 169 if (unlikely(should_fail_futex(!(op & FUTEX_PRIVATE_FLAG)))) in SYSCALL_DEFINE6() 173 ret = futex_init_timeout(cmd, op, &ts, &t); in SYSCALL_DEFINE6() 496 int ret, cmd = op & FUTEX_CMD_MASK; in SYSCALL_DEFINE6() [all …]
|
| A D | futex.h | 45 static inline unsigned int futex_to_flags(unsigned int op) in futex_to_flags() argument 49 if (!(op & FUTEX_PRIVATE_FLAG)) in futex_to_flags() 52 if (op & FUTEX_CLOCK_REALTIME) in futex_to_flags() 496 u32 __user *uaddr2, int nr_wake, int nr_wake2, int op);
|
| A D | waitwake.c | 204 unsigned int op = (encoded_op & 0x70000000) >> 28; in futex_atomic_op_inuser() local 224 ret = arch_futex_atomic_op_inuser(op, oparg, &oldval, uaddr); in futex_atomic_op_inuser() 252 int nr_wake, int nr_wake2, int op) in futex_wake_op() argument 273 op_ret = futex_atomic_op_inuser(op, uaddr2); in futex_wake_op()
|
| /kernel/locking/ |
| A D | spinlock.c | 67 #define BUILD_LOCK_OPS(op, locktype) \ argument 68 static void __lockfunc __raw_##op##_lock(locktype##_t *lock) \ 72 if (likely(do_raw_##op##_trylock(lock))) \ 76 arch_##op##_relax(&lock->raw_lock); \ 80 static unsigned long __lockfunc __raw_##op##_lock_irqsave(locktype##_t *lock) \ 87 if (likely(do_raw_##op##_trylock(lock))) \ 92 arch_##op##_relax(&lock->raw_lock); \ 98 static void __lockfunc __raw_##op##_lock_irq(locktype##_t *lock) \ 100 _raw_##op##_lock_irqsave(lock); \ 103 static void __lockfunc __raw_##op##_lock_bh(locktype##_t *lock) \ [all …]
|
| /kernel/bpf/ |
| A D | bpf_iter.c | 127 p = seq->op->start(seq, &seq->index); in bpf_seq_read() 132 seq->op->stop(seq, p); in bpf_seq_read() 137 err = seq->op->show(seq, p); in bpf_seq_read() 147 seq->op->stop(seq, p); in bpf_seq_read() 158 p = seq->op->next(seq, p, &seq->index); in bpf_seq_read() 162 seq->op->next); in bpf_seq_read() 178 seq->op->stop(seq, p); in bpf_seq_read() 184 err = seq->op->show(seq, p); in bpf_seq_read() 193 seq->op->stop(seq, p); in bpf_seq_read() 205 seq->op->stop(seq, NULL); in bpf_seq_read() [all …]
|
| A D | task_iter.c | 428 enum bpf_task_vma_iter_find_op op; in task_vma_seq_get_next() local 485 op = task_vma_iter_find_vma; in task_vma_seq_get_next() 492 op = task_vma_iter_next_vma; in task_vma_seq_get_next() 504 op = task_vma_iter_first_vma; in task_vma_seq_get_next() 512 op = task_vma_iter_find_vma; in task_vma_seq_get_next() 525 switch (op) { in task_vma_seq_get_next()
|
| /kernel/trace/rv/monitors/sleep/ |
| A D | sleep.c | 125 int op, cmd; in handle_sys_enter() local 146 op = args[1]; in handle_sys_enter() 147 cmd = op & FUTEX_CMD_MASK; in handle_sys_enter()
|
| /kernel/printk/ |
| A D | index.c | 151 static int pi_module_notify(struct notifier_block *nb, unsigned long op, in pi_module_notify() argument 156 switch (op) { in pi_module_notify()
|
| /kernel/kcsan/ |
| A D | core.c | 1193 #define DEFINE_TSAN_ATOMIC_RMW(op, bits, suffix) \ argument 1194 u##bits __tsan_atomic##bits##_##op(u##bits *ptr, u##bits v, int memorder); \ 1195 u##bits __tsan_atomic##bits##_##op(u##bits *ptr, u##bits v, int memorder) \ 1203 return __atomic_##op##suffix(ptr, v, memorder); \ 1205 EXPORT_SYMBOL(__tsan_atomic##bits##_##op)
|
| /kernel/sched/ |
| A D | ext.c | 12 #define SCX_OP_IDX(op) (offsetof(struct sched_ext_ops, op) / sizeof(void (*)(void))) argument 1207 #define SCX_HAS_OP(sch, op) test_bit(SCX_OP_IDX(op), (sch)->has_op) argument 1291 #define SCX_CALL_OP(sch, mask, op, rq, args...) \ argument 1297 (sch)->ops.op(args); \ 1300 (sch)->ops.op(args); \ 1306 #define SCX_CALL_OP_RET(sch, mask, op, rq, args...) \ argument 1308 __typeof__((sch)->ops.op(args)) __ret; \ 1314 __ret = (sch)->ops.op(args); \ 1317 __ret = (sch)->ops.op(args); \ 1339 SCX_CALL_OP((sch), mask, op, rq, task, ##args); \ [all …]
|