| /kernel/futex/ |
| A D | futex.h | 203 u32 bitset; 263 extern int futex_wait_setup(u32 __user *uaddr, u32 val, unsigned int flags, 273 static inline int futex_cmpxchg_value_locked(u32 *curval, u32 __user *uaddr, u32 uval, u32 newval) in futex_cmpxchg_value_locked() 301 static __always_inline int futex_get_value(u32 *dest, u32 __user *from) in futex_get_value() 303 u32 val; in futex_get_value() 318 static __always_inline int futex_put_value(u32 val, u32 __user *to) in futex_put_value() 332 static inline int futex_get_value_locked(u32 *dest, u32 __user *from) in futex_get_value_locked() 453 extern int futex_wait_requeue_pi(u32 __user *uaddr, unsigned int flags, u32 454 val, ktime_t *abs_time, u32 bitset, u32 __user 462 extern int __futex_wait(u32 __user *uaddr, unsigned int flags, u32 val, [all …]
|
| A D | pi.c | 203 static int attach_to_pi_state(u32 __user *uaddr, u32 uval, in attach_to_pi_state() 208 u32 uval2; in attach_to_pi_state() 321 static int handle_exit_race(u32 __user *uaddr, u32 uval, in handle_exit_race() 324 u32 uval2; in handle_exit_race() 411 static int attach_to_pi_owner(u32 __user *uaddr, u32 uval, union futex_key *key, in attach_to_pi_owner() 475 static int lock_pi_update_atomic(u32 __user *uaddr, u32 uval, u32 newval) in lock_pi_update_atomic() 478 u32 curval; in lock_pi_update_atomic() 522 u32 uval, newval, vpid = task_pid_vnr(task); in futex_lock_pi_atomic() 614 static int wake_futex_pi(u32 __user *uaddr, u32 uval, in wake_futex_pi() 621 u32 curval, newval; in wake_futex_pi() [all …]
|
| A D | waitwake.c | 155 int futex_wake(u32 __user *uaddr, unsigned int flags, int nr_wake, u32 bitset) in futex_wake() 251 int futex_wake_op(u32 __user *uaddr1, unsigned int flags, u32 __user *uaddr2, in futex_wake_op() 406 u32 uval; in futex_wait_multiple_setup() 444 u32 __user *uaddr = (u32 __user *)(unsigned long)vs[i].w.uaddr; in futex_wait_multiple_setup() 446 u32 val = vs[i].w.val; in futex_wait_multiple_setup() 591 int futex_wait_setup(u32 __user *uaddr, u32 val, unsigned int flags, in futex_wait_setup() 595 u32 uval; in futex_wait_setup() 666 int __futex_wait(u32 __user *uaddr, unsigned int flags, u32 val, in __futex_wait() 667 struct hrtimer_sleeper *to, u32 bitset) in __futex_wait() 706 int futex_wait(u32 __user *uaddr, unsigned int flags, u32 val, ktime_t *abs_time, u32 bitset) in futex_wait() [all …]
|
| A D | syscalls.c | 84 long do_futex(u32 __user *uaddr, int op, u32 val, ktime_t *timeout, in do_futex() 85 u32 __user *uaddr2, u32 val2, u32 val3) in do_futex() 133 static __always_inline bool futex_cmd_has_timeout(u32 cmd) in futex_cmd_has_timeout() 147 futex_init_timeout(u32 cmd, u32 op, struct timespec64 *ts, ktime_t *t) in futex_init_timeout() 160 SYSCALL_DEFINE6(futex, u32 __user *, uaddr, int, op, u32, val, in SYSCALL_DEFINE6() argument 162 u32 __user *, uaddr2, u32, val3) in SYSCALL_DEFINE6() 421 u32 cmpval; in SYSCALL_DEFINE4() 492 SYSCALL_DEFINE6(futex_time32, u32 __user *, uaddr, int, op, u32, val, in SYSCALL_DEFINE6() argument 493 const struct old_timespec32 __user *, utime, u32 __user *, uaddr2, in SYSCALL_DEFINE6() 494 u32, val3) in SYSCALL_DEFINE6()
|
| /kernel/bpf/ |
| A D | range_tree.c | 40 u32 rn_start; 41 u32 rn_last; /* inclusive */ 42 u32 __rn_subtree_last; 131 range_it_iter_first(struct range_tree *rt, u32 start, u32 last) in range_it_iter_first() 137 int range_tree_clear(struct range_tree *rt, u32 start, u32 len) in range_tree_clear() 139 u32 last = start + len - 1; in range_tree_clear() 145 u32 old_last = rn->rn_last; in range_tree_clear() 184 int is_range_tree_set(struct range_tree *rt, u32 start, u32 len) in is_range_tree_set() 186 u32 last = start + len - 1; in is_range_tree_set() 197 int range_tree_set(struct range_tree *rt, u32 start, u32 len) in range_tree_set() [all …]
|
| A D | btf.c | 613 u32 id, u32 *res_id) in btf_type_skip_modifiers() 629 u32 id, u32 *res_id) in btf_type_resolve_ptr() 641 u32 id, u32 *res_id) in btf_type_resolve_func_ptr() 1954 u32 *elem_id, u32 *total_nelems, u32 *type_id) in __btf_resolve_size() 2053 u32 *type_id, u32 *ret_size) in btf_type_id_size() 3488 u32 field_mask, u32 *seen_mask, in btf_get_field_type() 3559 u32 field_cnt, u32 repeat_cnt, u32 elem_size) in btf_repeat_fields() 3607 u32 off, u32 nelems, in btf_find_nested_struct() 3643 u32 off, u32 expected_size, in btf_find_field_one() 3644 u32 field_mask, u32 *seen_mask, in btf_find_field_one() [all …]
|
| A D | core.c | 256 u32 pages; in bpf_prog_realloc() 1787 DST = (u32) DST OP ((u32) SRC & 31); \ in ___bpf_prog_run() 1793 DST = (u32) DST OP (u32) IMM; \ in ___bpf_prog_run() 1801 DST = (u32) DST OP (u32) SRC; \ in ___bpf_prog_run() 1807 DST = (u32) DST OP (u32) IMM; \ in ___bpf_prog_run() 2222 (u32) BPF_R0, (u32) SRC); in ___bpf_prog_run() 2663 u32 *ids; in bpf_prog_array_copy_to_user() 2823 u32 *prog_ids, u32 request_cnt, in bpf_prog_array_copy_info() 2847 u32 i; in __bpf_free_used_maps() 2870 u32 i; in __bpf_free_used_btfs() [all …]
|
| A D | range_tree.h | 16 int range_tree_clear(struct range_tree *rt, u32 start, u32 len); 17 int range_tree_set(struct range_tree *rt, u32 start, u32 len); 18 int is_range_tree_set(struct range_tree *rt, u32 start, u32 len); 19 s64 range_tree_find(struct range_tree *rt, u32 len);
|
| A D | stackmap.c | 21 u32 hash; 22 u32 nr; 30 u32 n_buckets; 245 hash = jhash2((u32 *)ips, trace_len / sizeof(u32), 0); in __bpf_get_stackid() 470 u32 i; in __bpf_get_stack() 546 u32, size, u64, flags) in BPF_CALL_4() argument 653 u32 id = *(u32 *)key, trace_len; in bpf_stackmap_copy() 677 u32 id; in stack_map_get_next_key() 684 id = *(u32 *)key; in stack_map_get_next_key() 697 *(u32 *)next_key = id; in stack_map_get_next_key() [all …]
|
| A D | arraymap.c | 169 u32 index = *(u32 *)key; in array_map_lookup_elem() 244 u32 index = *(u32 *)key; in percpu_array_map_lookup_elem() 287 u32 index = *(u32 *)key; in percpu_array_map_lookup_percpu_elem() 301 u32 index = *(u32 *)key; in bpf_percpu_array_copy() 329 u32 index = key ? *(u32 *)key : U32_MAX; in array_map_get_next_key() 330 u32 *next = (u32 *)next_key; in array_map_get_next_key() 349 u32 index = *(u32 *)key; in array_map_update_elem() 388 u32 index = *(u32 *)key; in bpf_percpu_array_update() 509 u32 index = *(u32 *)key; in percpu_array_map_seq_show_elem() 881 u32 index = *(u32 *)key, ufd; in bpf_fd_array_map_update_elem() [all …]
|
| A D | bloom_filter.c | 17 u32 bitset_mask; 18 u32 hash_seed; 19 u32 nr_hash_funcs; 23 static u32 hash(struct bpf_bloom_filter *bloom, void *value, in hash() 24 u32 value_size, u32 index) in hash() 26 u32 h; in hash() 40 u32 i, h; in bloom_map_peek_elem() 55 u32 i, h; in bloom_map_push_elem() 97 u32 bitset_bytes, bitset_mask, nr_hash_funcs, nr_bits; in bloom_map_alloc() 128 check_mul_overflow(nr_bits / 5, (u32)7, &nr_bits) || in bloom_map_alloc()
|
| A D | reuseport_array.c | 43 if (attr->value_size != sizeof(u32) && in reuseport_array_alloc_check() 53 u32 index = *(u32 *)key; in reuseport_array_lookup_elem() 65 u32 index = *(u32 *)key; in reuseport_array_delete_elem() 98 u32 i; in reuseport_array_free() 192 u32 map_flags) in reuseport_array_update_check() 238 u32 index = *(u32 *)key; in bpf_fd_reuseport_array_update_elem() 320 u32 index = key ? *(u32 *)key : U32_MAX; in reuseport_array_get_next_key() 321 u32 *next = (u32 *)next_key; in reuseport_array_get_next_key()
|
| A D | cpumask.c | 32 static bool cpu_valid(u32 cpu) in cpu_valid() 121 __bpf_kfunc u32 bpf_cpumask_first(const struct cpumask *cpumask) in bpf_cpumask_first() 137 __bpf_kfunc u32 bpf_cpumask_first_zero(const struct cpumask *cpumask) in bpf_cpumask_first_zero() 154 __bpf_kfunc u32 bpf_cpumask_first_and(const struct cpumask *src1, in bpf_cpumask_first_and() 165 __bpf_kfunc void bpf_cpumask_set_cpu(u32 cpu, struct bpf_cpumask *cpumask) in bpf_cpumask_set_cpu() 178 __bpf_kfunc void bpf_cpumask_clear_cpu(u32 cpu, struct bpf_cpumask *cpumask) in bpf_cpumask_clear_cpu() 195 __bpf_kfunc bool bpf_cpumask_test_cpu(u32 cpu, const struct cpumask *cpumask) in bpf_cpumask_test_cpu() 212 __bpf_kfunc bool bpf_cpumask_test_and_set_cpu(u32 cpu, struct bpf_cpumask *cpumask) in bpf_cpumask_test_and_set_cpu() 405 __bpf_kfunc u32 bpf_cpumask_any_distribute(const struct cpumask *cpumask) in bpf_cpumask_any_distribute() 423 __bpf_kfunc u32 bpf_cpumask_any_and_distribute(const struct cpumask *src1, in bpf_cpumask_any_and_distribute() [all …]
|
| A D | cpumap.c | 142 u32 act, pass = 0; in cpu_map_bpf_prog_run_skb() 194 u32 act; in cpu_map_bpf_prog_run_xdp() 243 u32 xdp_n; 244 u32 skb_n; 292 u32 packets = 0; in cpu_map_kthread_run() 308 u32 i, n, m; in cpu_map_kthread_run() 431 u32 cpu) in __cpu_map_entry_alloc() 560 u32 key_cpu = *(u32 *)key; in cpu_map_delete_elem() 577 u32 key_cpu = *(u32 *)key; in cpu_map_update_elem() 611 u32 i; in cpu_map_free() [all …]
|
| A D | mprog.c | 8 u32 id_or_fd, u32 flags, in bpf_mprog_link() 31 u32 id_or_fd, u32 flags, in bpf_mprog_prog() 54 u32 id_or_fd, u32 flags, in bpf_mprog_tuple_relative() 129 struct bpf_tuple *ntuple, int idx, u32 flags) in bpf_mprog_insert() 229 u32 flags, u32 id_or_fd, u64 revision) in bpf_mprog_attach() 331 u32 flags, u32 id_or_fd, u64 revision) in bpf_mprog_detach() 398 u32 __user *uprog_flags, *ulink_flags; in bpf_mprog_query() 399 u32 __user *uprog_id, *ulink_id; in bpf_mprog_query() 403 const u32 flags = 0; in bpf_mprog_query() 404 u32 id, count = 0; in bpf_mprog_query()
|
| A D | bpf_lru_list.c | 356 u32 hash) in __local_list_add_pending() 358 *(u32 *)((void *)node + lru->hash_offset) = hash; in __local_list_add_pending() 405 u32 hash) in bpf_percpu_lru_pop_free() 437 u32 hash) in bpf_common_lru_pop_free() 565 u32 node_offset, u32 elem_size, in bpf_common_lru_populate() 566 u32 nr_elems) in bpf_common_lru_populate() 569 u32 i; in bpf_common_lru_populate() 586 u32 node_offset, u32 elem_size, in bpf_percpu_lru_populate() 587 u32 nr_elems) in bpf_percpu_lru_populate() 589 u32 i, pcpu_entries; in bpf_percpu_lru_populate() [all …]
|
| A D | helpers.c | 1032 *(u32 *)tmp_buf = *(u32 *)&cur_arg; in bpf_bprintf_prepare() 1033 *(u32 *)(tmp_buf + 4) = *((u32 *)&cur_arg + 1); in bpf_bprintf_prepare() 1035 *(u32 *)tmp_buf = (u32)(long)cur_arg; in bpf_bprintf_prepare() 1155 u32 idx; in bpf_timer_cb() 1195 u32 idx; in bpf_wq_work() 1707 enum bpf_dynptr_type type, u32 offset, u32 size) in bpf_dynptr_init() 2681 u32 len = buffer__szk; in bpf_dynptr_slice() 2797 u32 size; in bpf_dynptr_adjust() 2868 struct bpf_dynptr *src_ptr, u32 src_off, u32 size) in bpf_dynptr_copy() 2874 u32 off; in bpf_dynptr_copy() [all …]
|
| A D | log.c | 32 char __user *log_buf, u32 log_size) in bpf_vlog_init() 64 u32 new_n, n; in bpf_verifier_vlog() 82 new_n = min_t(u32, log->len_total - log->end_pos, n); in bpf_verifier_vlog() 94 u32 buf_start, buf_end; in bpf_verifier_vlog() 152 u32 pos; in bpf_vlog_reset() 228 u32 sublen; in bpf_vlog_finalize() 337 u32 nr_linfo; in find_linfo() 382 u32 insn_off, in verbose_linfo() 758 u32 frameno, bool print_all) in print_verifier_state() 863 static inline u32 vlog_alignment(u32 pos) in vlog_alignment() [all …]
|
| /kernel/locking/ |
| A D | qspinlock.h | 52 static inline __pure u32 encode_tail(int cpu, int idx) in encode_tail() 54 u32 tail; in encode_tail() 62 static inline __pure struct mcs_spinlock *decode_tail(u32 tail, in decode_tail() 114 static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) in xchg_tail() 120 return (u32)xchg_relaxed(&lock->tail, in xchg_tail() 158 static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) in xchg_tail() 160 u32 old, new; in xchg_tail() 184 static __always_inline u32 queued_fetch_set_pending_acquire(struct qspinlock *lock) in queued_fetch_set_pending_acquire()
|
| /kernel/trace/ |
| A D | trace_probe_tmpl.h | 17 *(u32 *)buf = (u32)val; in fetch_store_raw() 41 *(u32 *)buf <<= code->lshift; in fetch_apply_bitfield() 42 *(u32 *)buf >>= code->rshift; in fetch_apply_bitfield() 90 int maxlen = get_loc_len(*(u32 *)dest); in fetch_store_symstring() 128 u32 loc = 0; in process_fetch_insn_bottom() 182 loc = *(u32 *)dest; in process_fetch_insn_bottom() 186 loc = *(u32 *)dest; in process_fetch_insn_bottom() 190 loc = *(u32 *)dest; in process_fetch_insn_bottom() 221 dest += sizeof(u32); in process_fetch_insn_bottom() 222 *(u32 *)dest = update_data_loc(loc, ret); in process_fetch_insn_bottom() [all …]
|
| A D | bpf_trace.c | 327 u32, size) in BPF_CALL_3() argument 417 u32, data_len) in BPF_CALL_4() 798 u32 sig; 1683 u32 to_copy; in BPF_CALL_4() 2465 u32 cnt; 2466 u32 mods_cnt; 2539 u32 i; in kprobe_multi_put_modules() 2844 u32 i, err = 0; in get_modules_for_addrs() 2885 u32 i; in addrs_check_error_injection_list() 3072 u32 cnt; [all …]
|
| /kernel/ |
| A D | rseq.c | 40 u32 cpu_id_start, cpu_id, node_id, mm_cid; in rseq_validate_ro_fields() 164 u32 cpu_id = raw_smp_processor_id(); in rseq_update_cpu_node_id() 165 u32 node_id = cpu_to_node(cpu_id); in rseq_update_cpu_node_id() 166 u32 mm_cid = task_mm_cid(t); in rseq_update_cpu_node_id() 266 u32 __user *usig; in rseq_get_rseq_cs() 267 u32 sig; in rseq_get_rseq_cs() 298 usig = (u32 __user *)(unsigned long)(rseq_cs->abort_ip - sizeof(u32)); in rseq_get_rseq_cs() 312 static bool rseq_warn_flags(const char *str, u32 flags) in rseq_warn_flags() 314 u32 test_flags; in rseq_warn_flags() 329 u32 flags, event_mask; in rseq_need_restart() [all …]
|
| A D | user_namespace.c | 241 u32 count; 250 u32 first, last, id2; in cmp_map_id() 300 u32 first, last, id2; in map_id_range_down_base() 315 static u32 map_id_range_down(struct uid_gid_map *map, u32 id, u32 count) in map_id_range_down() 330 id = (u32) -1; in map_id_range_down() 335 u32 map_id_down(struct uid_gid_map *map, u32 id) in map_id_down() 349 u32 first, last, id2; in map_id_range_up_base() 381 u32 map_id_range_up(struct uid_gid_map *map, u32 id, u32 count) in map_id_range_up() 396 id = (u32) -1; in map_id_range_up() 401 u32 map_id_up(struct uid_gid_map *map, u32 id) in map_id_up() [all …]
|
| A D | audit.c | 105 u32 portid; 115 static u32 audit_rate_limit; 383 static int audit_log_config_change(char *function_name, u32 new, u32 old, in audit_log_config_change() 402 static int audit_do_config_change(char *function_name, u32 *to_change, u32 new) in audit_do_config_change() 405 u32 old = *to_change; in audit_do_config_change() 690 u32 portid; in auditd_send_unicast_skb() 844 u32 portid = 0; in kauditd_thread() 1119 u32 seq; in audit_get_feature() 1129 u32 old_lock, u32 new_lock, int res) in audit_log_feature_change() 1217 u32 seq; in audit_receive_msg() [all …]
|
| /kernel/gcov/ |
| A D | clang.c | 62 u32 checksum; 70 u32 ident; 71 u32 checksum; 72 u32 cfg_checksum; 74 u32 num_counters; 105 void llvm_gcda_start_file(const char *orig_filename, u32 version, u32 checksum) in llvm_gcda_start_file() 113 void llvm_gcda_emit_function(u32 ident, u32 func_checksum, u32 cfg_checksum) in llvm_gcda_emit_function() 128 void llvm_gcda_emit_arcs(u32 num_counters, u64 *counters) in llvm_gcda_emit_arcs() 279 u32 i; in gcov_info_add() 379 u32 i; in convert_to_gcda()
|