| /kernel/ |
| A D | iomem.c | 13 return (__force void *)ioremap_cache(offset, size); in arch_memremap_wb() 15 return (__force void *)ioremap(offset, size); in arch_memremap_wb() 31 unsigned long pfn = PHYS_PFN(offset); in try_ram_remap() 36 return __va(offset); in try_ram_remap() 71 int is_ram = region_intersects(offset, size, in memremap() 80 &offset, (unsigned long) size); in memremap() 93 addr = try_ram_remap(offset, size, flags); in memremap() 106 &offset, (unsigned long) size); in memremap() 111 addr = ioremap_wt(offset, size); in memremap() 114 addr = ioremap_wc(offset, size); in memremap() [all …]
|
| A D | kallsyms.c | 280 unsigned long *offset) in get_symbol_pos() argument 326 if (offset) in get_symbol_pos() 327 *offset = addr - symbol_start; in get_symbol_pos() 336 unsigned long *offset) in kallsyms_lookup_size_offset() argument 341 get_symbol_pos(addr, symbolsize, offset); in kallsyms_lookup_size_offset() 350 unsigned long *offset, char **modname, in kallsyms_lookup_buildid() argument 378 offset, modname, namebuf); in kallsyms_lookup_buildid() 382 offset, modname, namebuf); in kallsyms_lookup_buildid() 396 unsigned long *offset, in kallsyms_lookup() argument 432 unsigned long offset, size; in __sprint_symbol() local [all …]
|
| A D | context_tracking.c | 81 static noinstr void ct_kernel_exit_state(int offset) in ct_kernel_exit_state() argument 91 (void)ct_state_inc(offset); in ct_kernel_exit_state() 100 static noinstr void ct_kernel_enter_state(int offset) in ct_kernel_enter_state() argument 109 seq = ct_state_inc(offset); in ct_kernel_enter_state() 123 static void noinstr ct_kernel_exit(bool user, int offset) in ct_kernel_exit() argument 149 ct_kernel_exit_state(offset); in ct_kernel_exit() 162 static void noinstr ct_kernel_enter(bool user, int offset) in ct_kernel_enter() argument 177 ct_kernel_enter_state(offset); in ct_kernel_enter() 425 static __always_inline void ct_kernel_exit(bool user, int offset) { } in ct_kernel_exit() argument 426 static __always_inline void ct_kernel_enter(bool user, int offset) { } in ct_kernel_enter() argument
|
| A D | relay.c | 304 buf->offset = 0; in __relay_reset() 563 if (buf->offset != buf->chan->subbuf_size + 1) { in relay_switch_subbuf() 566 prev_padding = buf->chan->subbuf_size - buf->offset; in relay_switch_subbuf() 592 buf->offset = 0; in relay_switch_subbuf() 594 buf->offset = buf->chan->subbuf_size + 1; in relay_switch_subbuf() 600 if (unlikely(length + buf->offset > buf->chan->subbuf_size)) in relay_switch_subbuf() 816 buf->offset == buf->bytes_consumed) in relay_file_read_consume() 831 (buf->offset == subbuf_size)) in relay_file_read_consume() 852 if (unlikely(buf->offset > subbuf_size)) { in relay_file_read_avail() 871 if (buf->offset == subbuf_size && in relay_file_read_avail() [all …]
|
| /kernel/trace/ |
| A D | fgraph.c | 447 offset -= offs; in get_ret_stack() 485 offset--; in fgraph_retrieve_parent_data() 568 int offset; in ftrace_push_return_trace() local 633 return offset; in ftrace_push_return_trace() 657 int offset; in function_graph_enter_regs() local 669 if (offset < 0) in function_graph_enter_regs() 817 int offset; in __ftrace_return_to_handler() local 901 if (offset < 0) in ftrace_graph_get_ret_stack() 905 ret_stack = get_ret_stack(task, offset, &offset); in ftrace_graph_get_ret_stack() 924 if (offset < 0) in ftrace_graph_top_ret_addr() [all …]
|
| A D | trace_probe.c | 210 offset = 0; in __trace_probe_log_err() 235 if (!offset) in traceprobe_split_symbol_offset() 245 *offset = 0; in traceprobe_split_symbol_offset() 403 code->offset = 0; in check_prepare_btf_string_fetch() 805 code[1].offset = offset; in store_entry_arg_at() 896 return offset; in __store_entry_arg() 1090 long offset = 0; in parse_probe_arg() local 1163 code->offset = offset; in parse_probe_arg() 1214 code->offset = offset; in parse_probe_arg() 1863 long offset; in traceprobe_update_arg() local [all …]
|
| A D | trace_events_inject.c | 146 if (field->size + field->offset > size) in trace_get_entry_size() 147 size = field->size + field->offset; in trace_get_entry_size() 177 str_loc -= field->offset + field->size; in trace_alloc_entry() 179 str_item = (u32 *)(entry + field->offset); in trace_alloc_entry() 184 paddr = (char **)(entry + field->offset); in trace_alloc_entry() 238 str_loc -= field->offset + field->size; in parse_entry() 243 paddr = (char **)(entry + field->offset); in parse_entry() 251 memcpy(entry + field->offset, &tmp, 1); in parse_entry() 257 memcpy(entry + field->offset, &tmp, 2); in parse_entry() 263 memcpy(entry + field->offset, &tmp, 4); in parse_entry() [all …]
|
| A D | trace_probe_tmpl.h | 136 ret = probe_mem_read(&val, (void *)val + code->offset, in process_fetch_insn_bottom() 141 (void *)val + code->offset, sizeof(val)); in process_fetch_insn_bottom() 155 ret = fetch_store_strlen(val + code->offset); in process_fetch_insn_bottom() 159 ret = fetch_store_strlen_user(val + code->offset); in process_fetch_insn_bottom() 163 ret = fetch_store_symstrlen(val + code->offset); in process_fetch_insn_bottom() 176 probe_mem_read(dest, (void *)val + code->offset, code->size); in process_fetch_insn_bottom() 179 probe_mem_read_user(dest, (void *)val + code->offset, code->size); in process_fetch_insn_bottom() 183 ret = fetch_store_string(val + code->offset, dest, base); in process_fetch_insn_bottom() 187 ret = fetch_store_string_user(val + code->offset, dest, base); in process_fetch_insn_bottom() 191 ret = fetch_store_symstring(val + code->offset, dest, base); in process_fetch_insn_bottom() [all …]
|
| A D | trace_btf.c | 59 u32 offset; member 97 anon_stack[top++].offset = in btf_find_struct_member() 98 cur_offset + member->offset; in btf_find_struct_member() 112 cur_offset = anon_stack[top].offset; in btf_find_struct_member()
|
| A D | trace_probe.h | 77 u32 offset = get_loc_offs(loc); in update_data_loc() local 79 return make_data_loc(maxlen - consumed, offset + consumed); in update_data_loc() 124 int offset; member 231 unsigned int offset; /* Offset from argument entry */ member 424 int offset; member 452 extern int traceprobe_split_symbol_offset(char *symbol, long *offset); 454 char *buf, int offset); 476 size_t offset, struct trace_probe *tp); 579 void __trace_probe_log_err(int offset, int err);
|
| A D | trace_kprobe.c | 98 return tk->rp.kp.offset; in trace_kprobe_offset() 157 else if (tk->rp.kp.offset) in trace_kprobe_match_command_head() 207 addr += tk->rp.kp.offset; in trace_kprobe_address() 290 tk->rp.kp.offset = offs; in alloc_trace_kprobe() 444 unsigned long offset, size; in __within_notrace_func() local 450 addr -= offset; in __within_notrace_func() 877 long offset = 0; in trace_kprobe_create_internal() local 952 if (ret || offset < 0 || offset > UINT_MAX) { in trace_kprobe_create_internal() 1041 ctx->offset = 0; in trace_kprobe_create_internal() 1301 else if (tk->rp.kp.offset) in trace_kprobe_show() [all …]
|
| /kernel/entry/ |
| A D | syscall_user_dispatch.c | 40 if (likely(instruction_pointer(regs) - sd->offset < sd->len)) in syscall_user_dispatch() 73 unsigned long offset, unsigned long len, in task_set_syscall_user_dispatch() argument 78 if (offset || len || selector) in task_set_syscall_user_dispatch() 88 if (offset && offset + len <= offset) in task_set_syscall_user_dispatch() 92 if (len == 0 || offset + len <= offset) in task_set_syscall_user_dispatch() 98 offset = offset + len; in task_set_syscall_user_dispatch() 119 task->syscall_dispatch.offset = offset; in task_set_syscall_user_dispatch() 131 int set_syscall_user_dispatch(unsigned long mode, unsigned long offset, in set_syscall_user_dispatch() argument 134 return task_set_syscall_user_dispatch(current, mode, offset, len, selector); in set_syscall_user_dispatch() 151 cfg.offset = sd->offset; in syscall_user_dispatch_get_config() [all …]
|
| /kernel/time/ |
| A D | namespace.c | 26 ktime_t offset; in do_timens_ktime_to_host() local 44 if (tim < offset) { in do_timens_ktime_to_host() 51 tim = ktime_sub(tim, offset); in do_timens_ktime_to_host() 182 struct timens_offset *offset = vc->offset; in timens_setup_vdso_clock_data() local 188 offset[CLOCK_MONOTONIC] = monotonic; in timens_setup_vdso_clock_data() 189 offset[CLOCK_MONOTONIC_RAW] = monotonic; in timens_setup_vdso_clock_data() 191 offset[CLOCK_BOOTTIME] = boottime; in timens_setup_vdso_clock_data() 192 offset[CLOCK_BOOTTIME_ALARM] = boottime; in timens_setup_vdso_clock_data() 445 struct timespec64 *offset = NULL; in proc_timens_set_offset() local 452 offset = &time_ns->offsets.boottime; in proc_timens_set_offset() [all …]
|
| A D | timer_list.c | 109 (unsigned long long) ktime_to_ns(base->offset)); in print_base() 112 print_active_timers(m, base, now + ktime_to_ns(base->offset)); in print_base() 305 static void *move_iter(struct timer_list_iter *iter, loff_t offset) in move_iter() argument 307 for (; offset; offset--) { in move_iter() 324 static void *timer_list_start(struct seq_file *file, loff_t *offset) in timer_list_start() argument 328 if (!*offset) in timer_list_start() 332 return move_iter(iter, *offset); in timer_list_start() 335 static void *timer_list_next(struct seq_file *file, void *v, loff_t *offset) in timer_list_next() argument 338 ++*offset; in timer_list_next()
|
| A D | ntp.c | 131 return offset; in ntp_offset_chunk() 133 return shift_right(offset, SHIFT_PLL + ntpdata->time_constant); in ntp_offset_chunk() 219 return shift_right(offset, SHIFT_PLL + ntpdata->time_constant); in ntp_offset_chunk() 295 offset = clamp(offset, -USEC_PER_SEC, USEC_PER_SEC); in ntp_update_offset() 296 offset *= NSEC_PER_USEC; in ntp_update_offset() 300 offset = clamp(offset, -MAXPHASE, MAXPHASE); in ntp_update_offset() 313 offset64 = offset; in ntp_update_offset() 757 ntp_update_offset(ntpdata, txc->offset); in process_adjtimex_modes() 781 ntpdata->time_adjust = txc->offset; in ntp_adjtimex() 787 txc->offset = save_adjust; in ntp_adjtimex() [all …]
|
| A D | timekeeping.c | 1238 *cycles -= base->offset; in convert_cs_to_base() 2075 s64 offset, in timekeeping_apply_adjustment() argument 2084 offset = -offset; in timekeeping_apply_adjustment() 2087 offset *= mult_adj; in timekeeping_apply_adjustment() 2270 if (offset < interval) in logarithmic_accumulation() 2271 return offset; in logarithmic_accumulation() 2274 offset -= interval; in logarithmic_accumulation() 2294 return offset; in logarithmic_accumulation() 2316 orig_offset = offset; in __timekeeping_advance() 2335 offset = logarithmic_accumulation(tk, offset, shift, &clock_set); in __timekeeping_advance() [all …]
|
| /kernel/power/ |
| A D | user.c | 209 compat_loff_t offset; member 216 sector_t offset; in snapshot_set_swap_area() local 228 offset = swap_area.offset; in snapshot_set_swap_area() 235 offset = swap_area.offset; in snapshot_set_swap_area() 242 data->swap = swap_type_of(swdev, offset); in snapshot_set_swap_area() 255 sector_t offset; in snapshot_ioctl() local 371 offset = alloc_swapdev_block(data->swap); in snapshot_ioctl() 372 if (offset) { in snapshot_ioctl() 373 offset <<= PAGE_SHIFT; in snapshot_ioctl() 374 error = put_user(offset, (loff_t __user *)arg); in snapshot_ioctl()
|
| A D | swap.c | 180 unsigned long offset; in alloc_swapdev_block() local 183 if (offset) { in alloc_swapdev_block() 378 if (!offset) in write_page() 442 sector_t offset; in swap_write_page() local 453 if (!offset) in swap_write_page() 460 handle->cur_swap = offset; in swap_write_page() 1000 sector_t offset; in get_swap_reader() local 1009 offset = swsusp_header->image; in get_swap_reader() 1010 while (offset) { in get_swap_reader() 1044 sector_t offset; in swap_read_page() local [all …]
|
| /kernel/events/ |
| A D | uprobes.c | 74 loff_t offset; member 858 if (l_offset < r->offset) in uprobe_cmp() 861 if (l_offset > r->offset) in uprobe_cmp() 872 loff_t offset; member 895 .offset = offset, in find_uprobe_rcu() 992 uprobe->offset = offset; in alloc_uprobe() 1469 loff_t offset; in unapply_uprobe() local 1476 if (uprobe->offset < offset || in unapply_uprobe() 1477 uprobe->offset >= offset + vma->vm_end - vma->vm_start) in unapply_uprobe() 1501 if (max < u->offset) in find_node_in_range() [all …]
|
| /kernel/locking/ |
| A D | qspinlock_paravirt.h | 203 #define for_each_hash_entry(he, offset, hash) \ argument 204 for (hash &= ~(PV_HE_PER_LINE - 1), he = &pv_lock_hash[hash], offset = 0; \ 205 offset < (1 << pv_lock_hash_bits); \ 206 offset++, he = &pv_lock_hash[(hash + offset) & ((1 << pv_lock_hash_bits) - 1)]) 210 unsigned long offset, hash = hash_ptr(lock, pv_lock_hash_bits); in pv_hash() local 214 for_each_hash_entry(he, offset, hash) { in pv_hash() 238 unsigned long offset, hash = hash_ptr(lock, pv_lock_hash_bits); in pv_unhash() local 242 for_each_hash_entry(he, offset, hash) { in pv_unhash()
|
| /kernel/sched/ |
| A D | stats.c | 175 static void *schedstat_start(struct seq_file *file, loff_t *offset) in schedstat_start() argument 177 unsigned long n = *offset; in schedstat_start() 189 *offset = n + 1; in schedstat_start() 197 static void *schedstat_next(struct seq_file *file, void *data, loff_t *offset) in schedstat_next() argument 199 (*offset)++; in schedstat_next() 201 return schedstat_start(file, offset); in schedstat_next()
|
| A D | loadavg.c | 73 void get_avenrun(unsigned long *loads, unsigned long offset, int shift) in get_avenrun() argument 75 loads[0] = (avenrun[0] + offset) << shift; in get_avenrun() 76 loads[1] = (avenrun[1] + offset) << shift; in get_avenrun() 77 loads[2] = (avenrun[2] + offset) << shift; in get_avenrun()
|
| /kernel/bpf/ |
| A D | helpers.c | 1710 ptr->offset = offset; in bpf_dynptr_init() 1777 memmove(dst, src->data + src->offset + offset, len); in __bpf_dynptr_read() 1780 return __bpf_skb_load_bytes(src->data, src->offset + offset, dst, len); in __bpf_dynptr_read() 1782 return __bpf_xdp_load_bytes(src->data, src->offset + offset, dst, len); in __bpf_dynptr_read() 1830 memmove(dst->data + dst->offset + offset, src, len); in __bpf_dynptr_write() 1833 return __bpf_skb_store_bytes(dst->data, dst->offset + offset, src, len, in __bpf_dynptr_write() 1838 return __bpf_xdp_store_bytes(dst->data, dst->offset + offset, src, len); in __bpf_dynptr_write() 1882 return (unsigned long)(ptr->data + ptr->offset + offset); in BPF_CALL_3() 2696 return ptr->data + ptr->offset + offset; in bpf_dynptr_slice() 2701 return skb_pointer_if_linear(ptr->data, ptr->offset + offset, len); in bpf_dynptr_slice() [all …]
|
| /kernel/irq/ |
| A D | ipi.c | 26 unsigned int nr_irqs, offset; in irq_reserve_ipi() local 54 offset = 0; in irq_reserve_ipi() 64 offset = cpumask_first(dest); in irq_reserve_ipi() 69 next = cpumask_next_zero(offset, dest); in irq_reserve_ipi() 95 data->common->ipi_offset = offset; in irq_reserve_ipi()
|
| A D | irq_sim.c | 131 unsigned int offset = 0; in irq_sim_handle_irq() local 137 offset = find_next_bit(work_ctx->pending, in irq_sim_handle_irq() 138 work_ctx->irq_count, offset); in irq_sim_handle_irq() 139 clear_bit(offset, work_ctx->pending); in irq_sim_handle_irq() 140 irqnum = irq_find_mapping(work_ctx->domain, offset); in irq_sim_handle_irq()
|