| /kernel/time/ |
| A D | posix-cpu-timers.c | 496 if (timer->it.cpu.firing) { in posix_cpu_timer_del() 918 if (!it->expires) in check_cpu_itimer() 922 if (it->incr) in check_cpu_itimer() 923 it->expires += it->incr; in check_cpu_itimer() 925 it->expires = 0; in check_cpu_itimer() 933 if (it->expires && it->expires < *expires) in check_cpu_itimer() 934 *expires = it->expires; in check_cpu_itimer() 1481 struct itimerspec64 it; in do_cpu_nanosleep() local 1501 memset(&it, 0, sizeof(it)); in do_cpu_nanosleep() 1502 it.it_value = *rqtp; in do_cpu_nanosleep() [all …]
|
| A D | itimer.c | 51 struct cpu_itimer *it = &tsk->signal->it[clock_id]; in get_cpu_itimer() local 55 val = it->expires; in get_cpu_itimer() 56 interval = it->incr; in get_cpu_itimer() 193 struct cpu_itimer *it = &tsk->signal->it[clock_id]; in set_cpu_itimer() local 200 oval = it->expires; in set_cpu_itimer() 201 ointerval = it->incr; in set_cpu_itimer() 207 it->expires = nval; in set_cpu_itimer() 208 it->incr = ninterval; in set_cpu_itimer()
|
| A D | alarmtimer.c | 516 struct k_itimer *ptr = container_of(alarm, struct k_itimer, it.alarm.alarmtimer); in alarm_handle_timer() 528 struct alarm *alarm = &timr->it.alarm.alarmtimer; in alarm_timer_rearm() 541 struct alarm *alarm = &timr->it.alarm.alarmtimer; in alarm_timer_forward() 553 struct alarm *alarm = &timr->it.alarm.alarmtimer; in alarm_timer_remaining() 564 return alarm_try_to_cancel(&timr->it.alarm.alarmtimer); in alarm_timer_try_to_cancel() 577 hrtimer_cancel_wait_running(&timr->it.alarm.alarmtimer.timer); in alarm_timer_wait_running() 590 struct alarm *alarm = &timr->it.alarm.alarmtimer; in alarm_timer_arm() 598 alarm_start(&timr->it.alarm.alarmtimer, expires); in alarm_timer_arm() 670 alarm_init(&new_timer->it.alarm.alarmtimer, type, alarm_handle_timer); in alarm_timer_create()
|
| A D | time.c | 984 int get_itimerspec64(struct itimerspec64 *it, in get_itimerspec64() argument 989 ret = get_timespec64(&it->it_interval, &uit->it_interval); in get_itimerspec64() 993 ret = get_timespec64(&it->it_value, &uit->it_value); in get_itimerspec64() 1007 int put_itimerspec64(const struct itimerspec64 *it, in put_itimerspec64() argument 1012 ret = put_timespec64(&it->it_interval, &uit->it_interval); in put_itimerspec64() 1016 ret = put_timespec64(&it->it_value, &uit->it_value); in put_itimerspec64()
|
| A D | Kconfig | 106 rate, even when the CPU doesn't need it. 138 desired range of dynticks CPUs to use it. This is implemented at 172 for user context tracking and the subsystems that rely on it: RCU 188 We keep it around for a little while to enforce backward 210 it is good enough for the clocksource watchdog!
|
| A D | posix-timers.c | 300 struct hrtimer *timer = &timr->it.real.timer; in common_hrtimer_rearm() 377 struct k_itimer *timr = container_of(timer, struct k_itimer, it.real.timer); in posix_timer_fn() 461 hrtimer_setup(&new_timer->it.real.timer, posix_timer_fn, new_timer->it_clock, 0); in common_timer_create() 660 struct hrtimer *timer = &timr->it.real.timer; in common_hrtimer_remaining() 667 struct hrtimer *timer = &timr->it.real.timer; in common_hrtimer_forward() 809 struct hrtimer *timer = &timr->it.real.timer; in common_hrtimer_arm() 825 hrtimer_setup(&timr->it.real.timer, posix_timer_fn, timr->it_clock, mode); in common_hrtimer_arm() 837 return hrtimer_try_to_cancel(&timr->it.real.timer); in common_hrtimer_try_to_cancel() 842 hrtimer_cancel_wait_running(&timer->it.real.timer); in common_timer_wait_running()
|
| /kernel/bpf/ |
| A D | kmem_cache_iter.c | 24 __bpf_kfunc int bpf_iter_kmem_cache_new(struct bpf_iter_kmem_cache *it) in bpf_iter_kmem_cache_new() argument 26 struct bpf_iter_kmem_cache_kern *kit = (void *)it; in bpf_iter_kmem_cache_new() 28 BUILD_BUG_ON(sizeof(*kit) > sizeof(*it)); in bpf_iter_kmem_cache_new() 29 BUILD_BUG_ON(__alignof__(*kit) != __alignof__(*it)); in bpf_iter_kmem_cache_new() 35 __bpf_kfunc struct kmem_cache *bpf_iter_kmem_cache_next(struct bpf_iter_kmem_cache *it) in bpf_iter_kmem_cache_next() argument 37 struct bpf_iter_kmem_cache_kern *kit = (void *)it; in bpf_iter_kmem_cache_next() 80 __bpf_kfunc void bpf_iter_kmem_cache_destroy(struct bpf_iter_kmem_cache *it) in bpf_iter_kmem_cache_destroy() argument 82 struct bpf_iter_kmem_cache_kern *kit = (void *)it; in bpf_iter_kmem_cache_destroy() 111 struct bpf_iter_kmem_cache it; member 167 bpf_iter_kmem_cache_destroy(&p->it); in kmem_cache_iter_seq_stop() [all …]
|
| A D | dmabuf_iter.c | 119 __bpf_kfunc int bpf_iter_dmabuf_new(struct bpf_iter_dmabuf *it) in bpf_iter_dmabuf_new() argument 121 struct bpf_iter_dmabuf_kern *kit = (void *)it; in bpf_iter_dmabuf_new() 123 BUILD_BUG_ON(sizeof(*kit) > sizeof(*it)); in bpf_iter_dmabuf_new() 124 BUILD_BUG_ON(__alignof__(*kit) != __alignof__(*it)); in bpf_iter_dmabuf_new() 130 __bpf_kfunc struct dma_buf *bpf_iter_dmabuf_next(struct bpf_iter_dmabuf *it) in bpf_iter_dmabuf_next() argument 132 struct bpf_iter_dmabuf_kern *kit = (void *)it; in bpf_iter_dmabuf_next() 142 __bpf_kfunc void bpf_iter_dmabuf_destroy(struct bpf_iter_dmabuf *it) in bpf_iter_dmabuf_destroy() argument 144 struct bpf_iter_dmabuf_kern *kit = (void *)it; in bpf_iter_dmabuf_destroy()
|
| A D | task_iter.c | 818 __bpf_kfunc int bpf_iter_task_vma_new(struct bpf_iter_task_vma *it, in bpf_iter_task_vma_new() argument 821 struct bpf_iter_task_vma_kern *kit = (void *)it; in bpf_iter_task_vma_new() 864 struct bpf_iter_task_vma_kern *kit = (void *)it; in bpf_iter_task_vma_next() 873 struct bpf_iter_task_vma_kern *kit = (void *)it; in bpf_iter_task_vma_destroy() 899 struct bpf_iter_css_task_kern *kit = (void *)it; in bpf_iter_css_task_new() 923 struct bpf_iter_css_task_kern *kit = (void *)it; in bpf_iter_css_task_next() 932 struct bpf_iter_css_task_kern *kit = (void *)it; in bpf_iter_css_task_destroy() 965 __bpf_kfunc int bpf_iter_task_new(struct bpf_iter_task *it, in bpf_iter_task_new() argument 968 struct bpf_iter_task_kern *kit = (void *)it; in bpf_iter_task_new() 999 struct bpf_iter_task_kern *kit = (void *)it; in bpf_iter_task_next() [all …]
|
| A D | cgroup_iter.c | 310 __bpf_kfunc int bpf_iter_css_new(struct bpf_iter_css *it, in bpf_iter_css_new() argument 313 struct bpf_iter_css_kern *kit = (void *)it; in bpf_iter_css_new() 334 __bpf_kfunc struct cgroup_subsys_state *bpf_iter_css_next(struct bpf_iter_css *it) in bpf_iter_css_next() argument 336 struct bpf_iter_css_kern *kit = (void *)it; in bpf_iter_css_next() 355 __bpf_kfunc void bpf_iter_css_destroy(struct bpf_iter_css *it) in bpf_iter_css_destroy() argument
|
| A D | Kconfig | 80 still reenable it by setting it to 0 later on, or permanently 81 disable it by setting it to 1 (from which no other transition to
|
| A D | bpf_iter.c | 785 __bpf_kfunc int bpf_iter_num_new(struct bpf_iter_num *it, int start, int end) in bpf_iter_num_new() argument 787 struct bpf_iter_num_kern *s = (void *)it; in bpf_iter_num_new() 816 __bpf_kfunc int *bpf_iter_num_next(struct bpf_iter_num* it) in bpf_iter_num_next() argument 818 struct bpf_iter_num_kern *s = (void *)it; in bpf_iter_num_next() 835 __bpf_kfunc void bpf_iter_num_destroy(struct bpf_iter_num *it) in bpf_iter_num_destroy() argument 837 struct bpf_iter_num_kern *s = (void *)it; in bpf_iter_num_destroy()
|
| /kernel/cgroup/ |
| A D | legacy_freezer.c | 252 struct css_task_iter it; in update_if_frozen() local 275 css_task_iter_start(css, 0, &it); in update_if_frozen() 277 while ((task = css_task_iter_next(&it))) { in update_if_frozen() 284 css_task_iter_end(&it); in update_if_frozen() 316 struct css_task_iter it; in freeze_cgroup() local 319 css_task_iter_start(&freezer->css, 0, &it); in freeze_cgroup() 320 while ((task = css_task_iter_next(&it))) in freeze_cgroup() 322 css_task_iter_end(&it); in freeze_cgroup() 327 struct css_task_iter it; in unfreeze_cgroup() local 331 while ((task = css_task_iter_next(&it))) in unfreeze_cgroup() [all …]
|
| A D | cgroup.c | 4916 it->task_pos = it->cur_tasks_head->next; in css_task_iter_advance_css_set() 4948 it->task_pos = it->task_pos->next; in css_task_iter_skip() 4968 it->task_pos = it->task_pos->next; in css_task_iter_advance() 4970 if (it->task_pos == &it->cur_cset->tasks) { in css_task_iter_advance() 4971 it->cur_tasks_head = &it->cur_cset->mg_tasks; in css_task_iter_advance() 4972 it->task_pos = it->cur_tasks_head->next; in css_task_iter_advance() 4974 if (it->task_pos == &it->cur_cset->mg_tasks) { in css_task_iter_advance() 4976 it->task_pos = it->cur_tasks_head->next; in css_task_iter_advance() 4978 if (it->task_pos == &it->cur_cset->dying_tasks) in css_task_iter_advance() 5022 memset(it, 0, sizeof(*it)); in css_task_iter_start() [all …]
|
| A D | freezer.c | 176 struct css_task_iter it; in cgroup_do_freeze() local 193 css_task_iter_start(&cgrp->self, 0, &it); in cgroup_do_freeze() 194 while ((task = css_task_iter_next(&it))) { in cgroup_do_freeze() 203 css_task_iter_end(&it); in cgroup_do_freeze()
|
| A D | cgroup-v1.c | 107 struct css_task_iter it; in cgroup_transfer_tasks() local 137 css_task_iter_start(&from->self, 0, &it); in cgroup_transfer_tasks() 140 task = css_task_iter_next(&it); in cgroup_transfer_tasks() 145 css_task_iter_end(&it); in cgroup_transfer_tasks() 341 struct css_task_iter it; in pidlist_array_load() local 358 css_task_iter_start(&cgrp->self, 0, &it); in pidlist_array_load() 359 while ((tsk = css_task_iter_next(&it))) { in pidlist_array_load() 370 css_task_iter_end(&it); in pidlist_array_load() 722 struct css_task_iter it; in cgroupstats_build() local 743 css_task_iter_start(&cgrp->self, 0, &it); in cgroupstats_build() [all …]
|
| A D | cpuset-v1.c | 241 struct css_task_iter it; in cpuset1_update_tasks_flags() local 244 css_task_iter_start(&cs->css, 0, &it); in cpuset1_update_tasks_flags() 245 while ((task = css_task_iter_next(&it))) in cpuset1_update_tasks_flags() 247 css_task_iter_end(&it); in cpuset1_update_tasks_flags()
|
| /kernel/sched/ |
| A D | idle.c | 377 struct idle_timer *it = container_of(timer, struct idle_timer, timer); in idle_inject_timer_fn() local 379 WRITE_ONCE(it->done, 1); in idle_inject_timer_fn() 387 struct idle_timer it; in play_idle_precise() local 405 it.done = 0; in play_idle_precise() 406 hrtimer_setup_on_stack(&it.timer, idle_inject_timer_fn, CLOCK_MONOTONIC, in play_idle_precise() 408 hrtimer_start(&it.timer, ns_to_ktime(duration_ns), in play_idle_precise() 411 while (!READ_ONCE(it.done)) in play_idle_precise()
|
| /kernel/trace/ |
| A D | Kconfig | 267 enable it via the trace option funcgraph-retval. 278 enable it via the trace option funcgraph-retaddr. 477 but when it is running, it can cause the system to be 635 Otherwise keep it disabled. 867 selected by tracers that use it. 896 that reflects when it is enabled or not. 947 As it is a tight loop, it benchmarks as hot cache. That's fine because 1039 called outside of RCU, as if they are, it can cause a race. But it 1100 up since it runs this on every system call defined. 1157 Do not use it on production systems. [all …]
|
| /kernel/rcu/ |
| A D | Kconfig.debug | 16 Enable RCU lockdep checking for list usages. By default it is 19 false-positive splats, we keep it default disabled but once all 80 This information can be useful, but it does incur additional 95 can be useful, but it does incur additional overhead, overhead 113 Say M if you want to build it as a module instead. 151 as this one, but will override this if it exists. 168 containing only lockless code. After all, it is quite possible 208 grace periods, making them as short as it can. This limits 212 But in conjunction with tools like KASAN, it can be helpful
|
| /kernel/power/ |
| A D | Kconfig | 48 system and powers it off; and restores that checkpoint on reboot. 57 In principle it does not require ACPI or APM, although for example 58 ACPI will be used for the final steps when it is available. One 65 have it detect the saved image, restore memory state from it, and 157 non-zero numbered CPU, it may define ARCH_SUSPEND_NONZERO_CPU. This 177 enabled. In particular, only enable this if it is very common to be 239 make it wake up a few seconds later using an RTC wakeup alarm. 243 linked, ensuring that it's available when this test runs. 310 machine, reboot it and then run 334 manpage ("man 8 hdparm") for that), and it doesn't turn off
|
| /kernel/trace/rv/monitors/pagefault/ |
| A D | Kconfig | 19 This monitor does not affect execution speed while it is not running, 20 therefore it is safe to enable this in production kernel.
|
| /kernel/module/ |
| A D | Kconfig | 52 with loading modules when it shouldn't or if you want to help 109 Only enable this for debugging system functionality, never have it 145 kernel believes it is unsafe: the kernel will remove the module 146 without waiting for anyone to stop using it (using the -f option to 166 Saying Y here makes it sometimes possible to use modules 213 supports it. 245 sum of the source files which made it. This helps maintainers 353 Note that for modules inside an initrd or initramfs, it's more 415 but it creates consistency between symbols defining namespaces and 425 When kernel code requests a module, it does so by calling
|
| /kernel/ |
| A D | Kconfig.kexec | 27 but it is independent of the system firmware. And like a reboot 28 you can start any kernel with it, not just Linux. 57 there's a signature that we can check, then it must be valid. 125 PHYSICAL_START, or it must be built as a relocatable image
|
| /kernel/dma/ |
| A D | Kconfig | 98 starts with one memory pool at boot and it will allocate additional 197 default, but it can be enabled by passing cma=size[MG] to the kernel. 208 If 0 percent is selected, CMA is disabled by default, but it can be 237 for larger buffers it just a memory waste. With this parameter you can
|