| /kernel/ |
| A D | nsproxy.c | 156 if ((flags & CLONE_VM) || in copy_namespaces() 268 if (!flags || (flags & ~(CLONE_NEWNS | CLONE_NEWUTS | CLONE_NEWIPC | in check_setns_flags() 278 if (flags & CLONE_NEWPID) in check_setns_flags() 307 unsigned flags = nsset->flags; in put_nsset() local 315 if (nsset->fs && (flags & CLONE_NEWNS) && (flags & ~CLONE_NEWNS)) in put_nsset() 345 nsset->flags = flags; in prepare_nsset() 368 unsigned flags = nsset->flags; in validate_nsset() local 504 unsigned flags = nsset->flags; in commit_nsset() local 516 if ((flags & CLONE_NEWNS) && (flags & ~CLONE_NEWNS)) { in commit_nsset() 548 if (flags && (ns->ops->type != flags)) in SYSCALL_DEFINE2() [all …]
|
| A D | kexec.c | 25 unsigned long flags) in kimage_alloc_init() argument 29 bool kexec_on_panic = flags & KEXEC_ON_CRASH; in kimage_alloc_init() 107 if (flags & KEXEC_ON_CRASH) { in do_kexec_load() 121 if (flags & KEXEC_ON_CRASH) { in do_kexec_load() 134 if (flags & KEXEC_PRESERVE_CONTEXT) in do_kexec_load() 138 if ((flags & KEXEC_ON_CRASH) && arch_crash_hotplug_support(image, flags)) in do_kexec_load() 203 unsigned long flags) in kexec_load_check() argument 205 int image_type = (flags & KEXEC_ON_CRASH) ? in kexec_load_check() 230 if ((flags & KEXEC_FLAGS) != (flags & ~KEXEC_ARCH_MASK)) in kexec_load_check() 248 result = kexec_load_check(nr_segments, flags); in SYSCALL_DEFINE4() [all …]
|
| A D | kthread.c | 57 unsigned long flags; member 84 WARN_ON(!(k->flags & PF_KTHREAD)); in to_kthread() 596 unsigned long flags; in __kthread_bind_mask() local 606 p->flags |= PF_NO_SETAFFINITY; in __kthread_bind_mask() 860 unsigned long flags; in kthread_affine_preferred() local 1051 worker->flags = flags; in __kthread_create_worker_on_node() 1188 unsigned long flags; in kthread_queue_work() local 1214 unsigned long flags; in kthread_delayed_work_timer_fn() local 1286 unsigned long flags; in kthread_queue_delayed_work() local 1433 unsigned long flags; in kthread_mod_delayed_work() local [all …]
|
| A D | resource.c | 39 .flags = IORESOURCE_IO, 47 .flags = IORESOURCE_MEM, 320 return (p->flags & flags) == flags && (desc == IORES_DESC_NONE || desc == p->desc); in is_type_match() 375 .flags = p->flags, in find_next_iomem_res() 700 avail.flags = new->flags & ~IORESOURCE_UNSET; in __find_resource_space() 702 alloc.flags = avail.flags; in __find_resource_space() 1271 res->flags |= IORESOURCE_BUSY | flags; in __request_region_locked() 1293 if (conflict->flags & flags & IORESOURCE_MUXED) { in __request_region_locked() 1480 new_res->flags = res->flags; in release_mem_region_adjustable() 1506 return r1->flags == r2->flags && r1->end + 1 == r2->start && in system_ram_resources_mergeable() [all …]
|
| A D | iomem.c | 10 unsigned long flags) in arch_memremap_wb() argument 22 unsigned long flags) in arch_memremap_can_ram_remap() argument 29 unsigned long flags) in try_ram_remap() argument 75 if (!flags) in memremap() 85 if (flags & MEMREMAP_WB) { in memremap() 93 addr = try_ram_remap(offset, size, flags); in memremap() 95 addr = arch_memremap_wb(offset, size, flags); in memremap() 110 if (!addr && (flags & MEMREMAP_WT)) in memremap() 113 if (!addr && (flags & MEMREMAP_WC)) in memremap() 138 size_t size, unsigned long flags) in devm_memremap() argument [all …]
|
| /kernel/sched/ |
| A D | wait.c | 20 unsigned long flags; in add_wait_queue() local 31 unsigned long flags; in add_wait_queue_exclusive() local 42 unsigned long flags; in add_wait_queue_priority() local 71 unsigned long flags; in remove_wait_queue() local 105 unsigned flags = curr->flags; in __wake_up_common() local 121 unsigned long flags; in __wake_up_common_lock() local 250 unsigned long flags; in prepare_to_wait() local 265 unsigned long flags; in prepare_to_wait_exclusive() local 282 wq_entry->flags = flags; in init_wait_entry() 291 unsigned long flags; in prepare_to_wait_event() local [all …]
|
| A D | isolation.c | 24 unsigned long flags; member 91 if (!housekeeping.flags) in housekeeping_init() 120 if ((flags & HK_FLAG_KERNEL_NOISE) && !(housekeeping.flags & HK_FLAG_KERNEL_NOISE)) { in housekeeping_setup() 160 unsigned long iter_flags = flags & housekeeping.flags; in housekeeping_setup() 170 iter_flags = flags & ~housekeeping.flags; in housekeeping_setup() 176 if ((flags & HK_FLAG_KERNEL_NOISE) && !(housekeeping.flags & HK_FLAG_KERNEL_NOISE)) in housekeeping_setup() 179 housekeeping.flags |= flags; in housekeeping_setup() 192 unsigned long flags; in housekeeping_nohz_full_setup() local 202 unsigned long flags = 0; in housekeeping_isolcpus_setup() local 248 if (!flags) in housekeeping_isolcpus_setup() [all …]
|
| A D | ext_idle.c | 125 if (flags & SCX_PICK_IDLE_CORE) in pick_idle_cpu_in_node() 547 if (!(current->flags & PF_EXITING) && in scx_select_cpu_dfl() 605 if (flags & SCX_PICK_IDLE_CORE) { in scx_select_cpu_dfl() 645 cpu = scx_pick_idle_cpu(allowed, node, flags); in scx_select_cpu_dfl() 860 const struct cpumask *allowed, u64 flags) in select_cpu_from_kfunc() argument 909 allowed ?: p->cpus_ptr, flags); in select_cpu_from_kfunc() 1127 int node, u64 flags) in scx_bpf_pick_idle_cpu_node() argument 1159 u64 flags) in scx_bpf_pick_idle_cpu() argument 1194 int node, u64 flags) in scx_bpf_pick_any_cpu_node() argument 1206 if (flags & SCX_PICK_IDLE_IN_NODE) in scx_bpf_pick_any_cpu_node() [all …]
|
| /kernel/futex/ |
| A D | syscalls.c | 109 return futex_requeue(uaddr, flags, uaddr2, flags, val, val2, NULL, 0); in do_futex() 111 return futex_requeue(uaddr, flags, uaddr2, flags, val, val2, &val3, 0); in do_futex() 128 return futex_requeue(uaddr, flags, uaddr2, flags, val, val2, &val3, 1); in do_futex() 201 unsigned int flags; in futex_parse_waitv() local 209 flags = futex2_to_flags(aux.flags); in futex_parse_waitv() 216 futexv[i].w.flags = flags; in futex_parse_waitv() 299 if (flags) in SYSCALL_DEFINE5() 342 unsigned int, flags) in SYSCALL_DEFINE4() argument 347 flags = futex2_to_flags(flags); in SYSCALL_DEFINE4() 384 flags = futex2_to_flags(flags); in SYSCALL_DEFINE6() [all …]
|
| A D | futex.h | 47 unsigned int flags = FLAGS_SIZE_32; in futex_to_flags() local 50 flags |= FLAGS_SHARED; in futex_to_flags() 53 flags |= FLAGS_CLOCKRT; in futex_to_flags() 55 return flags; in futex_to_flags() 66 flags |= FLAGS_SHARED; in futex2_to_flags() 69 flags |= FLAGS_NUMA; in futex2_to_flags() 72 flags |= FLAGS_MPOL; in futex2_to_flags() 74 return flags; in futex2_to_flags() 98 if (flags & FLAGS_NUMA) { in futex_flags_valid() 112 int bits = 8 * futex_size(flags); in futex_validate_input() [all …]
|
| /kernel/rcu/ |
| A D | tree_exp.h | 80 unsigned long flags; in sync_exp_reset_tree_hotplug() local 136 unsigned long flags; in sync_exp_reset_tree() local 174 unsigned long flags; in sync_rcu_exp_done_unlocked() local 229 unsigned long flags; in rcu_report_exp_rnp() local 269 unsigned long flags; in rcu_report_exp_rdp() local 361 unsigned long flags; in __sync_rcu_exp_select_node_cpus() local 629 unsigned long flags; in synchronize_rcu_expedited_wait() local 748 unsigned long flags; in rcu_exp_handler() local 809 unsigned long flags; in rcu_print_task_exp_stall() local 835 unsigned long flags; in rcu_exp_print_detail_task_stall_rnp() local [all …]
|
| A D | tree_nocb.h | 230 unsigned long flags; in wake_nocb_gp() local 269 unsigned long flags; in wake_nocb_gp_defer() local 610 unsigned long flags; in nocb_gp_toggle_rdp() local 653 unsigned long flags; in nocb_gp_wait() local 881 unsigned long flags; in nocb_cb_wait() local 987 unsigned long flags; in do_nocb_deferred_wakeup_timer() local 1005 unsigned long flags; in do_nocb_deferred_wakeup() local 1025 unsigned long flags; in rcu_nocb_queue_toggle_rdp() local 1041 unsigned long flags; in rcu_nocb_rdp_deoffload_wait_cond() local 1058 unsigned long flags; in rcu_nocb_rdp_deoffload() local [all …]
|
| /kernel/bpf/ |
| A D | mprog.c | 8 u32 id_or_fd, u32 flags, in bpf_mprog_link() argument 12 bool id = flags & BPF_F_ID; in bpf_mprog_link() 31 u32 id_or_fd, u32 flags, in bpf_mprog_prog() argument 35 bool id = flags & BPF_F_ID; in bpf_mprog_prog() 58 bool id = flags & BPF_F_ID; in bpf_mprog_tuple_relative() 282 flags = BPF_F_AFTER; in bpf_mprog_attach() 380 flags = BPF_F_AFTER; in bpf_mprog_detach() 403 const u32 flags = 0; in bpf_mprog_query() local 414 if (copy_to_user(&uattr->query.attach_flags, &flags, sizeof(flags))) in bpf_mprog_query() 439 copy_to_user(uprog_flags + i, &flags, sizeof(flags))) in bpf_mprog_query() [all …]
|
| A D | stackmap.c | 234 bool user = flags & BPF_F_USER_STACK; in __bpf_get_stackid() 301 u64, flags) in BPF_CALL_3() argument 366 user = flags & BPF_F_USER_STACK; in BPF_CALL_3() 390 flags = (flags & ~BPF_F_SKIP_FIELD_MASK) | skip; in BPF_CALL_3() 497 u64, flags) in BPF_CALL_4() argument 513 u64, flags) in BPF_CALL_4() argument 529 u64 flags, bool may_fault) in __bpf_get_task_stack() argument 546 u32, size, u64, flags) in BPF_CALL_4() argument 563 u32, size, u64, flags) in BPF_CALL_4() argument 596 user = flags & BPF_F_USER_STACK; in BPF_CALL_4() [all …]
|
| A D | ringbuf.c | 246 u64 flags) in ringbuf_map_update_elem() argument 467 if (unlikely(flags)) in BPF_CALL_3() 504 if (flags & BPF_RB_FORCE_WAKEUP) in bpf_ringbuf_commit() 537 u64, flags) in BPF_CALL_4() argument 570 switch (flags) { in BPF_CALL_2() 598 if (unlikely(flags)) { in BPF_CALL_4() 689 sample_len = hdr_len & ~flags; in __bpf_user_ringbuf_peek() 705 if (flags & BPF_RINGBUF_DISCARD_BIT) { in __bpf_user_ringbuf_peek() 715 if (flags & BPF_RINGBUF_BUSY_BIT) in __bpf_user_ringbuf_peek() 747 if (unlikely(flags & ~wakeup_flags)) in BPF_CALL_4() [all …]
|
| /kernel/locking/ |
| A D | rwbase_rt.c | 188 unsigned long flags) in __rwbase_write_unlock() argument 204 unsigned long flags; in rwbase_write_unlock() local 206 raw_spin_lock_irqsave(&rtm->wait_lock, flags); in rwbase_write_unlock() 213 unsigned long flags; in rwbase_write_downgrade() local 215 raw_spin_lock_irqsave(&rtm->wait_lock, flags); in rwbase_write_downgrade() 241 unsigned long flags; in rwbase_write_lock() local 252 raw_spin_lock_irqsave(&rtm->wait_lock, flags); in rwbase_write_lock() 262 __rwbase_write_unlock(rwb, 0, flags); in rwbase_write_lock() 289 unsigned long flags; in rwbase_write_trylock() local 296 raw_spin_lock_irqsave(&rtm->wait_lock, flags); in rwbase_write_trylock() [all …]
|
| A D | semaphore.c | 93 unsigned long flags; in down() local 96 raw_spin_lock_irqsave(&sem->lock, flags); in down() 116 unsigned long flags; in down_interruptible() local 120 raw_spin_lock_irqsave(&sem->lock, flags); in down_interruptible() 143 unsigned long flags; in down_killable() local 147 raw_spin_lock_irqsave(&sem->lock, flags); in down_killable() 173 unsigned long flags; in down_trylock() local 176 raw_spin_lock_irqsave(&sem->lock, flags); in down_trylock() 198 unsigned long flags; in down_timeout() local 202 raw_spin_lock_irqsave(&sem->lock, flags); in down_timeout() [all …]
|
| /kernel/trace/ |
| A D | trace_functions_graph.c | 99 return (tracer_flags.val & flags) == flags; in tracer_flags_is_set() 115 struct trace_seq *s, u32 flags); 706 if (flags & TRACE_GRAPH_PRINT_CPU) in print_graph_irq() 901 struct trace_seq *s, u32 flags) in print_graph_entry_leaf() argument 965 cpu, iter->ent->pid, flags); in print_graph_entry_leaf() 1059 if (flags & TRACE_GRAPH_PRINT_CPU) in print_graph_prologue() 1235 u32 flags) in print_graph_return() argument 1312 cpu, pid, flags); in print_graph_return() 1473 if (flags & TRACE_GRAPH_PRINT_CPU) in print_lat_header() 1491 print_lat_header(s, flags); in __print_graph_headers_flags() [all …]
|
| /kernel/printk/ |
| A D | internal.h | 20 (con->flags & CON_NBCON) ? "" : "legacy ", \ 21 (con->flags & CON_BOOT) ? "boot" : "", \ 82 #define printk_safe_enter_irqsave(flags) \ argument 84 local_irq_save(flags); \ 91 local_irq_restore(flags); \ 99 enum printk_info_flags *flags); 123 if (!(flags & CON_ENABLED)) in console_is_usable() 126 if ((flags & CON_SUSPENDED)) in console_is_usable() 129 if (flags & CON_NBCON) { in console_is_usable() 190 #define printk_safe_enter_irqsave(flags) local_irq_save(flags) argument [all …]
|
| A D | printk.c | 318 unsigned long flags; in __down_trylock_console_sem() local 338 unsigned long flags; in __up_console_sem() local 1148 dest_r.info->flags = r->info->flags; in add_to_rb() 1181 unsigned long flags; in setup_log_buf() local 2007 unsigned long flags; in console_trylock_spinning() local 2198 if (flags) in printk_parse_prefix() 2342 r.info->flags = flags & 0x1f; in vprintk_store() 3088 unsigned long flags; in console_emit_next_record() local 3429 short flags; in __console_rewind_all() local 3984 unsigned long flags; in register_console() local [all …]
|
| /kernel/time/ |
| A D | clocksource.c | 205 cs->flags |= CLOCK_SOURCE_UNSTABLE; in __clocksource_unstable() 233 unsigned long flags; in clocksource_mark_unstable() local 421 cs->flags &= ~CLOCK_SOURCE_WATCHDOG; in clocksource_reset_watchdog() 479 cs->flags |= CLOCK_SOURCE_WATCHDOG; in clocksource_watchdog() 642 unsigned long flags; in clocksource_select_watchdog() local 691 unsigned long flags; in __clocksource_watchdog_kthread() local 1239 unsigned long flags; in __clocksource_register_scale() local 1258 clocksource_watchdog_lock(&flags); in __clocksource_register_scale() 1261 clocksource_watchdog_unlock(&flags); in __clocksource_register_scale() 1276 unsigned long flags; in clocksource_unbind() local [all …]
|
| A D | timer.c | 516 timer->flags = (timer->flags & ~TIMER_ARRAYMASK) | in timer_set_idx() 808 unsigned int flags, 853 unsigned int flags, in do_init_timer() argument 860 timer->flags = flags | raw_smp_processor_id(); in do_init_timer() 1005 if (timer->flags == tf) in lock_timer_base() 1302 unsigned long flags; in add_timer_on() local 1363 unsigned long flags; in __timer_delete() local 1454 unsigned long flags; in __try_to_del_timer_sync() local 1595 unsigned long flags; in __timer_delete_sync() local 1601 local_irq_save(flags); in __timer_delete_sync() [all …]
|
| /kernel/cgroup/ |
| A D | freezer.c | 24 set_bit(CGRP_FROZEN, &cgrp->flags); in cgroup_update_frozen_flag() 26 clear_bit(CGRP_FROZEN, &cgrp->flags); in cgroup_update_frozen_flag() 49 if (!test_bit(CGRP_FREEZE, &cgrp->flags) || in cgroup_propagate_frozen() 154 unsigned long flags; in cgroup_freeze_task() local 157 if (!lock_task_sighand(task, &flags)) in cgroup_freeze_task() 168 unlock_task_sighand(task, &flags); in cgroup_freeze_task() 183 set_bit(CGRP_FREEZE, &cgrp->flags); in cgroup_do_freeze() 185 clear_bit(CGRP_FREEZE, &cgrp->flags); in cgroup_do_freeze() 199 if (task->flags & PF_KTHREAD) in cgroup_do_freeze() 227 if (task->flags & PF_KTHREAD) in cgroup_freezer_migrate_task() [all …]
|
| /kernel/dma/ |
| A D | debug.c | 179 unsigned long flags; in driver_filter() local 256 *flags = __flags; in get_hash_bucket() 475 unsigned long flags; in active_cacheline_insert() local 497 unsigned long flags; in active_cacheline_remove() local 584 unsigned long flags; in add_dma_entry() local 659 unsigned long flags; in dma_entry_alloc() local 691 unsigned long flags; in dma_entry_free() local 717 unsigned long flags; in filter_read() local 739 unsigned long flags; in filter_write() local 826 unsigned long flags; in device_dma_allocations() local [all …]
|
| /kernel/irq/ |
| A D | manage.c | 1575 if (!((old->flags & new->flags) & IRQF_SHARED) || in __setup_irq() 1582 else if ((old->flags ^ new->flags) & IRQF_ONESHOT) in __setup_irq() 1770 irq, new->flags, new->name, old->flags, old->name); in __setup_irq() 1812 unsigned long flags; in __free_irq() local 1895 local_irq_save(flags); in __free_irq() 1897 local_irq_restore(flags); in __free_irq() 2116 action->flags = irqflags; in request_threaded_irq() 2142 unsigned long flags; in request_threaded_irq() local 2145 local_irq_save(flags); in request_threaded_irq() 2477 if (flags && flags != IRQF_TIMER) in __request_percpu_irq() [all …]
|