Searched refs:nid (Results 1 – 7 of 7) sorted by relevance
| /kernel/dma/ |
| A D | contiguous.c | 102 int nid, count = 0; in early_numa_cma() local 117 numa_cma_size[nid] = tmp; in early_numa_cma() 160 int nid; in dma_numa_cma_reserve() local 162 for_each_node(nid) { in dma_numa_cma_reserve() 167 if (!node_online(nid)) { in dma_numa_cma_reserve() 178 0, false, name, cma, nid); in dma_numa_cma_reserve() 181 ret, nid); in dma_numa_cma_reserve() 184 if (numa_cma_size[nid]) { in dma_numa_cma_reserve() 189 name, cma, nid); in dma_numa_cma_reserve() 192 ret, nid); in dma_numa_cma_reserve() [all …]
|
| /kernel/sched/ |
| A D | fair.c | 2507 int nid, ret, dist; in task_numa_migrate() local 2558 if (nid == env.src_nid || nid == p->numa_preferred_nid) in task_numa_migrate() 2575 env.dst_nid = nid; in task_numa_migrate() 2591 nid = env.src_nid; in task_numa_migrate() 2596 sched_setnuma(p, nid); in task_numa_migrate() 2653 int nid, active_nodes = 0; in numa_group_count_active_nodes() local 2802 return nid; in preferred_group_nid() 2867 nid = a; in preferred_group_nid() 2875 return nid; in preferred_group_nid() 2965 max_nid = nid; in task_numa_placement() [all …]
|
| A D | core.c | 3529 int nid = cpu_to_node(cpu); in select_fallback_rq() local 3539 if (nid != -1) { in select_fallback_rq() 3540 nodemask = cpumask_of_node(nid); in select_fallback_rq() 8117 void sched_setnuma(struct task_struct *p, int nid) in sched_setnuma() argument 8132 p->numa_preferred_nid = nid; in sched_setnuma()
|
| /kernel/ |
| A D | kexec_handover.c | 461 static phys_addr_t __init scratch_size_node(int nid) in scratch_size_node() argument 467 nid); in scratch_size_node() 488 int nid, i = 0; in kho_reserve_scratch() local 526 for_each_online_node(nid) { in kho_reserve_scratch() 527 size = scratch_size_node(nid); in kho_reserve_scratch() 530 nid, true); in kho_reserve_scratch()
|
| A D | padata.c | 447 int nworks, nid; in padata_do_multithreaded() local 487 nid = next_node_in(old_node, node_states[N_CPU]); in padata_do_multithreaded() 488 } while (!atomic_try_cmpxchg(&last_used_nid, &old_node, nid)); in padata_do_multithreaded() 489 queue_work_node(nid, system_unbound_wq, &pw->pw_work); in padata_do_multithreaded()
|
| /kernel/bpf/ |
| A D | syscall.c | 581 static struct page *__bpf_alloc_page(int nid) in __bpf_alloc_page() argument 584 return alloc_pages_nolock(nid, 0); in __bpf_alloc_page() 586 return alloc_pages_node(nid, in __bpf_alloc_page() 592 int bpf_map_alloc_pages(const struct bpf_map *map, int nid, in bpf_map_alloc_pages() argument 605 pg = __bpf_alloc_page(nid); in bpf_map_alloc_pages()
|
| /kernel/cgroup/ |
| A D | cpuset.c | 4278 bool cpuset_node_allowed(struct cgroup *cgroup, int nid) in cpuset_node_allowed() argument 4309 allowed = node_isset(nid, cs->effective_mems); in cpuset_node_allowed()
|
Completed in 57 milliseconds