Lines Matching refs:ng

1148 static inline unsigned long group_faults_priv(struct numa_group *ng);
1149 static inline unsigned long group_faults_shared(struct numa_group *ng);
1191 struct numa_group *ng; in task_scan_start() local
1195 ng = rcu_dereference(p->numa_group); in task_scan_start()
1196 if (ng) { in task_scan_start()
1197 unsigned long shared = group_faults_shared(ng); in task_scan_start()
1198 unsigned long private = group_faults_priv(ng); in task_scan_start()
1200 period *= refcount_read(&ng->refcount); in task_scan_start()
1213 struct numa_group *ng; in task_scan_max() local
1219 ng = deref_curr_numa_group(p); in task_scan_max()
1220 if (ng) { in task_scan_max()
1221 unsigned long shared = group_faults_shared(ng); in task_scan_max()
1222 unsigned long private = group_faults_priv(ng); in task_scan_max()
1225 period *= refcount_read(&ng->refcount); in task_scan_max()
1258 struct numa_group *ng; in task_numa_group_id() local
1262 ng = rcu_dereference(p->numa_group); in task_numa_group_id()
1263 if (ng) in task_numa_group_id()
1264 gid = ng->gid; in task_numa_group_id()
1292 struct numa_group *ng = deref_task_numa_group(p); in group_faults() local
1294 if (!ng) in group_faults()
1297 return ng->faults[task_faults_idx(NUMA_MEM, nid, 0)] + in group_faults()
1298 ng->faults[task_faults_idx(NUMA_MEM, nid, 1)]; in group_faults()
1307 static inline unsigned long group_faults_priv(struct numa_group *ng) in group_faults_priv() argument
1313 faults += ng->faults[task_faults_idx(NUMA_MEM, node, 1)]; in group_faults_priv()
1319 static inline unsigned long group_faults_shared(struct numa_group *ng) in group_faults_shared() argument
1325 faults += ng->faults[task_faults_idx(NUMA_MEM, node, 0)]; in group_faults_shared()
1338 static bool numa_is_active_node(int nid, struct numa_group *ng) in numa_is_active_node() argument
1340 return group_faults_cpu(ng, nid) * ACTIVE_NODE_FRACTION > ng->max_faults_cpu; in numa_is_active_node()
1437 struct numa_group *ng = deref_task_numa_group(p); in group_weight() local
1440 if (!ng) in group_weight()
1443 total_faults = ng->total_faults; in group_weight()
1571 struct numa_group *ng = deref_curr_numa_group(p); in should_numa_migrate_memory() local
1649 if (!ng) in should_numa_migrate_memory()
1656 if (group_faults_cpu(ng, dst_nid) > group_faults_cpu(ng, src_nid) * in should_numa_migrate_memory()
1668 return group_faults_cpu(ng, dst_nid) * group_faults(p, src_nid) * 3 > in should_numa_migrate_memory()
1669 group_faults_cpu(ng, src_nid) * group_faults(p, dst_nid) * 4; in should_numa_migrate_memory()
2179 struct numa_group *ng; in task_numa_migrate() local
2229 ng = deref_curr_numa_group(p); in task_numa_migrate()
2230 if (env.best_cpu == -1 || (ng && ng->active_nodes > 1)) { in task_numa_migrate()
2263 if (ng) { in task_numa_migrate()
2560 struct numa_group *ng; in task_numa_placement() local
2578 ng = deref_curr_numa_group(p); in task_numa_placement()
2579 if (ng) { in task_numa_placement()
2580 group_lock = &ng->lock; in task_numa_placement()
2621 if (ng) { in task_numa_placement()
2629 ng->faults[mem_idx] += diff; in task_numa_placement()
2630 ng->faults[cpu_idx] += f_diff; in task_numa_placement()
2631 ng->total_faults += diff; in task_numa_placement()
2632 group_faults += ng->faults[mem_idx]; in task_numa_placement()
2636 if (!ng) { in task_numa_placement()
2662 if (ng) { in task_numa_placement()
2663 numa_group_count_active_nodes(ng); in task_numa_placement()
2842 struct numa_group *ng; in task_numa_fault() local
2892 ng = deref_curr_numa_group(p); in task_numa_fault()
2893 if (!priv && !local && ng && ng->active_nodes > 1 && in task_numa_fault()
2894 numa_is_active_node(cpu_node, ng) && in task_numa_fault()
2895 numa_is_active_node(mem_node, ng)) in task_numa_fault()
12583 struct numa_group *ng; in show_numa_stats() local
12586 ng = rcu_dereference(p->numa_group); in show_numa_stats()
12592 if (ng) { in show_numa_stats()
12593 gsf = ng->faults[task_faults_idx(NUMA_MEM, node, 0)], in show_numa_stats()
12594 gpf = ng->faults[task_faults_idx(NUMA_MEM, node, 1)]; in show_numa_stats()