Lines Matching refs:ng
1112 static inline unsigned long group_faults_priv(struct numa_group *ng);
1113 static inline unsigned long group_faults_shared(struct numa_group *ng);
1155 struct numa_group *ng; in task_scan_start() local
1159 ng = rcu_dereference(p->numa_group); in task_scan_start()
1160 if (ng) { in task_scan_start()
1161 unsigned long shared = group_faults_shared(ng); in task_scan_start()
1162 unsigned long private = group_faults_priv(ng); in task_scan_start()
1164 period *= refcount_read(&ng->refcount); in task_scan_start()
1177 struct numa_group *ng; in task_scan_max() local
1183 ng = deref_curr_numa_group(p); in task_scan_max()
1184 if (ng) { in task_scan_max()
1185 unsigned long shared = group_faults_shared(ng); in task_scan_max()
1186 unsigned long private = group_faults_priv(ng); in task_scan_max()
1189 period *= refcount_read(&ng->refcount); in task_scan_max()
1222 struct numa_group *ng; in task_numa_group_id() local
1226 ng = rcu_dereference(p->numa_group); in task_numa_group_id()
1227 if (ng) in task_numa_group_id()
1228 gid = ng->gid; in task_numa_group_id()
1256 struct numa_group *ng = deref_task_numa_group(p); in group_faults() local
1258 if (!ng) in group_faults()
1261 return ng->faults[task_faults_idx(NUMA_MEM, nid, 0)] + in group_faults()
1262 ng->faults[task_faults_idx(NUMA_MEM, nid, 1)]; in group_faults()
1271 static inline unsigned long group_faults_priv(struct numa_group *ng) in group_faults_priv() argument
1277 faults += ng->faults[task_faults_idx(NUMA_MEM, node, 1)]; in group_faults_priv()
1283 static inline unsigned long group_faults_shared(struct numa_group *ng) in group_faults_shared() argument
1289 faults += ng->faults[task_faults_idx(NUMA_MEM, node, 0)]; in group_faults_shared()
1302 static bool numa_is_active_node(int nid, struct numa_group *ng) in numa_is_active_node() argument
1304 return group_faults_cpu(ng, nid) * ACTIVE_NODE_FRACTION > ng->max_faults_cpu; in numa_is_active_node()
1400 struct numa_group *ng = deref_task_numa_group(p); in group_weight() local
1403 if (!ng) in group_weight()
1406 total_faults = ng->total_faults; in group_weight()
1420 struct numa_group *ng = deref_curr_numa_group(p); in should_numa_migrate_memory() local
1463 if (!ng) in should_numa_migrate_memory()
1470 if (group_faults_cpu(ng, dst_nid) > group_faults_cpu(ng, src_nid) * in should_numa_migrate_memory()
1482 return group_faults_cpu(ng, dst_nid) * group_faults(p, src_nid) * 3 > in should_numa_migrate_memory()
1483 group_faults_cpu(ng, src_nid) * group_faults(p, dst_nid) * 4; in should_numa_migrate_memory()
1984 struct numa_group *ng; in task_numa_migrate() local
2032 ng = deref_curr_numa_group(p); in task_numa_migrate()
2033 if (env.best_cpu == -1 || (ng && ng->active_nodes > 1)) { in task_numa_migrate()
2066 if (ng) { in task_numa_migrate()
2363 struct numa_group *ng; in task_numa_placement() local
2381 ng = deref_curr_numa_group(p); in task_numa_placement()
2382 if (ng) { in task_numa_placement()
2383 group_lock = &ng->lock; in task_numa_placement()
2424 if (ng) { in task_numa_placement()
2432 ng->faults[mem_idx] += diff; in task_numa_placement()
2433 ng->faults_cpu[mem_idx] += f_diff; in task_numa_placement()
2434 ng->total_faults += diff; in task_numa_placement()
2435 group_faults += ng->faults[mem_idx]; in task_numa_placement()
2439 if (!ng) { in task_numa_placement()
2450 if (ng) { in task_numa_placement()
2451 numa_group_count_active_nodes(ng); in task_numa_placement()
2632 struct numa_group *ng; in task_numa_fault() local
2673 ng = deref_curr_numa_group(p); in task_numa_fault()
2674 if (!priv && !local && ng && ng->active_nodes > 1 && in task_numa_fault()
2675 numa_is_active_node(cpu_node, ng) && in task_numa_fault()
2676 numa_is_active_node(mem_node, ng)) in task_numa_fault()
11243 struct numa_group *ng; in show_numa_stats() local
11246 ng = rcu_dereference(p->numa_group); in show_numa_stats()
11252 if (ng) { in show_numa_stats()
11253 gsf = ng->faults[task_faults_idx(NUMA_MEM, node, 0)], in show_numa_stats()
11254 gpf = ng->faults[task_faults_idx(NUMA_MEM, node, 1)]; in show_numa_stats()