Lines Matching refs:ng
1103 static inline unsigned long group_faults_priv(struct numa_group *ng);
1104 static inline unsigned long group_faults_shared(struct numa_group *ng);
1146 struct numa_group *ng; in task_scan_start() local
1150 ng = rcu_dereference(p->numa_group); in task_scan_start()
1151 if (ng) { in task_scan_start()
1152 unsigned long shared = group_faults_shared(ng); in task_scan_start()
1153 unsigned long private = group_faults_priv(ng); in task_scan_start()
1155 period *= refcount_read(&ng->refcount); in task_scan_start()
1168 struct numa_group *ng; in task_scan_max() local
1174 ng = deref_curr_numa_group(p); in task_scan_max()
1175 if (ng) { in task_scan_max()
1176 unsigned long shared = group_faults_shared(ng); in task_scan_max()
1177 unsigned long private = group_faults_priv(ng); in task_scan_max()
1180 period *= refcount_read(&ng->refcount); in task_scan_max()
1213 struct numa_group *ng; in task_numa_group_id() local
1217 ng = rcu_dereference(p->numa_group); in task_numa_group_id()
1218 if (ng) in task_numa_group_id()
1219 gid = ng->gid; in task_numa_group_id()
1247 struct numa_group *ng = deref_task_numa_group(p); in group_faults() local
1249 if (!ng) in group_faults()
1252 return ng->faults[task_faults_idx(NUMA_MEM, nid, 0)] + in group_faults()
1253 ng->faults[task_faults_idx(NUMA_MEM, nid, 1)]; in group_faults()
1262 static inline unsigned long group_faults_priv(struct numa_group *ng) in group_faults_priv() argument
1268 faults += ng->faults[task_faults_idx(NUMA_MEM, node, 1)]; in group_faults_priv()
1274 static inline unsigned long group_faults_shared(struct numa_group *ng) in group_faults_shared() argument
1280 faults += ng->faults[task_faults_idx(NUMA_MEM, node, 0)]; in group_faults_shared()
1293 static bool numa_is_active_node(int nid, struct numa_group *ng) in numa_is_active_node() argument
1295 return group_faults_cpu(ng, nid) * ACTIVE_NODE_FRACTION > ng->max_faults_cpu; in numa_is_active_node()
1391 struct numa_group *ng = deref_task_numa_group(p); in group_weight() local
1394 if (!ng) in group_weight()
1397 total_faults = ng->total_faults; in group_weight()
1411 struct numa_group *ng = deref_curr_numa_group(p); in should_numa_migrate_memory() local
1454 if (!ng) in should_numa_migrate_memory()
1461 if (group_faults_cpu(ng, dst_nid) > group_faults_cpu(ng, src_nid) * in should_numa_migrate_memory()
1473 return group_faults_cpu(ng, dst_nid) * group_faults(p, src_nid) * 3 > in should_numa_migrate_memory()
1474 group_faults_cpu(ng, src_nid) * group_faults(p, dst_nid) * 4; in should_numa_migrate_memory()
1754 struct numa_group *ng; in task_numa_migrate() local
1802 ng = deref_curr_numa_group(p); in task_numa_migrate()
1803 if (env.best_cpu == -1 || (ng && ng->active_nodes > 1)) { in task_numa_migrate()
1836 if (ng) { in task_numa_migrate()
2131 struct numa_group *ng; in task_numa_placement() local
2149 ng = deref_curr_numa_group(p); in task_numa_placement()
2150 if (ng) { in task_numa_placement()
2151 group_lock = &ng->lock; in task_numa_placement()
2192 if (ng) { in task_numa_placement()
2200 ng->faults[mem_idx] += diff; in task_numa_placement()
2201 ng->faults_cpu[mem_idx] += f_diff; in task_numa_placement()
2202 ng->total_faults += diff; in task_numa_placement()
2203 group_faults += ng->faults[mem_idx]; in task_numa_placement()
2207 if (!ng) { in task_numa_placement()
2218 if (ng) { in task_numa_placement()
2219 numa_group_count_active_nodes(ng); in task_numa_placement()
2400 struct numa_group *ng; in task_numa_fault() local
2441 ng = deref_curr_numa_group(p); in task_numa_fault()
2442 if (!priv && !local && ng && ng->active_nodes > 1 && in task_numa_fault()
2443 numa_is_active_node(cpu_node, ng) && in task_numa_fault()
2444 numa_is_active_node(mem_node, ng)) in task_numa_fault()
10488 struct numa_group *ng; in show_numa_stats() local
10491 ng = rcu_dereference(p->numa_group); in show_numa_stats()
10497 if (ng) { in show_numa_stats()
10498 gsf = ng->faults[task_faults_idx(NUMA_MEM, node, 0)], in show_numa_stats()
10499 gpf = ng->faults[task_faults_idx(NUMA_MEM, node, 1)]; in show_numa_stats()