Lines Matching refs:ng
1136 static inline unsigned long group_faults_priv(struct numa_group *ng);
1137 static inline unsigned long group_faults_shared(struct numa_group *ng);
1179 struct numa_group *ng; in task_scan_start() local
1183 ng = rcu_dereference(p->numa_group); in task_scan_start()
1184 if (ng) { in task_scan_start()
1185 unsigned long shared = group_faults_shared(ng); in task_scan_start()
1186 unsigned long private = group_faults_priv(ng); in task_scan_start()
1188 period *= refcount_read(&ng->refcount); in task_scan_start()
1201 struct numa_group *ng; in task_scan_max() local
1207 ng = deref_curr_numa_group(p); in task_scan_max()
1208 if (ng) { in task_scan_max()
1209 unsigned long shared = group_faults_shared(ng); in task_scan_max()
1210 unsigned long private = group_faults_priv(ng); in task_scan_max()
1213 period *= refcount_read(&ng->refcount); in task_scan_max()
1246 struct numa_group *ng; in task_numa_group_id() local
1250 ng = rcu_dereference(p->numa_group); in task_numa_group_id()
1251 if (ng) in task_numa_group_id()
1252 gid = ng->gid; in task_numa_group_id()
1280 struct numa_group *ng = deref_task_numa_group(p); in group_faults() local
1282 if (!ng) in group_faults()
1285 return ng->faults[task_faults_idx(NUMA_MEM, nid, 0)] + in group_faults()
1286 ng->faults[task_faults_idx(NUMA_MEM, nid, 1)]; in group_faults()
1295 static inline unsigned long group_faults_priv(struct numa_group *ng) in group_faults_priv() argument
1301 faults += ng->faults[task_faults_idx(NUMA_MEM, node, 1)]; in group_faults_priv()
1307 static inline unsigned long group_faults_shared(struct numa_group *ng) in group_faults_shared() argument
1313 faults += ng->faults[task_faults_idx(NUMA_MEM, node, 0)]; in group_faults_shared()
1326 static bool numa_is_active_node(int nid, struct numa_group *ng) in numa_is_active_node() argument
1328 return group_faults_cpu(ng, nid) * ACTIVE_NODE_FRACTION > ng->max_faults_cpu; in numa_is_active_node()
1425 struct numa_group *ng = deref_task_numa_group(p); in group_weight() local
1428 if (!ng) in group_weight()
1431 total_faults = ng->total_faults; in group_weight()
1559 struct numa_group *ng = deref_curr_numa_group(p); in should_numa_migrate_memory() local
1637 if (!ng) in should_numa_migrate_memory()
1644 if (group_faults_cpu(ng, dst_nid) > group_faults_cpu(ng, src_nid) * in should_numa_migrate_memory()
1656 return group_faults_cpu(ng, dst_nid) * group_faults(p, src_nid) * 3 > in should_numa_migrate_memory()
1657 group_faults_cpu(ng, src_nid) * group_faults(p, dst_nid) * 4; in should_numa_migrate_memory()
2167 struct numa_group *ng; in task_numa_migrate() local
2217 ng = deref_curr_numa_group(p); in task_numa_migrate()
2218 if (env.best_cpu == -1 || (ng && ng->active_nodes > 1)) { in task_numa_migrate()
2251 if (ng) { in task_numa_migrate()
2548 struct numa_group *ng; in task_numa_placement() local
2566 ng = deref_curr_numa_group(p); in task_numa_placement()
2567 if (ng) { in task_numa_placement()
2568 group_lock = &ng->lock; in task_numa_placement()
2609 if (ng) { in task_numa_placement()
2617 ng->faults[mem_idx] += diff; in task_numa_placement()
2618 ng->faults[cpu_idx] += f_diff; in task_numa_placement()
2619 ng->total_faults += diff; in task_numa_placement()
2620 group_faults += ng->faults[mem_idx]; in task_numa_placement()
2624 if (!ng) { in task_numa_placement()
2650 if (ng) { in task_numa_placement()
2651 numa_group_count_active_nodes(ng); in task_numa_placement()
2830 struct numa_group *ng; in task_numa_fault() local
2880 ng = deref_curr_numa_group(p); in task_numa_fault()
2881 if (!priv && !local && ng && ng->active_nodes > 1 && in task_numa_fault()
2882 numa_is_active_node(cpu_node, ng) && in task_numa_fault()
2883 numa_is_active_node(mem_node, ng)) in task_numa_fault()
12212 struct numa_group *ng; in show_numa_stats() local
12215 ng = rcu_dereference(p->numa_group); in show_numa_stats()
12221 if (ng) { in show_numa_stats()
12222 gsf = ng->faults[task_faults_idx(NUMA_MEM, node, 0)], in show_numa_stats()
12223 gpf = ng->faults[task_faults_idx(NUMA_MEM, node, 1)]; in show_numa_stats()