Lines Matching refs:ng
1419 static inline unsigned long group_faults_priv(struct numa_group *ng);
1420 static inline unsigned long group_faults_shared(struct numa_group *ng);
1462 struct numa_group *ng; in task_scan_start() local
1466 ng = rcu_dereference(p->numa_group); in task_scan_start()
1467 if (ng) { in task_scan_start()
1468 unsigned long shared = group_faults_shared(ng); in task_scan_start()
1469 unsigned long private = group_faults_priv(ng); in task_scan_start()
1471 period *= refcount_read(&ng->refcount); in task_scan_start()
1484 struct numa_group *ng; in task_scan_max() local
1490 ng = deref_curr_numa_group(p); in task_scan_max()
1491 if (ng) { in task_scan_max()
1492 unsigned long shared = group_faults_shared(ng); in task_scan_max()
1493 unsigned long private = group_faults_priv(ng); in task_scan_max()
1496 period *= refcount_read(&ng->refcount); in task_scan_max()
1529 struct numa_group *ng; in task_numa_group_id() local
1533 ng = rcu_dereference(p->numa_group); in task_numa_group_id()
1534 if (ng) in task_numa_group_id()
1535 gid = ng->gid; in task_numa_group_id()
1563 struct numa_group *ng = deref_task_numa_group(p); in group_faults() local
1565 if (!ng) in group_faults()
1568 return ng->faults[task_faults_idx(NUMA_MEM, nid, 0)] + in group_faults()
1569 ng->faults[task_faults_idx(NUMA_MEM, nid, 1)]; in group_faults()
1578 static inline unsigned long group_faults_priv(struct numa_group *ng) in group_faults_priv() argument
1584 faults += ng->faults[task_faults_idx(NUMA_MEM, node, 1)]; in group_faults_priv()
1590 static inline unsigned long group_faults_shared(struct numa_group *ng) in group_faults_shared() argument
1596 faults += ng->faults[task_faults_idx(NUMA_MEM, node, 0)]; in group_faults_shared()
1609 static bool numa_is_active_node(int nid, struct numa_group *ng) in numa_is_active_node() argument
1611 return group_faults_cpu(ng, nid) * ACTIVE_NODE_FRACTION > ng->max_faults_cpu; in numa_is_active_node()
1708 struct numa_group *ng = deref_task_numa_group(p); in group_weight() local
1711 if (!ng) in group_weight()
1714 total_faults = ng->total_faults; in group_weight()
1842 struct numa_group *ng = deref_curr_numa_group(p); in should_numa_migrate_memory() local
1920 if (!ng) in should_numa_migrate_memory()
1927 if (group_faults_cpu(ng, dst_nid) > group_faults_cpu(ng, src_nid) * in should_numa_migrate_memory()
1939 return group_faults_cpu(ng, dst_nid) * group_faults(p, src_nid) * 3 > in should_numa_migrate_memory()
1940 group_faults_cpu(ng, src_nid) * group_faults(p, dst_nid) * 4; in should_numa_migrate_memory()
2433 struct numa_group *ng; in task_numa_migrate() local
2483 ng = deref_curr_numa_group(p); in task_numa_migrate()
2484 if (env.best_cpu == -1 || (ng && ng->active_nodes > 1)) { in task_numa_migrate()
2517 if (ng) { in task_numa_migrate()
2814 struct numa_group *ng; in task_numa_placement() local
2832 ng = deref_curr_numa_group(p); in task_numa_placement()
2833 if (ng) { in task_numa_placement()
2834 group_lock = &ng->lock; in task_numa_placement()
2875 if (ng) { in task_numa_placement()
2883 ng->faults[mem_idx] += diff; in task_numa_placement()
2884 ng->faults[cpu_idx] += f_diff; in task_numa_placement()
2885 ng->total_faults += diff; in task_numa_placement()
2886 group_faults += ng->faults[mem_idx]; in task_numa_placement()
2890 if (!ng) { in task_numa_placement()
2916 if (ng) { in task_numa_placement()
2917 numa_group_count_active_nodes(ng); in task_numa_placement()
3096 struct numa_group *ng; in task_numa_fault() local
3146 ng = deref_curr_numa_group(p); in task_numa_fault()
3147 if (!priv && !local && ng && ng->active_nodes > 1 && in task_numa_fault()
3148 numa_is_active_node(cpu_node, ng) && in task_numa_fault()
3149 numa_is_active_node(mem_node, ng)) in task_numa_fault()
12970 struct numa_group *ng; in show_numa_stats() local
12973 ng = rcu_dereference(p->numa_group); in show_numa_stats()
12979 if (ng) { in show_numa_stats()
12980 gsf = ng->faults[task_faults_idx(NUMA_MEM, node, 0)], in show_numa_stats()
12981 gpf = ng->faults[task_faults_idx(NUMA_MEM, node, 1)]; in show_numa_stats()