Lines Matching refs:sgc

78 				group->sgc->id,  in sched_domain_debug_one()
87 if (group->sgc->capacity != SCHED_CAPACITY_SCALE) in sched_domain_debug_one()
88 printk(KERN_CONT " cap=%lu", group->sgc->capacity); in sched_domain_debug_one()
569 if (free_sgc && atomic_dec_and_test(&sg->sgc->ref)) in free_sched_groups()
570 kfree(sg->sgc); in free_sched_groups()
909 sg->sgc = *per_cpu_ptr(sdd->sgc, cpu); in init_overlap_sched_group()
910 if (atomic_inc_return(&sg->sgc->ref) == 1) in init_overlap_sched_group()
921 sg->sgc->capacity = SCHED_CAPACITY_SCALE * cpumask_weight(sg_span); in init_overlap_sched_group()
922 sg->sgc->min_capacity = SCHED_CAPACITY_SCALE; in init_overlap_sched_group()
923 sg->sgc->max_capacity = SCHED_CAPACITY_SCALE; in init_overlap_sched_group()
1068 sg->sgc = *per_cpu_ptr(sdd->sgc, cpu); in get_group()
1073 WARN_ON(already_visited != (atomic_inc_return(&sg->sgc->ref) > 1)); in get_group()
1087 sg->sgc->capacity = SCHED_CAPACITY_SCALE * cpumask_weight(sched_group_span(sg)); in get_group()
1088 sg->sgc->min_capacity = SCHED_CAPACITY_SCALE; in get_group()
1089 sg->sgc->max_capacity = SCHED_CAPACITY_SCALE; in get_group()
1274 if (atomic_read(&(*per_cpu_ptr(sdd->sgc, cpu))->ref)) in claim_allocations()
1275 *per_cpu_ptr(sdd->sgc, cpu) = NULL; in claim_allocations()
1770 sdd->sgc = alloc_percpu(struct sched_group_capacity *); in __sdt_alloc()
1771 if (!sdd->sgc) in __sdt_alloc()
1778 struct sched_group_capacity *sgc; in __sdt_alloc() local
1803 sgc = kzalloc_node(sizeof(struct sched_group_capacity) + cpumask_size(), in __sdt_alloc()
1805 if (!sgc) in __sdt_alloc()
1809 sgc->id = j; in __sdt_alloc()
1812 *per_cpu_ptr(sdd->sgc, j) = sgc; in __sdt_alloc()
1841 if (sdd->sgc) in __sdt_free()
1842 kfree(*per_cpu_ptr(sdd->sgc, j)); in __sdt_free()
1850 free_percpu(sdd->sgc); in __sdt_free()
1851 sdd->sgc = NULL; in __sdt_free()