Lines Matching refs:nr_running
687 static u64 __sched_period(unsigned long nr_running) in __sched_period() argument
689 if (unlikely(nr_running > sched_nr_latency)) in __sched_period()
690 return nr_running * sysctl_sched_min_granularity; in __sched_period()
703 u64 slice = __sched_period(cfs_rq->nr_running + !se->on_rq); in sched_slice()
1511 unsigned int nr_running; member
1553 static inline long adjust_numa_imbalance(int imbalance, int nr_running);
1559 if ((ns->nr_running > ns->weight) && in numa_classify()
1564 if ((ns->nr_running < ns->weight) || in numa_classify()
1619 ns->nr_running += rq->cfs.h_nr_running; in update_numa_stats()
1622 if (find_idle && !rq->nr_running && idle_cpu(cpu)) { in update_numa_stats()
1930 src_running = env->src_stats.nr_running - 1; in task_numa_find_cpu()
1931 dst_running = env->dst_stats.nr_running + 1; in task_numa_find_cpu()
3000 cfs_rq->nr_running++; in account_entity_enqueue()
3013 cfs_rq->nr_running--; in account_entity_dequeue()
4245 if (cfs_rq->nr_running == 1 || cfs_bandwidth_used()) in enqueue_entity()
4248 if (cfs_rq->nr_running == 1) in enqueue_entity()
4542 if (cfs_rq->nr_running > 1) in entity_tick()
4719 if (cfs_rq->nr_running >= 1) in tg_unthrottle_up()
4891 if (rq->curr == rq->idle && rq->cfs.nr_running) in unthrottle_cfs_rq()
5075 if (!cfs_rq->runtime_enabled || cfs_rq->nr_running) in return_cfs_rq_runtime()
5422 if (cfs_rq_of(&curr->se)->nr_running < sched_nr_latency) in hrtick_update()
5458 return unlikely(rq->nr_running == rq->cfs.idle_h_nr_running && in sched_idle_rq()
5459 rq->nr_running); in sched_idle_rq()
5805 if (sync && cpu_rq(this_cpu)->nr_running == 1) in wake_affine_idle()
6250 this_rq()->nr_running <= 1) { in select_idle_sibling()
6827 if (rq->nr_running) in balance_fair()
6921 int scale = cfs_rq->nr_running >= sched_nr_latency; in check_preempt_wakeup()
7047 if (!cfs_rq->nr_running) in pick_next_task_fair()
7176 if (unlikely(rq->nr_running == 1)) in yield_task_fair()
7438 if (sched_feat(CACHE_HOT_BUDDY) && env->dst_rq->nr_running && in task_hot()
7479 if (env->src_rq->nr_running > env->src_rq->nr_preferred_running) in migrate_degrades_locality()
7661 if (env->idle != CPU_NOT_IDLE && env->src_rq->nr_running <= 1) in detach_tasks()
8382 int i, nr_running, local_group; in update_sg_lb_stats() local
8399 nr_running = rq->nr_running; in update_sg_lb_stats()
8400 sgs->sum_nr_running += nr_running; in update_sg_lb_stats()
8402 if (nr_running > 1) in update_sg_lb_stats()
8415 if (!nr_running && idle_cpu(i)) { in update_sg_lb_stats()
8585 if (rq->nr_running > rq->nr_numa_running) in fbq_classify_rq()
8587 if (rq->nr_running > rq->nr_preferred_running) in fbq_classify_rq()
8662 int i, nr_running; in update_sg_wakeup_stats() local
8676 nr_running = rq->nr_running - local; in update_sg_wakeup_stats()
8677 sgs->sum_nr_running += nr_running; in update_sg_wakeup_stats()
8682 if (!nr_running && idle_cpu_without(i, p)) in update_sg_wakeup_stats()
8992 static inline long adjust_numa_imbalance(int imbalance, int nr_running) in adjust_numa_imbalance() argument
9001 if (nr_running <= imbalance_min) in adjust_numa_imbalance()
9323 unsigned int nr_running; in find_busiest_queue() local
9352 nr_running = rq->cfs.h_nr_running; in find_busiest_queue()
9362 nr_running == 1) in find_busiest_queue()
9373 if (nr_running == 1 && load > env->imbalance && in find_busiest_queue()
9405 if (nr_running <= 1) in find_busiest_queue()
9415 if (busiest_nr < nr_running) { in find_busiest_queue()
9416 busiest_nr = nr_running; in find_busiest_queue()
9585 if (busiest->nr_running > 1) { in load_balance()
9593 env.loop_max = min(sysctl_sched_nr_migrate, busiest->nr_running); in load_balance()
9874 if (busiest_rq->nr_running <= 1) in active_load_balance_cpu_stop()
10149 if (rq->nr_running >= 2) { in nohz_balancer_kick()
10592 if (pulled_task || this_rq->nr_running > 0) in newidle_balance()
10616 if (this_rq->nr_running != this_rq->cfs.h_nr_running) in newidle_balance()
10756 if (rq->cfs.nr_running == 1) in prio_changed_fair()
11371 return rq ? rq->nr_running : -1; in sched_trace_rq_nr_running()