Lines Matching refs:dl

25 	return container_of(dl_se, struct task_struct, dl);  in dl_task_of()
30 return container_of(dl_rq, struct rq, dl); in rq_of_dl_rq()
38 return &rq->dl; in dl_rq_of_se()
69 return &cpu_rq(i)->dl.dl_bw; in dl_bw_of()
160 BUG_ON(p->dl.flags & SCHED_FLAG_SUGOV); in dl_change_utilization()
166 if (p->dl.dl_non_contending) { in dl_change_utilization()
167 sub_running_bw(&p->dl, &rq->dl); in dl_change_utilization()
168 p->dl.dl_non_contending = 0; in dl_change_utilization()
176 if (hrtimer_try_to_cancel(&p->dl.inactive_timer) == 1) in dl_change_utilization()
179 __sub_rq_bw(p->dl.dl_bw, &rq->dl); in dl_change_utilization()
180 __add_rq_bw(new_bw, &rq->dl); in dl_change_utilization()
239 struct sched_dl_entity *dl_se = &p->dl; in task_non_contending()
278 sub_rq_bw(&p->dl, &rq->dl); in task_non_contending()
280 __dl_sub(dl_b, p->dl.dl_bw, dl_bw_cpus(task_cpu(p))); in task_non_contending()
332 struct sched_dl_entity *dl_se = &p->dl; in is_leftmost()
447 struct dl_rq *dl_rq = &rq->dl; in enqueue_pushable_dl_task()
459 if (dl_entity_preempt(&p->dl, &entry->dl)) in enqueue_pushable_dl_task()
468 dl_rq->earliest_dl.next = p->dl.deadline; in enqueue_pushable_dl_task()
477 struct dl_rq *dl_rq = &rq->dl; in dequeue_pushable_dl_task()
488 struct task_struct, pushable_dl_tasks)->dl.deadline; in dequeue_pushable_dl_task()
498 return !RB_EMPTY_ROOT(&rq->dl.pushable_dl_tasks_root.rb_root); in has_pushable_dl_tasks()
561 if (p->dl.dl_non_contending || p->dl.dl_throttled) { in dl_task_offline_migration()
568 sub_running_bw(&p->dl, &rq->dl); in dl_task_offline_migration()
569 sub_rq_bw(&p->dl, &rq->dl); in dl_task_offline_migration()
571 add_rq_bw(&p->dl, &later_rq->dl); in dl_task_offline_migration()
572 add_running_bw(&p->dl, &later_rq->dl); in dl_task_offline_migration()
574 sub_rq_bw(&p->dl, &rq->dl); in dl_task_offline_migration()
575 add_rq_bw(&p->dl, &later_rq->dl); in dl_task_offline_migration()
585 __dl_sub(dl_b, p->dl.dl_bw, cpumask_weight(rq->rd->span)); in dl_task_offline_migration()
590 __dl_add(dl_b, p->dl.dl_bw, cpumask_weight(later_rq->rd->span)); in dl_task_offline_migration()
922 struct sched_dl_entity *dl_se = &p->dl; in start_dl_timer()
1155 u64 u_inact = rq->dl.this_bw - rq->dl.running_bw; /* Utot - Uact */ in grub_reclaim()
1157 u64 u_act_min = (dl_se->dl_bw * rq->dl.bw_ratio) >> RATIO_SHIFT; in grub_reclaim()
1167 if (u_inact + rq->dl.extra_bw > BW_UNIT - u_act_min) in grub_reclaim()
1170 u_act = BW_UNIT - u_inact - rq->dl.extra_bw; in grub_reclaim()
1182 struct sched_dl_entity *dl_se = &curr->dl; in update_curr_dl()
1228 &curr->dl); in update_curr_dl()
1252 if (!is_leftmost(curr, &rq->dl)) in update_curr_dl()
1300 sub_running_bw(&p->dl, dl_rq_of_se(&p->dl)); in inactive_task_timer()
1301 sub_rq_bw(&p->dl, dl_rq_of_se(&p->dl)); in inactive_task_timer()
1306 __dl_sub(dl_b, p->dl.dl_bw, dl_bw_cpus(task_cpu(p))); in inactive_task_timer()
1315 sub_running_bw(dl_se, &rq->dl); in inactive_task_timer()
1475 struct sched_dl_entity *pi_se = &p->dl; in enqueue_task_dl()
1485 if (pi_task && dl_prio(pi_task->normal_prio) && p->dl.dl_boosted) { in enqueue_task_dl()
1486 pi_se = &pi_task->dl; in enqueue_task_dl()
1495 BUG_ON(!p->dl.dl_boosted || flags != ENQUEUE_REPLENISH); in enqueue_task_dl()
1505 if (!p->dl.dl_throttled && !dl_is_implicit(&p->dl)) in enqueue_task_dl()
1506 dl_check_constrained_dl(&p->dl); in enqueue_task_dl()
1509 add_rq_bw(&p->dl, &rq->dl); in enqueue_task_dl()
1510 add_running_bw(&p->dl, &rq->dl); in enqueue_task_dl()
1525 if (p->dl.dl_throttled && !(flags & ENQUEUE_REPLENISH)) { in enqueue_task_dl()
1527 task_contending(&p->dl, flags); in enqueue_task_dl()
1532 enqueue_dl_entity(&p->dl, pi_se, flags); in enqueue_task_dl()
1540 dequeue_dl_entity(&p->dl); in __dequeue_task_dl()
1550 sub_running_bw(&p->dl, &rq->dl); in dequeue_task_dl()
1551 sub_rq_bw(&p->dl, &rq->dl); in dequeue_task_dl()
1585 rq->curr->dl.dl_yielded = 1; in yield_task_dl()
1626 !dl_entity_preempt(&p->dl, &curr->dl)) && in select_task_rq_dl()
1631 (dl_time_before(p->dl.deadline, in select_task_rq_dl()
1632 cpu_rq(target)->dl.earliest_dl.curr) || in select_task_rq_dl()
1633 (cpu_rq(target)->dl.dl_nr_running == 0))) in select_task_rq_dl()
1656 if (p->dl.dl_non_contending) { in migrate_task_rq_dl()
1657 sub_running_bw(&p->dl, &rq->dl); in migrate_task_rq_dl()
1658 p->dl.dl_non_contending = 0; in migrate_task_rq_dl()
1666 if (hrtimer_try_to_cancel(&p->dl.inactive_timer) == 1) in migrate_task_rq_dl()
1669 sub_rq_bw(&p->dl, &rq->dl); in migrate_task_rq_dl()
1696 if (!on_dl_rq(&p->dl) && need_pull_dl_task(rq, p)) { in balance_dl()
1719 if (dl_entity_preempt(&p->dl, &rq->curr->dl)) { in check_preempt_curr_dl()
1729 if ((p->dl.deadline == rq->curr->dl.deadline) && in check_preempt_curr_dl()
1738 hrtick_start(rq, p->dl.runtime); in start_hrtick_dl()
1777 struct dl_rq *dl_rq = &rq->dl; in pick_next_task_dl()
1797 if (on_dl_rq(&p->dl) && p->nr_cpus_allowed > 1) in put_prev_task_dl()
1819 if (hrtick_enabled(rq) && queued && p->dl.runtime > 0 && in task_tick_dl()
1820 is_leftmost(p, &rq->dl)) in task_tick_dl()
1851 struct rb_node *next_node = rq->dl.pushable_dl_tasks_root.rb_leftmost; in pick_earliest_pushable_dl_task()
1975 if (later_rq->dl.dl_nr_running && in find_lock_later_rq()
1976 !dl_time_before(task->dl.deadline, in find_lock_later_rq()
1977 later_rq->dl.earliest_dl.curr)) { in find_lock_later_rq()
2005 if (!later_rq->dl.dl_nr_running || in find_lock_later_rq()
2006 dl_time_before(task->dl.deadline, in find_lock_later_rq()
2007 later_rq->dl.earliest_dl.curr)) in find_lock_later_rq()
2025 p = rb_entry(rq->dl.pushable_dl_tasks_root.rb_leftmost, in pick_next_pushable_dl_task()
2049 if (!rq->dl.overloaded) in push_dl_task()
2066 dl_time_before(next_task->dl.deadline, rq->curr->dl.deadline) && in push_dl_task()
2158 if (this_rq->dl.dl_nr_running && in pull_dl_task()
2159 dl_time_before(this_rq->dl.earliest_dl.curr, in pull_dl_task()
2160 src_rq->dl.earliest_dl.next)) in pull_dl_task()
2170 if (src_rq->dl.dl_nr_running <= 1) in pull_dl_task()
2180 if (p && dl_time_before(p->dl.deadline, dmin) && in pull_dl_task()
2181 (!this_rq->dl.dl_nr_running || in pull_dl_task()
2182 dl_time_before(p->dl.deadline, in pull_dl_task()
2183 this_rq->dl.earliest_dl.curr))) { in pull_dl_task()
2191 if (dl_time_before(p->dl.deadline, in pull_dl_task()
2192 src_rq->curr->dl.deadline)) in pull_dl_task()
2200 dmin = p->dl.deadline; in pull_dl_task()
2223 !dl_entity_preempt(&p->dl, &rq->curr->dl))) { in task_woken_dl()
2254 __dl_sub(src_dl_b, p->dl.dl_bw, dl_bw_cpus(task_cpu(p))); in set_cpus_allowed_dl()
2264 if (rq->dl.overloaded) in rq_online_dl()
2268 if (rq->dl.dl_nr_running > 0) in rq_online_dl()
2269 cpudl_set(&rq->rd->cpudl, rq->cpu, rq->dl.earliest_dl.curr); in rq_online_dl()
2275 if (rq->dl.overloaded) in rq_offline_dl()
2304 __dl_add(dl_b, p->dl.dl_bw, cpumask_weight(rq->rd->span)); in dl_add_task_root_domain()
2333 if (task_on_rq_queued(p) && p->dl.dl_runtime) in switched_from_dl()
2343 if (p->dl.dl_non_contending) in switched_from_dl()
2344 sub_running_bw(&p->dl, &rq->dl); in switched_from_dl()
2345 sub_rq_bw(&p->dl, &rq->dl); in switched_from_dl()
2353 if (p->dl.dl_non_contending) in switched_from_dl()
2354 p->dl.dl_non_contending = 0; in switched_from_dl()
2361 if (!task_on_rq_queued(p) || rq->dl.dl_nr_running) in switched_from_dl()
2373 if (hrtimer_try_to_cancel(&p->dl.inactive_timer) == 1) in switched_to_dl()
2378 add_rq_bw(&p->dl, &rq->dl); in switched_to_dl()
2385 if (p->nr_cpus_allowed > 1 && rq->dl.overloaded) in switched_to_dl()
2410 if (!rq->dl.overloaded) in prio_changed_dl()
2418 if (dl_time_before(rq->dl.earliest_dl.curr, p->dl.deadline)) in prio_changed_dl()
2537 init_dl_rq_bw_ratio(&cpu_rq(cpu)->dl); in sched_dl_do_global()
2562 if (new_bw == p->dl.dl_bw && task_has_dl_policy(p)) in sched_dl_overflow()
2574 if (hrtimer_active(&p->dl.inactive_timer)) in sched_dl_overflow()
2575 __dl_sub(dl_b, p->dl.dl_bw, cpus); in sched_dl_overflow()
2579 !__dl_overflow(dl_b, cpus, p->dl.dl_bw, new_bw)) { in sched_dl_overflow()
2587 __dl_sub(dl_b, p->dl.dl_bw, cpus); in sched_dl_overflow()
2614 struct sched_dl_entity *dl_se = &p->dl; in __setparam_dl()
2626 struct sched_dl_entity *dl_se = &p->dl; in __getparam_dl()
2684 struct sched_dl_entity *dl_se = &p->dl; in __dl_clear_params()
2701 struct sched_dl_entity *dl_se = &p->dl; in dl_param_changed()
2727 overflow = __dl_overflow(dl_b, cpus, 0, p->dl.dl_bw); in dl_task_can_attach()
2737 __dl_add(dl_b, p->dl.dl_bw, cpus); in dl_task_can_attach()
2789 print_dl_rq(m, cpu, &cpu_rq(cpu)->dl); in print_dl_stats()