/Linux-v5.10/drivers/gpu/drm/scheduler/ |
D | sched_main.c | 72 static void drm_sched_rq_init(struct drm_gpu_scheduler *sched, in drm_sched_rq_init() argument 78 rq->sched = sched; in drm_sched_rq_init() 95 atomic_inc(&rq->sched->score); in drm_sched_rq_add_entity() 114 atomic_dec(&rq->sched->score); in drm_sched_rq_remove_entity() 176 struct drm_gpu_scheduler *sched = entity->rq->sched; in drm_sched_dependency_optimized() local 184 if (s_fence && s_fence->sched == sched) in drm_sched_dependency_optimized() 198 static void drm_sched_start_timeout(struct drm_gpu_scheduler *sched) in drm_sched_start_timeout() argument 200 if (sched->timeout != MAX_SCHEDULE_TIMEOUT && in drm_sched_start_timeout() 201 !list_empty(&sched->ring_mirror_list)) in drm_sched_start_timeout() 202 schedule_delayed_work(&sched->work_tdr, sched->timeout); in drm_sched_start_timeout() [all …]
|
D | sched_entity.c | 154 struct drm_gpu_scheduler *sched; in drm_sched_entity_flush() local 161 sched = entity->rq->sched; in drm_sched_entity_flush() 169 sched->job_scheduled, in drm_sched_entity_flush() 173 wait_event_killable(sched->job_scheduled, in drm_sched_entity_flush() 207 job->sched->ops->free_job(job); in drm_sched_entity_kill_jobs_cb() 260 struct drm_gpu_scheduler *sched = NULL; in drm_sched_entity_fini() local 263 sched = entity->rq->sched; in drm_sched_entity_fini() 271 if (sched) { in drm_sched_entity_fini() 332 drm_sched_wakeup(entity->rq->sched); in drm_sched_entity_wakeup() 362 struct drm_gpu_scheduler *sched = entity->rq->sched; in drm_sched_entity_add_dependency_cb() local [all …]
|
/Linux-v5.10/tools/perf/ |
D | builtin-sched.c | 146 int (*switch_event)(struct perf_sched *sched, struct evsel *evsel, 149 int (*runtime_event)(struct perf_sched *sched, struct evsel *evsel, 152 int (*wakeup_event)(struct perf_sched *sched, struct evsel *evsel, 156 int (*fork_event)(struct perf_sched *sched, union perf_event *event, 159 int (*migrate_task_event)(struct perf_sched *sched, 304 static void burn_nsecs(struct perf_sched *sched, u64 nsecs) in burn_nsecs() argument 310 } while (T1 + sched->run_measurement_overhead < T0 + nsecs); in burn_nsecs() 323 static void calibrate_run_measurement_overhead(struct perf_sched *sched) in calibrate_run_measurement_overhead() argument 330 burn_nsecs(sched, 0); in calibrate_run_measurement_overhead() 335 sched->run_measurement_overhead = min_delta; in calibrate_run_measurement_overhead() [all …]
|
/Linux-v5.10/tools/testing/selftests/ftrace/test.d/trigger/ |
D | trigger-filter.tc | 14 echo 'traceoff if child_pid == 0' > events/sched/sched_process_fork/trigger 23 ! echo 'traceoff if a' > events/sched/sched_process_fork/trigger 24 ! echo 'traceoff if common_pid=0' > events/sched/sched_process_fork/trigger 25 ! echo 'traceoff if common_pid==b' > events/sched/sched_process_fork/trigger 26 echo 'traceoff if common_pid == 0' > events/sched/sched_process_fork/trigger 27 echo '!traceoff' > events/sched/sched_process_fork/trigger 28 ! echo 'traceoff if common_pid == child_pid' > events/sched/sched_process_fork/trigger 29 echo 'traceoff if common_pid <= 0' > events/sched/sched_process_fork/trigger 30 echo '!traceoff' > events/sched/sched_process_fork/trigger 31 echo 'traceoff if common_pid >= 0' > events/sched/sched_process_fork/trigger [all …]
|
D | trigger-eventonoff.tc | 12 FEATURE=`grep enable_event events/sched/sched_process_fork/trigger` 19 echo 0 > events/sched/sched_switch/enable 20 echo 'enable_event:sched:sched_switch' > events/sched/sched_process_fork/trigger 22 if [ `cat events/sched/sched_switch/enable` != '1*' ]; then 29 echo 1 > events/sched/sched_switch/enable 30 echo 'disable_event:sched:sched_switch' > events/sched/sched_process_fork/trigger 32 if [ `cat events/sched/sched_switch/enable` != '0*' ]; then 39 ! echo 'enable_event:nogroup:noevent' > events/sched/sched_process_fork/trigger 40 ! echo 'disable_event+1' > events/sched/sched_process_fork/trigger 41 echo 'enable_event:sched:sched_switch' > events/sched/sched_process_fork/trigger [all …]
|
D | trigger-multihist.tc | 14 echo 'hist:keys=parent_pid:vals=child_pid' > events/sched/sched_process_fork/trigger 15 echo 'hist:keys=parent_comm:vals=child_pid' >> events/sched/sched_process_fork/trigger 17 grep parent_pid events/sched/sched_process_fork/hist > /dev/null || \ 19 grep child events/sched/sched_process_fork/hist > /dev/null || \ 22 grep "parent_comm: $COMM" events/sched/sched_process_fork/hist > /dev/null || \ 29 echo 'hist:name=test_hist:keys=common_pid' > events/sched/sched_process_fork/trigger 31 grep test_hist events/sched/sched_process_fork/hist > /dev/null || \ 36 echo 'hist:name=test_hist:keys=common_pid' > events/sched/sched_process_exit/trigger 38 grep test_hist events/sched/sched_process_exit/hist > /dev/null || \ 41 diffs=`diff events/sched/sched_process_exit/hist events/sched/sched_process_fork/hist | wc -l`
|
D | trigger-hist.tc | 14 echo 'hist:keys=parent_pid:vals=child_pid' > events/sched/sched_process_fork/trigger 16 grep parent_pid events/sched/sched_process_fork/hist > /dev/null || \ 18 grep child events/sched/sched_process_fork/hist > /dev/null || \ 25 echo 'hist:keys=parent_pid,child_pid' > events/sched/sched_process_fork/trigger 27 grep '^{ parent_pid:.*, child_pid:.*}' events/sched/sched_process_fork/hist > /dev/null || \ 34 echo 'hist:keys=parent_comm' > events/sched/sched_process_fork/trigger 37 grep "parent_comm: $COMM" events/sched/sched_process_fork/hist > /dev/null || \ 44 echo 'hist:keys=parent_pid,child_pid:sort=child_pid.ascending' > events/sched/sched_process_fork/tr… 55 events/sched/sched_process_fork/hist | cut -d: -f2 ` ||
|
D | trigger-traceonoff.tc | 13 echo 'traceoff' > events/sched/sched_process_fork/trigger 23 echo 'traceon' > events/sched/sched_process_fork/trigger 32 ! echo 'traceoff:badparam' > events/sched/sched_process_fork/trigger 33 ! echo 'traceoff+0' > events/sched/sched_process_fork/trigger 34 echo 'traceon' > events/sched/sched_process_fork/trigger 35 ! echo 'traceon' > events/sched/sched_process_fork/trigger 36 ! echo 'traceoff' > events/sched/sched_process_fork/trigger
|
D | trigger-snapshot.tc | 11 FEATURE=`grep snapshot events/sched/sched_process_fork/trigger` 19 echo 1 > events/sched/sched_process_fork/enable 21 echo 'snapshot:1' > events/sched/sched_process_fork/trigger 28 echo 0 > events/sched/sched_process_fork/enable 32 ! echo "snapshot+1" > events/sched/sched_process_fork/trigger 33 echo "snapshot" > events/sched/sched_process_fork/trigger 34 ! echo "snapshot" > events/sched/sched_process_fork/trigger
|
D | trigger-stacktrace.tc | 11 FEATURE=`grep stacktrace events/sched/sched_process_fork/trigger` 20 echo 'stacktrace' > events/sched/sched_process_fork/trigger 29 ! echo "stacktrace:foo" > events/sched/sched_process_fork/trigger 30 echo "stacktrace" > events/sched/sched_process_fork/trigger 31 ! echo "stacktrace" > events/sched/sched_process_fork/trigger
|
/Linux-v5.10/drivers/slimbus/ |
D | sched.c | 29 struct slim_sched *sched = &ctrl->sched; in slim_ctrl_clk_pause() local 38 mutex_lock(&sched->m_reconf); in slim_ctrl_clk_pause() 40 if (sched->clk_state == SLIM_CLK_ACTIVE) { in slim_ctrl_clk_pause() 41 mutex_unlock(&sched->m_reconf); in slim_ctrl_clk_pause() 49 ret = wait_for_completion_timeout(&sched->pause_comp, in slim_ctrl_clk_pause() 52 mutex_unlock(&sched->m_reconf); in slim_ctrl_clk_pause() 63 if (sched->clk_state == SLIM_CLK_PAUSED && ctrl->wakeup) in slim_ctrl_clk_pause() 66 sched->clk_state = SLIM_CLK_ACTIVE; in slim_ctrl_clk_pause() 67 mutex_unlock(&sched->m_reconf); in slim_ctrl_clk_pause() 73 if (ctrl->sched.clk_state == SLIM_CLK_PAUSED) { in slim_ctrl_clk_pause() [all …]
|
/Linux-v5.10/net/netfilter/ipvs/ |
D | ip_vs_sched.c | 61 struct ip_vs_scheduler *sched) in ip_vs_unbind_scheduler() argument 70 if (sched->done_service) in ip_vs_unbind_scheduler() 71 sched->done_service(svc); in ip_vs_unbind_scheduler() 81 struct ip_vs_scheduler *sched; in ip_vs_sched_getbyname() local 87 list_for_each_entry(sched, &ip_vs_schedulers, n_list) { in ip_vs_sched_getbyname() 91 if (sched->module && !try_module_get(sched->module)) { in ip_vs_sched_getbyname() 97 if (strcmp(sched_name, sched->name)==0) { in ip_vs_sched_getbyname() 100 return sched; in ip_vs_sched_getbyname() 102 module_put(sched->module); in ip_vs_sched_getbyname() 115 struct ip_vs_scheduler *sched; in ip_vs_scheduler_get() local [all …]
|
/Linux-v5.10/Documentation/scheduler/ |
D | index.rst | 10 sched-arch 11 sched-bwc 12 sched-deadline 13 sched-design-CFS 14 sched-domains 15 sched-capacity 16 sched-energy 17 sched-nice-design 18 sched-rt-group 19 sched-stats
|
/Linux-v5.10/crypto/ |
D | fcrypt.c | 54 __be32 sched[ROUNDS]; member 226 #define F_ENCRYPT(R, L, sched) \ argument 229 u.l = sched ^ R; \ 245 F_ENCRYPT(X.r, X.l, ctx->sched[0x0]); in fcrypt_encrypt() 246 F_ENCRYPT(X.l, X.r, ctx->sched[0x1]); in fcrypt_encrypt() 247 F_ENCRYPT(X.r, X.l, ctx->sched[0x2]); in fcrypt_encrypt() 248 F_ENCRYPT(X.l, X.r, ctx->sched[0x3]); in fcrypt_encrypt() 249 F_ENCRYPT(X.r, X.l, ctx->sched[0x4]); in fcrypt_encrypt() 250 F_ENCRYPT(X.l, X.r, ctx->sched[0x5]); in fcrypt_encrypt() 251 F_ENCRYPT(X.r, X.l, ctx->sched[0x6]); in fcrypt_encrypt() [all …]
|
/Linux-v5.10/drivers/net/wireless/ath/ath9k/ |
D | channel.c | 261 if (likely(sc->sched.channel_switch_time)) in ath_chanctx_check_active() 263 usecs_to_jiffies(sc->sched.channel_switch_time); in ath_chanctx_check_active() 311 ictx->flush_timeout = usecs_to_jiffies(sc->sched.channel_switch_time); in ath_chanctx_check_active() 382 mod_timer(&sc->sched.timer, jiffies + timeout); in ath_chanctx_setup_timer() 399 if (ctx->active && sc->sched.extend_absence) { in ath_chanctx_handle_bmiss() 401 sc->sched.extend_absence = false; in ath_chanctx_handle_bmiss() 408 if (ctx->active && sc->sched.beacon_miss >= 2) { in ath_chanctx_handle_bmiss() 410 sc->sched.extend_absence = true; in ath_chanctx_handle_bmiss() 423 avp->offchannel_duration = sc->sched.offchannel_duration; in ath_chanctx_offchannel_noa() 451 if (sc->sched.extend_absence) in ath_chanctx_set_periodic_noa() [all …]
|
/Linux-v5.10/include/drm/ |
D | gpu_scheduler.h | 120 struct drm_gpu_scheduler *sched; member 157 struct drm_gpu_scheduler *sched; member 192 struct drm_gpu_scheduler *sched; member 293 int drm_sched_init(struct drm_gpu_scheduler *sched, 298 void drm_sched_fini(struct drm_gpu_scheduler *sched); 307 void drm_sched_wakeup(struct drm_gpu_scheduler *sched); 308 void drm_sched_stop(struct drm_gpu_scheduler *sched, struct drm_sched_job *bad); 309 void drm_sched_start(struct drm_gpu_scheduler *sched, bool full_recovery); 310 void drm_sched_resubmit_jobs(struct drm_gpu_scheduler *sched); 314 void drm_sched_fault(struct drm_gpu_scheduler *sched); [all …]
|
/Linux-v5.10/tools/perf/Documentation/ |
D | perf-sched.txt | 1 perf-sched(1) 6 perf-sched - Tool to trace/measure scheduler properties (latencies) 11 'perf sched' {record|latency|map|replay|script|timehist} 15 There are several variants of 'perf sched': 17 'perf sched record <command>' to record the scheduling events 20 'perf sched latency' to report the per task scheduling latencies 23 'perf sched script' to see a detailed trace of the workload that 26 'perf sched replay' to simulate the workload that was recorded 27 via perf sched record. (this is done by starting up mockup threads 33 'perf sched map' to print a textual context-switching outline of [all …]
|
/Linux-v5.10/net/sctp/ |
D | stream_sched.c | 116 void sctp_sched_ops_register(enum sctp_sched_type sched, in sctp_sched_ops_register() argument 119 sctp_sched_ops[sched] = sched_ops; in sctp_sched_ops_register() 130 enum sctp_sched_type sched) in sctp_sched_set_sched() argument 132 struct sctp_sched_ops *n = sctp_sched_ops[sched]; in sctp_sched_set_sched() 133 struct sctp_sched_ops *old = asoc->outqueue.sched; in sctp_sched_set_sched() 141 if (sched > SCTP_SS_MAX) in sctp_sched_set_sched() 160 asoc->outqueue.sched = n; in sctp_sched_set_sched() 183 asoc->outqueue.sched = &sctp_sched_fcfs; /* Always safe */ in sctp_sched_set_sched() 193 if (asoc->outqueue.sched == sctp_sched_ops[i]) in sctp_sched_get_sched() 213 return asoc->outqueue.sched->set(&asoc->stream, sid, value, gfp); in sctp_sched_set_value() [all …]
|
/Linux-v5.10/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_job.c | 33 struct amdgpu_ring *ring = to_amdgpu_ring(s_job->sched); in amdgpu_job_timedout() 43 s_job->sched->name); in amdgpu_job_timedout() 49 job->base.sched->name, atomic_read(&ring->fence_drv.last_seq), in amdgpu_job_timedout() 57 drm_sched_suspend_timeout(&ring->sched); in amdgpu_job_timedout() 81 (*job)->base.sched = &adev->rings[0]->sched; in amdgpu_job_alloc() 113 struct amdgpu_ring *ring = to_amdgpu_ring(job->base.sched); in amdgpu_job_free_resources() 170 job->base.sched = &ring->sched; in amdgpu_job_submit_direct() 183 struct amdgpu_ring *ring = to_amdgpu_ring(s_entity->rq->sched); in amdgpu_job_dependency() 211 struct amdgpu_ring *ring = to_amdgpu_ring(sched_job->sched); in amdgpu_job_run() 247 void amdgpu_job_stop_all_jobs_on_sched(struct drm_gpu_scheduler *sched) in amdgpu_job_stop_all_jobs_on_sched() argument [all …]
|
/Linux-v5.10/tools/testing/selftests/ftrace/test.d/trigger/inter-event/ |
D | trigger-field-variable-support.tc | 14 echo 'hist:keys=comm:ts0=common_timestamp.usecs if comm=="ping"' > events/sched/sched_waking/trigger 15 …ext_pid,sched.sched_waking.prio,next_comm) if next_comm=="ping"' > events/sched/sched_switch/trigg… 23 if ! grep -q "synthetic_prio=prio" events/sched/sched_waking/hist; then 27 …xt_pid,sched.sched_waking.prio,next_comm) if next_comm=="ping"' >> events/sched/sched_switch/trigg… 29 if grep -q "synthetic_prio=prio" events/sched/sched_waking/hist; then
|
D | trigger-inter-event-combined-hist.tc | 20 echo 'hist:keys=pid:ts0=common_timestamp.usecs if comm=="ping"' > events/sched/sched_waking/trigger 21 …d.sched_waking).waking_latency($waking_lat,pid) if comm=="ping"' > events/sched/sched_wakeup/trigg… 25 echo 'hist:keys=pid:ts1=common_timestamp.usecs if comm=="ping"' >> events/sched/sched_wakeup/trigger 26 …keup).wakeup_latency($wakeup_lat,next_pid) if next_comm=="ping"' > events/sched/sched_switch/trigg…
|
/Linux-v5.10/tools/perf/scripts/python/bin/ |
D | sched-migration-record | 2 perf record -m 16384 -e sched:sched_wakeup -e sched:sched_wakeup_new -e sched:sched_switch -e sched…
|
/Linux-v5.10/arch/x86/events/ |
D | core.c | 756 static void perf_sched_init(struct perf_sched *sched, struct event_constraint **constraints, in perf_sched_init() argument 761 memset(sched, 0, sizeof(*sched)); in perf_sched_init() 762 sched->max_events = num; in perf_sched_init() 763 sched->max_weight = wmax; in perf_sched_init() 764 sched->max_gp = gpmax; in perf_sched_init() 765 sched->constraints = constraints; in perf_sched_init() 772 sched->state.event = idx; /* start with min weight */ in perf_sched_init() 773 sched->state.weight = wmin; in perf_sched_init() 774 sched->state.unassigned = num; in perf_sched_init() 777 static void perf_sched_save_state(struct perf_sched *sched) in perf_sched_save_state() argument [all …]
|
/Linux-v5.10/net/sched/ |
D | sch_taprio.c | 89 static ktime_t sched_base_time(const struct sched_gate_list *sched) in sched_base_time() argument 91 if (!sched) in sched_base_time() 94 return ns_to_ktime(sched->base_time); in sched_base_time() 113 struct sched_gate_list *sched = container_of(head, struct sched_gate_list, rcu); in taprio_free_sched_cb() local 116 if (!sched) in taprio_free_sched_cb() 119 list_for_each_entry_safe(entry, n, &sched->entries, list) { in taprio_free_sched_cb() 124 kfree(sched); in taprio_free_sched_cb() 142 static s32 get_cycle_time_elapsed(struct sched_gate_list *sched, ktime_t time) in get_cycle_time_elapsed() argument 147 time_since_sched_start = ktime_sub(time, sched->base_time); in get_cycle_time_elapsed() 148 div_s64_rem(time_since_sched_start, sched->cycle_time, &time_elapsed); in get_cycle_time_elapsed() [all …]
|
/Linux-v5.10/drivers/net/wireless/mediatek/mt76/mt7615/ |
D | sdio_txrx.c | 49 mutex_lock(&sdio->sched.lock); in mt7663s_refill_sched_quota() 50 sdio->sched.pse_mcu_quota += pse_mcu_quota; in mt7663s_refill_sched_quota() 51 sdio->sched.pse_data_quota += pse_data_quota; in mt7663s_refill_sched_quota() 52 sdio->sched.ple_data_quota += ple_data_quota; in mt7663s_refill_sched_quota() 53 mutex_unlock(&sdio->sched.lock); in mt7663s_refill_sched_quota() 146 pse_sz = DIV_ROUND_UP(buf_sz + sdio->sched.deficit, MT_PSE_PAGE_SZ); in mt7663s_tx_pick_quota() 149 if (sdio->sched.pse_mcu_quota < *pse_size + pse_sz) in mt7663s_tx_pick_quota() 152 if (sdio->sched.pse_data_quota < *pse_size + pse_sz || in mt7663s_tx_pick_quota() 153 sdio->sched.ple_data_quota < *ple_size) in mt7663s_tx_pick_quota() 166 mutex_lock(&sdio->sched.lock); in mt7663s_tx_update_quota() [all …]
|