/Linux-v6.1/include/drm/ |
D | gpu_scheduler.h | 84 * Runqueue on which this entity is currently scheduled. 96 * be scheduled on any scheduler on this list. 152 * &drm_sched_fence.scheduled uses the fence_context but 181 * Points to the finished fence of the last scheduled job. Only written 211 * struct drm_sched_rq - queue of entities to be scheduled. 215 * @entities: list of the entities to be scheduled. 216 * @current_entity: the entity which is to be scheduled. 234 * @scheduled: this fence is what will be signaled by the scheduler 235 * when the job is scheduled. 237 struct dma_fence scheduled; member [all …]
|
/Linux-v6.1/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
D | pno.h | 17 * brcmf_pno_start_sched_scan - initiate scheduled scan on device. 20 * @req: configuration parameters for scheduled scan. 26 * brcmf_pno_stop_sched_scan - terminate scheduled scan on device. 34 * brcmf_pno_wiphy_params - fill scheduled scan parameters in wiphy instance.
|
/Linux-v6.1/Documentation/powerpc/ |
D | pmu-ebb.rst | 44 user process. This means once an EBB event is scheduled on the PMU, no non-EBB 56 first will be scheduled and the other will be put in error state. See the 84 userspace is able to reliably determine which PMC the event is scheduled on. 95 guarantee that it has been scheduled on the PMU. To ensure that the EBB event 96 has been scheduled on the PMU, you must perform a read() on the event. If the 97 read() returns EOF, then the event has not been scheduled and EBBs are not
|
/Linux-v6.1/drivers/gpu/drm/scheduler/ |
D | sched_fence.c | 53 dma_fence_signal(&fence->scheduled); in drm_sched_fence_scheduled() 117 * Drop the extra reference from the scheduled fence to the base fence. 123 dma_fence_put(&fence->scheduled); in drm_sched_fence_release_finished() 141 return container_of(f, struct drm_sched_fence, scheduled); in to_drm_sched_fence() 172 dma_fence_init(&fence->scheduled, &drm_sched_fence_ops_scheduled, in drm_sched_fence_init()
|
D | sched_main.c | 28 * into software queues which are then scheduled on a hardware run queue. 41 * 4. Entities themselves maintain a queue of jobs that will be scheduled on 44 * The jobs in a entity are always scheduled in the order that they were pushed. 365 * jobs from it will not be scheduled further 547 guilty_context = s_job->s_fence->scheduled.context; in drm_sched_resubmit_jobs_ext() 550 if (found_guilty && s_job->s_fence->scheduled.context == guilty_context) in drm_sched_resubmit_jobs_ext() 838 /* make the scheduled timestamp more accurate */ in drm_sched_get_cleanup_job() 843 next->s_fence->scheduled.timestamp = in drm_sched_get_cleanup_job() 1114 if (bad->s_fence->scheduled.context == in drm_sched_increase_karma_ext()
|
/Linux-v6.1/drivers/gpu/drm/ |
D | drm_vblank_work.c | 99 * If @work is already scheduled, this function will reschedule said work 103 * %1 if @work was successfully (re)scheduled, %0 if it was either already 104 * scheduled or cancelled, or a negative error code on failure. 131 /* Already scheduled w/ same vbl count */ in drm_vblank_work_schedule() 172 * Cancel an already scheduled vblank work and wait for its 175 * On return, @work is guaranteed to no longer be scheduled or running, even 212 * drm_vblank_work_flush - wait for a scheduled vblank work to finish
|
/Linux-v6.1/drivers/staging/media/deprecated/saa7146/av7110/ |
D | Kconfig | 38 This driver is deprecated and is scheduled for removal by 55 This driver is deprecated and is scheduled for removal by 77 This driver is deprecated and is scheduled for removal by 103 This driver is deprecated and is scheduled for removal by
|
/Linux-v6.1/drivers/staging/media/deprecated/saa7146/saa7146/ |
D | Kconfig | 11 This driver is deprecated and is scheduled for removal by 25 This driver is deprecated and is scheduled for removal by 44 This driver is deprecated and is scheduled for removal by
|
/Linux-v6.1/drivers/staging/media/deprecated/vpfe_capture/ |
D | Kconfig | 16 This driver is deprecated and is scheduled for removal by 36 This driver is deprecated and is scheduled for removal by 54 This driver is deprecated and is scheduled for removal by
|
/Linux-v6.1/net/sctp/ |
D | stream_sched_prio.c | 51 /* Look into scheduled priorities first, as they are sorted and in sctp_sched_prio_get_head() 52 * we can find it fast IF it's scheduled. in sctp_sched_prio_get_head() 92 bool scheduled = false; in sctp_sched_prio_unsched() local 97 /* Scheduled */ in sctp_sched_prio_unsched() 98 scheduled = true; in sctp_sched_prio_unsched() 114 return scheduled; in sctp_sched_prio_unsched() 124 /* Nothing to do if already scheduled */ in sctp_sched_prio_sched()
|
/Linux-v6.1/drivers/net/wireless/intel/iwlwifi/fw/api/ |
D | time-event.h | 50 * the first fragment is scheduled. 52 * the first 2 fragments are scheduled. 58 * scheduled. 116 * the first fragment is scheduled. 118 * the first 2 fragments are scheduled. 124 * scheduled. 253 * @status: true if scheduled, false otherwise (not executed) 386 * Note: the session protection will always be scheduled to start as
|
/Linux-v6.1/drivers/staging/media/deprecated/saa7146/ttpci/ |
D | Kconfig | 33 This driver is deprecated and is scheduled for removal by 63 This driver is deprecated and is scheduled for removal by 89 This driver is deprecated and is scheduled for removal by
|
/Linux-v6.1/drivers/usb/host/ |
D | xhci-mtk.h | 60 * (@repeat==1) scheduled within the interval 72 * scheduled first time within the interval 74 * scheduled within a interval. in the simple algorithm, only 78 * @pkts: number of packets to be transferred in the scheduled uframes
|
/Linux-v6.1/include/linux/ |
D | posix-timers.h | 137 * @work: The task work to be scheduled 138 * @scheduled: @work has been scheduled already, no further processing 142 unsigned int scheduled; member
|
/Linux-v6.1/drivers/soc/fsl/dpio/ |
D | qbman-portal.h | 398 * qbman_swp_fq_schedule() - Move the fq to the scheduled state 400 * @fqid: the index of frame queue to be scheduled 413 * qbman_swp_fq_force() - Force the FQ to fully scheduled state 417 * Force eligible will force a tentatively-scheduled FQ to be fully-scheduled 450 * XOFF FQs will remain in the tenatively-scheduled state, even when 451 * non-empty, meaning they won't be selected for scheduled dequeuing. 452 * If a FQ is changed to XOFF after it had already become truly-scheduled
|
/Linux-v6.1/drivers/gpu/drm/i915/ |
D | i915_priolist_types.h | 23 /* Interactive workload, scheduled for immediate pageflipping */ 32 * another context. They get scheduled with their default priority and
|
D | i915_vma_resource.h | 39 * is not considered published until unbind is scheduled, and as such it 40 * is illegal to access this fence before scheduled unbind other than 45 * is scheduled. 67 * taken when the unbind is scheduled.
|
/Linux-v6.1/arch/alpha/kernel/ |
D | perf_event.c | 36 /* Number of events scheduled; also number entries valid in arrays below. */ 40 /* Events currently scheduled. */ 42 /* Event type of each scheduled event. */ 44 /* Current index of each scheduled event; if not yet determined 149 * Check that a group of events can be simultaneously scheduled on to the 369 * Check that a group of events can be simultaneously scheduled on to the PMU. 387 * If new events have been scheduled then update cpuc with the new 637 * scheduled on to the PMU. At that point the code to programme the in __hw_perf_event_init() 646 * be scheduled on to the PMU. in __hw_perf_event_init() 731 /* Update cpuc with information from any new scheduled events. */ in alpha_pmu_enable()
|
/Linux-v6.1/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_job.c | 182 struct dma_fence *fence = &leader->base.s_fence->scheduled; in amdgpu_job_set_gang_leader() 200 if (job->gang_submit != &job->base.s_fence->scheduled) in amdgpu_job_free() 322 /* Signal all jobs not yet scheduled */ in amdgpu_job_stop_all_jobs_on_sched() 330 dma_fence_signal(&s_fence->scheduled); in amdgpu_job_stop_all_jobs_on_sched() 338 /* Signal all jobs already scheduled to HW */ in amdgpu_job_stop_all_jobs_on_sched()
|
/Linux-v6.1/kernel/ |
D | workqueue_internal.h | 35 struct list_head scheduled; /* L: scheduled works */ member
|
/Linux-v6.1/arch/s390/pci/ |
D | pci_irq.c | 179 atomic_t scheduled; member 185 atomic_t *scheduled = data; in zpci_handle_remote_irq() local 189 } while (atomic_dec_return(scheduled)); in zpci_handle_remote_irq() 212 if (atomic_inc_return(&cpu_data->scheduled) > 1) in zpci_handle_fallback_irq() 215 INIT_CSD(&cpu_data->csd, zpci_handle_remote_irq, &cpu_data->scheduled); in zpci_handle_fallback_irq()
|
/Linux-v6.1/Documentation/admin-guide/hw-vuln/ |
D | core-scheduling.rst | 35 Using this feature, userspace defines groups of tasks that can be co-scheduled 122 tasks selected to be scheduled are of the highest priority in a core. However, 128 by the scheduler (idle thread is scheduled to run). 224 - Gang scheduling: Requirements for a group of tasks that needs to be scheduled
|
D | l1d_flush.rst | 35 performed when the task is scheduled out and the incoming task belongs to a 68 requested L1D flushing is scheduled on a SMT-enabled core the kernel sends
|
/Linux-v6.1/include/xen/interface/ |
D | vcpu.h | 82 /* VCPU is runnable, but not currently scheduled on any physical CPU. */ 103 * updated by the hypervisor each time the VCPU is scheduled. Thus 106 * VCPU was last scheduled to run.
|
/Linux-v6.1/drivers/usb/dwc2/ |
D | hcd_queue.c | 122 * @interval: How often we need to be scheduled for the reservation this 621 * already scheduled something on the low speed schedule and 931 /* If we scheduled all w/out breaking out then we're all good */ in dwc2_uframe_schedule_split() 1042 * dwc2_pick_first_frame() - Choose 1st frame for qh that's already scheduled 1044 * Takes a qh that has already been scheduled (which means we know we have the 1074 * NOTE: if we could quantify how long till we actually get scheduled in dwc2_pick_first_frame() 1093 * the first frame should be based on when the first scheduled in dwc2_pick_first_frame() 1132 * scheduled for. in dwc2_pick_first_frame() 1259 * Wait for the lock, or for us to be scheduled again. We in dwc2_unreserve_timer_fn() 1260 * could be scheduled again if: in dwc2_unreserve_timer_fn() [all …]
|