| /kernel/linux/linux-5.10/include/drm/ |
| D | gpu_scheduler.h | 55 * @rq: runqueue on which this entity is currently scheduled. 57 * Jobs from this entity can be scheduled on any scheduler 67 * The &drm_sched_fence.scheduled uses the 75 * @last_scheduled: points to the finished fence of the last scheduled job. 107 * struct drm_sched_rq - queue of entities to be scheduled. 111 * @entities: list of the entities to be scheduled. 112 * @current_entity: the entity which is to be scheduled. 130 * @scheduled: this fence is what will be signaled by the scheduler 131 * when the job is scheduled. 133 struct dma_fence scheduled; member [all …]
|
| /kernel/linux/linux-5.10/drivers/gpu/drm/scheduler/ |
| D | sched_fence.c | 53 int ret = dma_fence_signal(&fence->scheduled); in drm_sched_fence_scheduled() 56 DMA_FENCE_TRACE(&fence->scheduled, in drm_sched_fence_scheduled() 59 DMA_FENCE_TRACE(&fence->scheduled, in drm_sched_fence_scheduled() 122 * Drop the extra reference from the scheduled fence to the base fence. 128 dma_fence_put(&fence->scheduled); in drm_sched_fence_release_finished() 146 return container_of(f, struct drm_sched_fence, scheduled); in to_drm_sched_fence() 170 dma_fence_init(&fence->scheduled, &drm_sched_fence_ops_scheduled, in drm_sched_fence_create()
|
| /kernel/linux/linux-6.6/include/drm/ |
| D | gpu_scheduler.h | 102 * Runqueue on which this entity is currently scheduled. 114 * be scheduled on any scheduler on this list. 170 * &drm_sched_fence.scheduled uses the fence_context but 199 * Points to the finished fence of the last scheduled job. Only written 244 * struct drm_sched_rq - queue of entities to be scheduled. 248 * @entities: list of the entities to be scheduled. 249 * @current_entity: the entity which is to be scheduled. 269 * @scheduled: this fence is what will be signaled by the scheduler 270 * when the job is scheduled. 272 struct dma_fence scheduled; member [all …]
|
| /kernel/linux/linux-6.6/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
| D | pno.h | 17 * brcmf_pno_start_sched_scan - initiate scheduled scan on device. 20 * @req: configuration parameters for scheduled scan. 26 * brcmf_pno_stop_sched_scan - terminate scheduled scan on device. 34 * brcmf_pno_wiphy_params - fill scheduled scan parameters in wiphy instance.
|
| /kernel/linux/linux-5.10/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
| D | pno.h | 17 * brcmf_pno_start_sched_scan - initiate scheduled scan on device. 20 * @req: configuration parameters for scheduled scan. 26 * brcmf_pno_stop_sched_scan - terminate scheduled scan on device. 34 * brcmf_pno_wiphy_params - fill scheduled scan parameters in wiphy instance.
|
| /kernel/linux/linux-6.6/drivers/gpu/drm/scheduler/ |
| D | sched_fence.c | 68 /* Set the parent before signaling the scheduled fence, such that, in drm_sched_fence_scheduled() 70 * been scheduled (which is the case for drivers delegating waits in drm_sched_fence_scheduled() 77 dma_fence_signal(&fence->scheduled); in drm_sched_fence_scheduled() 143 * Drop the extra reference from the scheduled fence to the base fence. 149 dma_fence_put(&fence->scheduled); in drm_sched_fence_release_finished() 199 return container_of(f, struct drm_sched_fence, scheduled); in to_drm_sched_fence() 230 dma_fence_init(&fence->scheduled, &drm_sched_fence_ops_scheduled, in drm_sched_fence_init()
|
| D | sched_entity.c | 146 * drm_sched_entity_error - return error of last scheduled job 149 * Opportunistically return the error of the last scheduled job. Result can 190 if (s_fence && f == &s_fence->scheduled) { in drm_sched_entity_kill_jobs_cb() 191 /* The dependencies array had a reference on the scheduled in drm_sched_entity_kill_jobs_cb() 200 * had on the scheduled fence. in drm_sched_entity_kill_jobs_cb() 202 dma_fence_put(&s_fence->scheduled); in drm_sched_entity_kill_jobs_cb() 394 * Fence is a scheduled/finished fence from a job in drm_sched_entity_add_dependency_cb() 408 * it to be scheduled in drm_sched_entity_add_dependency_cb() 410 fence = dma_fence_get(&s_fence->scheduled); in drm_sched_entity_add_dependency_cb()
|
| /kernel/linux/linux-6.6/net/mptcp/ |
| D | sched.c | 119 bool scheduled) in mptcp_subflow_set_scheduled() argument 121 WRITE_ONCE(subflow->scheduled, scheduled); in mptcp_subflow_set_scheduled() 143 if (READ_ONCE(subflow->scheduled)) in mptcp_sched_get_send() 165 if (READ_ONCE(subflow->scheduled)) in mptcp_sched_get_retrans()
|
| /kernel/linux/linux-5.10/Documentation/powerpc/ |
| D | pmu-ebb.rst | 44 user process. This means once an EBB event is scheduled on the PMU, no non-EBB 56 first will be scheduled and the other will be put in error state. See the 84 userspace is able to reliably determine which PMC the event is scheduled on. 95 guarantee that it has been scheduled on the PMU. To ensure that the EBB event 96 has been scheduled on the PMU, you must perform a read() on the event. If the 97 read() returns EOF, then the event has not been scheduled and EBBs are not
|
| /kernel/linux/linux-6.6/Documentation/powerpc/ |
| D | pmu-ebb.rst | 44 user process. This means once an EBB event is scheduled on the PMU, no non-EBB 56 first will be scheduled and the other will be put in error state. See the 84 userspace is able to reliably determine which PMC the event is scheduled on. 95 guarantee that it has been scheduled on the PMU. To ensure that the EBB event 96 has been scheduled on the PMU, you must perform a read() on the event. If the 97 read() returns EOF, then the event has not been scheduled and EBBs are not
|
| /kernel/linux/linux-5.10/drivers/gpu/drm/ |
| D | drm_vblank_work.c | 99 * If @work is already scheduled, this function will reschedule said work 103 * %1 if @work was successfully (re)scheduled, %0 if it was either already 104 * scheduled or cancelled, or a negative error code on failure. 131 /* Already scheduled w/ same vbl count */ in drm_vblank_work_schedule() 172 * Cancel an already scheduled vblank work and wait for its 175 * On return, @work is guaranteed to no longer be scheduled or running, even 212 * drm_vblank_work_flush - wait for a scheduled vblank work to finish
|
| /kernel/linux/linux-6.6/drivers/gpu/drm/ |
| D | drm_vblank_work.c | 99 * If @work is already scheduled, this function will reschedule said work 103 * %1 if @work was successfully (re)scheduled, %0 if it was either already 104 * scheduled or cancelled, or a negative error code on failure. 131 /* Already scheduled w/ same vbl count */ in drm_vblank_work_schedule() 172 * Cancel an already scheduled vblank work and wait for its 175 * On return, @work is guaranteed to no longer be scheduled or running, even 212 * drm_vblank_work_flush - wait for a scheduled vblank work to finish
|
| /kernel/linux/linux-6.6/net/sctp/ |
| D | stream_sched_prio.c | 64 /* Look into scheduled priorities first, as they are sorted and in sctp_sched_prio_get_head() 65 * we can find it fast IF it's scheduled. in sctp_sched_prio_get_head() 105 bool scheduled = false; in sctp_sched_prio_unsched() local 110 /* Scheduled */ in sctp_sched_prio_unsched() 111 scheduled = true; in sctp_sched_prio_unsched() 127 return scheduled; in sctp_sched_prio_unsched() 137 /* Nothing to do if already scheduled */ in sctp_sched_prio_sched()
|
| /kernel/linux/linux-5.10/net/sctp/ |
| D | stream_sched_prio.c | 64 /* Look into scheduled priorities first, as they are sorted and in sctp_sched_prio_get_head() 65 * we can find it fast IF it's scheduled. in sctp_sched_prio_get_head() 105 bool scheduled = false; in sctp_sched_prio_unsched() local 110 /* Scheduled */ in sctp_sched_prio_unsched() 111 scheduled = true; in sctp_sched_prio_unsched() 127 return scheduled; in sctp_sched_prio_unsched() 137 /* Nothing to do if already scheduled */ in sctp_sched_prio_sched()
|
| /kernel/linux/linux-6.6/drivers/net/wireless/intel/iwlwifi/fw/api/ |
| D | time-event.h | 50 * the first fragment is scheduled. 52 * the first 2 fragments are scheduled. 58 * scheduled. 116 * the first fragment is scheduled. 118 * the first 2 fragments are scheduled. 124 * scheduled. 253 * @status: true if scheduled, false otherwise (not executed) 387 * Note: the session protection will always be scheduled to start as
|
| /kernel/linux/linux-5.10/drivers/usb/host/ |
| D | xhci-mtk.h | 54 * (@repeat==1) scheduled within the interval 64 * scheduled first time within the interval 66 * scheduled within a interval. in the simple algorithm, only 70 * @pkts: number of packets to be transferred in the scheduled uframes
|
| /kernel/linux/linux-6.6/drivers/usb/host/ |
| D | xhci-mtk.h | 60 * (@repeat==1) scheduled within the interval 72 * scheduled first time within the interval 74 * scheduled within a interval. in the simple algorithm, only 78 * @pkts: number of packets to be transferred in the scheduled uframes
|
| /kernel/linux/linux-5.10/drivers/net/wireless/intel/iwlwifi/fw/api/ |
| D | time-event.h | 108 * the first fragment is scheduled. 110 * the first 2 fragments are scheduled. 116 * scheduled. 174 * the first fragment is scheduled. 176 * the first 2 fragments are scheduled. 182 * scheduled. 311 * @status: true if scheduled, false otherwise (not executed) 444 * Note: the session protection will always be scheduled to start as
|
| /kernel/linux/linux-6.6/drivers/staging/media/deprecated/atmel/ |
| D | Kconfig | 21 This driver is deprecated and is scheduled for removal by 40 This driver is deprecated and is scheduled for removal by
|
| /kernel/linux/linux-6.6/drivers/infiniband/sw/rxe/ |
| D | rxe_task.c | 25 /* Check if task is idle i.e. not running, not scheduled in 54 * scheduled in the work queue. This routine is 170 "%ld tasks scheduled, %ld tasks done\n", in do_task() 226 /* now the task cannot be scheduled or run just wait in rxe_cleanup_task() 227 * for the previously scheduled tasks to finish. in rxe_cleanup_task()
|
| /kernel/linux/linux-6.6/drivers/gpu/drm/i915/ |
| D | i915_priolist_types.h | 23 /* Interactive workload, scheduled for immediate pageflipping */ 32 * another context. They get scheduled with their default priority and
|
| /kernel/linux/linux-6.6/drivers/soc/fsl/dpio/ |
| D | qbman-portal.h | 398 * qbman_swp_fq_schedule() - Move the fq to the scheduled state 400 * @fqid: the index of frame queue to be scheduled 413 * qbman_swp_fq_force() - Force the FQ to fully scheduled state 417 * Force eligible will force a tentatively-scheduled FQ to be fully-scheduled 450 * XOFF FQs will remain in the tenatively-scheduled state, even when 451 * non-empty, meaning they won't be selected for scheduled dequeuing. 452 * If a FQ is changed to XOFF after it had already become truly-scheduled
|
| /kernel/linux/linux-5.10/drivers/soc/fsl/dpio/ |
| D | qbman-portal.h | 391 * qbman_swp_fq_schedule() - Move the fq to the scheduled state 393 * @fqid: the index of frame queue to be scheduled 406 * qbman_swp_fq_force() - Force the FQ to fully scheduled state 410 * Force eligible will force a tentatively-scheduled FQ to be fully-scheduled 443 * XOFF FQs will remain in the tenatively-scheduled state, even when 444 * non-empty, meaning they won't be selected for scheduled dequeuing. 445 * If a FQ is changed to XOFF after it had already become truly-scheduled
|
| /kernel/linux/linux-5.10/drivers/gpu/drm/i915/ |
| D | i915_priolist_types.h | 23 /* Interactive workload, scheduled for immediate pageflipping */ 38 * another context. They get scheduled with their default priority and
|
| /kernel/linux/linux-5.10/include/linux/ |
| D | posix-timers.h | 134 * @work: The task work to be scheduled 136 * @scheduled: @work has been scheduled already, no further processing 141 unsigned int scheduled; member
|