/linux/drivers/md/dm-vdo/ |
H A D | action-manager.c | 237 * If the manager is not operating normally, the action will not be scheduled. 239 * Return: true if an action was scheduled. 283 * @parent: The object to notify once the action is complete or if the action can not be scheduled; 287 * action completes. If there is already a pending action, this action will not be scheduled, and, 291 * Return: true if the action was scheduled. 310 * @parent: The object to notify once the action is complete or if the action can not be scheduled; 315 * scheduled, and, if it has a parent, that parent will be notified. At least one of the preamble, 318 * Return: true if the action was scheduled. 341 * @parent: The object to notify once the action is complete or if the action can not be scheduled; 346 * scheduled, and, if it has a parent, that parent will be notified. At least one of the preamble, [all …]
|
H A D | action-manager.h | 21 * A scheduled action consists of four components: 66 * Return: true if an action was scheduled
|
/linux/include/drm/ |
H A D | gpu_scheduler.h | 104 * Runqueue on which this entity is currently scheduled. 116 * be scheduled on any scheduler on this list. 165 * &drm_sched_fence.scheduled uses the fence_context but 194 * Points to the finished fence of the last scheduled job. Only written 239 * struct drm_sched_rq - queue of entities to be scheduled. 243 * @current_entity: the entity which is to be scheduled. 244 * @entities: list of the entities to be scheduled. 266 * @scheduled: this fence is what will be signaled by the scheduler 267 * when the job is scheduled. 269 struct dma_fence scheduled; member [all …]
|
/linux/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
H A D | pno.h | 17 * brcmf_pno_start_sched_scan - initiate scheduled scan on device. 20 * @req: configuration parameters for scheduled scan. 26 * brcmf_pno_stop_sched_scan - terminate scheduled scan on device. 34 * brcmf_pno_wiphy_params - fill scheduled scan parameters in wiphy instance.
|
/linux/Documentation/arch/powerpc/ |
H A D | pmu-ebb.rst | 44 user process. This means once an EBB event is scheduled on the PMU, no non-EBB 56 first will be scheduled and the other will be put in error state. See the 84 userspace is able to reliably determine which PMC the event is scheduled on. 95 guarantee that it has been scheduled on the PMU. To ensure that the EBB event 96 has been scheduled on the PMU, you must perform a read() on the event. If the 97 read() returns EOF, then the event has not been scheduled and EBBs are not
|
/linux/include/linux/ |
H A D | posix-timers_types.h | 64 * @work: The task work to be scheduled 66 * @scheduled: @work has been scheduled already, no further processing 71 unsigned int scheduled; member
|
/linux/net/mptcp/ |
H A D | sched.c | 163 bool scheduled) in mptcp_sched_get_send() 165 WRITE_ONCE(subflow->scheduled, scheduled); in mptcp_sched_get_send() 186 if (READ_ONCE(subflow->scheduled)) in mptcp_sched_get_retrans() 206 if (READ_ONCE(subflow->scheduled)) in mptcp_sched_get_retrans() 152 mptcp_subflow_set_scheduled(struct mptcp_subflow_context * subflow,bool scheduled) mptcp_subflow_set_scheduled() argument
|
/linux/net/sctp/ |
H A D | stream_sched_prio.c | 64 /* Look into scheduled priorities first, as they are sorted and in sctp_sched_prio_get_head() 65 * we can find it fast IF it's scheduled. in sctp_sched_prio_get_head() 105 bool scheduled = false; in sctp_sched_prio_unsched() local 110 /* Scheduled */ in sctp_sched_prio_unsched() 111 scheduled = true; in sctp_sched_prio_unsched() 127 return scheduled; in sctp_sched_prio_unsched() 137 /* Nothing to do if already scheduled */ in sctp_sched_prio_sched()
|
/linux/drivers/net/wireless/intel/iwlwifi/fw/api/ |
H A D | time-event.h | 50 * the first fragment is scheduled. 52 * the first 2 fragments are scheduled. 58 * scheduled. 116 * the first fragment is scheduled. 118 * the first 2 fragments are scheduled. 124 * scheduled. 253 * @status: true if scheduled, false otherwise (not executed) 480 * Note: the session protection will always be scheduled to start as
|
/linux/drivers/gpu/drm/scheduler/ |
H A D | sched_entity.c | 155 * drm_sched_entity_error - return error of last scheduled job 158 * Opportunistically return the error of the last scheduled job. Result can 199 if (s_fence && f == &s_fence->scheduled) { in drm_sched_entity_kill_jobs_cb() 200 /* The dependencies array had a reference on the scheduled in drm_sched_entity_kill_jobs_cb() 209 * had on the scheduled fence. in drm_sched_entity_kill_jobs_cb() 211 dma_fence_put(&s_fence->scheduled); in drm_sched_entity_kill_jobs_cb() 403 * Fence is a scheduled/finished fence from a job in drm_sched_entity_add_dependency_cb() 417 * it to be scheduled in drm_sched_entity_add_dependency_cb() 419 fence = dma_fence_get(&s_fence->scheduled); in drm_sched_entity_add_dependency_cb()
|
/linux/drivers/infiniband/sw/rxe/ |
H A D | rxe_task.c | 25 /* Check if task is idle i.e. not running, not scheduled in 54 * scheduled in the work queue. This routine is 170 "%ld tasks scheduled, %ld tasks done\n", in do_task() 226 /* now the task cannot be scheduled or run just wait in rxe_cleanup_task() 227 * for the previously scheduled tasks to finish. in rxe_cleanup_task()
|
/linux/drivers/usb/host/ |
H A D | xhci-mtk.h | 71 * (@repeat==1) scheduled within the interval 82 * scheduled first time within the interval 84 * scheduled within a interval. in the simple algorithm, only 88 * @pkts: number of packets to be transferred in the scheduled uframes
|
/linux/drivers/soc/fsl/dpio/ |
H A D | qbman-portal.h | 398 * qbman_swp_fq_schedule() - Move the fq to the scheduled state 400 * @fqid: the index of frame queue to be scheduled 413 * qbman_swp_fq_force() - Force the FQ to fully scheduled state 417 * Force eligible will force a tentatively-scheduled FQ to be fully-scheduled 450 * XOFF FQs will remain in the tenatively-scheduled state, even when 451 * non-empty, meaning they won't be selected for scheduled dequeuing. 452 * If a FQ is changed to XOFF after it had already become truly-scheduled
|
/linux/drivers/staging/media/deprecated/atmel/ |
H A D | Kconfig | 21 This driver is deprecated and is scheduled for removal by 40 This driver is deprecated and is scheduled for removal by
|
/linux/arch/x86/kvm/vmx/ |
H A D | posted_intr.c | 20 * the vCPU is scheduled out and is blocking (e.g. in HLT) with IRQs enabled. 24 * vector is reset when the vCPU is scheduled in. 203 * scheduled out). in pi_enable_wakeup_handler() 236 * from KVM to be scheduled in. in vmx_vcpu_pi_put() 294 * blocking vCPUs may scheduled out without reconfiguring PID.NV to the wakeup
|
/linux/drivers/gpu/drm/i915/ |
H A D | i915_priolist_types.h | 23 /* Interactive workload, scheduled for immediate pageflipping */ 32 * another context. They get scheduled with their default priority and
|
H A D | i915_vma_resource.h | 60 * is not considered published until unbind is scheduled, and as such it 61 * is illegal to access this fence before scheduled unbind other than 66 * is scheduled. 92 * taken when the unbind is scheduled.
|
/linux/drivers/gpu/drm/scheduler/tests/ |
H A D | sched_tests.h | 208 * drm_mock_sched_job_wait_scheduled - Wait until a job is scheduled 213 * Returns: true if scheduled within the timeout provided, otherwise false 220 return dma_fence_wait_timeout(&job->base.s_fence->scheduled, in drm_mock_sched_job_wait_scheduled()
|
/linux/kernel/ |
H A D | async.c | 28 The async core will assign each scheduled event such a sequence cookie and 35 operations that were scheduled prior to the operation corresponding with the 38 Subsystem/driver initialization code that scheduled asynchronous probe 256 * If the asynchronous execution of @func is scheduled successfully, return
|
H A D | workqueue_internal.h | 42 struct list_head scheduled; /* L: scheduled works */ member
|
/linux/arch/s390/pci/ |
H A D | pci_irq.c | 179 atomic_t scheduled; member 185 atomic_t *scheduled = data; in zpci_handle_remote_irq() local 189 } while (atomic_dec_return(scheduled)); in zpci_handle_remote_irq() 212 if (atomic_inc_return(&cpu_data->scheduled) > 1) in zpci_handle_fallback_irq() 215 INIT_CSD(&cpu_data->csd, zpci_handle_remote_irq, &cpu_data->scheduled); in zpci_handle_fallback_irq()
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_job.c | 283 struct dma_fence *fence = &leader->base.s_fence->scheduled; in amdgpu_job_set_gang_leader() 303 if (job->gang_submit != &job->base.s_fence->scheduled) in amdgpu_job_free() 440 /* Signal all jobs not yet scheduled */ in amdgpu_job_stop_all_jobs_on_sched() 448 dma_fence_signal(&s_fence->scheduled); in amdgpu_job_stop_all_jobs_on_sched() 456 /* Signal all jobs already scheduled to HW */ in amdgpu_job_stop_all_jobs_on_sched()
|
/linux/Documentation/admin-guide/hw-vuln/ |
H A D | core-scheduling.rst | 35 Using this feature, userspace defines groups of tasks that can be co-scheduled 122 tasks selected to be scheduled are of the highest priority in a core. However, 128 by the scheduler (idle thread is scheduled to run). 224 - Gang scheduling: Requirements for a group of tasks that needs to be scheduled
|
/linux/drivers/usb/dwc2/ |
H A D | hcd_queue.c | 123 * @interval: How often we need to be scheduled for the reservation this 585 * already scheduled something on the low speed schedule and 895 /* If we scheduled all w/out breaking out then we're all good */ in dwc2_uframe_schedule_split() 1006 * dwc2_pick_first_frame() - Choose 1st frame for qh that's already scheduled 1008 * Takes a qh that has already been scheduled (which means we know we have the 1038 * NOTE: if we could quantify how long till we actually get scheduled in dwc2_pick_first_frame() 1057 * the first frame should be based on when the first scheduled in dwc2_pick_first_frame() 1096 * scheduled for. in dwc2_pick_first_frame() 1223 * Wait for the lock, or for us to be scheduled again. We in dwc2_unreserve_timer_fn() 1224 * could be scheduled again if: in dwc2_unreserve_timer_fn() [all …]
|
/linux/include/xen/interface/ |
H A D | vcpu.h | 82 /* VCPU is runnable, but not currently scheduled on any physical CPU. */ 103 * updated by the hypervisor each time the VCPU is scheduled. Thus 106 * VCPU was last scheduled to run.
|