Home
last modified time | relevance | path

Searched refs:runnable (Results 1 – 24 of 24) sorted by relevance

/linux/tools/verification/models/sched/
H A Dsssw.dot5 {node [shape = doublecircle] "runnable"};
6 {node [shape = circle] "runnable"};
10 "__init_runnable" -> "runnable";
11 "runnable" [label = "runnable", color = green3];
12 …"runnable" -> "runnable" [ label = "sched_set_state_runnable\nsched_wakeup\nsched_switch_in\nsched…
13 "runnable" -> "sleepable" [ label = "sched_set_state_sleepable" ];
14 "runnable" -> "sleeping" [ label = "sched_switch_blocking" ];
16 "signal_wakeup" -> "runnable" [ label = "signal_deliver" ];
20 "sleepable" -> "runnable" [ label = "sched_set_state_runnable\nsched_wakeup" ];
25 "sleeping" -> "runnable" [ label = "sched_wakeup" ];
[all …]
/linux/kernel/sched/
H A Dpelt.c104 unsigned long load, unsigned long runnable, int running) in accumulate_sum() argument
144 if (runnable) in accumulate_sum()
145 sa->runnable_sum += runnable * contrib << SCHED_CAPACITY_SHIFT; in accumulate_sum()
182 unsigned long load, unsigned long runnable, int running) in ___update_load_sum() argument
218 runnable = running = 0; in ___update_load_sum()
227 if (!accumulate_sum(delta, sa, load, runnable, running)) in ___update_load_sum()
H A Dext_internal.h389 void (*runnable)(struct task_struct *p, u64 enq_flags); member
426 void (*stopping)(struct task_struct *p, bool runnable);
H A Dfair.c2112 unsigned long runnable; member
2148 ((ns->compute_capacity * imbalance_pct) < (ns->runnable * 100)))) in numa_classify()
2153 ((ns->compute_capacity * imbalance_pct) > (ns->runnable * 100)))) in numa_classify()
2204 ns->runnable += cpu_runnable(rq); in update_numa_stats()
4382 __update_sa(&cfs_rq->avg, runnable, delta_avg, delta_sum); in update_tg_cfs_runnable()
4651 __update_sa(sa, runnable, -r, -r*divider); in update_cfs_rq_load_avg()
4737 __update_sa(&cfs_rq->avg, runnable, -se->avg.runnable_avg, -se->avg.runnable_sum); in detach_entity_load_avg()
7266 unsigned int runnable; in cpu_runnable_without() local
7273 runnable = READ_ONCE(cfs_rq->avg.runnable_avg); in cpu_runnable_without()
7276 lsub_positive(&runnable, p->se.avg.runnable_avg); in cpu_runnable_without()
[all …]
H A Dext.c1520 if (SCX_HAS_OP(sch, runnable) && !task_on_rq_migrating(p)) in enqueue_task_scx()
1521 SCX_CALL_OP_TASK(sch, SCX_KF_REST, runnable, rq, p, enq_flags); in enqueue_task_scx()
5529 static void sched_ext_ops__stopping(struct task_struct *p, bool runnable) {} in sched_ext_ops__stopping() argument
5566 .runnable = sched_ext_ops__runnable,
/linux/tools/sched_ext/
H A Dscx_flatcg.bpf.c13 * Let's say all three have runnable tasks. The total share that each of these
26 * and keep updating the eventual shares as the cgroups' runnable states change.
416 static void update_active_weight_sums(struct cgroup *cgrp, bool runnable) in update_active_weight_sums() argument
433 if (runnable) { in update_active_weight_sums()
444 * If @cgrp is becoming runnable, its hweight should be refreshed after in update_active_weight_sums()
450 if (!runnable) in update_active_weight_sums()
476 if (runnable) { in update_active_weight_sums()
503 if (runnable) in update_active_weight_sums()
539 void BPF_STRUCT_OPS(fcg_stopping, struct task_struct *p, bool runnable) in BPF_STRUCT_OPS() argument
952 .runnable
[all...]
H A Dscx_central.bpf.c246 void BPF_STRUCT_OPS(central_stopping, struct task_struct *p, bool runnable) in BPF_STRUCT_OPS() argument
/linux/Documentation/scheduler/
H A Dschedutil.rst35 Using this we track 2 key metrics: 'running' and 'runnable'. 'Running'
36 reflects the time an entity spends on the CPU, while 'runnable' reflects the
40 while 'runnable' will increase to reflect the amount of contention.
83 The result is that the above 'running' and 'runnable' metrics become invariant
104 A further runqueue wide sum (of runnable tasks) is maintained of:
H A Dsched-ext.rst20 a runnable task stalls, or on invoking the SysRq key sequence
64 detection of any internal error including stalled runnable tasks aborts the
210 from the global DSQ. If that doesn't yield a runnable task either,
279 * If the previous task is an SCX task and still runnable, keep executing
312 ops.runnable(); /* Task becomes ready to run */
314 while (task is runnable) {
323 while (task->scx.slice > 0 && task is runnable)
H A Dsched-eevdf.rst14 runnable tasks with the same priority. To do so, it assigns a virtual run
H A Dsched-capacity.rst243 accurately be predicted the moment a task first becomes runnable. The CFS class
/linux/tools/testing/selftests/sched_ext/
H A Dmaximal.bpf.c42 void BPF_STRUCT_OPS(maximal_stopping, struct task_struct *p, bool runnable) in BPF_STRUCT_OPS() argument
144 .runnable = (void *) maximal_runnable,
H A Dselect_cpu_vtime.bpf.c67 bool runnable) in BPF_STRUCT_OPS() argument
/linux/Documentation/timers/
H A Dno_hz.rst24 have only one runnable task (CONFIG_NO_HZ_FULL=y). Unless you
44 will frequently be multiple runnable tasks per CPU. In these cases,
107 If a CPU has only one runnable task, there is little point in sending it
109 Note that omitting scheduling-clock ticks for CPUs with only one runnable
113 sending scheduling-clock interrupts to CPUs with a single runnable task,
257 runnable task for a given CPU, even though there are a number
260 runnable high-priority SCHED_FIFO task and an arbitrary number
267 single runnable SCHED_FIFO task and multiple runnable SCHED_OTHER
270 And even when there are multiple runnable tasks on a given CPU,
/linux/Documentation/trace/rv/
H A Dmonitor_sched.rst67 The set non runnable on its own context (snroc) monitor ensures changes in a
270 is woken up or set to ``runnable``.
287 back to runnable, the resulting switch (if there) looks like a yield to the
292 This monitor doesn't include a running state, ``sleepable`` and ``runnable``
312 | | _blocking H runnable H | |
/linux/Documentation/tools/rv/
H A Drv-mon-sched.rst52 * snroc: set non runnable on its own context
/linux/drivers/gpu/drm/panthor/
H A Dpanthor_sched.c231 struct list_head runnable[PANTHOR_CSG_PRIORITY_COUNT]; member
2241 &sched->groups.runnable[group->priority]); in tick_ctx_cleanup()
2261 &sched->groups.runnable[group->priority]); in tick_ctx_cleanup()
2415 list_move_tail(&group->run_node, &sched->groups.runnable[prio]); in tick_ctx_apply()
2494 &sched->groups.runnable[prio], in tick_work()
2516 tick_ctx_pick_groups_from_list(sched, &ctx, &sched->groups.runnable[prio], in tick_work()
2636 &sched->groups.runnable[group->priority]); in sync_upd_work()
2676 struct list_head *queue = &sched->groups.runnable[group->priority]; in group_schedule_locked()
2770 &sched->groups.runnable[group->priority]); in panthor_group_start()
2962 for (i = 0; i < ARRAY_SIZE(sched->groups.runnable); i++) { in panthor_sched_pre_reset()
[all …]
/linux/Documentation/virt/
H A Dguest-halt-polling.rst18 even with other runnable tasks in the host.
/linux/Documentation/virt/kvm/
H A Dhalt-polling.rst18 interval or some other task on the runqueue is runnable the scheduler is
150 - Halt polling will only be conducted by the host when no other tasks are runnable on
H A Dapi.rst1592 whether the vcpu is runnable.
1617 whether the vcpu is runnable.
5761 KVM automatically accounts running and runnable time but blocked
6963 marking the exiting vCPU as runnable, or deny it and call KVM_RUN again.
/linux/Documentation/arch/s390/
H A Dvfio-ccw.rst331 space, and assemble a runnable kernel channel program by updating the
382 channel program, which becomes runnable for a real device.
/linux/Documentation/admin-guide/hw-vuln/
H A Dcore-scheduling.rst108 highest priority task with the same cookie is selected if there is one runnable
/linux/tools/perf/Documentation/
H A Dperf-sched.txt67 task scheduling delay (time between runnable and actually running) and
/linux/Documentation/locking/
H A Drt-mutex-design.rst69 of A. So now if B becomes runnable, it would not preempt C, since C now has