| /linux/tools/sched_ext/include/scx/ |
| H A D | compat.bpf.h | 34 void scx_bpf_dsq_move_set_vtime___new(struct bpf_iter_scx_dsq *it__iter, u64 vtime) __ksym __weak; 40 void scx_bpf_dispatch_from_dsq_set_vtime___old(struct bpf_iter_scx_dsq *it__iter, u64 vtime) __ksym… 56 #define scx_bpf_dsq_move_set_vtime(it__iter, vtime) \ argument 58 scx_bpf_dsq_move_set_vtime___new((it__iter), (vtime)) : \ 60 scx_bpf_dispatch_from_dsq_set_vtime___old((it__iter), (vtime)) : \ 237 void scx_bpf_dispatch_vtime___compat(struct task_struct *p, u64 dsq_id, u64 slice, u64 vtime, u64 e… 238 void scx_bpf_dsq_insert_vtime___compat(struct task_struct *p, u64 dsq_id, u64 slice, u64 vtime, u64… 281 scx_bpf_dsq_insert_vtime(struct task_struct *p, u64 dsq_id, u64 slice, u64 vtime, in scx_bpf_dsq_insert_vtime() argument 288 .vtime = vtime, in scx_bpf_dsq_insert_vtime() 294 scx_bpf_dsq_insert_vtime___compat(p, dsq_id, slice, vtime, in scx_bpf_dsq_insert_vtime() [all …]
|
| /linux/include/trace/events/ |
| H A D | iocost.h | 17 u64 last_period, u64 cur_period, u64 vtime), 19 TP_ARGS(iocg, path, now, last_period, cur_period, vtime), 29 __field(u64, vtime) 44 __entry->vtime = vtime; 52 "period=%llu->%llu vtime=%llu " 57 __entry->vtime, __entry->inuse, __entry->weight, 64 u64 last_period, u64 cur_period, u64 vtime), 66 TP_ARGS(iocg, path, now, last_period, cur_period, vtime) 71 u64 last_period, u64 cur_period, u64 vtime), [all...] |
| /linux/block/ |
| H A D | bfq-wf2q.c | 117 !bfq_gt(new_entity->start, st->vtime) in bfq_update_next_in_service() 662 !bfq_gt(last_idle->finish, st->vtime)) { in bfq_forget_idle() 667 st->vtime = last_idle->finish; in bfq_forget_idle() 670 if (first_idle && !bfq_gt(first_idle->finish, st->vtime)) in bfq_forget_idle() 771 entity->start = new_st->vtime; in __bfq_entity_update_weight_prio() 804 st->vtime += bfq_delta(served, st->wsum); in bfq_bfqq_served() 902 if (backshifted && bfq_gt(st->vtime, entity->finish)) { in bfq_update_fin_time_enqueue() 903 unsigned long delta = st->vtime - entity->finish; in bfq_update_fin_time_enqueue() 935 if (non_blocking_wait_rq && bfq_gt(st->vtime, entity->finish)) { in __bfq_activate_entity() 939 min_vstart = st->vtime; in __bfq_activate_entity() [all …]
|
| H A D | blk-iocost.c | 51 * The device virtual time (vtime) is used as the primary control metric. 73 * A given cgroup's vtime runs slower in inverse proportion to its hweight. 75 * against the device vtime - an IO which takes 10ms on the underlying 79 * vtime is running at a rate determined by its hweight. A cgroup tracks 80 * the vtime consumed by past IOs and can issue a new IO if doing so 81 * wouldn't outrun the current device vtime. Otherwise, the IO is 82 * suspended until the vtime has progressed enough to cover it. 97 * To slow down, we lower the vrate - the rate at which the device vtime 98 * passes compared to the wall clock. For example, if the vtime is running 144 * snapback mechanism in case the cgroup needs more IO vtime fo 504 atomic64_t vtime; global() member 1272 u64 vtime, vtarget; iocg_activate() local 1754 u64 vtime = atomic64_read(&iocg->vtime); hweight_after_donation() local 2211 u64 vtime = atomic64_read(&iocg->vtime); ioc_check_iocgs() local 2283 u64 vdone, vtime, usage_us; ioc_timer_fn() local 2467 adjust_inuse_and_calc_cost(struct ioc_gq * iocg,u64 vtime,u64 abs_cost,struct ioc_now * now) adjust_inuse_and_calc_cost() argument 2613 u64 abs_cost, cost, vtime; ioc_rqos_throttle() local 2747 u64 vtime, abs_cost, cost; ioc_rqos_merge() local [all...] |
| H A D | bfq-iosched.h | 65 u64 vtime; member
|
| /linux/tools/testing/selftests/sched_ext/ |
| H A D | select_cpu_vtime.bpf.c | 30 u64 vtime = p->scx.dsq_vtime; in task_vtime() local 32 if (vtime_before(vtime, vtime_now - SCX_SLICE_DFL)) in task_vtime() 35 return vtime; in task_vtime()
|
| /linux/tools/cgroup/ |
| H A D | iocost_monitor.py | 157 vtime = iocg.vtime.counter.value_() 161 self.inflight_pct = (vtime - vdone) * 100 / period_vtime
|
| /linux/tools/testing/selftests/x86/ |
| H A D | test_vsyscall.c | 60 const time_func_t vtime = (time_func_t)VSYS(0xffffffffff600400); variable 185 t_vsys = vtime(&t2_vsys); in test_time() 491 vtime(&tmp); in test_emulation()
|
| /linux/arch/s390/kernel/ |
| H A D | vtime.c | 97 static inline u64 scale_vtime(u64 vtime) in scale_vtime() argument 103 return vtime * mult / div; in scale_vtime() 104 return vtime; in scale_vtime()
|
| /linux/kernel/sched/ |
| H A D | ext.c | 499 u64 vtime; member 1040 * starving vtime-dispatched tasks by FIFO-dispatched tasks, we in dispatch_enqueue() 1041 * disallow any internal DSQ from doing vtime ordering of in dispatch_enqueue() 1044 scx_error(sch, "cannot use vtime ordering for built-in DSQs"); in dispatch_enqueue() 1066 * that @dsq->list is vtime ordered. in dispatch_enqueue() 5868 u64 dsq_id, u64 slice, u64 vtime, u64 enq_flags) in scx_dsq_insert_vtime() 5878 p->scx.dsq_vtime = vtime; in scx_dsq_insert_vtime() 5889 u64 vtime; 5894 * __scx_bpf_dsq_insert_vtime - Arg-wrapped vtime DSQ insertion 5899 * @args->vtime 5869 scx_dsq_insert_vtime(struct scx_sched * sch,struct task_struct * p,u64 dsq_id,u64 slice,u64 vtime,u64 enq_flags) scx_dsq_insert_vtime() argument 5890 u64 vtime; global() member 5944 scx_bpf_dsq_insert_vtime(struct task_struct * p,u64 dsq_id,u64 slice,u64 vtime,u64 enq_flags) scx_bpf_dsq_insert_vtime() argument 6187 scx_bpf_dsq_move_set_vtime(struct bpf_iter_scx_dsq * it__iter,u64 vtime) scx_bpf_dsq_move_set_vtime() argument 6437 scx_bpf_task_set_dsq_vtime(struct task_struct * p,u64 vtime) scx_bpf_task_set_dsq_vtime() argument [all...] |
| /linux/include/linux/ |
| H A D | sched.h | 385 struct vtime { struct 1115 struct vtime vtime; member
|
| /linux/kernel/ |
| H A D | fork.c | 2123 seqcount_init(&p->vtime.seqcount); in copy_process() 2124 p->vtime.starttime = 0; in copy_process() 2125 p->vtime.state = VTIME_INACTIVE; in copy_process()
|
| /linux/tools/sched_ext/ |
| H A D | README.md | 168 scx_simple can be run in either global weighted vtime mode, or FIFO mode.
|