| /linux/arch/arm64/kvm/ |
| H A D | pvtime.c | 16 u64 base = vcpu->arch.steal.base; in kvm_update_stolen_time() 17 u64 last_steal = vcpu->arch.steal.last_steal; in kvm_update_stolen_time() 19 u64 steal = 0; in kvm_update_stolen_time() local 26 if (!kvm_get_guest(kvm, base + offset, steal)) { in kvm_update_stolen_time() 27 steal = le64_to_cpu(steal); in kvm_update_stolen_time() 28 vcpu->arch.steal.last_steal = READ_ONCE(current->sched_info.run_delay); in kvm_update_stolen_time() 29 steal += vcpu->arch.steal.last_steal - last_steal; in kvm_update_stolen_time() 30 kvm_put_guest(kvm, base + offset, cpu_to_le64(steal)); in kvm_update_stolen_time() 43 if (vcpu->arch.steal.base != INVALID_GPA) in kvm_hypercall_pv_features() 55 u64 base = vcpu->arch.steal.base; in kvm_init_stolen_time() [all …]
|
| /linux/arch/riscv/kernel/ |
| H A D | paravirt.c | 95 __le64 steal; in pv_time_steal_clock() local 104 steal = READ_ONCE(st->steal); in pv_time_steal_clock() 109 return le64_to_cpu(steal); in pv_time_steal_clock()
|
| /linux/arch/loongarch/kernel/ |
| H A D | paravirt.c | 36 u64 steal; in paravt_steal_clock() local 44 steal = src->steal; in paravt_steal_clock() 49 return steal; in paravt_steal_clock()
|
| /linux/arch/s390/kernel/ |
| H A D | vtime.c | 204 u64 steal, avg_steal; in vtime_flush() local 209 steal = lc->steal_timer; in vtime_flush() 211 if ((s64) steal > 0) { in vtime_flush() 213 account_steal_time(cputime_to_nsecs(steal)); in vtime_flush() 214 avg_steal += steal; in vtime_flush()
|
| /linux/arch/x86/kernel/cpu/ |
| H A D | vmware.c | 229 struct vmware_steal_time *steal = &per_cpu(vmw_steal_time, cpu); in vmware_steal_clock() local 233 clock = READ_ONCE(steal->clock); in vmware_steal_clock() 238 initial_high = READ_ONCE(steal->clock_high); in vmware_steal_clock() 241 low = READ_ONCE(steal->clock_low); in vmware_steal_clock() 244 high = READ_ONCE(steal->clock_high); in vmware_steal_clock()
|
| /linux/drivers/gpu/drm/radeon/ |
| H A D | radeon_object.c | 540 int steal; in radeon_bo_get_surface_reg() local 553 steal = -1; in radeon_bo_get_surface_reg() 562 steal = i; in radeon_bo_get_surface_reg() 567 if (steal == -1) in radeon_bo_get_surface_reg() 570 reg = &rdev->surface_regs[steal]; in radeon_bo_get_surface_reg() 573 DRM_DEBUG("stealing surface reg %d from %p\n", steal, old_object); in radeon_bo_get_surface_reg() 576 i = steal; in radeon_bo_get_surface_reg()
|
| /linux/arch/x86/kernel/ |
| H A D | kvm.c | 408 u64 steal; in kvm_steal_clock() local 416 steal = src->steal; in kvm_steal_clock() 420 return steal; in kvm_steal_clock()
|
| /linux/arch/x86/include/uapi/asm/ |
| H A D | kvm_para.h | 63 __u64 steal; member
|
| /linux/Documentation/translations/zh_CN/admin-guide/ |
| H A D | cpu-load.rst | 12 avg-cpu: %user %nice %system %iowait %steal %idle
|
| /linux/Documentation/translations/zh_TW/admin-guide/ |
| H A D | cpu-load.rst | 18 avg-cpu: %user %nice %system %iowait %steal %idle
|
| /linux/arch/loongarch/kvm/ |
| H A D | vcpu.c | 163 u64 steal; in kvm_update_stolen_time() local 192 unsafe_get_user(steal, &st->steal, out); in kvm_update_stolen_time() 193 steal += current->sched_info.run_delay - vcpu->arch.st.last_steal; in kvm_update_stolen_time() 195 unsafe_put_user(steal, &st->steal, out); in kvm_update_stolen_time()
|
| /linux/Documentation/admin-guide/ |
| H A D | cpu-load.rst | 12 avg-cpu: %user %nice %system %iowait %steal %idle
|
| /linux/lib/raid6/ |
| H A D | altivec.uc | 21 * you can just "steal" the vec unit with enable_kernel_altivec() (but
|
| /linux/Documentation/virt/kvm/x86/ |
| H A D | cpuid.rst | 58 KVM_FEATURE_STEAL_TIME 5 steal time can be enabled by
|
| H A D | msr.rst | 268 __u64 steal; 295 steal: 298 reported as steal time.
|
| H A D | timekeeping.rst | 496 steal cycles from the naturally on X86 systems when SMM mode is used by the
|
| /linux/arch/arm64/include/asm/ |
| H A D | kvm_host.h | 892 } steal; member 1312 vcpu_arch->steal.base = INVALID_GPA; in kvm_arm_pvtime_vcpu_init() 1317 return (vcpu_arch->steal.base != INVALID_GPA); in kvm_arm_is_pvtime_enabled()
|
| /linux/arch/riscv/include/asm/ |
| H A D | sbi.h | 295 __le64 steal; member
|
| /linux/arch/loongarch/ |
| H A D | Kconfig | 697 bool "Paravirtual steal time accounting" 700 Select this option to enable fine granularity task steal time
|
| /linux/kernel/sched/ |
| H A D | core.c | 779 s64 __maybe_unused steal = 0, irq_delta = 0; in update_rq_clock_task() local 812 steal = prev_steal = paravirt_steal_clock(cpu_of(rq)); in update_rq_clock_task() 813 steal -= rq->prev_steal_time_rq; in update_rq_clock_task() 815 if (unlikely(steal > delta)) in update_rq_clock_task() 816 steal = delta; in update_rq_clock_task() 819 delta -= steal; in update_rq_clock_task() 826 if ((irq_delta + steal) && sched_feat(NONTASK_CAPACITY)) in update_rq_clock_task() 827 update_irq_load_avg(rq, irq_delta + steal); in update_rq_clock_task()
|
| /linux/Documentation/arch/x86/ |
| H A D | intel_txt.rst | 161 attempt to crash the system to gain control on reboot and steal
|
| /linux/Documentation/target/ |
| H A D | tcmu-design.rst | 192 command(a.k.a steal the original command's entry).
|
| /linux/arch/riscv/ |
| H A D | Kconfig | 1120 bool "Paravirtual steal time accounting" 1123 Select this option to enable fine granularity task steal time
|
| /linux/Documentation/filesystems/ |
| H A D | mount_api.rst | 503 of param->{string,blob,name,file}. Note that the function may steal and
|
| /linux/arch/arm/ |
| H A D | Kconfig | 1329 bool "Paravirtual steal time accounting" 1332 Select this option to enable fine granularity task steal time
|