| /linux/arch/x86/events/ |
| H A D | perf_event.h | 932 struct task_struct *task, bool sched_in); 1477 struct task_struct *task, bool sched_in); 1532 struct task_struct *task, bool sched_in); 1558 struct task_struct *task, bool sched_in) in amd_pmu_brs_sched_task() argument 1753 void intel_pmu_pebs_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in); 1768 struct task_struct *task, bool sched_in);
|
| H A D | core.c | 2730 struct task_struct *task, bool sched_in) in x86_pmu_sched_task() argument 2732 static_call_cond(x86_pmu_sched_task)(pmu_ctx, task, sched_in); in x86_pmu_sched_task()
|
| /linux/arch/s390/kernel/ |
| H A D | perf_pai.c | 674 struct task_struct *task, bool sched_in) in paicrypt_sched_task() argument 679 if (!sched_in) in paicrypt_sched_task() 738 struct task_struct *task, bool sched_in) in paiext_sched_task() argument 743 if (!sched_in) in paiext_sched_task()
|
| /linux/tools/perf/ |
| H A D | builtin-sched.c | 1145 struct thread *sched_out, *sched_in; in latency_switch_event() local 1165 sched_in = machine__findnew_thread(machine, -1, next_pid); in latency_switch_event() 1166 if (sched_out == NULL || sched_in == NULL) in latency_switch_event() 1182 in_events = thread_atoms_search(&sched->atom_root, sched_in, &sched->cmp_pid); in latency_switch_event() 1184 if (thread_atoms_insert(sched, sched_in)) in latency_switch_event() 1186 in_events = thread_atoms_search(&sched->atom_root, sched_in, &sched->cmp_pid); in latency_switch_event() 1202 thread__put(sched_in); in latency_switch_event() 1634 struct thread *sched_in, *sched_out; in map_switch_event() local 1676 sched_in = map__findnew_thread(sched, machine, -1, next_pid); in map_switch_event() 1678 if (sched_in in map_switch_event() [all...] |
| /linux/arch/powerpc/perf/ |
| H A D | core-book3s.c | 136 struct task_struct *task, bool sched_in) in power_pmu_sched_task() argument 451 struct task_struct *task, bool sched_in) in power_pmu_sched_task() argument 456 if (sched_in) in power_pmu_sched_task()
|
| /linux/kernel/events/ |
| H A D | core.c | 3751 struct task_struct *task, bool sched_in) in perf_ctx_sched_task_cb() argument 3760 pmu_ctx->pmu->sched_task(pmu_ctx, task, sched_in); in perf_ctx_sched_task_cb() 3895 struct task_struct *task, bool sched_in) in __perf_pmu_sched_task() argument 3909 pmu->sched_task(cpc->task_epc, task, sched_in); in __perf_pmu_sched_task() 3917 bool sched_in) in perf_pmu_sched_task() argument 3927 __perf_pmu_sched_task(cpc, sched_in ? next : prev, sched_in); in perf_pmu_sched_task() 3931 struct task_struct *next_prev, bool sched_in); 10090 struct task_struct *next_prev, bool sched_in) in perf_event_switch() argument 10102 .misc = sched_in ? 0 : PERF_RECORD_MISC_SWITCH_OUT, in perf_event_switch() 10110 if (!sched_in && task_is_runnable(task)) { in perf_event_switch()
|
| /linux/include/linux/ |
| H A D | perf_event.h | 515 struct task_struct *task, bool sched_in);
|
| /linux/arch/x86/events/intel/ |
| H A D | ds.c | 1592 void intel_pmu_pebs_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in intel_pmu_pebs_sched_task() argument 1596 if (!sched_in && pebs_needs_sched_cb(cpuc)) in intel_pmu_pebs_sched_task()
|
| H A D | core.c | 6234 struct task_struct *task, bool sched_in) in intel_pmu_sched_task() argument 6236 intel_pmu_pebs_sched_task(pmu_ctx, sched_in); in intel_pmu_sched_task() 6237 intel_pmu_lbr_sched_task(pmu_ctx, task, sched_in); in intel_pmu_sched_task()
|
| /linux/virt/kvm/ |
| H A D | kvm_main.c | 6525 kvm_preempt_ops.sched_in = kvm_sched_in; in kvm_init()
|
| /linux/kernel/sched/ |
| H A D | core.c | 4908 notifier->ops->sched_in(notifier, raw_smp_processor_id()); in __fire_sched_in_preempt_notifiers()
|