Home
last modified time | relevance | path

Searched refs:hybrid_pmu (Results 1 – 4 of 4) sorted by relevance

/linux/arch/x86/events/
H A Dcore.c211 cntr_mask |= x86_pmu.hybrid_pmu[i].cntr_mask64; in get_possible_counter_mask()
1925 if (!(x86_pmu.hybrid_pmu[i].pmu_type & pmu_attr->pmu_type)) in events_hybrid_sysfs_show()
1927 if (x86_pmu.hybrid_pmu[i].pmu_type & pmu->pmu_type) { in events_hybrid_sysfs_show()
2200 struct x86_hybrid_pmu *hybrid_pmu; in init_hw_perf_events() local
2204 hybrid_pmu = &x86_pmu.hybrid_pmu[i]; in init_hw_perf_events()
2206 hybrid_pmu->pmu = pmu; in init_hw_perf_events()
2207 hybrid_pmu->pmu.type = -1; in init_hw_perf_events()
2208 hybrid_pmu->pmu.attr_update = x86_pmu.attr_update; in init_hw_perf_events()
2209 hybrid_pmu->pmu.capabilities |= PERF_PMU_CAP_EXTENDED_HW_TYPE; in init_hw_perf_events()
2211 err = perf_pmu_register(&hybrid_pmu->pmu, hybrid_pmu->name, in init_hw_perf_events()
[all …]
H A Dperf_event.h769 static __always_inline struct x86_hybrid_pmu *hybrid_pmu(struct pmu *pmu) in hybrid_pmu() function
782 __Fp = &hybrid_pmu(_pmu)->_field; \
792 __Fp = &hybrid_pmu(_pmu)->_var; \
802 __Fp = hybrid_pmu(_pmu)->_field; \
1034 struct x86_hybrid_pmu *hybrid_pmu; member
/linux/arch/x86/events/intel/
H A Dcore.c4232 return hybrid_pmu(event->pmu)->pmu_type == hybrid_big; in require_mem_loads_aux_event()
5032 struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu); in adl_get_event_constraints()
5109 struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu); in mtl_get_event_constraints()
5122 struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu); in adl_hw_config()
5148 struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu); in arl_h_get_event_constraints()
5158 struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu); in arl_h_hw_config()
5729 enum hybrid_pmu_type pmu_type = x86_pmu.hybrid_pmu[i].pmu_type; in find_hybrid_pmu_for_cpu()
5733 return &x86_pmu.hybrid_pmu[i]; in find_hybrid_pmu_for_cpu()
5736 return &x86_pmu.hybrid_pmu[i]; in find_hybrid_pmu_for_cpu()
5740 return &x86_pmu.hybrid_pmu[i]; in find_hybrid_pmu_for_cpu()
[all …]
H A Dds.c149 data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX].pebs_data_source; in intel_pmu_pebs_data_source_adl()
153 data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX].pebs_data_source; in intel_pmu_pebs_data_source_adl()
172 data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX].pebs_data_source; in intel_pmu_pebs_data_source_mtl()
176 data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX].pebs_data_source; in intel_pmu_pebs_data_source_mtl()
187 data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_TINY_IDX].pebs_data_source; in intel_pmu_pebs_data_source_arl_h()
223 data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX].pebs_data_source; in intel_pmu_pebs_data_source_lnl()
226 data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX].pebs_data_source; in intel_pmu_pebs_data_source_lnl()
321 hybrid_pmu(event->pmu)->pmu_type == hybrid_big); in __grt_latency_data()
396 struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu); in lnl_latency_data()
406 struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu); in arl_h_latency_data()