Home
last modified time | relevance | path

Searched refs:used_mask (Results 1 – 25 of 28) sorted by relevance

12

/linux/arch/sh/kernel/
H A Dperf_event.c30 unsigned long used_mask[BITS_TO_LONGS(MAX_HWEVENTS)]; member
239 __clear_bit(event->hw.idx, cpuc->used_mask); in sh_pmu_del()
253 if (__test_and_set_bit(idx, cpuc->used_mask)) { in sh_pmu_add()
254 idx = find_first_zero_bit(cpuc->used_mask, sh_pmu->num_events); in sh_pmu_add()
258 __set_bit(idx, cpuc->used_mask); in sh_pmu_add()
/linux/drivers/perf/
H A Darm_dsu_pmu.c94 DECLARE_BITMAP(used_mask, DSU_PMU_MAX_HW_CNTRS);
287 unsigned long *used_mask = hw_events->used_mask; in dsu_pmu_write_counter()
290 if (test_and_set_bit(DSU_PMU_IDX_CYCLE_COUNTER, used_mask)) in dsu_pmu_get_event_idx()
295 idx = find_first_zero_bit(used_mask, dsu_pmu->num_counters); in dsu_pmu_get_event_idx()
298 set_bit(idx, hw_events->used_mask); in dsu_pmu_get_event_idx()
457 clear_bit(idx, hw_events->used_mask); in dsu_pmu_del()
468 if (bitmap_empty(dsu_pmu->hw_events.used_mask, DSU_PMU_MAX_HW_CNTRS)) in dsu_pmu_del()
515 memset(fake_hw.used_mask, 0, sizeof(fake_hw.used_mask));
296 unsigned long *used_mask = hw_events->used_mask; dsu_pmu_get_event_idx() local
[all...]
H A Dstarfive_starlink_pmu.c66 DECLARE_BITMAP(used_mask, STARLINK_PMU_MAX_COUNTERS);
296 unsigned long *used_mask = hw_events->used_mask; in starlink_pmu_add() local
310 idx = find_first_zero_bit(used_mask, n_events); in starlink_pmu_add()
315 set_bit(idx, used_mask); in starlink_pmu_add()
339 clear_bit(hwc->idx, hw_events->used_mask); in starlink_pmu_del()
453 int enabled = bitmap_weight(hw_events->used_mask, in starlink_pmu_pm_notify()
H A Darm_v6_pmu.c320 if (test_and_set_bit(ARMV6_CYCLE_COUNTER, cpuc->used_mask)) in armv6pmu_get_event_idx()
329 if (!test_and_set_bit(ARMV6_COUNTER1, cpuc->used_mask)) in armv6pmu_get_event_idx()
332 if (!test_and_set_bit(ARMV6_COUNTER0, cpuc->used_mask)) in armv6pmu_get_event_idx()
343 clear_bit(event->hw.idx, cpuc->used_mask); in armv6pmu_clear_event_idx()
H A Darm_xscale_pmu.c272 if (test_and_set_bit(XSCALE_CYCLE_COUNTER, cpuc->used_mask)) in xscale1pmu_get_event_idx()
277 if (!test_and_set_bit(XSCALE_COUNTER1, cpuc->used_mask)) in xscale1pmu_get_event_idx()
280 if (!test_and_set_bit(XSCALE_COUNTER0, cpuc->used_mask)) in xscale1pmu_get_event_idx()
290 clear_bit(event->hw.idx, cpuc->used_mask); in xscalepmu_clear_event_idx()
635 if (!test_and_set_bit(XSCALE_COUNTER3, cpuc->used_mask)) in xscale2pmu_get_event_idx()
637 else if (!test_and_set_bit(XSCALE_COUNTER2, cpuc->used_mask)) in xscale2pmu_get_event_idx()
H A Darm-cci.c70 unsigned long *used_mask; member
315 if (test_and_set_bit(CCI400_PMU_CYCLE_CNTR_IDX, hw->used_mask)) in cci400_get_event_idx()
322 if (!test_and_set_bit(idx, hw->used_mask)) in cci400_get_event_idx()
644 for_each_set_bit(i, cci_pmu->hw_events.used_mask, cci_pmu->num_cntrs) { in cci_pmu_sync_counters()
742 * caller. We cannot rely on the used_mask maintained by the CCI_PMU
797 if (!test_and_set_bit(idx, hw->used_mask)) in pmu_get_event_idx()
1086 bool enabled = !bitmap_empty(hw_events->used_mask, cci_pmu->num_cntrs); in hw_perf_event_destroy()
1210 clear_bit(idx, hw_events->used_mask);
1247 * used_mask for the purposes of validation. in validate_event()
1249 .used_mask
[all...]
H A Darm_pmuv3.c774 for_each_andnot_bit(i, cpu_pmu->cntr_mask, cpuc->used_mask, in armv8pmu_enable_user_access()
900 if (!test_and_set_bit(idx, cpuc->used_mask)) in armv8pmu_get_single_idx()
918 if (!test_and_set_bit(idx, cpuc->used_mask)) { in armv8pmu_get_chain_idx()
920 if (!test_and_set_bit(idx - 1, cpuc->used_mask)) in armv8pmu_get_chain_idx()
923 clear_bit(idx, cpuc->used_mask); in armv8pmu_get_chain_idx()
939 if (!test_and_set_bit(ARMV8_PMU_CYCLE_IDX, cpuc->used_mask)) in armv8pmu_get_event_idx()
956 if (!test_and_set_bit(ARMV8_PMU_INSTR_IDX, cpuc->used_mask)) in armv8pmu_get_event_idx()
974 clear_bit(idx, cpuc->used_mask); in armv8pmu_clear_event_idx()
976 clear_bit(idx - 1, cpuc->used_mask); in armv8pmu_clear_event_idx()
H A Darm_pmu.c400 memset(&fake_pmu.used_mask, 0, sizeof(fake_pmu.used_mask)); in validate_group()
525 bool enabled = !bitmap_empty(hw_events->used_mask, ARMPMU_MAX_HWEVENTS); in armpmu_enable()
775 bool enabled = !bitmap_empty(hw_events->used_mask, ARMPMU_MAX_HWEVENTS); in cpu_pm_pmu_notify()
H A Darm_v7_pmu.c999 if (test_and_set_bit(ARMV7_IDX_CYCLE_COUNTER, cpuc->used_mask)) in armv7pmu_get_event_idx()
1010 if (!test_and_set_bit(idx, cpuc->used_mask)) in armv7pmu_get_event_idx()
1021 clear_bit(event->hw.idx, cpuc->used_mask); in armv7pmu_clear_event_idx()
1575 if (test_and_set_bit(bit, cpuc->used_mask)) in krait_pmu_get_event_idx()
1581 clear_bit(bit, cpuc->used_mask); in krait_pmu_get_event_idx()
1599 clear_bit(bit, cpuc->used_mask); in krait_pmu_clear_event_idx()
1893 if (test_and_set_bit(bit, cpuc->used_mask)) in scorpion_pmu_get_event_idx()
1899 clear_bit(bit, cpuc->used_mask); in scorpion_pmu_get_event_idx()
1917 clear_bit(bit, cpuc->used_mask); in scorpion_pmu_clear_event_idx()
H A Dqcom_l3_pmu.c159 unsigned long used_mask[BITS_TO_LONGS(L3_NUM_COUNTERS)]; member
557 idx = bitmap_find_free_region(l3pmu->used_mask, L3_NUM_COUNTERS, order); in qcom_l3_cache__event_add()
584 bitmap_release_region(l3pmu->used_mask, hwc->idx, order); in qcom_l3_cache__event_del()
H A Darm_dmc620_pmu.c99 DECLARE_BITMAP(used_mask, DMC620_PMU_MAX_COUNTERS);
307 if (!test_and_set_bit(idx, dmc620_pmu->used_mask)) in dmc620_get_event_idx()
631 clear_bit(idx, dmc620_pmu->used_mask);
H A Dalibaba_uncore_drw_pmu.c101 DECLARE_BITMAP(used_mask, ALI_DRW_PMU_COMMON_MAX_COUNTERS);
272 if (!test_and_set_bit(idx, drw_pmu->used_mask)) in ali_drw_get_counter_idx()
653 clear_bit(idx, drw_pmu->used_mask); in ali_drw_pmu_del()
H A Dapple_m1_cpu_pmu.c483 if (!test_and_set_bit(idx, cpuc->used_mask)) in m1_pmu_get_event_idx()
493 clear_bit(event->hw.idx, cpuc->used_mask); in m1_pmu_clear_event_idx()
/linux/drivers/hwtracing/coresight/
H A Dcoresight-cti-platform.c139 tc->con_in->used_mask = 0xF0; /* sigs <4,5,6,7> */ in cti_plat_create_v8_etm_connection()
140 tc->con_out->used_mask = 0xF0; /* sigs <4,5,6,7> */ in cti_plat_create_v8_etm_connection()
197 tc->con_in->used_mask = 0x3; /* sigs <0 1> */ in cti_plat_create_v8_connections()
200 tc->con_out->used_mask = 0x7; /* sigs <0 1 2 > */ in cti_plat_create_v8_connections()
260 tgrp->used_mask |= BIT(values[idx]); in cti_plat_read_trig_group()
334 drvdata->config.trig_out_filter |= tg->used_mask; in cti_plat_process_filter_sigs()
H A Dcoresight-cti-core.c273 drvdata->config.trig_in_use |= tc->con_in->used_mask; in cti_add_connection_entry()
274 drvdata->config.trig_out_use |= tc->con_out->used_mask; in cti_add_connection_entry()
328 tc->con_in->used_mask = n_trig_mask; in cti_add_default_connection()
329 tc->con_out->used_mask = n_trig_mask; in cti_add_default_connection()
H A Dcoresight-cti.h70 u32 used_mask; member
/linux/arch/xtensa/kernel/
H A Dperf_event.c63 unsigned long used_mask[BITS_TO_LONGS(XCHAL_NUM_PERF_COUNTERS)]; member
296 if (__test_and_set_bit(idx, ev->used_mask)) { in xtensa_pmu_add()
297 idx = find_first_zero_bit(ev->used_mask, in xtensa_pmu_add()
302 __set_bit(idx, ev->used_mask); in xtensa_pmu_add()
321 __clear_bit(event->hw.idx, ev->used_mask); in xtensa_pmu_del()
374 for_each_set_bit(i, ev->used_mask, XCHAL_NUM_PERF_COUNTERS) { in xtensa_pmu_irq_handler()
/linux/drivers/perf/hisilicon/
H A Dhisi_uncore_pmu.c87 unsigned long *used_mask = hisi_pmu->pmu_events.used_mask; in hisi_uncore_pmu_get_event_idx() local
91 idx = find_first_zero_bit(used_mask, num_counters); in hisi_uncore_pmu_get_event_idx()
95 set_bit(idx, used_mask); in hisi_uncore_pmu_get_event_idx()
113 clear_bit(idx, hisi_pmu->pmu_events.used_mask); in hisi_uncore_pmu_clear_event_idx()
384 bool enabled = !bitmap_empty(hisi_pmu->pmu_events.used_mask, in hisi_uncore_pmu_enable()
H A Dhisi_uncore_ddrc_pmu.c170 unsigned long *used_mask = ddrc_pmu->pmu_events.used_mask; in hisi_ddrc_pmu_v1_get_event_idx() local
175 if (test_bit(idx, used_mask)) in hisi_ddrc_pmu_v1_get_event_idx()
178 set_bit(idx, used_mask); in hisi_ddrc_pmu_v1_get_event_idx()
H A Dhisi_uncore_pmu.h80 DECLARE_BITMAP(used_mask, HISI_MAX_COUNTERS);
/linux/arch/loongarch/kernel/
H A Dperf_event.c97 unsigned long used_mask[BITS_TO_LONGS(LOONGARCH_MAX_HWEVENTS)]; member
258 if (!test_and_set_bit(i, cpuc->used_mask)) in loongarch_pmu_alloc_counter()
430 clear_bit(idx, cpuc->used_mask); in loongarch_pmu_del()
509 if (test_bit(n, cpuc->used_mask)) { in pmu_handle_irq()
/linux/arch/arc/kernel/
H A Dperf_event.c211 unsigned long used_mask[BITS_TO_LONGS(ARC_PERF_MAX_COUNTERS)]; member
514 __clear_bit(event->hw.idx, pmu_cpu->used_mask); in arc_pmu_del()
528 idx = ffz(pmu_cpu->used_mask[0]); in arc_pmu_add()
532 __set_bit(idx, pmu_cpu->used_mask); in arc_pmu_add()
/linux/arch/x86/events/intel/
H A Dp4.c1226 static int p4_next_cntr(int thread, unsigned long *used_mask, in p4_next_cntr() argument
1233 if (j != -1 && !test_bit(j, used_mask)) in p4_next_cntr()
1242 unsigned long used_mask[BITS_TO_LONGS(X86_PMC_IDX_MAX)]; in p4_pmu_schedule_events() local
1252 bitmap_zero(used_mask, X86_PMC_IDX_MAX); in p4_pmu_schedule_events()
1282 cntr_idx = p4_next_cntr(thread, used_mask, bind); in p4_pmu_schedule_events()
1316 set_bit(cntr_idx, used_mask); in p4_pmu_schedule_events()
/linux/include/linux/perf/
H A Darm_pmu.h64 DECLARE_BITMAP(used_mask, ARMPMU_MAX_HWEVENTS);
/linux/arch/mips/kernel/
H A Dperf_event_mipsxx.c39 unsigned long used_mask[BITS_TO_LONGS(MIPS_MAX_HWEVENTS)]; member
341 !test_and_set_bit(i, cpuc->used_mask)) in mipsxx_pmu_alloc_counter()
545 clear_bit(idx, cpuc->used_mask); in mipspmu_del()
1601 if (!test_bit(n, cpuc->used_mask)) in mipsxx_pmu_handle_shared_irq()

12