/linux/arch/sh/kernel/ |
H A D | perf_event.c | 30 unsigned long used_mask[BITS_TO_LONGS(MAX_HWEVENTS)]; member 239 __clear_bit(event->hw.idx, cpuc->used_mask); in sh_pmu_del() 253 if (__test_and_set_bit(idx, cpuc->used_mask)) { in sh_pmu_add() 254 idx = find_first_zero_bit(cpuc->used_mask, sh_pmu->num_events); in sh_pmu_add() 258 __set_bit(idx, cpuc->used_mask); in sh_pmu_add()
|
/linux/drivers/perf/ |
H A D | arm_dsu_pmu.c | 94 DECLARE_BITMAP(used_mask, DSU_PMU_MAX_HW_CNTRS); 287 unsigned long *used_mask = hw_events->used_mask; in dsu_pmu_write_counter() 290 if (test_and_set_bit(DSU_PMU_IDX_CYCLE_COUNTER, used_mask)) in dsu_pmu_get_event_idx() 295 idx = find_first_zero_bit(used_mask, dsu_pmu->num_counters); in dsu_pmu_get_event_idx() 298 set_bit(idx, hw_events->used_mask); in dsu_pmu_get_event_idx() 457 clear_bit(idx, hw_events->used_mask); in dsu_pmu_del() 468 if (bitmap_empty(dsu_pmu->hw_events.used_mask, DSU_PMU_MAX_HW_CNTRS)) in dsu_pmu_del() 515 memset(fake_hw.used_mask, 0, sizeof(fake_hw.used_mask)); 296 unsigned long *used_mask = hw_events->used_mask; dsu_pmu_get_event_idx() local [all...] |
H A D | starfive_starlink_pmu.c | 66 DECLARE_BITMAP(used_mask, STARLINK_PMU_MAX_COUNTERS); 296 unsigned long *used_mask = hw_events->used_mask; in starlink_pmu_add() local 310 idx = find_first_zero_bit(used_mask, n_events); in starlink_pmu_add() 315 set_bit(idx, used_mask); in starlink_pmu_add() 339 clear_bit(hwc->idx, hw_events->used_mask); in starlink_pmu_del() 453 int enabled = bitmap_weight(hw_events->used_mask, in starlink_pmu_pm_notify()
|
H A D | arm_v6_pmu.c | 320 if (test_and_set_bit(ARMV6_CYCLE_COUNTER, cpuc->used_mask)) in armv6pmu_get_event_idx() 329 if (!test_and_set_bit(ARMV6_COUNTER1, cpuc->used_mask)) in armv6pmu_get_event_idx() 332 if (!test_and_set_bit(ARMV6_COUNTER0, cpuc->used_mask)) in armv6pmu_get_event_idx() 343 clear_bit(event->hw.idx, cpuc->used_mask); in armv6pmu_clear_event_idx()
|
H A D | arm_xscale_pmu.c | 272 if (test_and_set_bit(XSCALE_CYCLE_COUNTER, cpuc->used_mask)) in xscale1pmu_get_event_idx() 277 if (!test_and_set_bit(XSCALE_COUNTER1, cpuc->used_mask)) in xscale1pmu_get_event_idx() 280 if (!test_and_set_bit(XSCALE_COUNTER0, cpuc->used_mask)) in xscale1pmu_get_event_idx() 290 clear_bit(event->hw.idx, cpuc->used_mask); in xscalepmu_clear_event_idx() 635 if (!test_and_set_bit(XSCALE_COUNTER3, cpuc->used_mask)) in xscale2pmu_get_event_idx() 637 else if (!test_and_set_bit(XSCALE_COUNTER2, cpuc->used_mask)) in xscale2pmu_get_event_idx()
|
H A D | arm-cci.c | 70 unsigned long *used_mask; member 315 if (test_and_set_bit(CCI400_PMU_CYCLE_CNTR_IDX, hw->used_mask)) in cci400_get_event_idx() 322 if (!test_and_set_bit(idx, hw->used_mask)) in cci400_get_event_idx() 644 for_each_set_bit(i, cci_pmu->hw_events.used_mask, cci_pmu->num_cntrs) { in cci_pmu_sync_counters() 742 * caller. We cannot rely on the used_mask maintained by the CCI_PMU 797 if (!test_and_set_bit(idx, hw->used_mask)) in pmu_get_event_idx() 1086 bool enabled = !bitmap_empty(hw_events->used_mask, cci_pmu->num_cntrs); in hw_perf_event_destroy() 1210 clear_bit(idx, hw_events->used_mask); 1247 * used_mask for the purposes of validation. in validate_event() 1249 .used_mask [all...] |
H A D | arm_pmuv3.c | 774 for_each_andnot_bit(i, cpu_pmu->cntr_mask, cpuc->used_mask, in armv8pmu_enable_user_access() 900 if (!test_and_set_bit(idx, cpuc->used_mask)) in armv8pmu_get_single_idx() 918 if (!test_and_set_bit(idx, cpuc->used_mask)) { in armv8pmu_get_chain_idx() 920 if (!test_and_set_bit(idx - 1, cpuc->used_mask)) in armv8pmu_get_chain_idx() 923 clear_bit(idx, cpuc->used_mask); in armv8pmu_get_chain_idx() 939 if (!test_and_set_bit(ARMV8_PMU_CYCLE_IDX, cpuc->used_mask)) in armv8pmu_get_event_idx() 956 if (!test_and_set_bit(ARMV8_PMU_INSTR_IDX, cpuc->used_mask)) in armv8pmu_get_event_idx() 974 clear_bit(idx, cpuc->used_mask); in armv8pmu_clear_event_idx() 976 clear_bit(idx - 1, cpuc->used_mask); in armv8pmu_clear_event_idx()
|
H A D | arm_pmu.c | 400 memset(&fake_pmu.used_mask, 0, sizeof(fake_pmu.used_mask)); in validate_group() 525 bool enabled = !bitmap_empty(hw_events->used_mask, ARMPMU_MAX_HWEVENTS); in armpmu_enable() 775 bool enabled = !bitmap_empty(hw_events->used_mask, ARMPMU_MAX_HWEVENTS); in cpu_pm_pmu_notify()
|
H A D | arm_v7_pmu.c | 999 if (test_and_set_bit(ARMV7_IDX_CYCLE_COUNTER, cpuc->used_mask)) in armv7pmu_get_event_idx() 1010 if (!test_and_set_bit(idx, cpuc->used_mask)) in armv7pmu_get_event_idx() 1021 clear_bit(event->hw.idx, cpuc->used_mask); in armv7pmu_clear_event_idx() 1575 if (test_and_set_bit(bit, cpuc->used_mask)) in krait_pmu_get_event_idx() 1581 clear_bit(bit, cpuc->used_mask); in krait_pmu_get_event_idx() 1599 clear_bit(bit, cpuc->used_mask); in krait_pmu_clear_event_idx() 1893 if (test_and_set_bit(bit, cpuc->used_mask)) in scorpion_pmu_get_event_idx() 1899 clear_bit(bit, cpuc->used_mask); in scorpion_pmu_get_event_idx() 1917 clear_bit(bit, cpuc->used_mask); in scorpion_pmu_clear_event_idx()
|
H A D | qcom_l3_pmu.c | 159 unsigned long used_mask[BITS_TO_LONGS(L3_NUM_COUNTERS)]; member 557 idx = bitmap_find_free_region(l3pmu->used_mask, L3_NUM_COUNTERS, order); in qcom_l3_cache__event_add() 584 bitmap_release_region(l3pmu->used_mask, hwc->idx, order); in qcom_l3_cache__event_del()
|
H A D | arm_dmc620_pmu.c | 99 DECLARE_BITMAP(used_mask, DMC620_PMU_MAX_COUNTERS); 307 if (!test_and_set_bit(idx, dmc620_pmu->used_mask)) in dmc620_get_event_idx() 631 clear_bit(idx, dmc620_pmu->used_mask);
|
H A D | alibaba_uncore_drw_pmu.c | 101 DECLARE_BITMAP(used_mask, ALI_DRW_PMU_COMMON_MAX_COUNTERS); 272 if (!test_and_set_bit(idx, drw_pmu->used_mask)) in ali_drw_get_counter_idx() 653 clear_bit(idx, drw_pmu->used_mask); in ali_drw_pmu_del()
|
H A D | apple_m1_cpu_pmu.c | 483 if (!test_and_set_bit(idx, cpuc->used_mask)) in m1_pmu_get_event_idx() 493 clear_bit(event->hw.idx, cpuc->used_mask); in m1_pmu_clear_event_idx()
|
/linux/drivers/hwtracing/coresight/ |
H A D | coresight-cti-platform.c | 139 tc->con_in->used_mask = 0xF0; /* sigs <4,5,6,7> */ in cti_plat_create_v8_etm_connection() 140 tc->con_out->used_mask = 0xF0; /* sigs <4,5,6,7> */ in cti_plat_create_v8_etm_connection() 197 tc->con_in->used_mask = 0x3; /* sigs <0 1> */ in cti_plat_create_v8_connections() 200 tc->con_out->used_mask = 0x7; /* sigs <0 1 2 > */ in cti_plat_create_v8_connections() 260 tgrp->used_mask |= BIT(values[idx]); in cti_plat_read_trig_group() 334 drvdata->config.trig_out_filter |= tg->used_mask; in cti_plat_process_filter_sigs()
|
H A D | coresight-cti-core.c | 273 drvdata->config.trig_in_use |= tc->con_in->used_mask; in cti_add_connection_entry() 274 drvdata->config.trig_out_use |= tc->con_out->used_mask; in cti_add_connection_entry() 328 tc->con_in->used_mask = n_trig_mask; in cti_add_default_connection() 329 tc->con_out->used_mask = n_trig_mask; in cti_add_default_connection()
|
H A D | coresight-cti.h | 70 u32 used_mask; member
|
/linux/arch/xtensa/kernel/ |
H A D | perf_event.c | 63 unsigned long used_mask[BITS_TO_LONGS(XCHAL_NUM_PERF_COUNTERS)]; member 296 if (__test_and_set_bit(idx, ev->used_mask)) { in xtensa_pmu_add() 297 idx = find_first_zero_bit(ev->used_mask, in xtensa_pmu_add() 302 __set_bit(idx, ev->used_mask); in xtensa_pmu_add() 321 __clear_bit(event->hw.idx, ev->used_mask); in xtensa_pmu_del() 374 for_each_set_bit(i, ev->used_mask, XCHAL_NUM_PERF_COUNTERS) { in xtensa_pmu_irq_handler()
|
/linux/drivers/perf/hisilicon/ |
H A D | hisi_uncore_pmu.c | 87 unsigned long *used_mask = hisi_pmu->pmu_events.used_mask; in hisi_uncore_pmu_get_event_idx() local 91 idx = find_first_zero_bit(used_mask, num_counters); in hisi_uncore_pmu_get_event_idx() 95 set_bit(idx, used_mask); in hisi_uncore_pmu_get_event_idx() 113 clear_bit(idx, hisi_pmu->pmu_events.used_mask); in hisi_uncore_pmu_clear_event_idx() 384 bool enabled = !bitmap_empty(hisi_pmu->pmu_events.used_mask, in hisi_uncore_pmu_enable()
|
H A D | hisi_uncore_ddrc_pmu.c | 170 unsigned long *used_mask = ddrc_pmu->pmu_events.used_mask; in hisi_ddrc_pmu_v1_get_event_idx() local 175 if (test_bit(idx, used_mask)) in hisi_ddrc_pmu_v1_get_event_idx() 178 set_bit(idx, used_mask); in hisi_ddrc_pmu_v1_get_event_idx()
|
H A D | hisi_uncore_pmu.h | 80 DECLARE_BITMAP(used_mask, HISI_MAX_COUNTERS);
|
/linux/arch/loongarch/kernel/ |
H A D | perf_event.c | 97 unsigned long used_mask[BITS_TO_LONGS(LOONGARCH_MAX_HWEVENTS)]; member 258 if (!test_and_set_bit(i, cpuc->used_mask)) in loongarch_pmu_alloc_counter() 430 clear_bit(idx, cpuc->used_mask); in loongarch_pmu_del() 509 if (test_bit(n, cpuc->used_mask)) { in pmu_handle_irq()
|
/linux/arch/arc/kernel/ |
H A D | perf_event.c | 211 unsigned long used_mask[BITS_TO_LONGS(ARC_PERF_MAX_COUNTERS)]; member 514 __clear_bit(event->hw.idx, pmu_cpu->used_mask); in arc_pmu_del() 528 idx = ffz(pmu_cpu->used_mask[0]); in arc_pmu_add() 532 __set_bit(idx, pmu_cpu->used_mask); in arc_pmu_add()
|
/linux/arch/x86/events/intel/ |
H A D | p4.c | 1226 static int p4_next_cntr(int thread, unsigned long *used_mask, in p4_next_cntr() argument 1233 if (j != -1 && !test_bit(j, used_mask)) in p4_next_cntr() 1242 unsigned long used_mask[BITS_TO_LONGS(X86_PMC_IDX_MAX)]; in p4_pmu_schedule_events() local 1252 bitmap_zero(used_mask, X86_PMC_IDX_MAX); in p4_pmu_schedule_events() 1282 cntr_idx = p4_next_cntr(thread, used_mask, bind); in p4_pmu_schedule_events() 1316 set_bit(cntr_idx, used_mask); in p4_pmu_schedule_events()
|
/linux/include/linux/perf/ |
H A D | arm_pmu.h | 64 DECLARE_BITMAP(used_mask, ARMPMU_MAX_HWEVENTS);
|
/linux/arch/mips/kernel/ |
H A D | perf_event_mipsxx.c | 39 unsigned long used_mask[BITS_TO_LONGS(MIPS_MAX_HWEVENTS)]; member 341 !test_and_set_bit(i, cpuc->used_mask)) in mipsxx_pmu_alloc_counter() 545 clear_bit(idx, cpuc->used_mask); in mipspmu_del() 1601 if (!test_bit(n, cpuc->used_mask)) in mipsxx_pmu_handle_shared_irq()
|