| /linux/arch/powerpc/perf/ |
| H A D | hv-24x7.c | 690 size_t event_idx, in catalog_event_len_validate() argument 701 if (event_idx >= event_entry_count) { in catalog_event_len_validate() 709 event_idx); in catalog_event_len_validate() 717 event_idx, ev_len, event); in catalog_event_len_validate() 722 event_idx, ev_len, ev_end, end, in catalog_event_len_validate() 730 event_idx, event_data_bytes, event, end, in catalog_event_len_validate() 737 event_idx, event, ev_end, offset, calc_ev_end); in catalog_event_len_validate() 761 event_data_bytes, junk_events, event_idx, event_attr_ct, i, in create_events_from_catalog() local 866 for (junk_events = 0, event = event_data, event_idx = 0, attr_max = 0; in create_events_from_catalog() 868 event_idx++, event = (void *)event + ev_len) { in create_events_from_catalog() [all …]
|
| H A D | core-book3s.c | 2222 .event_idx = power_pmu_event_idx,
|
| /linux/arch/riscv/kvm/ |
| H A D | vcpu_pmu.c | 230 if (pmc->event_idx == SBI_PMU_EVENT_IDX_INVALID) in pmu_fw_ctr_read_hi() 233 fevent_code = get_event_code(pmc->event_idx); in pmu_fw_ctr_read_hi() 262 if (pmc->event_idx == SBI_PMU_EVENT_IDX_INVALID) in pmu_ctr_read() 265 fevent_code = get_event_code(pmc->event_idx); in pmu_ctr_read() 512 eidx = einfo[i].event_idx; in kvm_riscv_vcpu_pmu_event_info() 601 fevent_code = get_event_code(pmc->event_idx); in kvm_riscv_vcpu_pmu_ctr_start() 664 fevent_code = get_event_code(pmc->event_idx); in kvm_riscv_vcpu_pmu_ctr_stop() 708 pmc->event_idx = SBI_PMU_EVENT_IDX_INVALID; in kvm_riscv_vcpu_pmu_ctr_stop() 806 pmc->event_idx = eidx; in kvm_riscv_vcpu_pmu_ctr_cfg_match() 881 pmc->event_idx = SBI_PMU_EVENT_IDX_INVALID; in kvm_riscv_vcpu_pmu_init() [all …]
|
| /linux/drivers/perf/ |
| H A D | riscv_pmu_sbi.c | 120 uint32_t event_idx; member 317 event_info_shmem[count++].event_idx = pmu_hw_event_map[i].event_idx; in pmu_sbi_check_event_info() 322 event_info_shmem[count++].event_idx = in pmu_sbi_check_event_info() 323 pmu_cache_event_map[i][j][k].event_idx; in pmu_sbi_check_event_info() 341 pmu_hw_event_map[i].event_idx = -ENOENT; in pmu_sbi_check_event_info() 351 pmu_cache_event_map[i][j][k].event_idx = -ENOENT; in pmu_sbi_check_event_info() 368 0, cmask, 0, edata->event_idx, 0, 0); in pmu_sbi_check_event() 374 edata->event_idx = -ENOENT; in pmu_sbi_check_event() 424 ret = pmu_hw_event_map[config].event_idx; in riscv_pmu_get_event_info() 626 ret = pmu_cache_event_map[cache_type][cache_op][cache_result].event_idx; in pmu_event_find_cache()
|
| H A D | riscv_pmu.c | 413 .event_idx = riscv_pmu_event_idx, in riscv_pmu_alloc()
|
| /linux/drivers/net/ |
| H A D | amt.c | 906 index = (amt->event_idx + amt->nr_events) % AMT_MAX_EVENTS; in amt_queue_event() 910 amt->event_idx %= AMT_MAX_EVENTS; in amt_queue_event() 2882 event = amt->events[amt->event_idx].event; in amt_event_work() 2883 skb = amt->events[amt->event_idx].skb; in amt_event_work() 2884 amt->events[amt->event_idx].event = AMT_EVENT_NONE; in amt_event_work() 2885 amt->events[amt->event_idx].skb = NULL; in amt_event_work() 2887 amt->event_idx++; in amt_event_work() 2888 amt->event_idx %= AMT_MAX_EVENTS; in amt_event_work() 2995 amt->event_idx = 0; in amt_dev_open()
|
| /linux/include/net/ |
| H A D | amt.h | 369 u8 event_idx; member
|
| /linux/arch/riscv/include/asm/ |
| H A D | sbi.h | 166 u32 event_idx; member
|
| /linux/drivers/nvme/host/ |
| H A D | pci.c | 552 static inline int nvme_dbbuf_need_event(u16 event_idx, u16 new_idx, u16 old) in nvme_dbbuf_need_event() argument 554 return (u16)(new_idx - event_idx - 1) < (u16)(new_idx - old); in nvme_dbbuf_need_event() 562 u16 old_value, event_idx; in nvme_dbbuf_update_and_check_event() local 581 event_idx = le32_to_cpu(*dbbuf_ei); in nvme_dbbuf_update_and_check_event() 582 if (!nvme_dbbuf_need_event(event_idx, value, old_value)) in nvme_dbbuf_update_and_check_event()
|
| /linux/drivers/virtio/ |
| H A D | virtio_ring.c | 1931 u16 new, old, off_wrap, flags, wrap_counter, event_idx; in virtqueue_kick_prepare_packed() local 1967 event_idx = off_wrap & ~(1 << VRING_PACKED_EVENT_F_WRAP_CTR); in virtqueue_kick_prepare_packed() 1969 event_idx -= vq->packed.vring.num; in virtqueue_kick_prepare_packed() 1971 needs_kick = vring_need_event(event_idx, new, old); in virtqueue_kick_prepare_packed()
|
| /linux/include/linux/ |
| H A D | perf_event.h | 509 int (*event_idx) (struct perf_event *event); /*optional */ member
|
| /linux/arch/x86/events/ |
| H A D | core.c | 2797 .event_idx = x86_pmu_event_idx,
|
| /linux/kernel/events/ |
| H A D | core.c | 6790 return event->pmu->event_idx(event); in perf_event_index() 12878 if (!pmu->event_idx) in DEFINE_FREE() 12879 pmu->event_idx = perf_event_idx_default; in DEFINE_FREE()
|