Searched refs:mem_op (Results 1 – 11 of 11) sorted by relevance
| /linux/drivers/spi/ |
| H A D | spi-intel.c | 182 struct spi_mem_op mem_op; member 722 if (iop->mem_op.cmd.nbytes != op->cmd.nbytes || in intel_spi_cmp_mem_op() 723 iop->mem_op.cmd.buswidth != op->cmd.buswidth || in intel_spi_cmp_mem_op() 724 iop->mem_op.cmd.dtr != op->cmd.dtr) in intel_spi_cmp_mem_op() 727 if (iop->mem_op.addr.nbytes != op->addr.nbytes || in intel_spi_cmp_mem_op() 728 iop->mem_op.addr.dtr != op->addr.dtr) in intel_spi_cmp_mem_op() 731 if (iop->mem_op.data.dir != op->data.dir || in intel_spi_cmp_mem_op() 732 iop->mem_op.data.dtr != op->data.dtr) in intel_spi_cmp_mem_op() 735 if (iop->mem_op.data.dir != SPI_MEM_NO_DATA) { in intel_spi_cmp_mem_op() 736 if (iop->mem_op.data.buswidth != op->data.buswidth) in intel_spi_cmp_mem_op() [all …]
|
| /linux/arch/x86/events/amd/ |
| H A D | ibs.c | 841 data_src->mem_op = PERF_MEM_OP_NA; in perf_ibs_get_mem_op() 844 data_src->mem_op = PERF_MEM_OP_LOAD; in perf_ibs_get_mem_op() 846 data_src->mem_op = PERF_MEM_OP_STORE; in perf_ibs_get_mem_op() 924 if (data_src->mem_op != PERF_MEM_OP_LOAD) in perf_ibs_get_mem_lvl() 1002 data_src->mem_op != PERF_MEM_OP_LOAD || in perf_ibs_get_mem_snoop() 1105 if (data->data_src.mem_op != PERF_MEM_OP_LOAD && in perf_ibs_parse_ld_st_data() 1106 data->data_src.mem_op != PERF_MEM_OP_STORE) in perf_ibs_parse_ld_st_data() 1117 data->data_src.mem_op == PERF_MEM_OP_LOAD) { in perf_ibs_parse_ld_st_data()
|
| /linux/tools/perf/util/ |
| H A D | mem-events.c | 423 op = mem_info__const_data_src(mem_info)->mem_op; in perf_mem__op_scnprintf() 630 u64 op = data_src->mem_op; in c2c_decode_stats() 818 switch (src.mem_op) { in mem_stat_index() 826 if (src.mem_op & PERF_MEM_OP_PFETCH) in mem_stat_index() 828 if (src.mem_op & PERF_MEM_OP_EXEC) in mem_stat_index()
|
| H A D | bpf-filter.l | 107 mem_op { return sample_part(PBF_TERM_DATA_SRC, 1); }
|
| H A D | arm-spe.c | 918 if (data_src->mem_op == PERF_MEM_OP_LOAD) in arm_spe__synth_memory_level() 920 if (data_src->mem_op == PERF_MEM_OP_STORE) in arm_spe__synth_memory_level() 1004 data_src.mem_op = PERF_MEM_OP_LOAD; in arm_spe__synth_data_source() 1006 data_src.mem_op = PERF_MEM_OP_STORE; in arm_spe__synth_data_source() 1008 data_src.mem_op = PERF_MEM_OP_NA; in arm_spe__synth_data_source()
|
| /linux/tools/perf/util/bpf_skel/ |
| H A D | sample_filter.bpf.c | 59 __u64 mem_op:5, /* type of opcode */ member 147 return kctx->data->data_src.mem_op; in perf_get_sample()
|
| /linux/include/uapi/linux/ |
| H A D | perf_event.h | 1323 __u64 mem_op : 5, /* Type of opcode */ member 1352 mem_op : 5; /* Type of opcode */ member
|
| /linux/tools/include/uapi/linux/ |
| H A D | perf_event.h | 1323 __u64 mem_op : 5, /* Type of opcode */ member 1352 mem_op : 5; /* Type of opcode */ member
|
| /linux/tools/perf/Documentation/ |
| H A D | perf-record.txt | 183 --filter 'mem_op == load || mem_op == store, mem_lvl > l1' 188 mem_op and mem_lvl fields come from the (memory) data_source, it'd only 194 $ sudo perf record -e cycles --filter 'mem_op == load' 206 p_stage_cyc, mem_op, mem_lvl, mem_snoop, mem_remote, mem_lock, 214 na, load, store, pfetch, exec (for mem_op)
|
| /linux/arch/x86/events/intel/ |
| H A D | ds.c | 416 dse.mem_op = PERF_MEM_OP_STORE; in precise_datala_hsw() 418 dse.mem_op = PERF_MEM_OP_LOAD; in precise_datala_hsw() 532 src.mem_op = P(OP, STORE); in arw_latency_data() 567 src.mem_op = P(OP, STORE); in lnc_latency_data() 625 src.mem_op = P(OP, STORE); in pnc_latency_data() 711 src.mem_op = P(OP,STORE); in store_latency_data()
|
| /linux/arch/s390/kvm/ |
| H A D | kvm-s390.c | 2957 struct kvm_s390_mem_op mem_op; in kvm_arch_vm_ioctl() local 2959 if (copy_from_user(&mem_op, argp, sizeof(mem_op)) == 0) in kvm_arch_vm_ioctl() 2960 r = kvm_s390_vm_mem_op(kvm, &mem_op); in kvm_arch_vm_ioctl() 5538 struct kvm_s390_mem_op mem_op; in kvm_arch_vcpu_ioctl() local 5540 if (copy_from_user(&mem_op, argp, sizeof(mem_op)) == 0) in kvm_arch_vcpu_ioctl() 5541 r = kvm_s390_vcpu_memsida_op(vcpu, &mem_op); in kvm_arch_vcpu_ioctl()
|