1e2be04c7SGreg Kroah-Hartman /* SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note */ 2607ca46eSDavid Howells /* 3607ca46eSDavid Howells * Performance events: 4607ca46eSDavid Howells * 5607ca46eSDavid Howells * Copyright (C) 2008-2009, Thomas Gleixner <tglx@linutronix.de> 6607ca46eSDavid Howells * Copyright (C) 2008-2011, Red Hat, Inc., Ingo Molnar 7607ca46eSDavid Howells * Copyright (C) 2008-2011, Red Hat, Inc., Peter Zijlstra 8607ca46eSDavid Howells * 9607ca46eSDavid Howells * Data type definitions, declarations, prototypes. 10607ca46eSDavid Howells * 11607ca46eSDavid Howells * Started by: Thomas Gleixner and Ingo Molnar 12607ca46eSDavid Howells * 13607ca46eSDavid Howells * For licencing details see kernel-base/COPYING 14607ca46eSDavid Howells */ 15607ca46eSDavid Howells #ifndef _UAPI_LINUX_PERF_EVENT_H 16607ca46eSDavid Howells #define _UAPI_LINUX_PERF_EVENT_H 17607ca46eSDavid Howells 18607ca46eSDavid Howells #include <linux/types.h> 19607ca46eSDavid Howells #include <linux/ioctl.h> 20607ca46eSDavid Howells #include <asm/byteorder.h> 21607ca46eSDavid Howells 22607ca46eSDavid Howells /* 23607ca46eSDavid Howells * User-space ABI bits: 24607ca46eSDavid Howells */ 25607ca46eSDavid Howells 26607ca46eSDavid Howells /* 27607ca46eSDavid Howells * attr.type 28607ca46eSDavid Howells */ 29607ca46eSDavid Howells enum perf_type_id { 30607ca46eSDavid Howells PERF_TYPE_HARDWARE = 0, 31607ca46eSDavid Howells PERF_TYPE_SOFTWARE = 1, 32607ca46eSDavid Howells PERF_TYPE_TRACEPOINT = 2, 33607ca46eSDavid Howells PERF_TYPE_HW_CACHE = 3, 34607ca46eSDavid Howells PERF_TYPE_RAW = 4, 35607ca46eSDavid Howells PERF_TYPE_BREAKPOINT = 5, 36607ca46eSDavid Howells 37607ca46eSDavid Howells PERF_TYPE_MAX, /* non-ABI */ 38607ca46eSDavid Howells }; 39607ca46eSDavid Howells 40607ca46eSDavid Howells /* 41607ca46eSDavid Howells * Generalized performance event event_id types, used by the 42607ca46eSDavid Howells * attr.event_id parameter of the sys_perf_event_open() 43607ca46eSDavid Howells * syscall: 44607ca46eSDavid Howells */ 45607ca46eSDavid Howells enum perf_hw_id { 46607ca46eSDavid Howells /* 47607ca46eSDavid Howells * Common hardware events, generalized by the kernel: 48607ca46eSDavid Howells */ 49607ca46eSDavid Howells PERF_COUNT_HW_CPU_CYCLES = 0, 50607ca46eSDavid Howells PERF_COUNT_HW_INSTRUCTIONS = 1, 51607ca46eSDavid Howells PERF_COUNT_HW_CACHE_REFERENCES = 2, 52607ca46eSDavid Howells PERF_COUNT_HW_CACHE_MISSES = 3, 53607ca46eSDavid Howells PERF_COUNT_HW_BRANCH_INSTRUCTIONS = 4, 54607ca46eSDavid Howells PERF_COUNT_HW_BRANCH_MISSES = 5, 55607ca46eSDavid Howells PERF_COUNT_HW_BUS_CYCLES = 6, 56607ca46eSDavid Howells PERF_COUNT_HW_STALLED_CYCLES_FRONTEND = 7, 57607ca46eSDavid Howells PERF_COUNT_HW_STALLED_CYCLES_BACKEND = 8, 58607ca46eSDavid Howells PERF_COUNT_HW_REF_CPU_CYCLES = 9, 59607ca46eSDavid Howells 60607ca46eSDavid Howells PERF_COUNT_HW_MAX, /* non-ABI */ 61607ca46eSDavid Howells }; 62607ca46eSDavid Howells 63607ca46eSDavid Howells /* 64607ca46eSDavid Howells * Generalized hardware cache events: 65607ca46eSDavid Howells * 66607ca46eSDavid Howells * { L1-D, L1-I, LLC, ITLB, DTLB, BPU, NODE } x 67607ca46eSDavid Howells * { read, write, prefetch } x 68607ca46eSDavid Howells * { accesses, misses } 69607ca46eSDavid Howells */ 70607ca46eSDavid Howells enum perf_hw_cache_id { 71607ca46eSDavid Howells PERF_COUNT_HW_CACHE_L1D = 0, 72607ca46eSDavid Howells PERF_COUNT_HW_CACHE_L1I = 1, 73607ca46eSDavid Howells PERF_COUNT_HW_CACHE_LL = 2, 74607ca46eSDavid Howells PERF_COUNT_HW_CACHE_DTLB = 3, 75607ca46eSDavid Howells PERF_COUNT_HW_CACHE_ITLB = 4, 76607ca46eSDavid Howells PERF_COUNT_HW_CACHE_BPU = 5, 77607ca46eSDavid Howells PERF_COUNT_HW_CACHE_NODE = 6, 78607ca46eSDavid Howells 79607ca46eSDavid Howells PERF_COUNT_HW_CACHE_MAX, /* non-ABI */ 80607ca46eSDavid Howells }; 81607ca46eSDavid Howells 82607ca46eSDavid Howells enum perf_hw_cache_op_id { 83607ca46eSDavid Howells PERF_COUNT_HW_CACHE_OP_READ = 0, 84607ca46eSDavid Howells PERF_COUNT_HW_CACHE_OP_WRITE = 1, 85607ca46eSDavid Howells PERF_COUNT_HW_CACHE_OP_PREFETCH = 2, 86607ca46eSDavid Howells 87607ca46eSDavid Howells PERF_COUNT_HW_CACHE_OP_MAX, /* non-ABI */ 88607ca46eSDavid Howells }; 89607ca46eSDavid Howells 90607ca46eSDavid Howells enum perf_hw_cache_op_result_id { 91607ca46eSDavid Howells PERF_COUNT_HW_CACHE_RESULT_ACCESS = 0, 92607ca46eSDavid Howells PERF_COUNT_HW_CACHE_RESULT_MISS = 1, 93607ca46eSDavid Howells 94607ca46eSDavid Howells PERF_COUNT_HW_CACHE_RESULT_MAX, /* non-ABI */ 95607ca46eSDavid Howells }; 96607ca46eSDavid Howells 97607ca46eSDavid Howells /* 98607ca46eSDavid Howells * Special "software" events provided by the kernel, even if the hardware 99607ca46eSDavid Howells * does not support performance events. These events measure various 100607ca46eSDavid Howells * physical and sw events of the kernel (and allow the profiling of them as 101607ca46eSDavid Howells * well): 102607ca46eSDavid Howells */ 103607ca46eSDavid Howells enum perf_sw_ids { 104607ca46eSDavid Howells PERF_COUNT_SW_CPU_CLOCK = 0, 105607ca46eSDavid Howells PERF_COUNT_SW_TASK_CLOCK = 1, 106607ca46eSDavid Howells PERF_COUNT_SW_PAGE_FAULTS = 2, 107607ca46eSDavid Howells PERF_COUNT_SW_CONTEXT_SWITCHES = 3, 108607ca46eSDavid Howells PERF_COUNT_SW_CPU_MIGRATIONS = 4, 109607ca46eSDavid Howells PERF_COUNT_SW_PAGE_FAULTS_MIN = 5, 110607ca46eSDavid Howells PERF_COUNT_SW_PAGE_FAULTS_MAJ = 6, 111607ca46eSDavid Howells PERF_COUNT_SW_ALIGNMENT_FAULTS = 7, 112607ca46eSDavid Howells PERF_COUNT_SW_EMULATION_FAULTS = 8, 113fa0097eeSAdrian Hunter PERF_COUNT_SW_DUMMY = 9, 114a43eec30SAlexei Starovoitov PERF_COUNT_SW_BPF_OUTPUT = 10, 115607ca46eSDavid Howells 116607ca46eSDavid Howells PERF_COUNT_SW_MAX, /* non-ABI */ 117607ca46eSDavid Howells }; 118607ca46eSDavid Howells 119607ca46eSDavid Howells /* 120607ca46eSDavid Howells * Bits that can be set in attr.sample_type to request information 121607ca46eSDavid Howells * in the overflow packets. 122607ca46eSDavid Howells */ 123607ca46eSDavid Howells enum perf_event_sample_format { 124607ca46eSDavid Howells PERF_SAMPLE_IP = 1U << 0, 125607ca46eSDavid Howells PERF_SAMPLE_TID = 1U << 1, 126607ca46eSDavid Howells PERF_SAMPLE_TIME = 1U << 2, 127607ca46eSDavid Howells PERF_SAMPLE_ADDR = 1U << 3, 128607ca46eSDavid Howells PERF_SAMPLE_READ = 1U << 4, 129607ca46eSDavid Howells PERF_SAMPLE_CALLCHAIN = 1U << 5, 130607ca46eSDavid Howells PERF_SAMPLE_ID = 1U << 6, 131607ca46eSDavid Howells PERF_SAMPLE_CPU = 1U << 7, 132607ca46eSDavid Howells PERF_SAMPLE_PERIOD = 1U << 8, 133607ca46eSDavid Howells PERF_SAMPLE_STREAM_ID = 1U << 9, 134607ca46eSDavid Howells PERF_SAMPLE_RAW = 1U << 10, 135607ca46eSDavid Howells PERF_SAMPLE_BRANCH_STACK = 1U << 11, 136607ca46eSDavid Howells PERF_SAMPLE_REGS_USER = 1U << 12, 137607ca46eSDavid Howells PERF_SAMPLE_STACK_USER = 1U << 13, 138c3feedf2SAndi Kleen PERF_SAMPLE_WEIGHT = 1U << 14, 139d6be9ad6SStephane Eranian PERF_SAMPLE_DATA_SRC = 1U << 15, 140ff3d527cSAdrian Hunter PERF_SAMPLE_IDENTIFIER = 1U << 16, 141fdfbbd07SAndi Kleen PERF_SAMPLE_TRANSACTION = 1U << 17, 14260e2364eSStephane Eranian PERF_SAMPLE_REGS_INTR = 1U << 18, 143fc7ce9c7SKan Liang PERF_SAMPLE_PHYS_ADDR = 1U << 19, 144*a4faf00dSAlexander Shishkin PERF_SAMPLE_AUX = 1U << 20, 145607ca46eSDavid Howells 146*a4faf00dSAlexander Shishkin PERF_SAMPLE_MAX = 1U << 21, /* non-ABI */ 1476cbc304fSPeter Zijlstra 14809121255SPeter Zijlstra __PERF_SAMPLE_CALLCHAIN_EARLY = 1ULL << 63, /* non-ABI; internal use */ 149607ca46eSDavid Howells }; 150607ca46eSDavid Howells 151607ca46eSDavid Howells /* 152607ca46eSDavid Howells * values to program into branch_sample_type when PERF_SAMPLE_BRANCH is set 153607ca46eSDavid Howells * 154607ca46eSDavid Howells * If the user does not pass priv level information via branch_sample_type, 155607ca46eSDavid Howells * the kernel uses the event's priv level. Branch and event priv levels do 156607ca46eSDavid Howells * not have to match. Branch priv level is checked for permissions. 157607ca46eSDavid Howells * 158607ca46eSDavid Howells * The branch types can be combined, however BRANCH_ANY covers all types 159607ca46eSDavid Howells * of branches and therefore it supersedes all the other types. 160607ca46eSDavid Howells */ 16127ac905bSYan, Zheng enum perf_branch_sample_type_shift { 16227ac905bSYan, Zheng PERF_SAMPLE_BRANCH_USER_SHIFT = 0, /* user branches */ 16327ac905bSYan, Zheng PERF_SAMPLE_BRANCH_KERNEL_SHIFT = 1, /* kernel branches */ 16427ac905bSYan, Zheng PERF_SAMPLE_BRANCH_HV_SHIFT = 2, /* hypervisor branches */ 16527ac905bSYan, Zheng 16627ac905bSYan, Zheng PERF_SAMPLE_BRANCH_ANY_SHIFT = 3, /* any branch types */ 16727ac905bSYan, Zheng PERF_SAMPLE_BRANCH_ANY_CALL_SHIFT = 4, /* any call branch */ 16827ac905bSYan, Zheng PERF_SAMPLE_BRANCH_ANY_RETURN_SHIFT = 5, /* any return branch */ 16927ac905bSYan, Zheng PERF_SAMPLE_BRANCH_IND_CALL_SHIFT = 6, /* indirect calls */ 17027ac905bSYan, Zheng PERF_SAMPLE_BRANCH_ABORT_TX_SHIFT = 7, /* transaction aborts */ 17127ac905bSYan, Zheng PERF_SAMPLE_BRANCH_IN_TX_SHIFT = 8, /* in transaction */ 17227ac905bSYan, Zheng PERF_SAMPLE_BRANCH_NO_TX_SHIFT = 9, /* not in transaction */ 17327ac905bSYan, Zheng PERF_SAMPLE_BRANCH_COND_SHIFT = 10, /* conditional branches */ 17427ac905bSYan, Zheng 1752c44b193SPeter Zijlstra PERF_SAMPLE_BRANCH_CALL_STACK_SHIFT = 11, /* call/ret stack */ 176c9fdfa14SStephane Eranian PERF_SAMPLE_BRANCH_IND_JUMP_SHIFT = 12, /* indirect jumps */ 177c229bf9dSStephane Eranian PERF_SAMPLE_BRANCH_CALL_SHIFT = 13, /* direct call */ 1782c44b193SPeter Zijlstra 179b16a5b52SAndi Kleen PERF_SAMPLE_BRANCH_NO_FLAGS_SHIFT = 14, /* no flags */ 180b16a5b52SAndi Kleen PERF_SAMPLE_BRANCH_NO_CYCLES_SHIFT = 15, /* no cycles */ 181b16a5b52SAndi Kleen 182eb0baf8aSJin Yao PERF_SAMPLE_BRANCH_TYPE_SAVE_SHIFT = 16, /* save branch type */ 183eb0baf8aSJin Yao 18427ac905bSYan, Zheng PERF_SAMPLE_BRANCH_MAX_SHIFT /* non-ABI */ 18527ac905bSYan, Zheng }; 18627ac905bSYan, Zheng 187607ca46eSDavid Howells enum perf_branch_sample_type { 18827ac905bSYan, Zheng PERF_SAMPLE_BRANCH_USER = 1U << PERF_SAMPLE_BRANCH_USER_SHIFT, 18927ac905bSYan, Zheng PERF_SAMPLE_BRANCH_KERNEL = 1U << PERF_SAMPLE_BRANCH_KERNEL_SHIFT, 19027ac905bSYan, Zheng PERF_SAMPLE_BRANCH_HV = 1U << PERF_SAMPLE_BRANCH_HV_SHIFT, 191607ca46eSDavid Howells 19227ac905bSYan, Zheng PERF_SAMPLE_BRANCH_ANY = 1U << PERF_SAMPLE_BRANCH_ANY_SHIFT, 1932c44b193SPeter Zijlstra PERF_SAMPLE_BRANCH_ANY_CALL = 1U << PERF_SAMPLE_BRANCH_ANY_CALL_SHIFT, 1942c44b193SPeter Zijlstra PERF_SAMPLE_BRANCH_ANY_RETURN = 1U << PERF_SAMPLE_BRANCH_ANY_RETURN_SHIFT, 1952c44b193SPeter Zijlstra PERF_SAMPLE_BRANCH_IND_CALL = 1U << PERF_SAMPLE_BRANCH_IND_CALL_SHIFT, 1962c44b193SPeter Zijlstra PERF_SAMPLE_BRANCH_ABORT_TX = 1U << PERF_SAMPLE_BRANCH_ABORT_TX_SHIFT, 19727ac905bSYan, Zheng PERF_SAMPLE_BRANCH_IN_TX = 1U << PERF_SAMPLE_BRANCH_IN_TX_SHIFT, 19827ac905bSYan, Zheng PERF_SAMPLE_BRANCH_NO_TX = 1U << PERF_SAMPLE_BRANCH_NO_TX_SHIFT, 19927ac905bSYan, Zheng PERF_SAMPLE_BRANCH_COND = 1U << PERF_SAMPLE_BRANCH_COND_SHIFT, 200607ca46eSDavid Howells 2012c44b193SPeter Zijlstra PERF_SAMPLE_BRANCH_CALL_STACK = 1U << PERF_SAMPLE_BRANCH_CALL_STACK_SHIFT, 202c9fdfa14SStephane Eranian PERF_SAMPLE_BRANCH_IND_JUMP = 1U << PERF_SAMPLE_BRANCH_IND_JUMP_SHIFT, 203c229bf9dSStephane Eranian PERF_SAMPLE_BRANCH_CALL = 1U << PERF_SAMPLE_BRANCH_CALL_SHIFT, 2042c44b193SPeter Zijlstra 205b16a5b52SAndi Kleen PERF_SAMPLE_BRANCH_NO_FLAGS = 1U << PERF_SAMPLE_BRANCH_NO_FLAGS_SHIFT, 206b16a5b52SAndi Kleen PERF_SAMPLE_BRANCH_NO_CYCLES = 1U << PERF_SAMPLE_BRANCH_NO_CYCLES_SHIFT, 207b16a5b52SAndi Kleen 208eb0baf8aSJin Yao PERF_SAMPLE_BRANCH_TYPE_SAVE = 209eb0baf8aSJin Yao 1U << PERF_SAMPLE_BRANCH_TYPE_SAVE_SHIFT, 210eb0baf8aSJin Yao 21127ac905bSYan, Zheng PERF_SAMPLE_BRANCH_MAX = 1U << PERF_SAMPLE_BRANCH_MAX_SHIFT, 212607ca46eSDavid Howells }; 213607ca46eSDavid Howells 214eb0baf8aSJin Yao /* 215eb0baf8aSJin Yao * Common flow change classification 216eb0baf8aSJin Yao */ 217eb0baf8aSJin Yao enum { 218eb0baf8aSJin Yao PERF_BR_UNKNOWN = 0, /* unknown */ 219eb0baf8aSJin Yao PERF_BR_COND = 1, /* conditional */ 220eb0baf8aSJin Yao PERF_BR_UNCOND = 2, /* unconditional */ 221eb0baf8aSJin Yao PERF_BR_IND = 3, /* indirect */ 222eb0baf8aSJin Yao PERF_BR_CALL = 4, /* function call */ 223eb0baf8aSJin Yao PERF_BR_IND_CALL = 5, /* indirect function call */ 224eb0baf8aSJin Yao PERF_BR_RET = 6, /* function return */ 225eb0baf8aSJin Yao PERF_BR_SYSCALL = 7, /* syscall */ 226eb0baf8aSJin Yao PERF_BR_SYSRET = 8, /* syscall return */ 227eb0baf8aSJin Yao PERF_BR_COND_CALL = 9, /* conditional function call */ 228eb0baf8aSJin Yao PERF_BR_COND_RET = 10, /* conditional function return */ 229eb0baf8aSJin Yao PERF_BR_MAX, 230eb0baf8aSJin Yao }; 231eb0baf8aSJin Yao 232607ca46eSDavid Howells #define PERF_SAMPLE_BRANCH_PLM_ALL \ 233607ca46eSDavid Howells (PERF_SAMPLE_BRANCH_USER|\ 234607ca46eSDavid Howells PERF_SAMPLE_BRANCH_KERNEL|\ 235607ca46eSDavid Howells PERF_SAMPLE_BRANCH_HV) 236607ca46eSDavid Howells 237607ca46eSDavid Howells /* 238607ca46eSDavid Howells * Values to determine ABI of the registers dump. 239607ca46eSDavid Howells */ 240607ca46eSDavid Howells enum perf_sample_regs_abi { 241607ca46eSDavid Howells PERF_SAMPLE_REGS_ABI_NONE = 0, 242607ca46eSDavid Howells PERF_SAMPLE_REGS_ABI_32 = 1, 243607ca46eSDavid Howells PERF_SAMPLE_REGS_ABI_64 = 2, 244607ca46eSDavid Howells }; 245607ca46eSDavid Howells 246607ca46eSDavid Howells /* 247fdfbbd07SAndi Kleen * Values for the memory transaction event qualifier, mostly for 248fdfbbd07SAndi Kleen * abort events. Multiple bits can be set. 249fdfbbd07SAndi Kleen */ 250fdfbbd07SAndi Kleen enum { 251fdfbbd07SAndi Kleen PERF_TXN_ELISION = (1 << 0), /* From elision */ 252fdfbbd07SAndi Kleen PERF_TXN_TRANSACTION = (1 << 1), /* From transaction */ 253fdfbbd07SAndi Kleen PERF_TXN_SYNC = (1 << 2), /* Instruction is related */ 254fdfbbd07SAndi Kleen PERF_TXN_ASYNC = (1 << 3), /* Instruction not related */ 255fdfbbd07SAndi Kleen PERF_TXN_RETRY = (1 << 4), /* Retry possible */ 256fdfbbd07SAndi Kleen PERF_TXN_CONFLICT = (1 << 5), /* Conflict abort */ 257fdfbbd07SAndi Kleen PERF_TXN_CAPACITY_WRITE = (1 << 6), /* Capacity write abort */ 258fdfbbd07SAndi Kleen PERF_TXN_CAPACITY_READ = (1 << 7), /* Capacity read abort */ 259fdfbbd07SAndi Kleen 260fdfbbd07SAndi Kleen PERF_TXN_MAX = (1 << 8), /* non-ABI */ 261fdfbbd07SAndi Kleen 262fdfbbd07SAndi Kleen /* bits 32..63 are reserved for the abort code */ 263fdfbbd07SAndi Kleen 264fdfbbd07SAndi Kleen PERF_TXN_ABORT_MASK = (0xffffffffULL << 32), 265fdfbbd07SAndi Kleen PERF_TXN_ABORT_SHIFT = 32, 266fdfbbd07SAndi Kleen }; 267fdfbbd07SAndi Kleen 268fdfbbd07SAndi Kleen /* 269607ca46eSDavid Howells * The format of the data returned by read() on a perf event fd, 270607ca46eSDavid Howells * as specified by attr.read_format: 271607ca46eSDavid Howells * 272607ca46eSDavid Howells * struct read_format { 273607ca46eSDavid Howells * { u64 value; 274607ca46eSDavid Howells * { u64 time_enabled; } && PERF_FORMAT_TOTAL_TIME_ENABLED 275607ca46eSDavid Howells * { u64 time_running; } && PERF_FORMAT_TOTAL_TIME_RUNNING 276607ca46eSDavid Howells * { u64 id; } && PERF_FORMAT_ID 277607ca46eSDavid Howells * } && !PERF_FORMAT_GROUP 278607ca46eSDavid Howells * 279607ca46eSDavid Howells * { u64 nr; 280607ca46eSDavid Howells * { u64 time_enabled; } && PERF_FORMAT_TOTAL_TIME_ENABLED 281607ca46eSDavid Howells * { u64 time_running; } && PERF_FORMAT_TOTAL_TIME_RUNNING 282607ca46eSDavid Howells * { u64 value; 283607ca46eSDavid Howells * { u64 id; } && PERF_FORMAT_ID 284607ca46eSDavid Howells * } cntr[nr]; 285607ca46eSDavid Howells * } && PERF_FORMAT_GROUP 286607ca46eSDavid Howells * }; 287607ca46eSDavid Howells */ 288607ca46eSDavid Howells enum perf_event_read_format { 289607ca46eSDavid Howells PERF_FORMAT_TOTAL_TIME_ENABLED = 1U << 0, 290607ca46eSDavid Howells PERF_FORMAT_TOTAL_TIME_RUNNING = 1U << 1, 291607ca46eSDavid Howells PERF_FORMAT_ID = 1U << 2, 292607ca46eSDavid Howells PERF_FORMAT_GROUP = 1U << 3, 293607ca46eSDavid Howells 294607ca46eSDavid Howells PERF_FORMAT_MAX = 1U << 4, /* non-ABI */ 295607ca46eSDavid Howells }; 296607ca46eSDavid Howells 297607ca46eSDavid Howells #define PERF_ATTR_SIZE_VER0 64 /* sizeof first published struct */ 298607ca46eSDavid Howells #define PERF_ATTR_SIZE_VER1 72 /* add: config2 */ 299607ca46eSDavid Howells #define PERF_ATTR_SIZE_VER2 80 /* add: branch_sample_type */ 300607ca46eSDavid Howells #define PERF_ATTR_SIZE_VER3 96 /* add: sample_regs_user */ 301607ca46eSDavid Howells /* add: sample_stack_user */ 30260e2364eSStephane Eranian #define PERF_ATTR_SIZE_VER4 104 /* add: sample_regs_intr */ 3031a594131SAlexander Shishkin #define PERF_ATTR_SIZE_VER5 112 /* add: aux_watermark */ 304*a4faf00dSAlexander Shishkin #define PERF_ATTR_SIZE_VER6 120 /* add: aux_sample_size */ 305607ca46eSDavid Howells 306607ca46eSDavid Howells /* 307607ca46eSDavid Howells * Hardware event_id to monitor via a performance monitoring event: 30897c79a38SArnaldo Carvalho de Melo * 30997c79a38SArnaldo Carvalho de Melo * @sample_max_stack: Max number of frame pointers in a callchain, 31097c79a38SArnaldo Carvalho de Melo * should be < /proc/sys/kernel/perf_event_max_stack 311607ca46eSDavid Howells */ 312607ca46eSDavid Howells struct perf_event_attr { 313607ca46eSDavid Howells 314607ca46eSDavid Howells /* 315607ca46eSDavid Howells * Major type: hardware/software/tracepoint/etc. 316607ca46eSDavid Howells */ 317607ca46eSDavid Howells __u32 type; 318607ca46eSDavid Howells 319607ca46eSDavid Howells /* 320607ca46eSDavid Howells * Size of the attr structure, for fwd/bwd compat. 321607ca46eSDavid Howells */ 322607ca46eSDavid Howells __u32 size; 323607ca46eSDavid Howells 324607ca46eSDavid Howells /* 325607ca46eSDavid Howells * Type specific configuration information. 326607ca46eSDavid Howells */ 327607ca46eSDavid Howells __u64 config; 328607ca46eSDavid Howells 329607ca46eSDavid Howells union { 330607ca46eSDavid Howells __u64 sample_period; 331607ca46eSDavid Howells __u64 sample_freq; 332607ca46eSDavid Howells }; 333607ca46eSDavid Howells 334607ca46eSDavid Howells __u64 sample_type; 335607ca46eSDavid Howells __u64 read_format; 336607ca46eSDavid Howells 337607ca46eSDavid Howells __u64 disabled : 1, /* off by default */ 338607ca46eSDavid Howells inherit : 1, /* children inherit it */ 339607ca46eSDavid Howells pinned : 1, /* must always be on PMU */ 340607ca46eSDavid Howells exclusive : 1, /* only group on PMU */ 341607ca46eSDavid Howells exclude_user : 1, /* don't count user */ 342607ca46eSDavid Howells exclude_kernel : 1, /* ditto kernel */ 343607ca46eSDavid Howells exclude_hv : 1, /* ditto hypervisor */ 344607ca46eSDavid Howells exclude_idle : 1, /* don't count when idle */ 345607ca46eSDavid Howells mmap : 1, /* include mmap data */ 346607ca46eSDavid Howells comm : 1, /* include comm data */ 347607ca46eSDavid Howells freq : 1, /* use freq, not period */ 348607ca46eSDavid Howells inherit_stat : 1, /* per task counts */ 349607ca46eSDavid Howells enable_on_exec : 1, /* next exec enables */ 350607ca46eSDavid Howells task : 1, /* trace fork/exit */ 351607ca46eSDavid Howells watermark : 1, /* wakeup_watermark */ 352607ca46eSDavid Howells /* 353607ca46eSDavid Howells * precise_ip: 354607ca46eSDavid Howells * 355607ca46eSDavid Howells * 0 - SAMPLE_IP can have arbitrary skid 356607ca46eSDavid Howells * 1 - SAMPLE_IP must have constant skid 357607ca46eSDavid Howells * 2 - SAMPLE_IP requested to have 0 skid 358607ca46eSDavid Howells * 3 - SAMPLE_IP must have 0 skid 359607ca46eSDavid Howells * 360607ca46eSDavid Howells * See also PERF_RECORD_MISC_EXACT_IP 361607ca46eSDavid Howells */ 362607ca46eSDavid Howells precise_ip : 2, /* skid constraint */ 363607ca46eSDavid Howells mmap_data : 1, /* non-exec mmap data */ 364607ca46eSDavid Howells sample_id_all : 1, /* sample_type all events */ 365607ca46eSDavid Howells 366607ca46eSDavid Howells exclude_host : 1, /* don't count in host */ 367607ca46eSDavid Howells exclude_guest : 1, /* don't count in guest */ 368607ca46eSDavid Howells 369607ca46eSDavid Howells exclude_callchain_kernel : 1, /* exclude kernel callchains */ 370607ca46eSDavid Howells exclude_callchain_user : 1, /* exclude user callchains */ 37113d7a241SStephane Eranian mmap2 : 1, /* include mmap with inode data */ 37282b89778SAdrian Hunter comm_exec : 1, /* flag comm events that are due to an exec */ 37334f43927SPeter Zijlstra use_clockid : 1, /* use @clockid for time fields */ 37445ac1403SAdrian Hunter context_switch : 1, /* context switch data */ 3759ecda41aSWang Nan write_backward : 1, /* Write ring buffer from end to beginning */ 376e4222673SHari Bathini namespaces : 1, /* include namespaces data */ 37776193a94SSong Liu ksymbol : 1, /* include ksymbol events */ 3786ee52e2aSSong Liu bpf_event : 1, /* include bpf events */ 379ab43762eSAlexander Shishkin aux_output : 1, /* generate AUX records instead of events */ 380ab43762eSAlexander Shishkin __reserved_1 : 32; 381607ca46eSDavid Howells 382607ca46eSDavid Howells union { 383607ca46eSDavid Howells __u32 wakeup_events; /* wakeup every n events */ 384607ca46eSDavid Howells __u32 wakeup_watermark; /* bytes before wakeup */ 385607ca46eSDavid Howells }; 386607ca46eSDavid Howells 387607ca46eSDavid Howells __u32 bp_type; 388607ca46eSDavid Howells union { 389607ca46eSDavid Howells __u64 bp_addr; 39065074d43SSong Liu __u64 kprobe_func; /* for perf_kprobe */ 39165074d43SSong Liu __u64 uprobe_path; /* for perf_uprobe */ 392607ca46eSDavid Howells __u64 config1; /* extension of config */ 393607ca46eSDavid Howells }; 394607ca46eSDavid Howells union { 395607ca46eSDavid Howells __u64 bp_len; 39665074d43SSong Liu __u64 kprobe_addr; /* when kprobe_func == NULL */ 39765074d43SSong Liu __u64 probe_offset; /* for perf_[k,u]probe */ 398607ca46eSDavid Howells __u64 config2; /* extension of config1 */ 399607ca46eSDavid Howells }; 400607ca46eSDavid Howells __u64 branch_sample_type; /* enum perf_branch_sample_type */ 401607ca46eSDavid Howells 402607ca46eSDavid Howells /* 403607ca46eSDavid Howells * Defines set of user regs to dump on samples. 404607ca46eSDavid Howells * See asm/perf_regs.h for details. 405607ca46eSDavid Howells */ 406607ca46eSDavid Howells __u64 sample_regs_user; 407607ca46eSDavid Howells 408607ca46eSDavid Howells /* 409607ca46eSDavid Howells * Defines size of the user stack to dump on samples. 410607ca46eSDavid Howells */ 411607ca46eSDavid Howells __u32 sample_stack_user; 412607ca46eSDavid Howells 41334f43927SPeter Zijlstra __s32 clockid; 41460e2364eSStephane Eranian /* 41560e2364eSStephane Eranian * Defines set of regs to dump for each sample 41660e2364eSStephane Eranian * state captured on: 41760e2364eSStephane Eranian * - precise = 0: PMU interrupt 41860e2364eSStephane Eranian * - precise > 0: sampled instruction 41960e2364eSStephane Eranian * 42060e2364eSStephane Eranian * See asm/perf_regs.h for details. 42160e2364eSStephane Eranian */ 42260e2364eSStephane Eranian __u64 sample_regs_intr; 4231a594131SAlexander Shishkin 4241a594131SAlexander Shishkin /* 4251a594131SAlexander Shishkin * Wakeup watermark for AUX area 4261a594131SAlexander Shishkin */ 4271a594131SAlexander Shishkin __u32 aux_watermark; 42897c79a38SArnaldo Carvalho de Melo __u16 sample_max_stack; 429*a4faf00dSAlexander Shishkin __u16 __reserved_2; 430*a4faf00dSAlexander Shishkin __u32 aux_sample_size; 431*a4faf00dSAlexander Shishkin __u32 __reserved_3; 432607ca46eSDavid Howells }; 433607ca46eSDavid Howells 434f371b304SYonghong Song /* 435f371b304SYonghong Song * Structure used by below PERF_EVENT_IOC_QUERY_BPF command 436f371b304SYonghong Song * to query bpf programs attached to the same perf tracepoint 437f371b304SYonghong Song * as the given perf event. 438f371b304SYonghong Song */ 439f371b304SYonghong Song struct perf_event_query_bpf { 440f371b304SYonghong Song /* 441f371b304SYonghong Song * The below ids array length 442f371b304SYonghong Song */ 443f371b304SYonghong Song __u32 ids_len; 444f371b304SYonghong Song /* 445f371b304SYonghong Song * Set by the kernel to indicate the number of 446f371b304SYonghong Song * available programs 447f371b304SYonghong Song */ 448f371b304SYonghong Song __u32 prog_cnt; 449f371b304SYonghong Song /* 450f371b304SYonghong Song * User provided buffer to store program ids 451f371b304SYonghong Song */ 452f371b304SYonghong Song __u32 ids[0]; 453f371b304SYonghong Song }; 454f371b304SYonghong Song 455607ca46eSDavid Howells /* 456607ca46eSDavid Howells * Ioctls that can be done on a perf event fd: 457607ca46eSDavid Howells */ 458607ca46eSDavid Howells #define PERF_EVENT_IOC_ENABLE _IO ('$', 0) 459607ca46eSDavid Howells #define PERF_EVENT_IOC_DISABLE _IO ('$', 1) 460607ca46eSDavid Howells #define PERF_EVENT_IOC_REFRESH _IO ('$', 2) 461607ca46eSDavid Howells #define PERF_EVENT_IOC_RESET _IO ('$', 3) 462607ca46eSDavid Howells #define PERF_EVENT_IOC_PERIOD _IOW('$', 4, __u64) 463607ca46eSDavid Howells #define PERF_EVENT_IOC_SET_OUTPUT _IO ('$', 5) 464607ca46eSDavid Howells #define PERF_EVENT_IOC_SET_FILTER _IOW('$', 6, char *) 465a8e0108cSVince Weaver #define PERF_EVENT_IOC_ID _IOR('$', 7, __u64 *) 4662541517cSAlexei Starovoitov #define PERF_EVENT_IOC_SET_BPF _IOW('$', 8, __u32) 46786e7972fSWang Nan #define PERF_EVENT_IOC_PAUSE_OUTPUT _IOW('$', 9, __u32) 468f371b304SYonghong Song #define PERF_EVENT_IOC_QUERY_BPF _IOWR('$', 10, struct perf_event_query_bpf *) 46932ff77e8SMilind Chabbi #define PERF_EVENT_IOC_MODIFY_ATTRIBUTES _IOW('$', 11, struct perf_event_attr *) 470607ca46eSDavid Howells 471607ca46eSDavid Howells enum perf_event_ioc_flags { 472607ca46eSDavid Howells PERF_IOC_FLAG_GROUP = 1U << 0, 473607ca46eSDavid Howells }; 474607ca46eSDavid Howells 475607ca46eSDavid Howells /* 476607ca46eSDavid Howells * Structure of the page that can be mapped via mmap 477607ca46eSDavid Howells */ 478607ca46eSDavid Howells struct perf_event_mmap_page { 479607ca46eSDavid Howells __u32 version; /* version number of this structure */ 480607ca46eSDavid Howells __u32 compat_version; /* lowest version this is compat with */ 481607ca46eSDavid Howells 482607ca46eSDavid Howells /* 483607ca46eSDavid Howells * Bits needed to read the hw events in user-space. 484607ca46eSDavid Howells * 485b438b1abSAndy Lutomirski * u32 seq, time_mult, time_shift, index, width; 486607ca46eSDavid Howells * u64 count, enabled, running; 487607ca46eSDavid Howells * u64 cyc, time_offset; 488607ca46eSDavid Howells * s64 pmc = 0; 489607ca46eSDavid Howells * 490607ca46eSDavid Howells * do { 491607ca46eSDavid Howells * seq = pc->lock; 492607ca46eSDavid Howells * barrier() 493607ca46eSDavid Howells * 494607ca46eSDavid Howells * enabled = pc->time_enabled; 495607ca46eSDavid Howells * running = pc->time_running; 496607ca46eSDavid Howells * 497607ca46eSDavid Howells * if (pc->cap_usr_time && enabled != running) { 498607ca46eSDavid Howells * cyc = rdtsc(); 499607ca46eSDavid Howells * time_offset = pc->time_offset; 500607ca46eSDavid Howells * time_mult = pc->time_mult; 501607ca46eSDavid Howells * time_shift = pc->time_shift; 502607ca46eSDavid Howells * } 503607ca46eSDavid Howells * 504b438b1abSAndy Lutomirski * index = pc->index; 505607ca46eSDavid Howells * count = pc->offset; 506b438b1abSAndy Lutomirski * if (pc->cap_user_rdpmc && index) { 507607ca46eSDavid Howells * width = pc->pmc_width; 508b438b1abSAndy Lutomirski * pmc = rdpmc(index - 1); 509607ca46eSDavid Howells * } 510607ca46eSDavid Howells * 511607ca46eSDavid Howells * barrier(); 512607ca46eSDavid Howells * } while (pc->lock != seq); 513607ca46eSDavid Howells * 514607ca46eSDavid Howells * NOTE: for obvious reason this only works on self-monitoring 515607ca46eSDavid Howells * processes. 516607ca46eSDavid Howells */ 517607ca46eSDavid Howells __u32 lock; /* seqlock for synchronization */ 518607ca46eSDavid Howells __u32 index; /* hardware event identifier */ 519607ca46eSDavid Howells __s64 offset; /* add to hardware event value */ 520607ca46eSDavid Howells __u64 time_enabled; /* time event active */ 521607ca46eSDavid Howells __u64 time_running; /* time event on cpu */ 522607ca46eSDavid Howells union { 523607ca46eSDavid Howells __u64 capabilities; 524860f085bSAdrian Hunter struct { 525fa731587SPeter Zijlstra __u64 cap_bit0 : 1, /* Always 0, deprecated, see commit 860f085b74e9 */ 526fa731587SPeter Zijlstra cap_bit0_is_deprecated : 1, /* Always 1, signals that bit 0 is zero */ 527fa731587SPeter Zijlstra 528fa731587SPeter Zijlstra cap_user_rdpmc : 1, /* The RDPMC instruction can be used to read counts */ 529fa731587SPeter Zijlstra cap_user_time : 1, /* The time_* fields are used */ 530fa731587SPeter Zijlstra cap_user_time_zero : 1, /* The time_zero field is used */ 531fa731587SPeter Zijlstra cap_____res : 59; 532607ca46eSDavid Howells }; 533860f085bSAdrian Hunter }; 534607ca46eSDavid Howells 535607ca46eSDavid Howells /* 536b438b1abSAndy Lutomirski * If cap_user_rdpmc this field provides the bit-width of the value 537607ca46eSDavid Howells * read using the rdpmc() or equivalent instruction. This can be used 538607ca46eSDavid Howells * to sign extend the result like: 539607ca46eSDavid Howells * 540607ca46eSDavid Howells * pmc <<= 64 - width; 541607ca46eSDavid Howells * pmc >>= 64 - width; // signed shift right 542607ca46eSDavid Howells * count += pmc; 543607ca46eSDavid Howells */ 544607ca46eSDavid Howells __u16 pmc_width; 545607ca46eSDavid Howells 546607ca46eSDavid Howells /* 547607ca46eSDavid Howells * If cap_usr_time the below fields can be used to compute the time 548607ca46eSDavid Howells * delta since time_enabled (in ns) using rdtsc or similar. 549607ca46eSDavid Howells * 550607ca46eSDavid Howells * u64 quot, rem; 551607ca46eSDavid Howells * u64 delta; 552607ca46eSDavid Howells * 553607ca46eSDavid Howells * quot = (cyc >> time_shift); 554b9511cd7SAdrian Hunter * rem = cyc & (((u64)1 << time_shift) - 1); 555607ca46eSDavid Howells * delta = time_offset + quot * time_mult + 556607ca46eSDavid Howells * ((rem * time_mult) >> time_shift); 557607ca46eSDavid Howells * 558607ca46eSDavid Howells * Where time_offset,time_mult,time_shift and cyc are read in the 559607ca46eSDavid Howells * seqcount loop described above. This delta can then be added to 560b438b1abSAndy Lutomirski * enabled and possible running (if index), improving the scaling: 561607ca46eSDavid Howells * 562607ca46eSDavid Howells * enabled += delta; 563b438b1abSAndy Lutomirski * if (index) 564607ca46eSDavid Howells * running += delta; 565607ca46eSDavid Howells * 566607ca46eSDavid Howells * quot = count / running; 567607ca46eSDavid Howells * rem = count % running; 568607ca46eSDavid Howells * count = quot * enabled + (rem * enabled) / running; 569607ca46eSDavid Howells */ 570607ca46eSDavid Howells __u16 time_shift; 571607ca46eSDavid Howells __u32 time_mult; 572607ca46eSDavid Howells __u64 time_offset; 573c73deb6aSAdrian Hunter /* 574c73deb6aSAdrian Hunter * If cap_usr_time_zero, the hardware clock (e.g. TSC) can be calculated 575c73deb6aSAdrian Hunter * from sample timestamps. 576c73deb6aSAdrian Hunter * 577c73deb6aSAdrian Hunter * time = timestamp - time_zero; 578c73deb6aSAdrian Hunter * quot = time / time_mult; 579c73deb6aSAdrian Hunter * rem = time % time_mult; 580c73deb6aSAdrian Hunter * cyc = (quot << time_shift) + (rem << time_shift) / time_mult; 581c73deb6aSAdrian Hunter * 582c73deb6aSAdrian Hunter * And vice versa: 583c73deb6aSAdrian Hunter * 584c73deb6aSAdrian Hunter * quot = cyc >> time_shift; 585b9511cd7SAdrian Hunter * rem = cyc & (((u64)1 << time_shift) - 1); 586c73deb6aSAdrian Hunter * timestamp = time_zero + quot * time_mult + 587c73deb6aSAdrian Hunter * ((rem * time_mult) >> time_shift); 588c73deb6aSAdrian Hunter */ 589c73deb6aSAdrian Hunter __u64 time_zero; 590fa731587SPeter Zijlstra __u32 size; /* Header size up to __reserved[] fields. */ 591607ca46eSDavid Howells 592607ca46eSDavid Howells /* 593607ca46eSDavid Howells * Hole for extension of the self monitor capabilities 594607ca46eSDavid Howells */ 595607ca46eSDavid Howells 596fa731587SPeter Zijlstra __u8 __reserved[118*8+4]; /* align to 1k. */ 597607ca46eSDavid Howells 598607ca46eSDavid Howells /* 599607ca46eSDavid Howells * Control data for the mmap() data buffer. 600607ca46eSDavid Howells * 601bf378d34SPeter Zijlstra * User-space reading the @data_head value should issue an smp_rmb(), 602bf378d34SPeter Zijlstra * after reading this value. 603607ca46eSDavid Howells * 604607ca46eSDavid Howells * When the mapping is PROT_WRITE the @data_tail value should be 605bf378d34SPeter Zijlstra * written by userspace to reflect the last read data, after issueing 606bf378d34SPeter Zijlstra * an smp_mb() to separate the data read from the ->data_tail store. 607bf378d34SPeter Zijlstra * In this case the kernel will not over-write unread data. 608bf378d34SPeter Zijlstra * 609bf378d34SPeter Zijlstra * See perf_output_put_handle() for the data ordering. 610e8c6deacSAlexander Shishkin * 611e8c6deacSAlexander Shishkin * data_{offset,size} indicate the location and size of the perf record 612e8c6deacSAlexander Shishkin * buffer within the mmapped area. 613607ca46eSDavid Howells */ 614607ca46eSDavid Howells __u64 data_head; /* head in the data section */ 615607ca46eSDavid Howells __u64 data_tail; /* user-space written tail */ 616e8c6deacSAlexander Shishkin __u64 data_offset; /* where the buffer starts */ 617e8c6deacSAlexander Shishkin __u64 data_size; /* data buffer size */ 61845bfb2e5SPeter Zijlstra 61945bfb2e5SPeter Zijlstra /* 62045bfb2e5SPeter Zijlstra * AUX area is defined by aux_{offset,size} fields that should be set 62145bfb2e5SPeter Zijlstra * by the userspace, so that 62245bfb2e5SPeter Zijlstra * 62345bfb2e5SPeter Zijlstra * aux_offset >= data_offset + data_size 62445bfb2e5SPeter Zijlstra * 62545bfb2e5SPeter Zijlstra * prior to mmap()ing it. Size of the mmap()ed area should be aux_size. 62645bfb2e5SPeter Zijlstra * 62745bfb2e5SPeter Zijlstra * Ring buffer pointers aux_{head,tail} have the same semantics as 62845bfb2e5SPeter Zijlstra * data_{head,tail} and same ordering rules apply. 62945bfb2e5SPeter Zijlstra */ 63045bfb2e5SPeter Zijlstra __u64 aux_head; 63145bfb2e5SPeter Zijlstra __u64 aux_tail; 63245bfb2e5SPeter Zijlstra __u64 aux_offset; 63345bfb2e5SPeter Zijlstra __u64 aux_size; 634607ca46eSDavid Howells }; 635607ca46eSDavid Howells 636607ca46eSDavid Howells #define PERF_RECORD_MISC_CPUMODE_MASK (7 << 0) 637607ca46eSDavid Howells #define PERF_RECORD_MISC_CPUMODE_UNKNOWN (0 << 0) 638607ca46eSDavid Howells #define PERF_RECORD_MISC_KERNEL (1 << 0) 639607ca46eSDavid Howells #define PERF_RECORD_MISC_USER (2 << 0) 640607ca46eSDavid Howells #define PERF_RECORD_MISC_HYPERVISOR (3 << 0) 641607ca46eSDavid Howells #define PERF_RECORD_MISC_GUEST_KERNEL (4 << 0) 642607ca46eSDavid Howells #define PERF_RECORD_MISC_GUEST_USER (5 << 0) 643607ca46eSDavid Howells 64482b89778SAdrian Hunter /* 645930e6fcdSKan Liang * Indicates that /proc/PID/maps parsing are truncated by time out. 646930e6fcdSKan Liang */ 647930e6fcdSKan Liang #define PERF_RECORD_MISC_PROC_MAP_PARSE_TIMEOUT (1 << 12) 648930e6fcdSKan Liang /* 649972c1488SJiri Olsa * Following PERF_RECORD_MISC_* are used on different 650972c1488SJiri Olsa * events, so can reuse the same bit position: 651972c1488SJiri Olsa * 652972c1488SJiri Olsa * PERF_RECORD_MISC_MMAP_DATA - PERF_RECORD_MMAP* events 653972c1488SJiri Olsa * PERF_RECORD_MISC_COMM_EXEC - PERF_RECORD_COMM event 6544f8f382eSDavid Miller * PERF_RECORD_MISC_FORK_EXEC - PERF_RECORD_FORK event (perf internal) 655972c1488SJiri Olsa * PERF_RECORD_MISC_SWITCH_OUT - PERF_RECORD_SWITCH* events 65682b89778SAdrian Hunter */ 6572fe85427SStephane Eranian #define PERF_RECORD_MISC_MMAP_DATA (1 << 13) 65882b89778SAdrian Hunter #define PERF_RECORD_MISC_COMM_EXEC (1 << 13) 6594f8f382eSDavid Miller #define PERF_RECORD_MISC_FORK_EXEC (1 << 13) 66045ac1403SAdrian Hunter #define PERF_RECORD_MISC_SWITCH_OUT (1 << 13) 661607ca46eSDavid Howells /* 662101592b4SAlexey Budankov * These PERF_RECORD_MISC_* flags below are safely reused 663101592b4SAlexey Budankov * for the following events: 664101592b4SAlexey Budankov * 665101592b4SAlexey Budankov * PERF_RECORD_MISC_EXACT_IP - PERF_RECORD_SAMPLE of precise events 666101592b4SAlexey Budankov * PERF_RECORD_MISC_SWITCH_OUT_PREEMPT - PERF_RECORD_SWITCH* events 667101592b4SAlexey Budankov * 668101592b4SAlexey Budankov * 669101592b4SAlexey Budankov * PERF_RECORD_MISC_EXACT_IP: 670607ca46eSDavid Howells * Indicates that the content of PERF_SAMPLE_IP points to 671607ca46eSDavid Howells * the actual instruction that triggered the event. See also 672607ca46eSDavid Howells * perf_event_attr::precise_ip. 673101592b4SAlexey Budankov * 674101592b4SAlexey Budankov * PERF_RECORD_MISC_SWITCH_OUT_PREEMPT: 675101592b4SAlexey Budankov * Indicates that thread was preempted in TASK_RUNNING state. 676607ca46eSDavid Howells */ 677607ca46eSDavid Howells #define PERF_RECORD_MISC_EXACT_IP (1 << 14) 678101592b4SAlexey Budankov #define PERF_RECORD_MISC_SWITCH_OUT_PREEMPT (1 << 14) 679607ca46eSDavid Howells /* 680607ca46eSDavid Howells * Reserve the last bit to indicate some extended misc field 681607ca46eSDavid Howells */ 682607ca46eSDavid Howells #define PERF_RECORD_MISC_EXT_RESERVED (1 << 15) 683607ca46eSDavid Howells 684607ca46eSDavid Howells struct perf_event_header { 685607ca46eSDavid Howells __u32 type; 686607ca46eSDavid Howells __u16 misc; 687607ca46eSDavid Howells __u16 size; 688607ca46eSDavid Howells }; 689607ca46eSDavid Howells 690e4222673SHari Bathini struct perf_ns_link_info { 691e4222673SHari Bathini __u64 dev; 692e4222673SHari Bathini __u64 ino; 693e4222673SHari Bathini }; 694e4222673SHari Bathini 695e4222673SHari Bathini enum { 696e4222673SHari Bathini NET_NS_INDEX = 0, 697e4222673SHari Bathini UTS_NS_INDEX = 1, 698e4222673SHari Bathini IPC_NS_INDEX = 2, 699e4222673SHari Bathini PID_NS_INDEX = 3, 700e4222673SHari Bathini USER_NS_INDEX = 4, 701e4222673SHari Bathini MNT_NS_INDEX = 5, 702e4222673SHari Bathini CGROUP_NS_INDEX = 6, 703e4222673SHari Bathini 704e4222673SHari Bathini NR_NAMESPACES, /* number of available namespaces */ 705e4222673SHari Bathini }; 706e4222673SHari Bathini 707607ca46eSDavid Howells enum perf_event_type { 708607ca46eSDavid Howells 709607ca46eSDavid Howells /* 710607ca46eSDavid Howells * If perf_event_attr.sample_id_all is set then all event types will 711607ca46eSDavid Howells * have the sample_type selected fields related to where/when 712ff3d527cSAdrian Hunter * (identity) an event took place (TID, TIME, ID, STREAM_ID, CPU, 713ff3d527cSAdrian Hunter * IDENTIFIER) described in PERF_RECORD_SAMPLE below, it will be stashed 714ff3d527cSAdrian Hunter * just after the perf_event_header and the fields already present for 715ff3d527cSAdrian Hunter * the existing fields, i.e. at the end of the payload. That way a newer 716ff3d527cSAdrian Hunter * perf.data file will be supported by older perf tools, with these new 717ff3d527cSAdrian Hunter * optional fields being ignored. 718607ca46eSDavid Howells * 719a5cdd40cSPeter Zijlstra * struct sample_id { 720a5cdd40cSPeter Zijlstra * { u32 pid, tid; } && PERF_SAMPLE_TID 721a5cdd40cSPeter Zijlstra * { u64 time; } && PERF_SAMPLE_TIME 722a5cdd40cSPeter Zijlstra * { u64 id; } && PERF_SAMPLE_ID 723a5cdd40cSPeter Zijlstra * { u64 stream_id;} && PERF_SAMPLE_STREAM_ID 724a5cdd40cSPeter Zijlstra * { u32 cpu, res; } && PERF_SAMPLE_CPU 725ff3d527cSAdrian Hunter * { u64 id; } && PERF_SAMPLE_IDENTIFIER 726a5cdd40cSPeter Zijlstra * } && perf_event_attr::sample_id_all 727ff3d527cSAdrian Hunter * 728ff3d527cSAdrian Hunter * Note that PERF_SAMPLE_IDENTIFIER duplicates PERF_SAMPLE_ID. The 729ff3d527cSAdrian Hunter * advantage of PERF_SAMPLE_IDENTIFIER is that its position is fixed 730ff3d527cSAdrian Hunter * relative to header.size. 731a5cdd40cSPeter Zijlstra */ 732a5cdd40cSPeter Zijlstra 733a5cdd40cSPeter Zijlstra /* 734607ca46eSDavid Howells * The MMAP events record the PROT_EXEC mappings so that we can 735607ca46eSDavid Howells * correlate userspace IPs to code. They have the following structure: 736607ca46eSDavid Howells * 737607ca46eSDavid Howells * struct { 738607ca46eSDavid Howells * struct perf_event_header header; 739607ca46eSDavid Howells * 740607ca46eSDavid Howells * u32 pid, tid; 741607ca46eSDavid Howells * u64 addr; 742607ca46eSDavid Howells * u64 len; 743607ca46eSDavid Howells * u64 pgoff; 744607ca46eSDavid Howells * char filename[]; 745c5ecceefSPeter Zijlstra * struct sample_id sample_id; 746607ca46eSDavid Howells * }; 747607ca46eSDavid Howells */ 748607ca46eSDavid Howells PERF_RECORD_MMAP = 1, 749607ca46eSDavid Howells 750607ca46eSDavid Howells /* 751607ca46eSDavid Howells * struct { 752607ca46eSDavid Howells * struct perf_event_header header; 753607ca46eSDavid Howells * u64 id; 754607ca46eSDavid Howells * u64 lost; 755a5cdd40cSPeter Zijlstra * struct sample_id sample_id; 756607ca46eSDavid Howells * }; 757607ca46eSDavid Howells */ 758607ca46eSDavid Howells PERF_RECORD_LOST = 2, 759607ca46eSDavid Howells 760607ca46eSDavid Howells /* 761607ca46eSDavid Howells * struct { 762607ca46eSDavid Howells * struct perf_event_header header; 763607ca46eSDavid Howells * 764607ca46eSDavid Howells * u32 pid, tid; 765607ca46eSDavid Howells * char comm[]; 766a5cdd40cSPeter Zijlstra * struct sample_id sample_id; 767607ca46eSDavid Howells * }; 768607ca46eSDavid Howells */ 769607ca46eSDavid Howells PERF_RECORD_COMM = 3, 770607ca46eSDavid Howells 771607ca46eSDavid Howells /* 772607ca46eSDavid Howells * struct { 773607ca46eSDavid Howells * struct perf_event_header header; 774607ca46eSDavid Howells * u32 pid, ppid; 775607ca46eSDavid Howells * u32 tid, ptid; 776607ca46eSDavid Howells * u64 time; 777a5cdd40cSPeter Zijlstra * struct sample_id sample_id; 778607ca46eSDavid Howells * }; 779607ca46eSDavid Howells */ 780607ca46eSDavid Howells PERF_RECORD_EXIT = 4, 781607ca46eSDavid Howells 782607ca46eSDavid Howells /* 783607ca46eSDavid Howells * struct { 784607ca46eSDavid Howells * struct perf_event_header header; 785607ca46eSDavid Howells * u64 time; 786607ca46eSDavid Howells * u64 id; 787607ca46eSDavid Howells * u64 stream_id; 788a5cdd40cSPeter Zijlstra * struct sample_id sample_id; 789607ca46eSDavid Howells * }; 790607ca46eSDavid Howells */ 791607ca46eSDavid Howells PERF_RECORD_THROTTLE = 5, 792607ca46eSDavid Howells PERF_RECORD_UNTHROTTLE = 6, 793607ca46eSDavid Howells 794607ca46eSDavid Howells /* 795607ca46eSDavid Howells * struct { 796607ca46eSDavid Howells * struct perf_event_header header; 797607ca46eSDavid Howells * u32 pid, ppid; 798607ca46eSDavid Howells * u32 tid, ptid; 799607ca46eSDavid Howells * u64 time; 800a5cdd40cSPeter Zijlstra * struct sample_id sample_id; 801607ca46eSDavid Howells * }; 802607ca46eSDavid Howells */ 803607ca46eSDavid Howells PERF_RECORD_FORK = 7, 804607ca46eSDavid Howells 805607ca46eSDavid Howells /* 806607ca46eSDavid Howells * struct { 807607ca46eSDavid Howells * struct perf_event_header header; 808607ca46eSDavid Howells * u32 pid, tid; 809607ca46eSDavid Howells * 810607ca46eSDavid Howells * struct read_format values; 811a5cdd40cSPeter Zijlstra * struct sample_id sample_id; 812607ca46eSDavid Howells * }; 813607ca46eSDavid Howells */ 814607ca46eSDavid Howells PERF_RECORD_READ = 8, 815607ca46eSDavid Howells 816607ca46eSDavid Howells /* 817607ca46eSDavid Howells * struct { 818607ca46eSDavid Howells * struct perf_event_header header; 819607ca46eSDavid Howells * 820ff3d527cSAdrian Hunter * # 821ff3d527cSAdrian Hunter * # Note that PERF_SAMPLE_IDENTIFIER duplicates PERF_SAMPLE_ID. 822ff3d527cSAdrian Hunter * # The advantage of PERF_SAMPLE_IDENTIFIER is that its position 823ff3d527cSAdrian Hunter * # is fixed relative to header. 824ff3d527cSAdrian Hunter * # 825ff3d527cSAdrian Hunter * 826ff3d527cSAdrian Hunter * { u64 id; } && PERF_SAMPLE_IDENTIFIER 827607ca46eSDavid Howells * { u64 ip; } && PERF_SAMPLE_IP 828607ca46eSDavid Howells * { u32 pid, tid; } && PERF_SAMPLE_TID 829607ca46eSDavid Howells * { u64 time; } && PERF_SAMPLE_TIME 830607ca46eSDavid Howells * { u64 addr; } && PERF_SAMPLE_ADDR 831607ca46eSDavid Howells * { u64 id; } && PERF_SAMPLE_ID 832607ca46eSDavid Howells * { u64 stream_id;} && PERF_SAMPLE_STREAM_ID 833607ca46eSDavid Howells * { u32 cpu, res; } && PERF_SAMPLE_CPU 834607ca46eSDavid Howells * { u64 period; } && PERF_SAMPLE_PERIOD 835607ca46eSDavid Howells * 836607ca46eSDavid Howells * { struct read_format values; } && PERF_SAMPLE_READ 837607ca46eSDavid Howells * 838607ca46eSDavid Howells * { u64 nr, 839607ca46eSDavid Howells * u64 ips[nr]; } && PERF_SAMPLE_CALLCHAIN 840607ca46eSDavid Howells * 841607ca46eSDavid Howells * # 842607ca46eSDavid Howells * # The RAW record below is opaque data wrt the ABI 843607ca46eSDavid Howells * # 844607ca46eSDavid Howells * # That is, the ABI doesn't make any promises wrt to 845607ca46eSDavid Howells * # the stability of its content, it may vary depending 846607ca46eSDavid Howells * # on event, hardware, kernel version and phase of 847607ca46eSDavid Howells * # the moon. 848607ca46eSDavid Howells * # 849607ca46eSDavid Howells * # In other words, PERF_SAMPLE_RAW contents are not an ABI. 850607ca46eSDavid Howells * # 851607ca46eSDavid Howells * 852607ca46eSDavid Howells * { u32 size; 853607ca46eSDavid Howells * char data[size];}&& PERF_SAMPLE_RAW 854607ca46eSDavid Howells * 855b878e7fbSVince Weaver * { u64 nr; 856607ca46eSDavid Howells * { u64 from, to, flags } lbr[nr];} && PERF_SAMPLE_BRANCH_STACK 857607ca46eSDavid Howells * 858607ca46eSDavid Howells * { u64 abi; # enum perf_sample_regs_abi 859607ca46eSDavid Howells * u64 regs[weight(mask)]; } && PERF_SAMPLE_REGS_USER 860607ca46eSDavid Howells * 861607ca46eSDavid Howells * { u64 size; 862607ca46eSDavid Howells * char data[size]; 863607ca46eSDavid Howells * u64 dyn_size; } && PERF_SAMPLE_STACK_USER 864c3feedf2SAndi Kleen * 865c3feedf2SAndi Kleen * { u64 weight; } && PERF_SAMPLE_WEIGHT 866d6be9ad6SStephane Eranian * { u64 data_src; } && PERF_SAMPLE_DATA_SRC 867189b84fbSVince Weaver * { u64 transaction; } && PERF_SAMPLE_TRANSACTION 86860e2364eSStephane Eranian * { u64 abi; # enum perf_sample_regs_abi 86960e2364eSStephane Eranian * u64 regs[weight(mask)]; } && PERF_SAMPLE_REGS_INTR 870fc7ce9c7SKan Liang * { u64 phys_addr;} && PERF_SAMPLE_PHYS_ADDR 871*a4faf00dSAlexander Shishkin * { u64 size; 872*a4faf00dSAlexander Shishkin * char data[size]; } && PERF_SAMPLE_AUX 873607ca46eSDavid Howells * }; 874607ca46eSDavid Howells */ 875607ca46eSDavid Howells PERF_RECORD_SAMPLE = 9, 876607ca46eSDavid Howells 87713d7a241SStephane Eranian /* 87813d7a241SStephane Eranian * The MMAP2 records are an augmented version of MMAP, they add 87913d7a241SStephane Eranian * maj, min, ino numbers to be used to uniquely identify each mapping 88013d7a241SStephane Eranian * 88113d7a241SStephane Eranian * struct { 88213d7a241SStephane Eranian * struct perf_event_header header; 88313d7a241SStephane Eranian * 88413d7a241SStephane Eranian * u32 pid, tid; 88513d7a241SStephane Eranian * u64 addr; 88613d7a241SStephane Eranian * u64 len; 88713d7a241SStephane Eranian * u64 pgoff; 88813d7a241SStephane Eranian * u32 maj; 88913d7a241SStephane Eranian * u32 min; 89013d7a241SStephane Eranian * u64 ino; 89113d7a241SStephane Eranian * u64 ino_generation; 892f972eb63SPeter Zijlstra * u32 prot, flags; 89313d7a241SStephane Eranian * char filename[]; 89413d7a241SStephane Eranian * struct sample_id sample_id; 89513d7a241SStephane Eranian * }; 89613d7a241SStephane Eranian */ 89713d7a241SStephane Eranian PERF_RECORD_MMAP2 = 10, 89813d7a241SStephane Eranian 89968db7e98SAlexander Shishkin /* 90068db7e98SAlexander Shishkin * Records that new data landed in the AUX buffer part. 90168db7e98SAlexander Shishkin * 90268db7e98SAlexander Shishkin * struct { 90368db7e98SAlexander Shishkin * struct perf_event_header header; 90468db7e98SAlexander Shishkin * 90568db7e98SAlexander Shishkin * u64 aux_offset; 90668db7e98SAlexander Shishkin * u64 aux_size; 90768db7e98SAlexander Shishkin * u64 flags; 90868db7e98SAlexander Shishkin * struct sample_id sample_id; 90968db7e98SAlexander Shishkin * }; 91068db7e98SAlexander Shishkin */ 91168db7e98SAlexander Shishkin PERF_RECORD_AUX = 11, 91268db7e98SAlexander Shishkin 913ec0d7729SAlexander Shishkin /* 914ec0d7729SAlexander Shishkin * Indicates that instruction trace has started 915ec0d7729SAlexander Shishkin * 916ec0d7729SAlexander Shishkin * struct { 917ec0d7729SAlexander Shishkin * struct perf_event_header header; 918ec0d7729SAlexander Shishkin * u32 pid; 919ec0d7729SAlexander Shishkin * u32 tid; 92081df978cSJiri Olsa * struct sample_id sample_id; 921ec0d7729SAlexander Shishkin * }; 922ec0d7729SAlexander Shishkin */ 923ec0d7729SAlexander Shishkin PERF_RECORD_ITRACE_START = 12, 924ec0d7729SAlexander Shishkin 925f38b0dbbSKan Liang /* 926f38b0dbbSKan Liang * Records the dropped/lost sample number. 927f38b0dbbSKan Liang * 928f38b0dbbSKan Liang * struct { 929f38b0dbbSKan Liang * struct perf_event_header header; 930f38b0dbbSKan Liang * 931f38b0dbbSKan Liang * u64 lost; 932f38b0dbbSKan Liang * struct sample_id sample_id; 933f38b0dbbSKan Liang * }; 934f38b0dbbSKan Liang */ 935f38b0dbbSKan Liang PERF_RECORD_LOST_SAMPLES = 13, 936f38b0dbbSKan Liang 93745ac1403SAdrian Hunter /* 93845ac1403SAdrian Hunter * Records a context switch in or out (flagged by 93945ac1403SAdrian Hunter * PERF_RECORD_MISC_SWITCH_OUT). See also 94045ac1403SAdrian Hunter * PERF_RECORD_SWITCH_CPU_WIDE. 94145ac1403SAdrian Hunter * 94245ac1403SAdrian Hunter * struct { 94345ac1403SAdrian Hunter * struct perf_event_header header; 94445ac1403SAdrian Hunter * struct sample_id sample_id; 94545ac1403SAdrian Hunter * }; 94645ac1403SAdrian Hunter */ 94745ac1403SAdrian Hunter PERF_RECORD_SWITCH = 14, 94845ac1403SAdrian Hunter 94945ac1403SAdrian Hunter /* 95045ac1403SAdrian Hunter * CPU-wide version of PERF_RECORD_SWITCH with next_prev_pid and 95145ac1403SAdrian Hunter * next_prev_tid that are the next (switching out) or previous 95245ac1403SAdrian Hunter * (switching in) pid/tid. 95345ac1403SAdrian Hunter * 95445ac1403SAdrian Hunter * struct { 95545ac1403SAdrian Hunter * struct perf_event_header header; 95645ac1403SAdrian Hunter * u32 next_prev_pid; 95745ac1403SAdrian Hunter * u32 next_prev_tid; 95845ac1403SAdrian Hunter * struct sample_id sample_id; 95945ac1403SAdrian Hunter * }; 96045ac1403SAdrian Hunter */ 96145ac1403SAdrian Hunter PERF_RECORD_SWITCH_CPU_WIDE = 15, 96245ac1403SAdrian Hunter 963e4222673SHari Bathini /* 964e4222673SHari Bathini * struct { 965e4222673SHari Bathini * struct perf_event_header header; 966e4222673SHari Bathini * u32 pid; 967e4222673SHari Bathini * u32 tid; 968e4222673SHari Bathini * u64 nr_namespaces; 969e4222673SHari Bathini * { u64 dev, inode; } [nr_namespaces]; 970e4222673SHari Bathini * struct sample_id sample_id; 971e4222673SHari Bathini * }; 972e4222673SHari Bathini */ 973e4222673SHari Bathini PERF_RECORD_NAMESPACES = 16, 974e4222673SHari Bathini 97576193a94SSong Liu /* 97676193a94SSong Liu * Record ksymbol register/unregister events: 97776193a94SSong Liu * 97876193a94SSong Liu * struct { 97976193a94SSong Liu * struct perf_event_header header; 98076193a94SSong Liu * u64 addr; 98176193a94SSong Liu * u32 len; 98276193a94SSong Liu * u16 ksym_type; 98376193a94SSong Liu * u16 flags; 98476193a94SSong Liu * char name[]; 98576193a94SSong Liu * struct sample_id sample_id; 98676193a94SSong Liu * }; 98776193a94SSong Liu */ 98876193a94SSong Liu PERF_RECORD_KSYMBOL = 17, 98976193a94SSong Liu 9906ee52e2aSSong Liu /* 9916ee52e2aSSong Liu * Record bpf events: 9926ee52e2aSSong Liu * enum perf_bpf_event_type { 9936ee52e2aSSong Liu * PERF_BPF_EVENT_UNKNOWN = 0, 9946ee52e2aSSong Liu * PERF_BPF_EVENT_PROG_LOAD = 1, 9956ee52e2aSSong Liu * PERF_BPF_EVENT_PROG_UNLOAD = 2, 9966ee52e2aSSong Liu * }; 9976ee52e2aSSong Liu * 9986ee52e2aSSong Liu * struct { 9996ee52e2aSSong Liu * struct perf_event_header header; 10006ee52e2aSSong Liu * u16 type; 10016ee52e2aSSong Liu * u16 flags; 10026ee52e2aSSong Liu * u32 id; 10036ee52e2aSSong Liu * u8 tag[BPF_TAG_SIZE]; 10046ee52e2aSSong Liu * struct sample_id sample_id; 10056ee52e2aSSong Liu * }; 10066ee52e2aSSong Liu */ 10076ee52e2aSSong Liu PERF_RECORD_BPF_EVENT = 18, 10086ee52e2aSSong Liu 1009607ca46eSDavid Howells PERF_RECORD_MAX, /* non-ABI */ 1010607ca46eSDavid Howells }; 1011607ca46eSDavid Howells 101276193a94SSong Liu enum perf_record_ksymbol_type { 101376193a94SSong Liu PERF_RECORD_KSYMBOL_TYPE_UNKNOWN = 0, 101476193a94SSong Liu PERF_RECORD_KSYMBOL_TYPE_BPF = 1, 101576193a94SSong Liu PERF_RECORD_KSYMBOL_TYPE_MAX /* non-ABI */ 101676193a94SSong Liu }; 101776193a94SSong Liu 101876193a94SSong Liu #define PERF_RECORD_KSYMBOL_FLAGS_UNREGISTER (1 << 0) 101976193a94SSong Liu 10206ee52e2aSSong Liu enum perf_bpf_event_type { 10216ee52e2aSSong Liu PERF_BPF_EVENT_UNKNOWN = 0, 10226ee52e2aSSong Liu PERF_BPF_EVENT_PROG_LOAD = 1, 10236ee52e2aSSong Liu PERF_BPF_EVENT_PROG_UNLOAD = 2, 10246ee52e2aSSong Liu PERF_BPF_EVENT_MAX, /* non-ABI */ 10256ee52e2aSSong Liu }; 10266ee52e2aSSong Liu 1027607ca46eSDavid Howells #define PERF_MAX_STACK_DEPTH 127 1028c85b0334SArnaldo Carvalho de Melo #define PERF_MAX_CONTEXTS_PER_STACK 8 1029607ca46eSDavid Howells 1030607ca46eSDavid Howells enum perf_callchain_context { 1031607ca46eSDavid Howells PERF_CONTEXT_HV = (__u64)-32, 1032607ca46eSDavid Howells PERF_CONTEXT_KERNEL = (__u64)-128, 1033607ca46eSDavid Howells PERF_CONTEXT_USER = (__u64)-512, 1034607ca46eSDavid Howells 1035607ca46eSDavid Howells PERF_CONTEXT_GUEST = (__u64)-2048, 1036607ca46eSDavid Howells PERF_CONTEXT_GUEST_KERNEL = (__u64)-2176, 1037607ca46eSDavid Howells PERF_CONTEXT_GUEST_USER = (__u64)-2560, 1038607ca46eSDavid Howells 1039607ca46eSDavid Howells PERF_CONTEXT_MAX = (__u64)-4095, 1040607ca46eSDavid Howells }; 1041607ca46eSDavid Howells 104268db7e98SAlexander Shishkin /** 104368db7e98SAlexander Shishkin * PERF_RECORD_AUX::flags bits 104468db7e98SAlexander Shishkin */ 104568db7e98SAlexander Shishkin #define PERF_AUX_FLAG_TRUNCATED 0x01 /* record was truncated to fit */ 10462023a0d2SAlexander Shishkin #define PERF_AUX_FLAG_OVERWRITE 0x02 /* snapshot from overwrite mode */ 1047ae0c2d99SAlexander Shishkin #define PERF_AUX_FLAG_PARTIAL 0x04 /* record contains gaps */ 1048085b3062SWill Deacon #define PERF_AUX_FLAG_COLLISION 0x08 /* sample collided with another */ 104968db7e98SAlexander Shishkin 1050643fd0b9SPeter Zijlstra #define PERF_FLAG_FD_NO_GROUP (1UL << 0) 1051643fd0b9SPeter Zijlstra #define PERF_FLAG_FD_OUTPUT (1UL << 1) 1052643fd0b9SPeter Zijlstra #define PERF_FLAG_PID_CGROUP (1UL << 2) /* pid=cgroup id, per-cpu mode only */ 1053643fd0b9SPeter Zijlstra #define PERF_FLAG_FD_CLOEXEC (1UL << 3) /* O_CLOEXEC */ 1054607ca46eSDavid Howells 10558c5073dbSSukadev Bhattiprolu #if defined(__LITTLE_ENDIAN_BITFIELD) 1056d6be9ad6SStephane Eranian union perf_mem_data_src { 1057d6be9ad6SStephane Eranian __u64 val; 1058d6be9ad6SStephane Eranian struct { 1059d6be9ad6SStephane Eranian __u64 mem_op:5, /* type of opcode */ 1060d6be9ad6SStephane Eranian mem_lvl:14, /* memory hierarchy level */ 1061d6be9ad6SStephane Eranian mem_snoop:5, /* snoop mode */ 1062d6be9ad6SStephane Eranian mem_lock:2, /* lock instr */ 1063d6be9ad6SStephane Eranian mem_dtlb:7, /* tlb access */ 10646ae5fa61SAndi Kleen mem_lvl_num:4, /* memory hierarchy level number */ 10656ae5fa61SAndi Kleen mem_remote:1, /* remote */ 10666ae5fa61SAndi Kleen mem_snoopx:2, /* snoop mode, ext */ 10676ae5fa61SAndi Kleen mem_rsvd:24; 1068d6be9ad6SStephane Eranian }; 1069d6be9ad6SStephane Eranian }; 10708c5073dbSSukadev Bhattiprolu #elif defined(__BIG_ENDIAN_BITFIELD) 10718c5073dbSSukadev Bhattiprolu union perf_mem_data_src { 10728c5073dbSSukadev Bhattiprolu __u64 val; 10738c5073dbSSukadev Bhattiprolu struct { 10746ae5fa61SAndi Kleen __u64 mem_rsvd:24, 10756ae5fa61SAndi Kleen mem_snoopx:2, /* snoop mode, ext */ 10766ae5fa61SAndi Kleen mem_remote:1, /* remote */ 10776ae5fa61SAndi Kleen mem_lvl_num:4, /* memory hierarchy level number */ 10788c5073dbSSukadev Bhattiprolu mem_dtlb:7, /* tlb access */ 10798c5073dbSSukadev Bhattiprolu mem_lock:2, /* lock instr */ 10808c5073dbSSukadev Bhattiprolu mem_snoop:5, /* snoop mode */ 10818c5073dbSSukadev Bhattiprolu mem_lvl:14, /* memory hierarchy level */ 10828c5073dbSSukadev Bhattiprolu mem_op:5; /* type of opcode */ 10838c5073dbSSukadev Bhattiprolu }; 10848c5073dbSSukadev Bhattiprolu }; 10858c5073dbSSukadev Bhattiprolu #else 10868c5073dbSSukadev Bhattiprolu #error "Unknown endianness" 10878c5073dbSSukadev Bhattiprolu #endif 1088d6be9ad6SStephane Eranian 1089d6be9ad6SStephane Eranian /* type of opcode (load/store/prefetch,code) */ 1090d6be9ad6SStephane Eranian #define PERF_MEM_OP_NA 0x01 /* not available */ 1091d6be9ad6SStephane Eranian #define PERF_MEM_OP_LOAD 0x02 /* load instruction */ 1092d6be9ad6SStephane Eranian #define PERF_MEM_OP_STORE 0x04 /* store instruction */ 1093d6be9ad6SStephane Eranian #define PERF_MEM_OP_PFETCH 0x08 /* prefetch */ 1094d6be9ad6SStephane Eranian #define PERF_MEM_OP_EXEC 0x10 /* code (execution) */ 1095d6be9ad6SStephane Eranian #define PERF_MEM_OP_SHIFT 0 1096d6be9ad6SStephane Eranian 1097d6be9ad6SStephane Eranian /* memory hierarchy (memory level, hit or miss) */ 1098d6be9ad6SStephane Eranian #define PERF_MEM_LVL_NA 0x01 /* not available */ 1099d6be9ad6SStephane Eranian #define PERF_MEM_LVL_HIT 0x02 /* hit level */ 1100d6be9ad6SStephane Eranian #define PERF_MEM_LVL_MISS 0x04 /* miss level */ 1101d6be9ad6SStephane Eranian #define PERF_MEM_LVL_L1 0x08 /* L1 */ 1102d6be9ad6SStephane Eranian #define PERF_MEM_LVL_LFB 0x10 /* Line Fill Buffer */ 1103cc2f5a8aSStephane Eranian #define PERF_MEM_LVL_L2 0x20 /* L2 */ 1104cc2f5a8aSStephane Eranian #define PERF_MEM_LVL_L3 0x40 /* L3 */ 1105d6be9ad6SStephane Eranian #define PERF_MEM_LVL_LOC_RAM 0x80 /* Local DRAM */ 1106d6be9ad6SStephane Eranian #define PERF_MEM_LVL_REM_RAM1 0x100 /* Remote DRAM (1 hop) */ 1107d6be9ad6SStephane Eranian #define PERF_MEM_LVL_REM_RAM2 0x200 /* Remote DRAM (2 hops) */ 1108d6be9ad6SStephane Eranian #define PERF_MEM_LVL_REM_CCE1 0x400 /* Remote Cache (1 hop) */ 1109d6be9ad6SStephane Eranian #define PERF_MEM_LVL_REM_CCE2 0x800 /* Remote Cache (2 hops) */ 1110d6be9ad6SStephane Eranian #define PERF_MEM_LVL_IO 0x1000 /* I/O memory */ 1111d6be9ad6SStephane Eranian #define PERF_MEM_LVL_UNC 0x2000 /* Uncached memory */ 1112d6be9ad6SStephane Eranian #define PERF_MEM_LVL_SHIFT 5 1113d6be9ad6SStephane Eranian 11146ae5fa61SAndi Kleen #define PERF_MEM_REMOTE_REMOTE 0x01 /* Remote */ 11156ae5fa61SAndi Kleen #define PERF_MEM_REMOTE_SHIFT 37 11166ae5fa61SAndi Kleen 11176ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_L1 0x01 /* L1 */ 11186ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_L2 0x02 /* L2 */ 11196ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_L3 0x03 /* L3 */ 11206ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_L4 0x04 /* L4 */ 11216ae5fa61SAndi Kleen /* 5-0xa available */ 11226ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_ANY_CACHE 0x0b /* Any cache */ 11236ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_LFB 0x0c /* LFB */ 11246ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_RAM 0x0d /* RAM */ 11256ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_PMEM 0x0e /* PMEM */ 11266ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_NA 0x0f /* N/A */ 11276ae5fa61SAndi Kleen 11286ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_SHIFT 33 11296ae5fa61SAndi Kleen 1130d6be9ad6SStephane Eranian /* snoop mode */ 1131d6be9ad6SStephane Eranian #define PERF_MEM_SNOOP_NA 0x01 /* not available */ 1132d6be9ad6SStephane Eranian #define PERF_MEM_SNOOP_NONE 0x02 /* no snoop */ 1133d6be9ad6SStephane Eranian #define PERF_MEM_SNOOP_HIT 0x04 /* snoop hit */ 1134d6be9ad6SStephane Eranian #define PERF_MEM_SNOOP_MISS 0x08 /* snoop miss */ 1135d6be9ad6SStephane Eranian #define PERF_MEM_SNOOP_HITM 0x10 /* snoop hit modified */ 1136d6be9ad6SStephane Eranian #define PERF_MEM_SNOOP_SHIFT 19 1137d6be9ad6SStephane Eranian 11386ae5fa61SAndi Kleen #define PERF_MEM_SNOOPX_FWD 0x01 /* forward */ 11396ae5fa61SAndi Kleen /* 1 free */ 11406ae5fa61SAndi Kleen #define PERF_MEM_SNOOPX_SHIFT 37 11416ae5fa61SAndi Kleen 1142d6be9ad6SStephane Eranian /* locked instruction */ 1143d6be9ad6SStephane Eranian #define PERF_MEM_LOCK_NA 0x01 /* not available */ 1144d6be9ad6SStephane Eranian #define PERF_MEM_LOCK_LOCKED 0x02 /* locked transaction */ 1145d6be9ad6SStephane Eranian #define PERF_MEM_LOCK_SHIFT 24 1146d6be9ad6SStephane Eranian 1147d6be9ad6SStephane Eranian /* TLB access */ 1148d6be9ad6SStephane Eranian #define PERF_MEM_TLB_NA 0x01 /* not available */ 1149d6be9ad6SStephane Eranian #define PERF_MEM_TLB_HIT 0x02 /* hit level */ 1150d6be9ad6SStephane Eranian #define PERF_MEM_TLB_MISS 0x04 /* miss level */ 1151d6be9ad6SStephane Eranian #define PERF_MEM_TLB_L1 0x08 /* L1 */ 1152d6be9ad6SStephane Eranian #define PERF_MEM_TLB_L2 0x10 /* L2 */ 1153d6be9ad6SStephane Eranian #define PERF_MEM_TLB_WK 0x20 /* Hardware Walker*/ 1154d6be9ad6SStephane Eranian #define PERF_MEM_TLB_OS 0x40 /* OS fault handler */ 1155d6be9ad6SStephane Eranian #define PERF_MEM_TLB_SHIFT 26 1156d6be9ad6SStephane Eranian 1157d6be9ad6SStephane Eranian #define PERF_MEM_S(a, s) \ 11580d9dfc23SMike Frysinger (((__u64)PERF_MEM_##a##_##s) << PERF_MEM_##a##_SHIFT) 1159d6be9ad6SStephane Eranian 1160274481deSVince Weaver /* 1161274481deSVince Weaver * single taken branch record layout: 1162274481deSVince Weaver * 1163274481deSVince Weaver * from: source instruction (may not always be a branch insn) 1164274481deSVince Weaver * to: branch target 1165274481deSVince Weaver * mispred: branch target was mispredicted 1166274481deSVince Weaver * predicted: branch target was predicted 1167274481deSVince Weaver * 1168274481deSVince Weaver * support for mispred, predicted is optional. In case it 1169274481deSVince Weaver * is not supported mispred = predicted = 0. 1170274481deSVince Weaver * 1171274481deSVince Weaver * in_tx: running in a hardware transaction 1172274481deSVince Weaver * abort: aborting a hardware transaction 117371ef3c6bSAndi Kleen * cycles: cycles from last branch (or 0 if not supported) 1174eb0baf8aSJin Yao * type: branch type 1175274481deSVince Weaver */ 1176274481deSVince Weaver struct perf_branch_entry { 1177274481deSVince Weaver __u64 from; 1178274481deSVince Weaver __u64 to; 1179274481deSVince Weaver __u64 mispred:1, /* target mispredicted */ 1180274481deSVince Weaver predicted:1,/* target predicted */ 1181274481deSVince Weaver in_tx:1, /* in transaction */ 1182274481deSVince Weaver abort:1, /* transaction abort */ 118371ef3c6bSAndi Kleen cycles:16, /* cycle count to last branch */ 1184eb0baf8aSJin Yao type:4, /* branch type */ 1185eb0baf8aSJin Yao reserved:40; 1186274481deSVince Weaver }; 1187274481deSVince Weaver 1188607ca46eSDavid Howells #endif /* _UAPI_LINUX_PERF_EVENT_H */ 1189