1e2be04c7SGreg Kroah-Hartman /* SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note */ 2607ca46eSDavid Howells /* 3607ca46eSDavid Howells * Performance events: 4607ca46eSDavid Howells * 5607ca46eSDavid Howells * Copyright (C) 2008-2009, Thomas Gleixner <tglx@linutronix.de> 6607ca46eSDavid Howells * Copyright (C) 2008-2011, Red Hat, Inc., Ingo Molnar 7607ca46eSDavid Howells * Copyright (C) 2008-2011, Red Hat, Inc., Peter Zijlstra 8607ca46eSDavid Howells * 9607ca46eSDavid Howells * Data type definitions, declarations, prototypes. 10607ca46eSDavid Howells * 11607ca46eSDavid Howells * Started by: Thomas Gleixner and Ingo Molnar 12607ca46eSDavid Howells * 13607ca46eSDavid Howells * For licencing details see kernel-base/COPYING 14607ca46eSDavid Howells */ 15607ca46eSDavid Howells #ifndef _UAPI_LINUX_PERF_EVENT_H 16607ca46eSDavid Howells #define _UAPI_LINUX_PERF_EVENT_H 17607ca46eSDavid Howells 18607ca46eSDavid Howells #include <linux/types.h> 19607ca46eSDavid Howells #include <linux/ioctl.h> 20607ca46eSDavid Howells #include <asm/byteorder.h> 21607ca46eSDavid Howells 22607ca46eSDavid Howells /* 23607ca46eSDavid Howells * User-space ABI bits: 24607ca46eSDavid Howells */ 25607ca46eSDavid Howells 26607ca46eSDavid Howells /* 27607ca46eSDavid Howells * attr.type 28607ca46eSDavid Howells */ 29607ca46eSDavid Howells enum perf_type_id { 30607ca46eSDavid Howells PERF_TYPE_HARDWARE = 0, 31607ca46eSDavid Howells PERF_TYPE_SOFTWARE = 1, 32607ca46eSDavid Howells PERF_TYPE_TRACEPOINT = 2, 33607ca46eSDavid Howells PERF_TYPE_HW_CACHE = 3, 34607ca46eSDavid Howells PERF_TYPE_RAW = 4, 35607ca46eSDavid Howells PERF_TYPE_BREAKPOINT = 5, 36607ca46eSDavid Howells 37607ca46eSDavid Howells PERF_TYPE_MAX, /* non-ABI */ 38607ca46eSDavid Howells }; 39607ca46eSDavid Howells 40607ca46eSDavid Howells /* 41607ca46eSDavid Howells * Generalized performance event event_id types, used by the 42607ca46eSDavid Howells * attr.event_id parameter of the sys_perf_event_open() 43607ca46eSDavid Howells * syscall: 44607ca46eSDavid Howells */ 45607ca46eSDavid Howells enum perf_hw_id { 46607ca46eSDavid Howells /* 47607ca46eSDavid Howells * Common hardware events, generalized by the kernel: 48607ca46eSDavid Howells */ 49607ca46eSDavid Howells PERF_COUNT_HW_CPU_CYCLES = 0, 50607ca46eSDavid Howells PERF_COUNT_HW_INSTRUCTIONS = 1, 51607ca46eSDavid Howells PERF_COUNT_HW_CACHE_REFERENCES = 2, 52607ca46eSDavid Howells PERF_COUNT_HW_CACHE_MISSES = 3, 53607ca46eSDavid Howells PERF_COUNT_HW_BRANCH_INSTRUCTIONS = 4, 54607ca46eSDavid Howells PERF_COUNT_HW_BRANCH_MISSES = 5, 55607ca46eSDavid Howells PERF_COUNT_HW_BUS_CYCLES = 6, 56607ca46eSDavid Howells PERF_COUNT_HW_STALLED_CYCLES_FRONTEND = 7, 57607ca46eSDavid Howells PERF_COUNT_HW_STALLED_CYCLES_BACKEND = 8, 58607ca46eSDavid Howells PERF_COUNT_HW_REF_CPU_CYCLES = 9, 59607ca46eSDavid Howells 60607ca46eSDavid Howells PERF_COUNT_HW_MAX, /* non-ABI */ 61607ca46eSDavid Howells }; 62607ca46eSDavid Howells 63607ca46eSDavid Howells /* 64607ca46eSDavid Howells * Generalized hardware cache events: 65607ca46eSDavid Howells * 66607ca46eSDavid Howells * { L1-D, L1-I, LLC, ITLB, DTLB, BPU, NODE } x 67607ca46eSDavid Howells * { read, write, prefetch } x 68607ca46eSDavid Howells * { accesses, misses } 69607ca46eSDavid Howells */ 70607ca46eSDavid Howells enum perf_hw_cache_id { 71607ca46eSDavid Howells PERF_COUNT_HW_CACHE_L1D = 0, 72607ca46eSDavid Howells PERF_COUNT_HW_CACHE_L1I = 1, 73607ca46eSDavid Howells PERF_COUNT_HW_CACHE_LL = 2, 74607ca46eSDavid Howells PERF_COUNT_HW_CACHE_DTLB = 3, 75607ca46eSDavid Howells PERF_COUNT_HW_CACHE_ITLB = 4, 76607ca46eSDavid Howells PERF_COUNT_HW_CACHE_BPU = 5, 77607ca46eSDavid Howells PERF_COUNT_HW_CACHE_NODE = 6, 78607ca46eSDavid Howells 79607ca46eSDavid Howells PERF_COUNT_HW_CACHE_MAX, /* non-ABI */ 80607ca46eSDavid Howells }; 81607ca46eSDavid Howells 82607ca46eSDavid Howells enum perf_hw_cache_op_id { 83607ca46eSDavid Howells PERF_COUNT_HW_CACHE_OP_READ = 0, 84607ca46eSDavid Howells PERF_COUNT_HW_CACHE_OP_WRITE = 1, 85607ca46eSDavid Howells PERF_COUNT_HW_CACHE_OP_PREFETCH = 2, 86607ca46eSDavid Howells 87607ca46eSDavid Howells PERF_COUNT_HW_CACHE_OP_MAX, /* non-ABI */ 88607ca46eSDavid Howells }; 89607ca46eSDavid Howells 90607ca46eSDavid Howells enum perf_hw_cache_op_result_id { 91607ca46eSDavid Howells PERF_COUNT_HW_CACHE_RESULT_ACCESS = 0, 92607ca46eSDavid Howells PERF_COUNT_HW_CACHE_RESULT_MISS = 1, 93607ca46eSDavid Howells 94607ca46eSDavid Howells PERF_COUNT_HW_CACHE_RESULT_MAX, /* non-ABI */ 95607ca46eSDavid Howells }; 96607ca46eSDavid Howells 97607ca46eSDavid Howells /* 98607ca46eSDavid Howells * Special "software" events provided by the kernel, even if the hardware 99607ca46eSDavid Howells * does not support performance events. These events measure various 100607ca46eSDavid Howells * physical and sw events of the kernel (and allow the profiling of them as 101607ca46eSDavid Howells * well): 102607ca46eSDavid Howells */ 103607ca46eSDavid Howells enum perf_sw_ids { 104607ca46eSDavid Howells PERF_COUNT_SW_CPU_CLOCK = 0, 105607ca46eSDavid Howells PERF_COUNT_SW_TASK_CLOCK = 1, 106607ca46eSDavid Howells PERF_COUNT_SW_PAGE_FAULTS = 2, 107607ca46eSDavid Howells PERF_COUNT_SW_CONTEXT_SWITCHES = 3, 108607ca46eSDavid Howells PERF_COUNT_SW_CPU_MIGRATIONS = 4, 109607ca46eSDavid Howells PERF_COUNT_SW_PAGE_FAULTS_MIN = 5, 110607ca46eSDavid Howells PERF_COUNT_SW_PAGE_FAULTS_MAJ = 6, 111607ca46eSDavid Howells PERF_COUNT_SW_ALIGNMENT_FAULTS = 7, 112607ca46eSDavid Howells PERF_COUNT_SW_EMULATION_FAULTS = 8, 113fa0097eeSAdrian Hunter PERF_COUNT_SW_DUMMY = 9, 114a43eec30SAlexei Starovoitov PERF_COUNT_SW_BPF_OUTPUT = 10, 115607ca46eSDavid Howells 116607ca46eSDavid Howells PERF_COUNT_SW_MAX, /* non-ABI */ 117607ca46eSDavid Howells }; 118607ca46eSDavid Howells 119607ca46eSDavid Howells /* 120607ca46eSDavid Howells * Bits that can be set in attr.sample_type to request information 121607ca46eSDavid Howells * in the overflow packets. 122607ca46eSDavid Howells */ 123607ca46eSDavid Howells enum perf_event_sample_format { 124607ca46eSDavid Howells PERF_SAMPLE_IP = 1U << 0, 125607ca46eSDavid Howells PERF_SAMPLE_TID = 1U << 1, 126607ca46eSDavid Howells PERF_SAMPLE_TIME = 1U << 2, 127607ca46eSDavid Howells PERF_SAMPLE_ADDR = 1U << 3, 128607ca46eSDavid Howells PERF_SAMPLE_READ = 1U << 4, 129607ca46eSDavid Howells PERF_SAMPLE_CALLCHAIN = 1U << 5, 130607ca46eSDavid Howells PERF_SAMPLE_ID = 1U << 6, 131607ca46eSDavid Howells PERF_SAMPLE_CPU = 1U << 7, 132607ca46eSDavid Howells PERF_SAMPLE_PERIOD = 1U << 8, 133607ca46eSDavid Howells PERF_SAMPLE_STREAM_ID = 1U << 9, 134607ca46eSDavid Howells PERF_SAMPLE_RAW = 1U << 10, 135607ca46eSDavid Howells PERF_SAMPLE_BRANCH_STACK = 1U << 11, 136607ca46eSDavid Howells PERF_SAMPLE_REGS_USER = 1U << 12, 137607ca46eSDavid Howells PERF_SAMPLE_STACK_USER = 1U << 13, 138c3feedf2SAndi Kleen PERF_SAMPLE_WEIGHT = 1U << 14, 139d6be9ad6SStephane Eranian PERF_SAMPLE_DATA_SRC = 1U << 15, 140ff3d527cSAdrian Hunter PERF_SAMPLE_IDENTIFIER = 1U << 16, 141fdfbbd07SAndi Kleen PERF_SAMPLE_TRANSACTION = 1U << 17, 14260e2364eSStephane Eranian PERF_SAMPLE_REGS_INTR = 1U << 18, 143fc7ce9c7SKan Liang PERF_SAMPLE_PHYS_ADDR = 1U << 19, 144a4faf00dSAlexander Shishkin PERF_SAMPLE_AUX = 1U << 20, 1456546b19fSNamhyung Kim PERF_SAMPLE_CGROUP = 1U << 21, 146607ca46eSDavid Howells 1476546b19fSNamhyung Kim PERF_SAMPLE_MAX = 1U << 22, /* non-ABI */ 1486cbc304fSPeter Zijlstra 14909121255SPeter Zijlstra __PERF_SAMPLE_CALLCHAIN_EARLY = 1ULL << 63, /* non-ABI; internal use */ 150607ca46eSDavid Howells }; 151607ca46eSDavid Howells 152607ca46eSDavid Howells /* 153607ca46eSDavid Howells * values to program into branch_sample_type when PERF_SAMPLE_BRANCH is set 154607ca46eSDavid Howells * 155607ca46eSDavid Howells * If the user does not pass priv level information via branch_sample_type, 156607ca46eSDavid Howells * the kernel uses the event's priv level. Branch and event priv levels do 157607ca46eSDavid Howells * not have to match. Branch priv level is checked for permissions. 158607ca46eSDavid Howells * 159607ca46eSDavid Howells * The branch types can be combined, however BRANCH_ANY covers all types 160607ca46eSDavid Howells * of branches and therefore it supersedes all the other types. 161607ca46eSDavid Howells */ 16227ac905bSYan, Zheng enum perf_branch_sample_type_shift { 16327ac905bSYan, Zheng PERF_SAMPLE_BRANCH_USER_SHIFT = 0, /* user branches */ 16427ac905bSYan, Zheng PERF_SAMPLE_BRANCH_KERNEL_SHIFT = 1, /* kernel branches */ 16527ac905bSYan, Zheng PERF_SAMPLE_BRANCH_HV_SHIFT = 2, /* hypervisor branches */ 16627ac905bSYan, Zheng 16727ac905bSYan, Zheng PERF_SAMPLE_BRANCH_ANY_SHIFT = 3, /* any branch types */ 16827ac905bSYan, Zheng PERF_SAMPLE_BRANCH_ANY_CALL_SHIFT = 4, /* any call branch */ 16927ac905bSYan, Zheng PERF_SAMPLE_BRANCH_ANY_RETURN_SHIFT = 5, /* any return branch */ 17027ac905bSYan, Zheng PERF_SAMPLE_BRANCH_IND_CALL_SHIFT = 6, /* indirect calls */ 17127ac905bSYan, Zheng PERF_SAMPLE_BRANCH_ABORT_TX_SHIFT = 7, /* transaction aborts */ 17227ac905bSYan, Zheng PERF_SAMPLE_BRANCH_IN_TX_SHIFT = 8, /* in transaction */ 17327ac905bSYan, Zheng PERF_SAMPLE_BRANCH_NO_TX_SHIFT = 9, /* not in transaction */ 17427ac905bSYan, Zheng PERF_SAMPLE_BRANCH_COND_SHIFT = 10, /* conditional branches */ 17527ac905bSYan, Zheng 1762c44b193SPeter Zijlstra PERF_SAMPLE_BRANCH_CALL_STACK_SHIFT = 11, /* call/ret stack */ 177c9fdfa14SStephane Eranian PERF_SAMPLE_BRANCH_IND_JUMP_SHIFT = 12, /* indirect jumps */ 178c229bf9dSStephane Eranian PERF_SAMPLE_BRANCH_CALL_SHIFT = 13, /* direct call */ 1792c44b193SPeter Zijlstra 180b16a5b52SAndi Kleen PERF_SAMPLE_BRANCH_NO_FLAGS_SHIFT = 14, /* no flags */ 181b16a5b52SAndi Kleen PERF_SAMPLE_BRANCH_NO_CYCLES_SHIFT = 15, /* no cycles */ 182b16a5b52SAndi Kleen 183eb0baf8aSJin Yao PERF_SAMPLE_BRANCH_TYPE_SAVE_SHIFT = 16, /* save branch type */ 184eb0baf8aSJin Yao 185bbfd5e4fSKan Liang PERF_SAMPLE_BRANCH_HW_INDEX_SHIFT = 17, /* save low level index of raw branch records */ 186bbfd5e4fSKan Liang 18727ac905bSYan, Zheng PERF_SAMPLE_BRANCH_MAX_SHIFT /* non-ABI */ 18827ac905bSYan, Zheng }; 18927ac905bSYan, Zheng 190607ca46eSDavid Howells enum perf_branch_sample_type { 19127ac905bSYan, Zheng PERF_SAMPLE_BRANCH_USER = 1U << PERF_SAMPLE_BRANCH_USER_SHIFT, 19227ac905bSYan, Zheng PERF_SAMPLE_BRANCH_KERNEL = 1U << PERF_SAMPLE_BRANCH_KERNEL_SHIFT, 19327ac905bSYan, Zheng PERF_SAMPLE_BRANCH_HV = 1U << PERF_SAMPLE_BRANCH_HV_SHIFT, 194607ca46eSDavid Howells 19527ac905bSYan, Zheng PERF_SAMPLE_BRANCH_ANY = 1U << PERF_SAMPLE_BRANCH_ANY_SHIFT, 1962c44b193SPeter Zijlstra PERF_SAMPLE_BRANCH_ANY_CALL = 1U << PERF_SAMPLE_BRANCH_ANY_CALL_SHIFT, 1972c44b193SPeter Zijlstra PERF_SAMPLE_BRANCH_ANY_RETURN = 1U << PERF_SAMPLE_BRANCH_ANY_RETURN_SHIFT, 1982c44b193SPeter Zijlstra PERF_SAMPLE_BRANCH_IND_CALL = 1U << PERF_SAMPLE_BRANCH_IND_CALL_SHIFT, 1992c44b193SPeter Zijlstra PERF_SAMPLE_BRANCH_ABORT_TX = 1U << PERF_SAMPLE_BRANCH_ABORT_TX_SHIFT, 20027ac905bSYan, Zheng PERF_SAMPLE_BRANCH_IN_TX = 1U << PERF_SAMPLE_BRANCH_IN_TX_SHIFT, 20127ac905bSYan, Zheng PERF_SAMPLE_BRANCH_NO_TX = 1U << PERF_SAMPLE_BRANCH_NO_TX_SHIFT, 20227ac905bSYan, Zheng PERF_SAMPLE_BRANCH_COND = 1U << PERF_SAMPLE_BRANCH_COND_SHIFT, 203607ca46eSDavid Howells 2042c44b193SPeter Zijlstra PERF_SAMPLE_BRANCH_CALL_STACK = 1U << PERF_SAMPLE_BRANCH_CALL_STACK_SHIFT, 205c9fdfa14SStephane Eranian PERF_SAMPLE_BRANCH_IND_JUMP = 1U << PERF_SAMPLE_BRANCH_IND_JUMP_SHIFT, 206c229bf9dSStephane Eranian PERF_SAMPLE_BRANCH_CALL = 1U << PERF_SAMPLE_BRANCH_CALL_SHIFT, 2072c44b193SPeter Zijlstra 208b16a5b52SAndi Kleen PERF_SAMPLE_BRANCH_NO_FLAGS = 1U << PERF_SAMPLE_BRANCH_NO_FLAGS_SHIFT, 209b16a5b52SAndi Kleen PERF_SAMPLE_BRANCH_NO_CYCLES = 1U << PERF_SAMPLE_BRANCH_NO_CYCLES_SHIFT, 210b16a5b52SAndi Kleen 211eb0baf8aSJin Yao PERF_SAMPLE_BRANCH_TYPE_SAVE = 212eb0baf8aSJin Yao 1U << PERF_SAMPLE_BRANCH_TYPE_SAVE_SHIFT, 213eb0baf8aSJin Yao 214bbfd5e4fSKan Liang PERF_SAMPLE_BRANCH_HW_INDEX = 1U << PERF_SAMPLE_BRANCH_HW_INDEX_SHIFT, 215bbfd5e4fSKan Liang 21627ac905bSYan, Zheng PERF_SAMPLE_BRANCH_MAX = 1U << PERF_SAMPLE_BRANCH_MAX_SHIFT, 217607ca46eSDavid Howells }; 218607ca46eSDavid Howells 219eb0baf8aSJin Yao /* 220eb0baf8aSJin Yao * Common flow change classification 221eb0baf8aSJin Yao */ 222eb0baf8aSJin Yao enum { 223eb0baf8aSJin Yao PERF_BR_UNKNOWN = 0, /* unknown */ 224eb0baf8aSJin Yao PERF_BR_COND = 1, /* conditional */ 225eb0baf8aSJin Yao PERF_BR_UNCOND = 2, /* unconditional */ 226eb0baf8aSJin Yao PERF_BR_IND = 3, /* indirect */ 227eb0baf8aSJin Yao PERF_BR_CALL = 4, /* function call */ 228eb0baf8aSJin Yao PERF_BR_IND_CALL = 5, /* indirect function call */ 229eb0baf8aSJin Yao PERF_BR_RET = 6, /* function return */ 230eb0baf8aSJin Yao PERF_BR_SYSCALL = 7, /* syscall */ 231eb0baf8aSJin Yao PERF_BR_SYSRET = 8, /* syscall return */ 232eb0baf8aSJin Yao PERF_BR_COND_CALL = 9, /* conditional function call */ 233eb0baf8aSJin Yao PERF_BR_COND_RET = 10, /* conditional function return */ 234eb0baf8aSJin Yao PERF_BR_MAX, 235eb0baf8aSJin Yao }; 236eb0baf8aSJin Yao 237607ca46eSDavid Howells #define PERF_SAMPLE_BRANCH_PLM_ALL \ 238607ca46eSDavid Howells (PERF_SAMPLE_BRANCH_USER|\ 239607ca46eSDavid Howells PERF_SAMPLE_BRANCH_KERNEL|\ 240607ca46eSDavid Howells PERF_SAMPLE_BRANCH_HV) 241607ca46eSDavid Howells 242607ca46eSDavid Howells /* 243607ca46eSDavid Howells * Values to determine ABI of the registers dump. 244607ca46eSDavid Howells */ 245607ca46eSDavid Howells enum perf_sample_regs_abi { 246607ca46eSDavid Howells PERF_SAMPLE_REGS_ABI_NONE = 0, 247607ca46eSDavid Howells PERF_SAMPLE_REGS_ABI_32 = 1, 248607ca46eSDavid Howells PERF_SAMPLE_REGS_ABI_64 = 2, 249607ca46eSDavid Howells }; 250607ca46eSDavid Howells 251607ca46eSDavid Howells /* 252fdfbbd07SAndi Kleen * Values for the memory transaction event qualifier, mostly for 253fdfbbd07SAndi Kleen * abort events. Multiple bits can be set. 254fdfbbd07SAndi Kleen */ 255fdfbbd07SAndi Kleen enum { 256fdfbbd07SAndi Kleen PERF_TXN_ELISION = (1 << 0), /* From elision */ 257fdfbbd07SAndi Kleen PERF_TXN_TRANSACTION = (1 << 1), /* From transaction */ 258fdfbbd07SAndi Kleen PERF_TXN_SYNC = (1 << 2), /* Instruction is related */ 259fdfbbd07SAndi Kleen PERF_TXN_ASYNC = (1 << 3), /* Instruction not related */ 260fdfbbd07SAndi Kleen PERF_TXN_RETRY = (1 << 4), /* Retry possible */ 261fdfbbd07SAndi Kleen PERF_TXN_CONFLICT = (1 << 5), /* Conflict abort */ 262fdfbbd07SAndi Kleen PERF_TXN_CAPACITY_WRITE = (1 << 6), /* Capacity write abort */ 263fdfbbd07SAndi Kleen PERF_TXN_CAPACITY_READ = (1 << 7), /* Capacity read abort */ 264fdfbbd07SAndi Kleen 265fdfbbd07SAndi Kleen PERF_TXN_MAX = (1 << 8), /* non-ABI */ 266fdfbbd07SAndi Kleen 267fdfbbd07SAndi Kleen /* bits 32..63 are reserved for the abort code */ 268fdfbbd07SAndi Kleen 269fdfbbd07SAndi Kleen PERF_TXN_ABORT_MASK = (0xffffffffULL << 32), 270fdfbbd07SAndi Kleen PERF_TXN_ABORT_SHIFT = 32, 271fdfbbd07SAndi Kleen }; 272fdfbbd07SAndi Kleen 273fdfbbd07SAndi Kleen /* 274607ca46eSDavid Howells * The format of the data returned by read() on a perf event fd, 275607ca46eSDavid Howells * as specified by attr.read_format: 276607ca46eSDavid Howells * 277607ca46eSDavid Howells * struct read_format { 278607ca46eSDavid Howells * { u64 value; 279607ca46eSDavid Howells * { u64 time_enabled; } && PERF_FORMAT_TOTAL_TIME_ENABLED 280607ca46eSDavid Howells * { u64 time_running; } && PERF_FORMAT_TOTAL_TIME_RUNNING 281607ca46eSDavid Howells * { u64 id; } && PERF_FORMAT_ID 282607ca46eSDavid Howells * } && !PERF_FORMAT_GROUP 283607ca46eSDavid Howells * 284607ca46eSDavid Howells * { u64 nr; 285607ca46eSDavid Howells * { u64 time_enabled; } && PERF_FORMAT_TOTAL_TIME_ENABLED 286607ca46eSDavid Howells * { u64 time_running; } && PERF_FORMAT_TOTAL_TIME_RUNNING 287607ca46eSDavid Howells * { u64 value; 288607ca46eSDavid Howells * { u64 id; } && PERF_FORMAT_ID 289607ca46eSDavid Howells * } cntr[nr]; 290607ca46eSDavid Howells * } && PERF_FORMAT_GROUP 291607ca46eSDavid Howells * }; 292607ca46eSDavid Howells */ 293607ca46eSDavid Howells enum perf_event_read_format { 294607ca46eSDavid Howells PERF_FORMAT_TOTAL_TIME_ENABLED = 1U << 0, 295607ca46eSDavid Howells PERF_FORMAT_TOTAL_TIME_RUNNING = 1U << 1, 296607ca46eSDavid Howells PERF_FORMAT_ID = 1U << 2, 297607ca46eSDavid Howells PERF_FORMAT_GROUP = 1U << 3, 298607ca46eSDavid Howells 299607ca46eSDavid Howells PERF_FORMAT_MAX = 1U << 4, /* non-ABI */ 300607ca46eSDavid Howells }; 301607ca46eSDavid Howells 302607ca46eSDavid Howells #define PERF_ATTR_SIZE_VER0 64 /* sizeof first published struct */ 303607ca46eSDavid Howells #define PERF_ATTR_SIZE_VER1 72 /* add: config2 */ 304607ca46eSDavid Howells #define PERF_ATTR_SIZE_VER2 80 /* add: branch_sample_type */ 305607ca46eSDavid Howells #define PERF_ATTR_SIZE_VER3 96 /* add: sample_regs_user */ 306607ca46eSDavid Howells /* add: sample_stack_user */ 30760e2364eSStephane Eranian #define PERF_ATTR_SIZE_VER4 104 /* add: sample_regs_intr */ 3081a594131SAlexander Shishkin #define PERF_ATTR_SIZE_VER5 112 /* add: aux_watermark */ 309a4faf00dSAlexander Shishkin #define PERF_ATTR_SIZE_VER6 120 /* add: aux_sample_size */ 310607ca46eSDavid Howells 311607ca46eSDavid Howells /* 312607ca46eSDavid Howells * Hardware event_id to monitor via a performance monitoring event: 31397c79a38SArnaldo Carvalho de Melo * 31497c79a38SArnaldo Carvalho de Melo * @sample_max_stack: Max number of frame pointers in a callchain, 31597c79a38SArnaldo Carvalho de Melo * should be < /proc/sys/kernel/perf_event_max_stack 316607ca46eSDavid Howells */ 317607ca46eSDavid Howells struct perf_event_attr { 318607ca46eSDavid Howells 319607ca46eSDavid Howells /* 320607ca46eSDavid Howells * Major type: hardware/software/tracepoint/etc. 321607ca46eSDavid Howells */ 322607ca46eSDavid Howells __u32 type; 323607ca46eSDavid Howells 324607ca46eSDavid Howells /* 325607ca46eSDavid Howells * Size of the attr structure, for fwd/bwd compat. 326607ca46eSDavid Howells */ 327607ca46eSDavid Howells __u32 size; 328607ca46eSDavid Howells 329607ca46eSDavid Howells /* 330607ca46eSDavid Howells * Type specific configuration information. 331607ca46eSDavid Howells */ 332607ca46eSDavid Howells __u64 config; 333607ca46eSDavid Howells 334607ca46eSDavid Howells union { 335607ca46eSDavid Howells __u64 sample_period; 336607ca46eSDavid Howells __u64 sample_freq; 337607ca46eSDavid Howells }; 338607ca46eSDavid Howells 339607ca46eSDavid Howells __u64 sample_type; 340607ca46eSDavid Howells __u64 read_format; 341607ca46eSDavid Howells 342607ca46eSDavid Howells __u64 disabled : 1, /* off by default */ 343607ca46eSDavid Howells inherit : 1, /* children inherit it */ 344607ca46eSDavid Howells pinned : 1, /* must always be on PMU */ 345607ca46eSDavid Howells exclusive : 1, /* only group on PMU */ 346607ca46eSDavid Howells exclude_user : 1, /* don't count user */ 347607ca46eSDavid Howells exclude_kernel : 1, /* ditto kernel */ 348607ca46eSDavid Howells exclude_hv : 1, /* ditto hypervisor */ 349607ca46eSDavid Howells exclude_idle : 1, /* don't count when idle */ 350607ca46eSDavid Howells mmap : 1, /* include mmap data */ 351607ca46eSDavid Howells comm : 1, /* include comm data */ 352607ca46eSDavid Howells freq : 1, /* use freq, not period */ 353607ca46eSDavid Howells inherit_stat : 1, /* per task counts */ 354607ca46eSDavid Howells enable_on_exec : 1, /* next exec enables */ 355607ca46eSDavid Howells task : 1, /* trace fork/exit */ 356607ca46eSDavid Howells watermark : 1, /* wakeup_watermark */ 357607ca46eSDavid Howells /* 358607ca46eSDavid Howells * precise_ip: 359607ca46eSDavid Howells * 360607ca46eSDavid Howells * 0 - SAMPLE_IP can have arbitrary skid 361607ca46eSDavid Howells * 1 - SAMPLE_IP must have constant skid 362607ca46eSDavid Howells * 2 - SAMPLE_IP requested to have 0 skid 363607ca46eSDavid Howells * 3 - SAMPLE_IP must have 0 skid 364607ca46eSDavid Howells * 365607ca46eSDavid Howells * See also PERF_RECORD_MISC_EXACT_IP 366607ca46eSDavid Howells */ 367607ca46eSDavid Howells precise_ip : 2, /* skid constraint */ 368607ca46eSDavid Howells mmap_data : 1, /* non-exec mmap data */ 369607ca46eSDavid Howells sample_id_all : 1, /* sample_type all events */ 370607ca46eSDavid Howells 371607ca46eSDavid Howells exclude_host : 1, /* don't count in host */ 372607ca46eSDavid Howells exclude_guest : 1, /* don't count in guest */ 373607ca46eSDavid Howells 374607ca46eSDavid Howells exclude_callchain_kernel : 1, /* exclude kernel callchains */ 375607ca46eSDavid Howells exclude_callchain_user : 1, /* exclude user callchains */ 37613d7a241SStephane Eranian mmap2 : 1, /* include mmap with inode data */ 37782b89778SAdrian Hunter comm_exec : 1, /* flag comm events that are due to an exec */ 37834f43927SPeter Zijlstra use_clockid : 1, /* use @clockid for time fields */ 37945ac1403SAdrian Hunter context_switch : 1, /* context switch data */ 3809ecda41aSWang Nan write_backward : 1, /* Write ring buffer from end to beginning */ 381e4222673SHari Bathini namespaces : 1, /* include namespaces data */ 38276193a94SSong Liu ksymbol : 1, /* include ksymbol events */ 3836ee52e2aSSong Liu bpf_event : 1, /* include bpf events */ 384ab43762eSAlexander Shishkin aux_output : 1, /* generate AUX records instead of events */ 38596aaab68SNamhyung Kim cgroup : 1, /* include cgroup events */ 386*e17d43b9SAdrian Hunter text_poke : 1, /* include text poke events */ 387*e17d43b9SAdrian Hunter __reserved_1 : 30; 388607ca46eSDavid Howells 389607ca46eSDavid Howells union { 390607ca46eSDavid Howells __u32 wakeup_events; /* wakeup every n events */ 391607ca46eSDavid Howells __u32 wakeup_watermark; /* bytes before wakeup */ 392607ca46eSDavid Howells }; 393607ca46eSDavid Howells 394607ca46eSDavid Howells __u32 bp_type; 395607ca46eSDavid Howells union { 396607ca46eSDavid Howells __u64 bp_addr; 39765074d43SSong Liu __u64 kprobe_func; /* for perf_kprobe */ 39865074d43SSong Liu __u64 uprobe_path; /* for perf_uprobe */ 399607ca46eSDavid Howells __u64 config1; /* extension of config */ 400607ca46eSDavid Howells }; 401607ca46eSDavid Howells union { 402607ca46eSDavid Howells __u64 bp_len; 40365074d43SSong Liu __u64 kprobe_addr; /* when kprobe_func == NULL */ 40465074d43SSong Liu __u64 probe_offset; /* for perf_[k,u]probe */ 405607ca46eSDavid Howells __u64 config2; /* extension of config1 */ 406607ca46eSDavid Howells }; 407607ca46eSDavid Howells __u64 branch_sample_type; /* enum perf_branch_sample_type */ 408607ca46eSDavid Howells 409607ca46eSDavid Howells /* 410607ca46eSDavid Howells * Defines set of user regs to dump on samples. 411607ca46eSDavid Howells * See asm/perf_regs.h for details. 412607ca46eSDavid Howells */ 413607ca46eSDavid Howells __u64 sample_regs_user; 414607ca46eSDavid Howells 415607ca46eSDavid Howells /* 416607ca46eSDavid Howells * Defines size of the user stack to dump on samples. 417607ca46eSDavid Howells */ 418607ca46eSDavid Howells __u32 sample_stack_user; 419607ca46eSDavid Howells 42034f43927SPeter Zijlstra __s32 clockid; 42160e2364eSStephane Eranian /* 42260e2364eSStephane Eranian * Defines set of regs to dump for each sample 42360e2364eSStephane Eranian * state captured on: 42460e2364eSStephane Eranian * - precise = 0: PMU interrupt 42560e2364eSStephane Eranian * - precise > 0: sampled instruction 42660e2364eSStephane Eranian * 42760e2364eSStephane Eranian * See asm/perf_regs.h for details. 42860e2364eSStephane Eranian */ 42960e2364eSStephane Eranian __u64 sample_regs_intr; 4301a594131SAlexander Shishkin 4311a594131SAlexander Shishkin /* 4321a594131SAlexander Shishkin * Wakeup watermark for AUX area 4331a594131SAlexander Shishkin */ 4341a594131SAlexander Shishkin __u32 aux_watermark; 43597c79a38SArnaldo Carvalho de Melo __u16 sample_max_stack; 436a4faf00dSAlexander Shishkin __u16 __reserved_2; 437a4faf00dSAlexander Shishkin __u32 aux_sample_size; 438a4faf00dSAlexander Shishkin __u32 __reserved_3; 439607ca46eSDavid Howells }; 440607ca46eSDavid Howells 441f371b304SYonghong Song /* 442f371b304SYonghong Song * Structure used by below PERF_EVENT_IOC_QUERY_BPF command 443f371b304SYonghong Song * to query bpf programs attached to the same perf tracepoint 444f371b304SYonghong Song * as the given perf event. 445f371b304SYonghong Song */ 446f371b304SYonghong Song struct perf_event_query_bpf { 447f371b304SYonghong Song /* 448f371b304SYonghong Song * The below ids array length 449f371b304SYonghong Song */ 450f371b304SYonghong Song __u32 ids_len; 451f371b304SYonghong Song /* 452f371b304SYonghong Song * Set by the kernel to indicate the number of 453f371b304SYonghong Song * available programs 454f371b304SYonghong Song */ 455f371b304SYonghong Song __u32 prog_cnt; 456f371b304SYonghong Song /* 457f371b304SYonghong Song * User provided buffer to store program ids 458f371b304SYonghong Song */ 459f371b304SYonghong Song __u32 ids[0]; 460f371b304SYonghong Song }; 461f371b304SYonghong Song 462607ca46eSDavid Howells /* 463607ca46eSDavid Howells * Ioctls that can be done on a perf event fd: 464607ca46eSDavid Howells */ 465607ca46eSDavid Howells #define PERF_EVENT_IOC_ENABLE _IO ('$', 0) 466607ca46eSDavid Howells #define PERF_EVENT_IOC_DISABLE _IO ('$', 1) 467607ca46eSDavid Howells #define PERF_EVENT_IOC_REFRESH _IO ('$', 2) 468607ca46eSDavid Howells #define PERF_EVENT_IOC_RESET _IO ('$', 3) 469607ca46eSDavid Howells #define PERF_EVENT_IOC_PERIOD _IOW('$', 4, __u64) 470607ca46eSDavid Howells #define PERF_EVENT_IOC_SET_OUTPUT _IO ('$', 5) 471607ca46eSDavid Howells #define PERF_EVENT_IOC_SET_FILTER _IOW('$', 6, char *) 472a8e0108cSVince Weaver #define PERF_EVENT_IOC_ID _IOR('$', 7, __u64 *) 4732541517cSAlexei Starovoitov #define PERF_EVENT_IOC_SET_BPF _IOW('$', 8, __u32) 47486e7972fSWang Nan #define PERF_EVENT_IOC_PAUSE_OUTPUT _IOW('$', 9, __u32) 475f371b304SYonghong Song #define PERF_EVENT_IOC_QUERY_BPF _IOWR('$', 10, struct perf_event_query_bpf *) 47632ff77e8SMilind Chabbi #define PERF_EVENT_IOC_MODIFY_ATTRIBUTES _IOW('$', 11, struct perf_event_attr *) 477607ca46eSDavid Howells 478607ca46eSDavid Howells enum perf_event_ioc_flags { 479607ca46eSDavid Howells PERF_IOC_FLAG_GROUP = 1U << 0, 480607ca46eSDavid Howells }; 481607ca46eSDavid Howells 482607ca46eSDavid Howells /* 483607ca46eSDavid Howells * Structure of the page that can be mapped via mmap 484607ca46eSDavid Howells */ 485607ca46eSDavid Howells struct perf_event_mmap_page { 486607ca46eSDavid Howells __u32 version; /* version number of this structure */ 487607ca46eSDavid Howells __u32 compat_version; /* lowest version this is compat with */ 488607ca46eSDavid Howells 489607ca46eSDavid Howells /* 490607ca46eSDavid Howells * Bits needed to read the hw events in user-space. 491607ca46eSDavid Howells * 492b438b1abSAndy Lutomirski * u32 seq, time_mult, time_shift, index, width; 493607ca46eSDavid Howells * u64 count, enabled, running; 494607ca46eSDavid Howells * u64 cyc, time_offset; 495607ca46eSDavid Howells * s64 pmc = 0; 496607ca46eSDavid Howells * 497607ca46eSDavid Howells * do { 498607ca46eSDavid Howells * seq = pc->lock; 499607ca46eSDavid Howells * barrier() 500607ca46eSDavid Howells * 501607ca46eSDavid Howells * enabled = pc->time_enabled; 502607ca46eSDavid Howells * running = pc->time_running; 503607ca46eSDavid Howells * 504607ca46eSDavid Howells * if (pc->cap_usr_time && enabled != running) { 505607ca46eSDavid Howells * cyc = rdtsc(); 506607ca46eSDavid Howells * time_offset = pc->time_offset; 507607ca46eSDavid Howells * time_mult = pc->time_mult; 508607ca46eSDavid Howells * time_shift = pc->time_shift; 509607ca46eSDavid Howells * } 510607ca46eSDavid Howells * 511b438b1abSAndy Lutomirski * index = pc->index; 512607ca46eSDavid Howells * count = pc->offset; 513b438b1abSAndy Lutomirski * if (pc->cap_user_rdpmc && index) { 514607ca46eSDavid Howells * width = pc->pmc_width; 515b438b1abSAndy Lutomirski * pmc = rdpmc(index - 1); 516607ca46eSDavid Howells * } 517607ca46eSDavid Howells * 518607ca46eSDavid Howells * barrier(); 519607ca46eSDavid Howells * } while (pc->lock != seq); 520607ca46eSDavid Howells * 521607ca46eSDavid Howells * NOTE: for obvious reason this only works on self-monitoring 522607ca46eSDavid Howells * processes. 523607ca46eSDavid Howells */ 524607ca46eSDavid Howells __u32 lock; /* seqlock for synchronization */ 525607ca46eSDavid Howells __u32 index; /* hardware event identifier */ 526607ca46eSDavid Howells __s64 offset; /* add to hardware event value */ 527607ca46eSDavid Howells __u64 time_enabled; /* time event active */ 528607ca46eSDavid Howells __u64 time_running; /* time event on cpu */ 529607ca46eSDavid Howells union { 530607ca46eSDavid Howells __u64 capabilities; 531860f085bSAdrian Hunter struct { 532fa731587SPeter Zijlstra __u64 cap_bit0 : 1, /* Always 0, deprecated, see commit 860f085b74e9 */ 533fa731587SPeter Zijlstra cap_bit0_is_deprecated : 1, /* Always 1, signals that bit 0 is zero */ 534fa731587SPeter Zijlstra 535fa731587SPeter Zijlstra cap_user_rdpmc : 1, /* The RDPMC instruction can be used to read counts */ 536fa731587SPeter Zijlstra cap_user_time : 1, /* The time_* fields are used */ 537fa731587SPeter Zijlstra cap_user_time_zero : 1, /* The time_zero field is used */ 538fa731587SPeter Zijlstra cap_____res : 59; 539607ca46eSDavid Howells }; 540860f085bSAdrian Hunter }; 541607ca46eSDavid Howells 542607ca46eSDavid Howells /* 543b438b1abSAndy Lutomirski * If cap_user_rdpmc this field provides the bit-width of the value 544607ca46eSDavid Howells * read using the rdpmc() or equivalent instruction. This can be used 545607ca46eSDavid Howells * to sign extend the result like: 546607ca46eSDavid Howells * 547607ca46eSDavid Howells * pmc <<= 64 - width; 548607ca46eSDavid Howells * pmc >>= 64 - width; // signed shift right 549607ca46eSDavid Howells * count += pmc; 550607ca46eSDavid Howells */ 551607ca46eSDavid Howells __u16 pmc_width; 552607ca46eSDavid Howells 553607ca46eSDavid Howells /* 554607ca46eSDavid Howells * If cap_usr_time the below fields can be used to compute the time 555607ca46eSDavid Howells * delta since time_enabled (in ns) using rdtsc or similar. 556607ca46eSDavid Howells * 557607ca46eSDavid Howells * u64 quot, rem; 558607ca46eSDavid Howells * u64 delta; 559607ca46eSDavid Howells * 560607ca46eSDavid Howells * quot = (cyc >> time_shift); 561b9511cd7SAdrian Hunter * rem = cyc & (((u64)1 << time_shift) - 1); 562607ca46eSDavid Howells * delta = time_offset + quot * time_mult + 563607ca46eSDavid Howells * ((rem * time_mult) >> time_shift); 564607ca46eSDavid Howells * 565607ca46eSDavid Howells * Where time_offset,time_mult,time_shift and cyc are read in the 566607ca46eSDavid Howells * seqcount loop described above. This delta can then be added to 567b438b1abSAndy Lutomirski * enabled and possible running (if index), improving the scaling: 568607ca46eSDavid Howells * 569607ca46eSDavid Howells * enabled += delta; 570b438b1abSAndy Lutomirski * if (index) 571607ca46eSDavid Howells * running += delta; 572607ca46eSDavid Howells * 573607ca46eSDavid Howells * quot = count / running; 574607ca46eSDavid Howells * rem = count % running; 575607ca46eSDavid Howells * count = quot * enabled + (rem * enabled) / running; 576607ca46eSDavid Howells */ 577607ca46eSDavid Howells __u16 time_shift; 578607ca46eSDavid Howells __u32 time_mult; 579607ca46eSDavid Howells __u64 time_offset; 580c73deb6aSAdrian Hunter /* 581c73deb6aSAdrian Hunter * If cap_usr_time_zero, the hardware clock (e.g. TSC) can be calculated 582c73deb6aSAdrian Hunter * from sample timestamps. 583c73deb6aSAdrian Hunter * 584c73deb6aSAdrian Hunter * time = timestamp - time_zero; 585c73deb6aSAdrian Hunter * quot = time / time_mult; 586c73deb6aSAdrian Hunter * rem = time % time_mult; 587c73deb6aSAdrian Hunter * cyc = (quot << time_shift) + (rem << time_shift) / time_mult; 588c73deb6aSAdrian Hunter * 589c73deb6aSAdrian Hunter * And vice versa: 590c73deb6aSAdrian Hunter * 591c73deb6aSAdrian Hunter * quot = cyc >> time_shift; 592b9511cd7SAdrian Hunter * rem = cyc & (((u64)1 << time_shift) - 1); 593c73deb6aSAdrian Hunter * timestamp = time_zero + quot * time_mult + 594c73deb6aSAdrian Hunter * ((rem * time_mult) >> time_shift); 595c73deb6aSAdrian Hunter */ 596c73deb6aSAdrian Hunter __u64 time_zero; 597fa731587SPeter Zijlstra __u32 size; /* Header size up to __reserved[] fields. */ 598607ca46eSDavid Howells 599607ca46eSDavid Howells /* 600607ca46eSDavid Howells * Hole for extension of the self monitor capabilities 601607ca46eSDavid Howells */ 602607ca46eSDavid Howells 603fa731587SPeter Zijlstra __u8 __reserved[118*8+4]; /* align to 1k. */ 604607ca46eSDavid Howells 605607ca46eSDavid Howells /* 606607ca46eSDavid Howells * Control data for the mmap() data buffer. 607607ca46eSDavid Howells * 608bf378d34SPeter Zijlstra * User-space reading the @data_head value should issue an smp_rmb(), 609bf378d34SPeter Zijlstra * after reading this value. 610607ca46eSDavid Howells * 611607ca46eSDavid Howells * When the mapping is PROT_WRITE the @data_tail value should be 612bf378d34SPeter Zijlstra * written by userspace to reflect the last read data, after issueing 613bf378d34SPeter Zijlstra * an smp_mb() to separate the data read from the ->data_tail store. 614bf378d34SPeter Zijlstra * In this case the kernel will not over-write unread data. 615bf378d34SPeter Zijlstra * 616bf378d34SPeter Zijlstra * See perf_output_put_handle() for the data ordering. 617e8c6deacSAlexander Shishkin * 618e8c6deacSAlexander Shishkin * data_{offset,size} indicate the location and size of the perf record 619e8c6deacSAlexander Shishkin * buffer within the mmapped area. 620607ca46eSDavid Howells */ 621607ca46eSDavid Howells __u64 data_head; /* head in the data section */ 622607ca46eSDavid Howells __u64 data_tail; /* user-space written tail */ 623e8c6deacSAlexander Shishkin __u64 data_offset; /* where the buffer starts */ 624e8c6deacSAlexander Shishkin __u64 data_size; /* data buffer size */ 62545bfb2e5SPeter Zijlstra 62645bfb2e5SPeter Zijlstra /* 62745bfb2e5SPeter Zijlstra * AUX area is defined by aux_{offset,size} fields that should be set 62845bfb2e5SPeter Zijlstra * by the userspace, so that 62945bfb2e5SPeter Zijlstra * 63045bfb2e5SPeter Zijlstra * aux_offset >= data_offset + data_size 63145bfb2e5SPeter Zijlstra * 63245bfb2e5SPeter Zijlstra * prior to mmap()ing it. Size of the mmap()ed area should be aux_size. 63345bfb2e5SPeter Zijlstra * 63445bfb2e5SPeter Zijlstra * Ring buffer pointers aux_{head,tail} have the same semantics as 63545bfb2e5SPeter Zijlstra * data_{head,tail} and same ordering rules apply. 63645bfb2e5SPeter Zijlstra */ 63745bfb2e5SPeter Zijlstra __u64 aux_head; 63845bfb2e5SPeter Zijlstra __u64 aux_tail; 63945bfb2e5SPeter Zijlstra __u64 aux_offset; 64045bfb2e5SPeter Zijlstra __u64 aux_size; 641607ca46eSDavid Howells }; 642607ca46eSDavid Howells 643607ca46eSDavid Howells #define PERF_RECORD_MISC_CPUMODE_MASK (7 << 0) 644607ca46eSDavid Howells #define PERF_RECORD_MISC_CPUMODE_UNKNOWN (0 << 0) 645607ca46eSDavid Howells #define PERF_RECORD_MISC_KERNEL (1 << 0) 646607ca46eSDavid Howells #define PERF_RECORD_MISC_USER (2 << 0) 647607ca46eSDavid Howells #define PERF_RECORD_MISC_HYPERVISOR (3 << 0) 648607ca46eSDavid Howells #define PERF_RECORD_MISC_GUEST_KERNEL (4 << 0) 649607ca46eSDavid Howells #define PERF_RECORD_MISC_GUEST_USER (5 << 0) 650607ca46eSDavid Howells 65182b89778SAdrian Hunter /* 652930e6fcdSKan Liang * Indicates that /proc/PID/maps parsing are truncated by time out. 653930e6fcdSKan Liang */ 654930e6fcdSKan Liang #define PERF_RECORD_MISC_PROC_MAP_PARSE_TIMEOUT (1 << 12) 655930e6fcdSKan Liang /* 656972c1488SJiri Olsa * Following PERF_RECORD_MISC_* are used on different 657972c1488SJiri Olsa * events, so can reuse the same bit position: 658972c1488SJiri Olsa * 659972c1488SJiri Olsa * PERF_RECORD_MISC_MMAP_DATA - PERF_RECORD_MMAP* events 660972c1488SJiri Olsa * PERF_RECORD_MISC_COMM_EXEC - PERF_RECORD_COMM event 6614f8f382eSDavid Miller * PERF_RECORD_MISC_FORK_EXEC - PERF_RECORD_FORK event (perf internal) 662972c1488SJiri Olsa * PERF_RECORD_MISC_SWITCH_OUT - PERF_RECORD_SWITCH* events 66382b89778SAdrian Hunter */ 6642fe85427SStephane Eranian #define PERF_RECORD_MISC_MMAP_DATA (1 << 13) 66582b89778SAdrian Hunter #define PERF_RECORD_MISC_COMM_EXEC (1 << 13) 6664f8f382eSDavid Miller #define PERF_RECORD_MISC_FORK_EXEC (1 << 13) 66745ac1403SAdrian Hunter #define PERF_RECORD_MISC_SWITCH_OUT (1 << 13) 668607ca46eSDavid Howells /* 669101592b4SAlexey Budankov * These PERF_RECORD_MISC_* flags below are safely reused 670101592b4SAlexey Budankov * for the following events: 671101592b4SAlexey Budankov * 672101592b4SAlexey Budankov * PERF_RECORD_MISC_EXACT_IP - PERF_RECORD_SAMPLE of precise events 673101592b4SAlexey Budankov * PERF_RECORD_MISC_SWITCH_OUT_PREEMPT - PERF_RECORD_SWITCH* events 674101592b4SAlexey Budankov * 675101592b4SAlexey Budankov * 676101592b4SAlexey Budankov * PERF_RECORD_MISC_EXACT_IP: 677607ca46eSDavid Howells * Indicates that the content of PERF_SAMPLE_IP points to 678607ca46eSDavid Howells * the actual instruction that triggered the event. See also 679607ca46eSDavid Howells * perf_event_attr::precise_ip. 680101592b4SAlexey Budankov * 681101592b4SAlexey Budankov * PERF_RECORD_MISC_SWITCH_OUT_PREEMPT: 682101592b4SAlexey Budankov * Indicates that thread was preempted in TASK_RUNNING state. 683607ca46eSDavid Howells */ 684607ca46eSDavid Howells #define PERF_RECORD_MISC_EXACT_IP (1 << 14) 685101592b4SAlexey Budankov #define PERF_RECORD_MISC_SWITCH_OUT_PREEMPT (1 << 14) 686607ca46eSDavid Howells /* 687607ca46eSDavid Howells * Reserve the last bit to indicate some extended misc field 688607ca46eSDavid Howells */ 689607ca46eSDavid Howells #define PERF_RECORD_MISC_EXT_RESERVED (1 << 15) 690607ca46eSDavid Howells 691607ca46eSDavid Howells struct perf_event_header { 692607ca46eSDavid Howells __u32 type; 693607ca46eSDavid Howells __u16 misc; 694607ca46eSDavid Howells __u16 size; 695607ca46eSDavid Howells }; 696607ca46eSDavid Howells 697e4222673SHari Bathini struct perf_ns_link_info { 698e4222673SHari Bathini __u64 dev; 699e4222673SHari Bathini __u64 ino; 700e4222673SHari Bathini }; 701e4222673SHari Bathini 702e4222673SHari Bathini enum { 703e4222673SHari Bathini NET_NS_INDEX = 0, 704e4222673SHari Bathini UTS_NS_INDEX = 1, 705e4222673SHari Bathini IPC_NS_INDEX = 2, 706e4222673SHari Bathini PID_NS_INDEX = 3, 707e4222673SHari Bathini USER_NS_INDEX = 4, 708e4222673SHari Bathini MNT_NS_INDEX = 5, 709e4222673SHari Bathini CGROUP_NS_INDEX = 6, 710e4222673SHari Bathini 711e4222673SHari Bathini NR_NAMESPACES, /* number of available namespaces */ 712e4222673SHari Bathini }; 713e4222673SHari Bathini 714607ca46eSDavid Howells enum perf_event_type { 715607ca46eSDavid Howells 716607ca46eSDavid Howells /* 717607ca46eSDavid Howells * If perf_event_attr.sample_id_all is set then all event types will 718607ca46eSDavid Howells * have the sample_type selected fields related to where/when 719ff3d527cSAdrian Hunter * (identity) an event took place (TID, TIME, ID, STREAM_ID, CPU, 720ff3d527cSAdrian Hunter * IDENTIFIER) described in PERF_RECORD_SAMPLE below, it will be stashed 721ff3d527cSAdrian Hunter * just after the perf_event_header and the fields already present for 722ff3d527cSAdrian Hunter * the existing fields, i.e. at the end of the payload. That way a newer 723ff3d527cSAdrian Hunter * perf.data file will be supported by older perf tools, with these new 724ff3d527cSAdrian Hunter * optional fields being ignored. 725607ca46eSDavid Howells * 726a5cdd40cSPeter Zijlstra * struct sample_id { 727a5cdd40cSPeter Zijlstra * { u32 pid, tid; } && PERF_SAMPLE_TID 728a5cdd40cSPeter Zijlstra * { u64 time; } && PERF_SAMPLE_TIME 729a5cdd40cSPeter Zijlstra * { u64 id; } && PERF_SAMPLE_ID 730a5cdd40cSPeter Zijlstra * { u64 stream_id;} && PERF_SAMPLE_STREAM_ID 731a5cdd40cSPeter Zijlstra * { u32 cpu, res; } && PERF_SAMPLE_CPU 732ff3d527cSAdrian Hunter * { u64 id; } && PERF_SAMPLE_IDENTIFIER 733a5cdd40cSPeter Zijlstra * } && perf_event_attr::sample_id_all 734ff3d527cSAdrian Hunter * 735ff3d527cSAdrian Hunter * Note that PERF_SAMPLE_IDENTIFIER duplicates PERF_SAMPLE_ID. The 736ff3d527cSAdrian Hunter * advantage of PERF_SAMPLE_IDENTIFIER is that its position is fixed 737ff3d527cSAdrian Hunter * relative to header.size. 738a5cdd40cSPeter Zijlstra */ 739a5cdd40cSPeter Zijlstra 740a5cdd40cSPeter Zijlstra /* 741607ca46eSDavid Howells * The MMAP events record the PROT_EXEC mappings so that we can 742607ca46eSDavid Howells * correlate userspace IPs to code. They have the following structure: 743607ca46eSDavid Howells * 744607ca46eSDavid Howells * struct { 745607ca46eSDavid Howells * struct perf_event_header header; 746607ca46eSDavid Howells * 747607ca46eSDavid Howells * u32 pid, tid; 748607ca46eSDavid Howells * u64 addr; 749607ca46eSDavid Howells * u64 len; 750607ca46eSDavid Howells * u64 pgoff; 751607ca46eSDavid Howells * char filename[]; 752c5ecceefSPeter Zijlstra * struct sample_id sample_id; 753607ca46eSDavid Howells * }; 754607ca46eSDavid Howells */ 755607ca46eSDavid Howells PERF_RECORD_MMAP = 1, 756607ca46eSDavid Howells 757607ca46eSDavid Howells /* 758607ca46eSDavid Howells * struct { 759607ca46eSDavid Howells * struct perf_event_header header; 760607ca46eSDavid Howells * u64 id; 761607ca46eSDavid Howells * u64 lost; 762a5cdd40cSPeter Zijlstra * struct sample_id sample_id; 763607ca46eSDavid Howells * }; 764607ca46eSDavid Howells */ 765607ca46eSDavid Howells PERF_RECORD_LOST = 2, 766607ca46eSDavid Howells 767607ca46eSDavid Howells /* 768607ca46eSDavid Howells * struct { 769607ca46eSDavid Howells * struct perf_event_header header; 770607ca46eSDavid Howells * 771607ca46eSDavid Howells * u32 pid, tid; 772607ca46eSDavid Howells * char comm[]; 773a5cdd40cSPeter Zijlstra * struct sample_id sample_id; 774607ca46eSDavid Howells * }; 775607ca46eSDavid Howells */ 776607ca46eSDavid Howells PERF_RECORD_COMM = 3, 777607ca46eSDavid Howells 778607ca46eSDavid Howells /* 779607ca46eSDavid Howells * struct { 780607ca46eSDavid Howells * struct perf_event_header header; 781607ca46eSDavid Howells * u32 pid, ppid; 782607ca46eSDavid Howells * u32 tid, ptid; 783607ca46eSDavid Howells * u64 time; 784a5cdd40cSPeter Zijlstra * struct sample_id sample_id; 785607ca46eSDavid Howells * }; 786607ca46eSDavid Howells */ 787607ca46eSDavid Howells PERF_RECORD_EXIT = 4, 788607ca46eSDavid Howells 789607ca46eSDavid Howells /* 790607ca46eSDavid Howells * struct { 791607ca46eSDavid Howells * struct perf_event_header header; 792607ca46eSDavid Howells * u64 time; 793607ca46eSDavid Howells * u64 id; 794607ca46eSDavid Howells * u64 stream_id; 795a5cdd40cSPeter Zijlstra * struct sample_id sample_id; 796607ca46eSDavid Howells * }; 797607ca46eSDavid Howells */ 798607ca46eSDavid Howells PERF_RECORD_THROTTLE = 5, 799607ca46eSDavid Howells PERF_RECORD_UNTHROTTLE = 6, 800607ca46eSDavid Howells 801607ca46eSDavid Howells /* 802607ca46eSDavid Howells * struct { 803607ca46eSDavid Howells * struct perf_event_header header; 804607ca46eSDavid Howells * u32 pid, ppid; 805607ca46eSDavid Howells * u32 tid, ptid; 806607ca46eSDavid Howells * u64 time; 807a5cdd40cSPeter Zijlstra * struct sample_id sample_id; 808607ca46eSDavid Howells * }; 809607ca46eSDavid Howells */ 810607ca46eSDavid Howells PERF_RECORD_FORK = 7, 811607ca46eSDavid Howells 812607ca46eSDavid Howells /* 813607ca46eSDavid Howells * struct { 814607ca46eSDavid Howells * struct perf_event_header header; 815607ca46eSDavid Howells * u32 pid, tid; 816607ca46eSDavid Howells * 817607ca46eSDavid Howells * struct read_format values; 818a5cdd40cSPeter Zijlstra * struct sample_id sample_id; 819607ca46eSDavid Howells * }; 820607ca46eSDavid Howells */ 821607ca46eSDavid Howells PERF_RECORD_READ = 8, 822607ca46eSDavid Howells 823607ca46eSDavid Howells /* 824607ca46eSDavid Howells * struct { 825607ca46eSDavid Howells * struct perf_event_header header; 826607ca46eSDavid Howells * 827ff3d527cSAdrian Hunter * # 828ff3d527cSAdrian Hunter * # Note that PERF_SAMPLE_IDENTIFIER duplicates PERF_SAMPLE_ID. 829ff3d527cSAdrian Hunter * # The advantage of PERF_SAMPLE_IDENTIFIER is that its position 830ff3d527cSAdrian Hunter * # is fixed relative to header. 831ff3d527cSAdrian Hunter * # 832ff3d527cSAdrian Hunter * 833ff3d527cSAdrian Hunter * { u64 id; } && PERF_SAMPLE_IDENTIFIER 834607ca46eSDavid Howells * { u64 ip; } && PERF_SAMPLE_IP 835607ca46eSDavid Howells * { u32 pid, tid; } && PERF_SAMPLE_TID 836607ca46eSDavid Howells * { u64 time; } && PERF_SAMPLE_TIME 837607ca46eSDavid Howells * { u64 addr; } && PERF_SAMPLE_ADDR 838607ca46eSDavid Howells * { u64 id; } && PERF_SAMPLE_ID 839607ca46eSDavid Howells * { u64 stream_id;} && PERF_SAMPLE_STREAM_ID 840607ca46eSDavid Howells * { u32 cpu, res; } && PERF_SAMPLE_CPU 841607ca46eSDavid Howells * { u64 period; } && PERF_SAMPLE_PERIOD 842607ca46eSDavid Howells * 843607ca46eSDavid Howells * { struct read_format values; } && PERF_SAMPLE_READ 844607ca46eSDavid Howells * 845607ca46eSDavid Howells * { u64 nr, 846607ca46eSDavid Howells * u64 ips[nr]; } && PERF_SAMPLE_CALLCHAIN 847607ca46eSDavid Howells * 848607ca46eSDavid Howells * # 849607ca46eSDavid Howells * # The RAW record below is opaque data wrt the ABI 850607ca46eSDavid Howells * # 851607ca46eSDavid Howells * # That is, the ABI doesn't make any promises wrt to 852607ca46eSDavid Howells * # the stability of its content, it may vary depending 853607ca46eSDavid Howells * # on event, hardware, kernel version and phase of 854607ca46eSDavid Howells * # the moon. 855607ca46eSDavid Howells * # 856607ca46eSDavid Howells * # In other words, PERF_SAMPLE_RAW contents are not an ABI. 857607ca46eSDavid Howells * # 858607ca46eSDavid Howells * 859607ca46eSDavid Howells * { u32 size; 860607ca46eSDavid Howells * char data[size];}&& PERF_SAMPLE_RAW 861607ca46eSDavid Howells * 862b878e7fbSVince Weaver * { u64 nr; 863bbfd5e4fSKan Liang * { u64 hw_idx; } && PERF_SAMPLE_BRANCH_HW_INDEX 864bbfd5e4fSKan Liang * { u64 from, to, flags } lbr[nr]; 865bbfd5e4fSKan Liang * } && PERF_SAMPLE_BRANCH_STACK 866607ca46eSDavid Howells * 867607ca46eSDavid Howells * { u64 abi; # enum perf_sample_regs_abi 868607ca46eSDavid Howells * u64 regs[weight(mask)]; } && PERF_SAMPLE_REGS_USER 869607ca46eSDavid Howells * 870607ca46eSDavid Howells * { u64 size; 871607ca46eSDavid Howells * char data[size]; 872607ca46eSDavid Howells * u64 dyn_size; } && PERF_SAMPLE_STACK_USER 873c3feedf2SAndi Kleen * 874c3feedf2SAndi Kleen * { u64 weight; } && PERF_SAMPLE_WEIGHT 875d6be9ad6SStephane Eranian * { u64 data_src; } && PERF_SAMPLE_DATA_SRC 876189b84fbSVince Weaver * { u64 transaction; } && PERF_SAMPLE_TRANSACTION 87760e2364eSStephane Eranian * { u64 abi; # enum perf_sample_regs_abi 87860e2364eSStephane Eranian * u64 regs[weight(mask)]; } && PERF_SAMPLE_REGS_INTR 879fc7ce9c7SKan Liang * { u64 phys_addr;} && PERF_SAMPLE_PHYS_ADDR 880a4faf00dSAlexander Shishkin * { u64 size; 881a4faf00dSAlexander Shishkin * char data[size]; } && PERF_SAMPLE_AUX 882607ca46eSDavid Howells * }; 883607ca46eSDavid Howells */ 884607ca46eSDavid Howells PERF_RECORD_SAMPLE = 9, 885607ca46eSDavid Howells 88613d7a241SStephane Eranian /* 88713d7a241SStephane Eranian * The MMAP2 records are an augmented version of MMAP, they add 88813d7a241SStephane Eranian * maj, min, ino numbers to be used to uniquely identify each mapping 88913d7a241SStephane Eranian * 89013d7a241SStephane Eranian * struct { 89113d7a241SStephane Eranian * struct perf_event_header header; 89213d7a241SStephane Eranian * 89313d7a241SStephane Eranian * u32 pid, tid; 89413d7a241SStephane Eranian * u64 addr; 89513d7a241SStephane Eranian * u64 len; 89613d7a241SStephane Eranian * u64 pgoff; 89713d7a241SStephane Eranian * u32 maj; 89813d7a241SStephane Eranian * u32 min; 89913d7a241SStephane Eranian * u64 ino; 90013d7a241SStephane Eranian * u64 ino_generation; 901f972eb63SPeter Zijlstra * u32 prot, flags; 90213d7a241SStephane Eranian * char filename[]; 90313d7a241SStephane Eranian * struct sample_id sample_id; 90413d7a241SStephane Eranian * }; 90513d7a241SStephane Eranian */ 90613d7a241SStephane Eranian PERF_RECORD_MMAP2 = 10, 90713d7a241SStephane Eranian 90868db7e98SAlexander Shishkin /* 90968db7e98SAlexander Shishkin * Records that new data landed in the AUX buffer part. 91068db7e98SAlexander Shishkin * 91168db7e98SAlexander Shishkin * struct { 91268db7e98SAlexander Shishkin * struct perf_event_header header; 91368db7e98SAlexander Shishkin * 91468db7e98SAlexander Shishkin * u64 aux_offset; 91568db7e98SAlexander Shishkin * u64 aux_size; 91668db7e98SAlexander Shishkin * u64 flags; 91768db7e98SAlexander Shishkin * struct sample_id sample_id; 91868db7e98SAlexander Shishkin * }; 91968db7e98SAlexander Shishkin */ 92068db7e98SAlexander Shishkin PERF_RECORD_AUX = 11, 92168db7e98SAlexander Shishkin 922ec0d7729SAlexander Shishkin /* 923ec0d7729SAlexander Shishkin * Indicates that instruction trace has started 924ec0d7729SAlexander Shishkin * 925ec0d7729SAlexander Shishkin * struct { 926ec0d7729SAlexander Shishkin * struct perf_event_header header; 927ec0d7729SAlexander Shishkin * u32 pid; 928ec0d7729SAlexander Shishkin * u32 tid; 92981df978cSJiri Olsa * struct sample_id sample_id; 930ec0d7729SAlexander Shishkin * }; 931ec0d7729SAlexander Shishkin */ 932ec0d7729SAlexander Shishkin PERF_RECORD_ITRACE_START = 12, 933ec0d7729SAlexander Shishkin 934f38b0dbbSKan Liang /* 935f38b0dbbSKan Liang * Records the dropped/lost sample number. 936f38b0dbbSKan Liang * 937f38b0dbbSKan Liang * struct { 938f38b0dbbSKan Liang * struct perf_event_header header; 939f38b0dbbSKan Liang * 940f38b0dbbSKan Liang * u64 lost; 941f38b0dbbSKan Liang * struct sample_id sample_id; 942f38b0dbbSKan Liang * }; 943f38b0dbbSKan Liang */ 944f38b0dbbSKan Liang PERF_RECORD_LOST_SAMPLES = 13, 945f38b0dbbSKan Liang 94645ac1403SAdrian Hunter /* 94745ac1403SAdrian Hunter * Records a context switch in or out (flagged by 94845ac1403SAdrian Hunter * PERF_RECORD_MISC_SWITCH_OUT). See also 94945ac1403SAdrian Hunter * PERF_RECORD_SWITCH_CPU_WIDE. 95045ac1403SAdrian Hunter * 95145ac1403SAdrian Hunter * struct { 95245ac1403SAdrian Hunter * struct perf_event_header header; 95345ac1403SAdrian Hunter * struct sample_id sample_id; 95445ac1403SAdrian Hunter * }; 95545ac1403SAdrian Hunter */ 95645ac1403SAdrian Hunter PERF_RECORD_SWITCH = 14, 95745ac1403SAdrian Hunter 95845ac1403SAdrian Hunter /* 95945ac1403SAdrian Hunter * CPU-wide version of PERF_RECORD_SWITCH with next_prev_pid and 96045ac1403SAdrian Hunter * next_prev_tid that are the next (switching out) or previous 96145ac1403SAdrian Hunter * (switching in) pid/tid. 96245ac1403SAdrian Hunter * 96345ac1403SAdrian Hunter * struct { 96445ac1403SAdrian Hunter * struct perf_event_header header; 96545ac1403SAdrian Hunter * u32 next_prev_pid; 96645ac1403SAdrian Hunter * u32 next_prev_tid; 96745ac1403SAdrian Hunter * struct sample_id sample_id; 96845ac1403SAdrian Hunter * }; 96945ac1403SAdrian Hunter */ 97045ac1403SAdrian Hunter PERF_RECORD_SWITCH_CPU_WIDE = 15, 97145ac1403SAdrian Hunter 972e4222673SHari Bathini /* 973e4222673SHari Bathini * struct { 974e4222673SHari Bathini * struct perf_event_header header; 975e4222673SHari Bathini * u32 pid; 976e4222673SHari Bathini * u32 tid; 977e4222673SHari Bathini * u64 nr_namespaces; 978e4222673SHari Bathini * { u64 dev, inode; } [nr_namespaces]; 979e4222673SHari Bathini * struct sample_id sample_id; 980e4222673SHari Bathini * }; 981e4222673SHari Bathini */ 982e4222673SHari Bathini PERF_RECORD_NAMESPACES = 16, 983e4222673SHari Bathini 98476193a94SSong Liu /* 98576193a94SSong Liu * Record ksymbol register/unregister events: 98676193a94SSong Liu * 98776193a94SSong Liu * struct { 98876193a94SSong Liu * struct perf_event_header header; 98976193a94SSong Liu * u64 addr; 99076193a94SSong Liu * u32 len; 99176193a94SSong Liu * u16 ksym_type; 99276193a94SSong Liu * u16 flags; 99376193a94SSong Liu * char name[]; 99476193a94SSong Liu * struct sample_id sample_id; 99576193a94SSong Liu * }; 99676193a94SSong Liu */ 99776193a94SSong Liu PERF_RECORD_KSYMBOL = 17, 99876193a94SSong Liu 9996ee52e2aSSong Liu /* 10006ee52e2aSSong Liu * Record bpf events: 10016ee52e2aSSong Liu * enum perf_bpf_event_type { 10026ee52e2aSSong Liu * PERF_BPF_EVENT_UNKNOWN = 0, 10036ee52e2aSSong Liu * PERF_BPF_EVENT_PROG_LOAD = 1, 10046ee52e2aSSong Liu * PERF_BPF_EVENT_PROG_UNLOAD = 2, 10056ee52e2aSSong Liu * }; 10066ee52e2aSSong Liu * 10076ee52e2aSSong Liu * struct { 10086ee52e2aSSong Liu * struct perf_event_header header; 10096ee52e2aSSong Liu * u16 type; 10106ee52e2aSSong Liu * u16 flags; 10116ee52e2aSSong Liu * u32 id; 10126ee52e2aSSong Liu * u8 tag[BPF_TAG_SIZE]; 10136ee52e2aSSong Liu * struct sample_id sample_id; 10146ee52e2aSSong Liu * }; 10156ee52e2aSSong Liu */ 10166ee52e2aSSong Liu PERF_RECORD_BPF_EVENT = 18, 10176ee52e2aSSong Liu 101896aaab68SNamhyung Kim /* 101996aaab68SNamhyung Kim * struct { 102096aaab68SNamhyung Kim * struct perf_event_header header; 102196aaab68SNamhyung Kim * u64 id; 102296aaab68SNamhyung Kim * char path[]; 102396aaab68SNamhyung Kim * struct sample_id sample_id; 102496aaab68SNamhyung Kim * }; 102596aaab68SNamhyung Kim */ 102696aaab68SNamhyung Kim PERF_RECORD_CGROUP = 19, 102796aaab68SNamhyung Kim 1028*e17d43b9SAdrian Hunter /* 1029*e17d43b9SAdrian Hunter * Records changes to kernel text i.e. self-modified code. 'old_len' is 1030*e17d43b9SAdrian Hunter * the number of old bytes, 'new_len' is the number of new bytes. Either 1031*e17d43b9SAdrian Hunter * 'old_len' or 'new_len' may be zero to indicate, for example, the 1032*e17d43b9SAdrian Hunter * addition or removal of a trampoline. 'bytes' contains the old bytes 1033*e17d43b9SAdrian Hunter * followed immediately by the new bytes. 1034*e17d43b9SAdrian Hunter * 1035*e17d43b9SAdrian Hunter * struct { 1036*e17d43b9SAdrian Hunter * struct perf_event_header header; 1037*e17d43b9SAdrian Hunter * u64 addr; 1038*e17d43b9SAdrian Hunter * u16 old_len; 1039*e17d43b9SAdrian Hunter * u16 new_len; 1040*e17d43b9SAdrian Hunter * u8 bytes[]; 1041*e17d43b9SAdrian Hunter * struct sample_id sample_id; 1042*e17d43b9SAdrian Hunter * }; 1043*e17d43b9SAdrian Hunter */ 1044*e17d43b9SAdrian Hunter PERF_RECORD_TEXT_POKE = 20, 1045*e17d43b9SAdrian Hunter 1046607ca46eSDavid Howells PERF_RECORD_MAX, /* non-ABI */ 1047607ca46eSDavid Howells }; 1048607ca46eSDavid Howells 104976193a94SSong Liu enum perf_record_ksymbol_type { 105076193a94SSong Liu PERF_RECORD_KSYMBOL_TYPE_UNKNOWN = 0, 105176193a94SSong Liu PERF_RECORD_KSYMBOL_TYPE_BPF = 1, 105276193a94SSong Liu PERF_RECORD_KSYMBOL_TYPE_MAX /* non-ABI */ 105376193a94SSong Liu }; 105476193a94SSong Liu 105576193a94SSong Liu #define PERF_RECORD_KSYMBOL_FLAGS_UNREGISTER (1 << 0) 105676193a94SSong Liu 10576ee52e2aSSong Liu enum perf_bpf_event_type { 10586ee52e2aSSong Liu PERF_BPF_EVENT_UNKNOWN = 0, 10596ee52e2aSSong Liu PERF_BPF_EVENT_PROG_LOAD = 1, 10606ee52e2aSSong Liu PERF_BPF_EVENT_PROG_UNLOAD = 2, 10616ee52e2aSSong Liu PERF_BPF_EVENT_MAX, /* non-ABI */ 10626ee52e2aSSong Liu }; 10636ee52e2aSSong Liu 1064607ca46eSDavid Howells #define PERF_MAX_STACK_DEPTH 127 1065c85b0334SArnaldo Carvalho de Melo #define PERF_MAX_CONTEXTS_PER_STACK 8 1066607ca46eSDavid Howells 1067607ca46eSDavid Howells enum perf_callchain_context { 1068607ca46eSDavid Howells PERF_CONTEXT_HV = (__u64)-32, 1069607ca46eSDavid Howells PERF_CONTEXT_KERNEL = (__u64)-128, 1070607ca46eSDavid Howells PERF_CONTEXT_USER = (__u64)-512, 1071607ca46eSDavid Howells 1072607ca46eSDavid Howells PERF_CONTEXT_GUEST = (__u64)-2048, 1073607ca46eSDavid Howells PERF_CONTEXT_GUEST_KERNEL = (__u64)-2176, 1074607ca46eSDavid Howells PERF_CONTEXT_GUEST_USER = (__u64)-2560, 1075607ca46eSDavid Howells 1076607ca46eSDavid Howells PERF_CONTEXT_MAX = (__u64)-4095, 1077607ca46eSDavid Howells }; 1078607ca46eSDavid Howells 107968db7e98SAlexander Shishkin /** 108068db7e98SAlexander Shishkin * PERF_RECORD_AUX::flags bits 108168db7e98SAlexander Shishkin */ 108268db7e98SAlexander Shishkin #define PERF_AUX_FLAG_TRUNCATED 0x01 /* record was truncated to fit */ 10832023a0d2SAlexander Shishkin #define PERF_AUX_FLAG_OVERWRITE 0x02 /* snapshot from overwrite mode */ 1084ae0c2d99SAlexander Shishkin #define PERF_AUX_FLAG_PARTIAL 0x04 /* record contains gaps */ 1085085b3062SWill Deacon #define PERF_AUX_FLAG_COLLISION 0x08 /* sample collided with another */ 108668db7e98SAlexander Shishkin 1087643fd0b9SPeter Zijlstra #define PERF_FLAG_FD_NO_GROUP (1UL << 0) 1088643fd0b9SPeter Zijlstra #define PERF_FLAG_FD_OUTPUT (1UL << 1) 1089643fd0b9SPeter Zijlstra #define PERF_FLAG_PID_CGROUP (1UL << 2) /* pid=cgroup id, per-cpu mode only */ 1090643fd0b9SPeter Zijlstra #define PERF_FLAG_FD_CLOEXEC (1UL << 3) /* O_CLOEXEC */ 1091607ca46eSDavid Howells 10928c5073dbSSukadev Bhattiprolu #if defined(__LITTLE_ENDIAN_BITFIELD) 1093d6be9ad6SStephane Eranian union perf_mem_data_src { 1094d6be9ad6SStephane Eranian __u64 val; 1095d6be9ad6SStephane Eranian struct { 1096d6be9ad6SStephane Eranian __u64 mem_op:5, /* type of opcode */ 1097d6be9ad6SStephane Eranian mem_lvl:14, /* memory hierarchy level */ 1098d6be9ad6SStephane Eranian mem_snoop:5, /* snoop mode */ 1099d6be9ad6SStephane Eranian mem_lock:2, /* lock instr */ 1100d6be9ad6SStephane Eranian mem_dtlb:7, /* tlb access */ 11016ae5fa61SAndi Kleen mem_lvl_num:4, /* memory hierarchy level number */ 11026ae5fa61SAndi Kleen mem_remote:1, /* remote */ 11036ae5fa61SAndi Kleen mem_snoopx:2, /* snoop mode, ext */ 11046ae5fa61SAndi Kleen mem_rsvd:24; 1105d6be9ad6SStephane Eranian }; 1106d6be9ad6SStephane Eranian }; 11078c5073dbSSukadev Bhattiprolu #elif defined(__BIG_ENDIAN_BITFIELD) 11088c5073dbSSukadev Bhattiprolu union perf_mem_data_src { 11098c5073dbSSukadev Bhattiprolu __u64 val; 11108c5073dbSSukadev Bhattiprolu struct { 11116ae5fa61SAndi Kleen __u64 mem_rsvd:24, 11126ae5fa61SAndi Kleen mem_snoopx:2, /* snoop mode, ext */ 11136ae5fa61SAndi Kleen mem_remote:1, /* remote */ 11146ae5fa61SAndi Kleen mem_lvl_num:4, /* memory hierarchy level number */ 11158c5073dbSSukadev Bhattiprolu mem_dtlb:7, /* tlb access */ 11168c5073dbSSukadev Bhattiprolu mem_lock:2, /* lock instr */ 11178c5073dbSSukadev Bhattiprolu mem_snoop:5, /* snoop mode */ 11188c5073dbSSukadev Bhattiprolu mem_lvl:14, /* memory hierarchy level */ 11198c5073dbSSukadev Bhattiprolu mem_op:5; /* type of opcode */ 11208c5073dbSSukadev Bhattiprolu }; 11218c5073dbSSukadev Bhattiprolu }; 11228c5073dbSSukadev Bhattiprolu #else 11238c5073dbSSukadev Bhattiprolu #error "Unknown endianness" 11248c5073dbSSukadev Bhattiprolu #endif 1125d6be9ad6SStephane Eranian 1126d6be9ad6SStephane Eranian /* type of opcode (load/store/prefetch,code) */ 1127d6be9ad6SStephane Eranian #define PERF_MEM_OP_NA 0x01 /* not available */ 1128d6be9ad6SStephane Eranian #define PERF_MEM_OP_LOAD 0x02 /* load instruction */ 1129d6be9ad6SStephane Eranian #define PERF_MEM_OP_STORE 0x04 /* store instruction */ 1130d6be9ad6SStephane Eranian #define PERF_MEM_OP_PFETCH 0x08 /* prefetch */ 1131d6be9ad6SStephane Eranian #define PERF_MEM_OP_EXEC 0x10 /* code (execution) */ 1132d6be9ad6SStephane Eranian #define PERF_MEM_OP_SHIFT 0 1133d6be9ad6SStephane Eranian 1134d6be9ad6SStephane Eranian /* memory hierarchy (memory level, hit or miss) */ 1135d6be9ad6SStephane Eranian #define PERF_MEM_LVL_NA 0x01 /* not available */ 1136d6be9ad6SStephane Eranian #define PERF_MEM_LVL_HIT 0x02 /* hit level */ 1137d6be9ad6SStephane Eranian #define PERF_MEM_LVL_MISS 0x04 /* miss level */ 1138d6be9ad6SStephane Eranian #define PERF_MEM_LVL_L1 0x08 /* L1 */ 1139d6be9ad6SStephane Eranian #define PERF_MEM_LVL_LFB 0x10 /* Line Fill Buffer */ 1140cc2f5a8aSStephane Eranian #define PERF_MEM_LVL_L2 0x20 /* L2 */ 1141cc2f5a8aSStephane Eranian #define PERF_MEM_LVL_L3 0x40 /* L3 */ 1142d6be9ad6SStephane Eranian #define PERF_MEM_LVL_LOC_RAM 0x80 /* Local DRAM */ 1143d6be9ad6SStephane Eranian #define PERF_MEM_LVL_REM_RAM1 0x100 /* Remote DRAM (1 hop) */ 1144d6be9ad6SStephane Eranian #define PERF_MEM_LVL_REM_RAM2 0x200 /* Remote DRAM (2 hops) */ 1145d6be9ad6SStephane Eranian #define PERF_MEM_LVL_REM_CCE1 0x400 /* Remote Cache (1 hop) */ 1146d6be9ad6SStephane Eranian #define PERF_MEM_LVL_REM_CCE2 0x800 /* Remote Cache (2 hops) */ 1147d6be9ad6SStephane Eranian #define PERF_MEM_LVL_IO 0x1000 /* I/O memory */ 1148d6be9ad6SStephane Eranian #define PERF_MEM_LVL_UNC 0x2000 /* Uncached memory */ 1149d6be9ad6SStephane Eranian #define PERF_MEM_LVL_SHIFT 5 1150d6be9ad6SStephane Eranian 11516ae5fa61SAndi Kleen #define PERF_MEM_REMOTE_REMOTE 0x01 /* Remote */ 11526ae5fa61SAndi Kleen #define PERF_MEM_REMOTE_SHIFT 37 11536ae5fa61SAndi Kleen 11546ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_L1 0x01 /* L1 */ 11556ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_L2 0x02 /* L2 */ 11566ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_L3 0x03 /* L3 */ 11576ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_L4 0x04 /* L4 */ 11586ae5fa61SAndi Kleen /* 5-0xa available */ 11596ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_ANY_CACHE 0x0b /* Any cache */ 11606ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_LFB 0x0c /* LFB */ 11616ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_RAM 0x0d /* RAM */ 11626ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_PMEM 0x0e /* PMEM */ 11636ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_NA 0x0f /* N/A */ 11646ae5fa61SAndi Kleen 11656ae5fa61SAndi Kleen #define PERF_MEM_LVLNUM_SHIFT 33 11666ae5fa61SAndi Kleen 1167d6be9ad6SStephane Eranian /* snoop mode */ 1168d6be9ad6SStephane Eranian #define PERF_MEM_SNOOP_NA 0x01 /* not available */ 1169d6be9ad6SStephane Eranian #define PERF_MEM_SNOOP_NONE 0x02 /* no snoop */ 1170d6be9ad6SStephane Eranian #define PERF_MEM_SNOOP_HIT 0x04 /* snoop hit */ 1171d6be9ad6SStephane Eranian #define PERF_MEM_SNOOP_MISS 0x08 /* snoop miss */ 1172d6be9ad6SStephane Eranian #define PERF_MEM_SNOOP_HITM 0x10 /* snoop hit modified */ 1173d6be9ad6SStephane Eranian #define PERF_MEM_SNOOP_SHIFT 19 1174d6be9ad6SStephane Eranian 11756ae5fa61SAndi Kleen #define PERF_MEM_SNOOPX_FWD 0x01 /* forward */ 11766ae5fa61SAndi Kleen /* 1 free */ 11776ae5fa61SAndi Kleen #define PERF_MEM_SNOOPX_SHIFT 37 11786ae5fa61SAndi Kleen 1179d6be9ad6SStephane Eranian /* locked instruction */ 1180d6be9ad6SStephane Eranian #define PERF_MEM_LOCK_NA 0x01 /* not available */ 1181d6be9ad6SStephane Eranian #define PERF_MEM_LOCK_LOCKED 0x02 /* locked transaction */ 1182d6be9ad6SStephane Eranian #define PERF_MEM_LOCK_SHIFT 24 1183d6be9ad6SStephane Eranian 1184d6be9ad6SStephane Eranian /* TLB access */ 1185d6be9ad6SStephane Eranian #define PERF_MEM_TLB_NA 0x01 /* not available */ 1186d6be9ad6SStephane Eranian #define PERF_MEM_TLB_HIT 0x02 /* hit level */ 1187d6be9ad6SStephane Eranian #define PERF_MEM_TLB_MISS 0x04 /* miss level */ 1188d6be9ad6SStephane Eranian #define PERF_MEM_TLB_L1 0x08 /* L1 */ 1189d6be9ad6SStephane Eranian #define PERF_MEM_TLB_L2 0x10 /* L2 */ 1190d6be9ad6SStephane Eranian #define PERF_MEM_TLB_WK 0x20 /* Hardware Walker*/ 1191d6be9ad6SStephane Eranian #define PERF_MEM_TLB_OS 0x40 /* OS fault handler */ 1192d6be9ad6SStephane Eranian #define PERF_MEM_TLB_SHIFT 26 1193d6be9ad6SStephane Eranian 1194d6be9ad6SStephane Eranian #define PERF_MEM_S(a, s) \ 11950d9dfc23SMike Frysinger (((__u64)PERF_MEM_##a##_##s) << PERF_MEM_##a##_SHIFT) 1196d6be9ad6SStephane Eranian 1197274481deSVince Weaver /* 1198274481deSVince Weaver * single taken branch record layout: 1199274481deSVince Weaver * 1200274481deSVince Weaver * from: source instruction (may not always be a branch insn) 1201274481deSVince Weaver * to: branch target 1202274481deSVince Weaver * mispred: branch target was mispredicted 1203274481deSVince Weaver * predicted: branch target was predicted 1204274481deSVince Weaver * 1205274481deSVince Weaver * support for mispred, predicted is optional. In case it 1206274481deSVince Weaver * is not supported mispred = predicted = 0. 1207274481deSVince Weaver * 1208274481deSVince Weaver * in_tx: running in a hardware transaction 1209274481deSVince Weaver * abort: aborting a hardware transaction 121071ef3c6bSAndi Kleen * cycles: cycles from last branch (or 0 if not supported) 1211eb0baf8aSJin Yao * type: branch type 1212274481deSVince Weaver */ 1213274481deSVince Weaver struct perf_branch_entry { 1214274481deSVince Weaver __u64 from; 1215274481deSVince Weaver __u64 to; 1216274481deSVince Weaver __u64 mispred:1, /* target mispredicted */ 1217274481deSVince Weaver predicted:1,/* target predicted */ 1218274481deSVince Weaver in_tx:1, /* in transaction */ 1219274481deSVince Weaver abort:1, /* transaction abort */ 122071ef3c6bSAndi Kleen cycles:16, /* cycle count to last branch */ 1221eb0baf8aSJin Yao type:4, /* branch type */ 1222eb0baf8aSJin Yao reserved:40; 1223274481deSVince Weaver }; 1224274481deSVince Weaver 1225607ca46eSDavid Howells #endif /* _UAPI_LINUX_PERF_EVENT_H */ 1226