/linux/include/trace/events/ |
H A D | kmem.h | 14 TP_PROTO(unsigned long call_site, 20 TP_ARGS(call_site, ptr, s, gfp_flags, node), 23 __field( unsigned long, call_site ) 33 __entry->call_site = call_site; 45 (void *)__entry->call_site, 56 TP_PROTO(unsigned long call_site, 63 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node), 66 __field( unsigned long, call_site ) 75 __entry->call_site = call_site; 84 (void *)__entry->call_site, [all …]
|
H A D | percpu.h | 13 TP_PROTO(unsigned long call_site, 18 TP_ARGS(call_site, reserved, is_atomic, size, align, base_addr, off, 22 __field( unsigned long, call_site ) 34 __entry->call_site = call_site; 47 (void *)__entry->call_site,
|
/linux/Documentation/trace/ |
H A D | histogram.rst | 231 field:unsigned long call_site; offset:8; size:8; signed:0; 241 # echo 'hist:key=call_site:val=bytes_req.buckets=32' > \ 245 call_site field of the kmalloc event as the key for the table, which 246 just means that each unique call_site address will have an entry 248 the hist trigger that for each unique entry (call_site) in the 250 requested by that call_site. 257 # trigger info: hist:keys=call_site:vals=bytes_req:sort=hitcount:size=2048 [active] 259 { call_site: 18446744072106379007 } hitcount: 1 bytes_req: 176 260 { call_site: 18446744071579557049 } hitcount: 1 bytes_req: 1024 261 { call_site: 18446744071580608289 } hitcount: 1 bytes_req: 16384 [all …]
|
H A D | events-kmem.rst | 21 kmalloc call_site=%lx ptr=%p bytes_req=%zu bytes_alloc=%zu gfp_flags=%s 22 kmalloc_node call_site=%lx ptr=%p bytes_req=%zu bytes_alloc=%zu gfp_flags=%s node=%d 23 kfree call_site=%lx ptr=%p 36 kmem_cache_alloc call_site=%lx ptr=%p bytes_req=%zu bytes_alloc=%zu gfp_flags=%s 37 kmem_cache_alloc_node call_site=%lx ptr=%p bytes_req=%zu bytes_alloc=%zu gfp_flags=%s node=%d 38 kmem_cache_free call_site=%lx ptr=%p 43 but the call_site can usually be used to extrapolate that information.
|
H A D | histogram-design.rst | 279 …# echo 'hist:keys=common_pid,call_site.sym:values=bytes_req,bytes_alloc,hitcount' >> events/kmem/k… 298 …# trigger info: hist:keys=common_pid,call_site.sym:vals=hitcount,bytes_req,bytes_alloc:sort=hitcou… 345 ftrace_event_field name: call_site 352 …# echo '!hist:keys=common_pid,call_site.sym:values=bytes_req,bytes_alloc,hitcount' >> events/kmem/…
|
H A D | events.rst | 212 call_site.function == security_prepare_creds 214 The above will filter when the field "call_site" falls on the address within 215 "security_prepare_creds". That is, it will compare the value of "call_site" and
|
/linux/tools/perf/ |
H A D | builtin-kmem.c | 62 u64 call_site; member 87 static int insert_alloc_stat(unsigned long call_site, unsigned long ptr, in insert_alloc_stat() argument 125 data->call_site = call_site; in insert_alloc_stat() 132 static int insert_caller_stat(unsigned long call_site, in insert_caller_stat() argument 143 if (call_site > data->call_site) in insert_caller_stat() 145 else if (call_site < data->call_site) in insert_caller_stat() 151 if (data && data->call_site == call_site) { in insert_caller_stat() 161 data->call_site = call_site; in insert_caller_stat() 177 call_site = evsel__intval(evsel, sample, "call_site"); in evsel__process_alloc_event() local 181 if (insert_alloc_stat(call_site, ptr, bytes_req, bytes_alloc, sample->cpu) || in evsel__process_alloc_event() [all …]
|
/linux/tools/perf/scripts/python/ |
H A D | check-perf-trace.py | 42 common_callchain, call_site, ptr, bytes_req, bytes_alloc, argument 51 (call_site, ptr, bytes_req, bytes_alloc,
|
/linux/tools/perf/scripts/perl/ |
H A D | check-perf-trace.pl | 47 $call_site, $ptr, $bytes_req, $bytes_alloc, 57 $call_site, $ptr, $bytes_req, $bytes_alloc,
|
/linux/tools/testing/selftests/bpf/progs/ |
H A D | bench_local_storage_create.c | 33 int BPF_PROG(kmalloc, unsigned long call_site, const void *ptr, in BPF_PROG() argument
|