1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * ring buffer based function tracer 4 * 5 * Copyright (C) 2007-2012 Steven Rostedt <srostedt@redhat.com> 6 * Copyright (C) 2008 Ingo Molnar <mingo@redhat.com> 7 * 8 * Originally taken from the RT patch by: 9 * Arnaldo Carvalho de Melo <acme@redhat.com> 10 * 11 * Based on code from the latency_tracer, that is: 12 * Copyright (C) 2004-2006 Ingo Molnar 13 * Copyright (C) 2004 Nadia Yvette Chambers 14 */ 15 #include <linux/ring_buffer.h> 16 #include <generated/utsrelease.h> 17 #include <linux/stacktrace.h> 18 #include <linux/writeback.h> 19 #include <linux/kallsyms.h> 20 #include <linux/security.h> 21 #include <linux/seq_file.h> 22 #include <linux/notifier.h> 23 #include <linux/irqflags.h> 24 #include <linux/debugfs.h> 25 #include <linux/tracefs.h> 26 #include <linux/pagemap.h> 27 #include <linux/hardirq.h> 28 #include <linux/linkage.h> 29 #include <linux/uaccess.h> 30 #include <linux/vmalloc.h> 31 #include <linux/ftrace.h> 32 #include <linux/module.h> 33 #include <linux/percpu.h> 34 #include <linux/splice.h> 35 #include <linux/kdebug.h> 36 #include <linux/string.h> 37 #include <linux/mount.h> 38 #include <linux/rwsem.h> 39 #include <linux/slab.h> 40 #include <linux/ctype.h> 41 #include <linux/init.h> 42 #include <linux/poll.h> 43 #include <linux/nmi.h> 44 #include <linux/fs.h> 45 #include <linux/trace.h> 46 #include <linux/sched/clock.h> 47 #include <linux/sched/rt.h> 48 #include <linux/fsnotify.h> 49 #include <linux/irq_work.h> 50 #include <linux/workqueue.h> 51 52 #include "trace.h" 53 #include "trace_output.h" 54 55 /* 56 * On boot up, the ring buffer is set to the minimum size, so that 57 * we do not waste memory on systems that are not using tracing. 58 */ 59 bool ring_buffer_expanded; 60 61 /* 62 * We need to change this state when a selftest is running. 63 * A selftest will lurk into the ring-buffer to count the 64 * entries inserted during the selftest although some concurrent 65 * insertions into the ring-buffer such as trace_printk could occurred 66 * at the same time, giving false positive or negative results. 67 */ 68 static bool __read_mostly tracing_selftest_running; 69 70 /* 71 * If boot-time tracing including tracers/events via kernel cmdline 72 * is running, we do not want to run SELFTEST. 73 */ 74 bool __read_mostly tracing_selftest_disabled; 75 76 #ifdef CONFIG_FTRACE_STARTUP_TEST 77 void __init disable_tracing_selftest(const char *reason) 78 { 79 if (!tracing_selftest_disabled) { 80 tracing_selftest_disabled = true; 81 pr_info("Ftrace startup test is disabled due to %s\n", reason); 82 } 83 } 84 #endif 85 86 /* Pipe tracepoints to printk */ 87 struct trace_iterator *tracepoint_print_iter; 88 int tracepoint_printk; 89 static DEFINE_STATIC_KEY_FALSE(tracepoint_printk_key); 90 91 /* For tracers that don't implement custom flags */ 92 static struct tracer_opt dummy_tracer_opt[] = { 93 { } 94 }; 95 96 static int 97 dummy_set_flag(struct trace_array *tr, u32 old_flags, u32 bit, int set) 98 { 99 return 0; 100 } 101 102 /* 103 * To prevent the comm cache from being overwritten when no 104 * tracing is active, only save the comm when a trace event 105 * occurred. 106 */ 107 static DEFINE_PER_CPU(bool, trace_taskinfo_save); 108 109 /* 110 * Kill all tracing for good (never come back). 111 * It is initialized to 1 but will turn to zero if the initialization 112 * of the tracer is successful. But that is the only place that sets 113 * this back to zero. 114 */ 115 static int tracing_disabled = 1; 116 117 cpumask_var_t __read_mostly tracing_buffer_mask; 118 119 /* 120 * ftrace_dump_on_oops - variable to dump ftrace buffer on oops 121 * 122 * If there is an oops (or kernel panic) and the ftrace_dump_on_oops 123 * is set, then ftrace_dump is called. This will output the contents 124 * of the ftrace buffers to the console. This is very useful for 125 * capturing traces that lead to crashes and outputing it to a 126 * serial console. 127 * 128 * It is default off, but you can enable it with either specifying 129 * "ftrace_dump_on_oops" in the kernel command line, or setting 130 * /proc/sys/kernel/ftrace_dump_on_oops 131 * Set 1 if you want to dump buffers of all CPUs 132 * Set 2 if you want to dump the buffer of the CPU that triggered oops 133 */ 134 135 enum ftrace_dump_mode ftrace_dump_on_oops; 136 137 /* When set, tracing will stop when a WARN*() is hit */ 138 int __disable_trace_on_warning; 139 140 #ifdef CONFIG_TRACE_EVAL_MAP_FILE 141 /* Map of enums to their values, for "eval_map" file */ 142 struct trace_eval_map_head { 143 struct module *mod; 144 unsigned long length; 145 }; 146 147 union trace_eval_map_item; 148 149 struct trace_eval_map_tail { 150 /* 151 * "end" is first and points to NULL as it must be different 152 * than "mod" or "eval_string" 153 */ 154 union trace_eval_map_item *next; 155 const char *end; /* points to NULL */ 156 }; 157 158 static DEFINE_MUTEX(trace_eval_mutex); 159 160 /* 161 * The trace_eval_maps are saved in an array with two extra elements, 162 * one at the beginning, and one at the end. The beginning item contains 163 * the count of the saved maps (head.length), and the module they 164 * belong to if not built in (head.mod). The ending item contains a 165 * pointer to the next array of saved eval_map items. 166 */ 167 union trace_eval_map_item { 168 struct trace_eval_map map; 169 struct trace_eval_map_head head; 170 struct trace_eval_map_tail tail; 171 }; 172 173 static union trace_eval_map_item *trace_eval_maps; 174 #endif /* CONFIG_TRACE_EVAL_MAP_FILE */ 175 176 int tracing_set_tracer(struct trace_array *tr, const char *buf); 177 static void ftrace_trace_userstack(struct trace_array *tr, 178 struct trace_buffer *buffer, 179 unsigned long flags, int pc); 180 181 #define MAX_TRACER_SIZE 100 182 static char bootup_tracer_buf[MAX_TRACER_SIZE] __initdata; 183 static char *default_bootup_tracer; 184 185 static bool allocate_snapshot; 186 187 static int __init set_cmdline_ftrace(char *str) 188 { 189 strlcpy(bootup_tracer_buf, str, MAX_TRACER_SIZE); 190 default_bootup_tracer = bootup_tracer_buf; 191 /* We are using ftrace early, expand it */ 192 ring_buffer_expanded = true; 193 return 1; 194 } 195 __setup("ftrace=", set_cmdline_ftrace); 196 197 static int __init set_ftrace_dump_on_oops(char *str) 198 { 199 if (*str++ != '=' || !*str) { 200 ftrace_dump_on_oops = DUMP_ALL; 201 return 1; 202 } 203 204 if (!strcmp("orig_cpu", str)) { 205 ftrace_dump_on_oops = DUMP_ORIG; 206 return 1; 207 } 208 209 return 0; 210 } 211 __setup("ftrace_dump_on_oops", set_ftrace_dump_on_oops); 212 213 static int __init stop_trace_on_warning(char *str) 214 { 215 if ((strcmp(str, "=0") != 0 && strcmp(str, "=off") != 0)) 216 __disable_trace_on_warning = 1; 217 return 1; 218 } 219 __setup("traceoff_on_warning", stop_trace_on_warning); 220 221 static int __init boot_alloc_snapshot(char *str) 222 { 223 allocate_snapshot = true; 224 /* We also need the main ring buffer expanded */ 225 ring_buffer_expanded = true; 226 return 1; 227 } 228 __setup("alloc_snapshot", boot_alloc_snapshot); 229 230 231 static char trace_boot_options_buf[MAX_TRACER_SIZE] __initdata; 232 233 static int __init set_trace_boot_options(char *str) 234 { 235 strlcpy(trace_boot_options_buf, str, MAX_TRACER_SIZE); 236 return 0; 237 } 238 __setup("trace_options=", set_trace_boot_options); 239 240 static char trace_boot_clock_buf[MAX_TRACER_SIZE] __initdata; 241 static char *trace_boot_clock __initdata; 242 243 static int __init set_trace_boot_clock(char *str) 244 { 245 strlcpy(trace_boot_clock_buf, str, MAX_TRACER_SIZE); 246 trace_boot_clock = trace_boot_clock_buf; 247 return 0; 248 } 249 __setup("trace_clock=", set_trace_boot_clock); 250 251 static int __init set_tracepoint_printk(char *str) 252 { 253 if ((strcmp(str, "=0") != 0 && strcmp(str, "=off") != 0)) 254 tracepoint_printk = 1; 255 return 1; 256 } 257 __setup("tp_printk", set_tracepoint_printk); 258 259 unsigned long long ns2usecs(u64 nsec) 260 { 261 nsec += 500; 262 do_div(nsec, 1000); 263 return nsec; 264 } 265 266 static void 267 trace_process_export(struct trace_export *export, 268 struct ring_buffer_event *event, int flag) 269 { 270 struct trace_entry *entry; 271 unsigned int size = 0; 272 273 if (export->flags & flag) { 274 entry = ring_buffer_event_data(event); 275 size = ring_buffer_event_length(event); 276 export->write(export, entry, size); 277 } 278 } 279 280 static DEFINE_MUTEX(ftrace_export_lock); 281 282 static struct trace_export __rcu *ftrace_exports_list __read_mostly; 283 284 static DEFINE_STATIC_KEY_FALSE(trace_function_exports_enabled); 285 static DEFINE_STATIC_KEY_FALSE(trace_event_exports_enabled); 286 static DEFINE_STATIC_KEY_FALSE(trace_marker_exports_enabled); 287 288 static inline void ftrace_exports_enable(struct trace_export *export) 289 { 290 if (export->flags & TRACE_EXPORT_FUNCTION) 291 static_branch_inc(&trace_function_exports_enabled); 292 293 if (export->flags & TRACE_EXPORT_EVENT) 294 static_branch_inc(&trace_event_exports_enabled); 295 296 if (export->flags & TRACE_EXPORT_MARKER) 297 static_branch_inc(&trace_marker_exports_enabled); 298 } 299 300 static inline void ftrace_exports_disable(struct trace_export *export) 301 { 302 if (export->flags & TRACE_EXPORT_FUNCTION) 303 static_branch_dec(&trace_function_exports_enabled); 304 305 if (export->flags & TRACE_EXPORT_EVENT) 306 static_branch_dec(&trace_event_exports_enabled); 307 308 if (export->flags & TRACE_EXPORT_MARKER) 309 static_branch_dec(&trace_marker_exports_enabled); 310 } 311 312 static void ftrace_exports(struct ring_buffer_event *event, int flag) 313 { 314 struct trace_export *export; 315 316 preempt_disable_notrace(); 317 318 export = rcu_dereference_raw_check(ftrace_exports_list); 319 while (export) { 320 trace_process_export(export, event, flag); 321 export = rcu_dereference_raw_check(export->next); 322 } 323 324 preempt_enable_notrace(); 325 } 326 327 static inline void 328 add_trace_export(struct trace_export **list, struct trace_export *export) 329 { 330 rcu_assign_pointer(export->next, *list); 331 /* 332 * We are entering export into the list but another 333 * CPU might be walking that list. We need to make sure 334 * the export->next pointer is valid before another CPU sees 335 * the export pointer included into the list. 336 */ 337 rcu_assign_pointer(*list, export); 338 } 339 340 static inline int 341 rm_trace_export(struct trace_export **list, struct trace_export *export) 342 { 343 struct trace_export **p; 344 345 for (p = list; *p != NULL; p = &(*p)->next) 346 if (*p == export) 347 break; 348 349 if (*p != export) 350 return -1; 351 352 rcu_assign_pointer(*p, (*p)->next); 353 354 return 0; 355 } 356 357 static inline void 358 add_ftrace_export(struct trace_export **list, struct trace_export *export) 359 { 360 ftrace_exports_enable(export); 361 362 add_trace_export(list, export); 363 } 364 365 static inline int 366 rm_ftrace_export(struct trace_export **list, struct trace_export *export) 367 { 368 int ret; 369 370 ret = rm_trace_export(list, export); 371 ftrace_exports_disable(export); 372 373 return ret; 374 } 375 376 int register_ftrace_export(struct trace_export *export) 377 { 378 if (WARN_ON_ONCE(!export->write)) 379 return -1; 380 381 mutex_lock(&ftrace_export_lock); 382 383 add_ftrace_export(&ftrace_exports_list, export); 384 385 mutex_unlock(&ftrace_export_lock); 386 387 return 0; 388 } 389 EXPORT_SYMBOL_GPL(register_ftrace_export); 390 391 int unregister_ftrace_export(struct trace_export *export) 392 { 393 int ret; 394 395 mutex_lock(&ftrace_export_lock); 396 397 ret = rm_ftrace_export(&ftrace_exports_list, export); 398 399 mutex_unlock(&ftrace_export_lock); 400 401 return ret; 402 } 403 EXPORT_SYMBOL_GPL(unregister_ftrace_export); 404 405 /* trace_flags holds trace_options default values */ 406 #define TRACE_DEFAULT_FLAGS \ 407 (FUNCTION_DEFAULT_FLAGS | \ 408 TRACE_ITER_PRINT_PARENT | TRACE_ITER_PRINTK | \ 409 TRACE_ITER_ANNOTATE | TRACE_ITER_CONTEXT_INFO | \ 410 TRACE_ITER_RECORD_CMD | TRACE_ITER_OVERWRITE | \ 411 TRACE_ITER_IRQ_INFO | TRACE_ITER_MARKERS) 412 413 /* trace_options that are only supported by global_trace */ 414 #define TOP_LEVEL_TRACE_FLAGS (TRACE_ITER_PRINTK | \ 415 TRACE_ITER_PRINTK_MSGONLY | TRACE_ITER_RECORD_CMD) 416 417 /* trace_flags that are default zero for instances */ 418 #define ZEROED_TRACE_FLAGS \ 419 (TRACE_ITER_EVENT_FORK | TRACE_ITER_FUNC_FORK) 420 421 /* 422 * The global_trace is the descriptor that holds the top-level tracing 423 * buffers for the live tracing. 424 */ 425 static struct trace_array global_trace = { 426 .trace_flags = TRACE_DEFAULT_FLAGS, 427 }; 428 429 LIST_HEAD(ftrace_trace_arrays); 430 431 int trace_array_get(struct trace_array *this_tr) 432 { 433 struct trace_array *tr; 434 int ret = -ENODEV; 435 436 mutex_lock(&trace_types_lock); 437 list_for_each_entry(tr, &ftrace_trace_arrays, list) { 438 if (tr == this_tr) { 439 tr->ref++; 440 ret = 0; 441 break; 442 } 443 } 444 mutex_unlock(&trace_types_lock); 445 446 return ret; 447 } 448 449 static void __trace_array_put(struct trace_array *this_tr) 450 { 451 WARN_ON(!this_tr->ref); 452 this_tr->ref--; 453 } 454 455 /** 456 * trace_array_put - Decrement the reference counter for this trace array. 457 * 458 * NOTE: Use this when we no longer need the trace array returned by 459 * trace_array_get_by_name(). This ensures the trace array can be later 460 * destroyed. 461 * 462 */ 463 void trace_array_put(struct trace_array *this_tr) 464 { 465 if (!this_tr) 466 return; 467 468 mutex_lock(&trace_types_lock); 469 __trace_array_put(this_tr); 470 mutex_unlock(&trace_types_lock); 471 } 472 EXPORT_SYMBOL_GPL(trace_array_put); 473 474 int tracing_check_open_get_tr(struct trace_array *tr) 475 { 476 int ret; 477 478 ret = security_locked_down(LOCKDOWN_TRACEFS); 479 if (ret) 480 return ret; 481 482 if (tracing_disabled) 483 return -ENODEV; 484 485 if (tr && trace_array_get(tr) < 0) 486 return -ENODEV; 487 488 return 0; 489 } 490 491 int call_filter_check_discard(struct trace_event_call *call, void *rec, 492 struct trace_buffer *buffer, 493 struct ring_buffer_event *event) 494 { 495 if (unlikely(call->flags & TRACE_EVENT_FL_FILTERED) && 496 !filter_match_preds(call->filter, rec)) { 497 __trace_event_discard_commit(buffer, event); 498 return 1; 499 } 500 501 return 0; 502 } 503 504 void trace_free_pid_list(struct trace_pid_list *pid_list) 505 { 506 vfree(pid_list->pids); 507 kfree(pid_list); 508 } 509 510 /** 511 * trace_find_filtered_pid - check if a pid exists in a filtered_pid list 512 * @filtered_pids: The list of pids to check 513 * @search_pid: The PID to find in @filtered_pids 514 * 515 * Returns true if @search_pid is fonud in @filtered_pids, and false otherwis. 516 */ 517 bool 518 trace_find_filtered_pid(struct trace_pid_list *filtered_pids, pid_t search_pid) 519 { 520 /* 521 * If pid_max changed after filtered_pids was created, we 522 * by default ignore all pids greater than the previous pid_max. 523 */ 524 if (search_pid >= filtered_pids->pid_max) 525 return false; 526 527 return test_bit(search_pid, filtered_pids->pids); 528 } 529 530 /** 531 * trace_ignore_this_task - should a task be ignored for tracing 532 * @filtered_pids: The list of pids to check 533 * @task: The task that should be ignored if not filtered 534 * 535 * Checks if @task should be traced or not from @filtered_pids. 536 * Returns true if @task should *NOT* be traced. 537 * Returns false if @task should be traced. 538 */ 539 bool 540 trace_ignore_this_task(struct trace_pid_list *filtered_pids, 541 struct trace_pid_list *filtered_no_pids, 542 struct task_struct *task) 543 { 544 /* 545 * If filterd_no_pids is not empty, and the task's pid is listed 546 * in filtered_no_pids, then return true. 547 * Otherwise, if filtered_pids is empty, that means we can 548 * trace all tasks. If it has content, then only trace pids 549 * within filtered_pids. 550 */ 551 552 return (filtered_pids && 553 !trace_find_filtered_pid(filtered_pids, task->pid)) || 554 (filtered_no_pids && 555 trace_find_filtered_pid(filtered_no_pids, task->pid)); 556 } 557 558 /** 559 * trace_filter_add_remove_task - Add or remove a task from a pid_list 560 * @pid_list: The list to modify 561 * @self: The current task for fork or NULL for exit 562 * @task: The task to add or remove 563 * 564 * If adding a task, if @self is defined, the task is only added if @self 565 * is also included in @pid_list. This happens on fork and tasks should 566 * only be added when the parent is listed. If @self is NULL, then the 567 * @task pid will be removed from the list, which would happen on exit 568 * of a task. 569 */ 570 void trace_filter_add_remove_task(struct trace_pid_list *pid_list, 571 struct task_struct *self, 572 struct task_struct *task) 573 { 574 if (!pid_list) 575 return; 576 577 /* For forks, we only add if the forking task is listed */ 578 if (self) { 579 if (!trace_find_filtered_pid(pid_list, self->pid)) 580 return; 581 } 582 583 /* Sorry, but we don't support pid_max changing after setting */ 584 if (task->pid >= pid_list->pid_max) 585 return; 586 587 /* "self" is set for forks, and NULL for exits */ 588 if (self) 589 set_bit(task->pid, pid_list->pids); 590 else 591 clear_bit(task->pid, pid_list->pids); 592 } 593 594 /** 595 * trace_pid_next - Used for seq_file to get to the next pid of a pid_list 596 * @pid_list: The pid list to show 597 * @v: The last pid that was shown (+1 the actual pid to let zero be displayed) 598 * @pos: The position of the file 599 * 600 * This is used by the seq_file "next" operation to iterate the pids 601 * listed in a trace_pid_list structure. 602 * 603 * Returns the pid+1 as we want to display pid of zero, but NULL would 604 * stop the iteration. 605 */ 606 void *trace_pid_next(struct trace_pid_list *pid_list, void *v, loff_t *pos) 607 { 608 unsigned long pid = (unsigned long)v; 609 610 (*pos)++; 611 612 /* pid already is +1 of the actual prevous bit */ 613 pid = find_next_bit(pid_list->pids, pid_list->pid_max, pid); 614 615 /* Return pid + 1 to allow zero to be represented */ 616 if (pid < pid_list->pid_max) 617 return (void *)(pid + 1); 618 619 return NULL; 620 } 621 622 /** 623 * trace_pid_start - Used for seq_file to start reading pid lists 624 * @pid_list: The pid list to show 625 * @pos: The position of the file 626 * 627 * This is used by seq_file "start" operation to start the iteration 628 * of listing pids. 629 * 630 * Returns the pid+1 as we want to display pid of zero, but NULL would 631 * stop the iteration. 632 */ 633 void *trace_pid_start(struct trace_pid_list *pid_list, loff_t *pos) 634 { 635 unsigned long pid; 636 loff_t l = 0; 637 638 pid = find_first_bit(pid_list->pids, pid_list->pid_max); 639 if (pid >= pid_list->pid_max) 640 return NULL; 641 642 /* Return pid + 1 so that zero can be the exit value */ 643 for (pid++; pid && l < *pos; 644 pid = (unsigned long)trace_pid_next(pid_list, (void *)pid, &l)) 645 ; 646 return (void *)pid; 647 } 648 649 /** 650 * trace_pid_show - show the current pid in seq_file processing 651 * @m: The seq_file structure to write into 652 * @v: A void pointer of the pid (+1) value to display 653 * 654 * Can be directly used by seq_file operations to display the current 655 * pid value. 656 */ 657 int trace_pid_show(struct seq_file *m, void *v) 658 { 659 unsigned long pid = (unsigned long)v - 1; 660 661 seq_printf(m, "%lu\n", pid); 662 return 0; 663 } 664 665 /* 128 should be much more than enough */ 666 #define PID_BUF_SIZE 127 667 668 int trace_pid_write(struct trace_pid_list *filtered_pids, 669 struct trace_pid_list **new_pid_list, 670 const char __user *ubuf, size_t cnt) 671 { 672 struct trace_pid_list *pid_list; 673 struct trace_parser parser; 674 unsigned long val; 675 int nr_pids = 0; 676 ssize_t read = 0; 677 ssize_t ret = 0; 678 loff_t pos; 679 pid_t pid; 680 681 if (trace_parser_get_init(&parser, PID_BUF_SIZE + 1)) 682 return -ENOMEM; 683 684 /* 685 * Always recreate a new array. The write is an all or nothing 686 * operation. Always create a new array when adding new pids by 687 * the user. If the operation fails, then the current list is 688 * not modified. 689 */ 690 pid_list = kmalloc(sizeof(*pid_list), GFP_KERNEL); 691 if (!pid_list) { 692 trace_parser_put(&parser); 693 return -ENOMEM; 694 } 695 696 pid_list->pid_max = READ_ONCE(pid_max); 697 698 /* Only truncating will shrink pid_max */ 699 if (filtered_pids && filtered_pids->pid_max > pid_list->pid_max) 700 pid_list->pid_max = filtered_pids->pid_max; 701 702 pid_list->pids = vzalloc((pid_list->pid_max + 7) >> 3); 703 if (!pid_list->pids) { 704 trace_parser_put(&parser); 705 kfree(pid_list); 706 return -ENOMEM; 707 } 708 709 if (filtered_pids) { 710 /* copy the current bits to the new max */ 711 for_each_set_bit(pid, filtered_pids->pids, 712 filtered_pids->pid_max) { 713 set_bit(pid, pid_list->pids); 714 nr_pids++; 715 } 716 } 717 718 while (cnt > 0) { 719 720 pos = 0; 721 722 ret = trace_get_user(&parser, ubuf, cnt, &pos); 723 if (ret < 0 || !trace_parser_loaded(&parser)) 724 break; 725 726 read += ret; 727 ubuf += ret; 728 cnt -= ret; 729 730 ret = -EINVAL; 731 if (kstrtoul(parser.buffer, 0, &val)) 732 break; 733 if (val >= pid_list->pid_max) 734 break; 735 736 pid = (pid_t)val; 737 738 set_bit(pid, pid_list->pids); 739 nr_pids++; 740 741 trace_parser_clear(&parser); 742 ret = 0; 743 } 744 trace_parser_put(&parser); 745 746 if (ret < 0) { 747 trace_free_pid_list(pid_list); 748 return ret; 749 } 750 751 if (!nr_pids) { 752 /* Cleared the list of pids */ 753 trace_free_pid_list(pid_list); 754 read = ret; 755 pid_list = NULL; 756 } 757 758 *new_pid_list = pid_list; 759 760 return read; 761 } 762 763 static u64 buffer_ftrace_now(struct array_buffer *buf, int cpu) 764 { 765 u64 ts; 766 767 /* Early boot up does not have a buffer yet */ 768 if (!buf->buffer) 769 return trace_clock_local(); 770 771 ts = ring_buffer_time_stamp(buf->buffer, cpu); 772 ring_buffer_normalize_time_stamp(buf->buffer, cpu, &ts); 773 774 return ts; 775 } 776 777 u64 ftrace_now(int cpu) 778 { 779 return buffer_ftrace_now(&global_trace.array_buffer, cpu); 780 } 781 782 /** 783 * tracing_is_enabled - Show if global_trace has been disabled 784 * 785 * Shows if the global trace has been enabled or not. It uses the 786 * mirror flag "buffer_disabled" to be used in fast paths such as for 787 * the irqsoff tracer. But it may be inaccurate due to races. If you 788 * need to know the accurate state, use tracing_is_on() which is a little 789 * slower, but accurate. 790 */ 791 int tracing_is_enabled(void) 792 { 793 /* 794 * For quick access (irqsoff uses this in fast path), just 795 * return the mirror variable of the state of the ring buffer. 796 * It's a little racy, but we don't really care. 797 */ 798 smp_rmb(); 799 return !global_trace.buffer_disabled; 800 } 801 802 /* 803 * trace_buf_size is the size in bytes that is allocated 804 * for a buffer. Note, the number of bytes is always rounded 805 * to page size. 806 * 807 * This number is purposely set to a low number of 16384. 808 * If the dump on oops happens, it will be much appreciated 809 * to not have to wait for all that output. Anyway this can be 810 * boot time and run time configurable. 811 */ 812 #define TRACE_BUF_SIZE_DEFAULT 1441792UL /* 16384 * 88 (sizeof(entry)) */ 813 814 static unsigned long trace_buf_size = TRACE_BUF_SIZE_DEFAULT; 815 816 /* trace_types holds a link list of available tracers. */ 817 static struct tracer *trace_types __read_mostly; 818 819 /* 820 * trace_types_lock is used to protect the trace_types list. 821 */ 822 DEFINE_MUTEX(trace_types_lock); 823 824 /* 825 * serialize the access of the ring buffer 826 * 827 * ring buffer serializes readers, but it is low level protection. 828 * The validity of the events (which returns by ring_buffer_peek() ..etc) 829 * are not protected by ring buffer. 830 * 831 * The content of events may become garbage if we allow other process consumes 832 * these events concurrently: 833 * A) the page of the consumed events may become a normal page 834 * (not reader page) in ring buffer, and this page will be rewrited 835 * by events producer. 836 * B) The page of the consumed events may become a page for splice_read, 837 * and this page will be returned to system. 838 * 839 * These primitives allow multi process access to different cpu ring buffer 840 * concurrently. 841 * 842 * These primitives don't distinguish read-only and read-consume access. 843 * Multi read-only access are also serialized. 844 */ 845 846 #ifdef CONFIG_SMP 847 static DECLARE_RWSEM(all_cpu_access_lock); 848 static DEFINE_PER_CPU(struct mutex, cpu_access_lock); 849 850 static inline void trace_access_lock(int cpu) 851 { 852 if (cpu == RING_BUFFER_ALL_CPUS) { 853 /* gain it for accessing the whole ring buffer. */ 854 down_write(&all_cpu_access_lock); 855 } else { 856 /* gain it for accessing a cpu ring buffer. */ 857 858 /* Firstly block other trace_access_lock(RING_BUFFER_ALL_CPUS). */ 859 down_read(&all_cpu_access_lock); 860 861 /* Secondly block other access to this @cpu ring buffer. */ 862 mutex_lock(&per_cpu(cpu_access_lock, cpu)); 863 } 864 } 865 866 static inline void trace_access_unlock(int cpu) 867 { 868 if (cpu == RING_BUFFER_ALL_CPUS) { 869 up_write(&all_cpu_access_lock); 870 } else { 871 mutex_unlock(&per_cpu(cpu_access_lock, cpu)); 872 up_read(&all_cpu_access_lock); 873 } 874 } 875 876 static inline void trace_access_lock_init(void) 877 { 878 int cpu; 879 880 for_each_possible_cpu(cpu) 881 mutex_init(&per_cpu(cpu_access_lock, cpu)); 882 } 883 884 #else 885 886 static DEFINE_MUTEX(access_lock); 887 888 static inline void trace_access_lock(int cpu) 889 { 890 (void)cpu; 891 mutex_lock(&access_lock); 892 } 893 894 static inline void trace_access_unlock(int cpu) 895 { 896 (void)cpu; 897 mutex_unlock(&access_lock); 898 } 899 900 static inline void trace_access_lock_init(void) 901 { 902 } 903 904 #endif 905 906 #ifdef CONFIG_STACKTRACE 907 static void __ftrace_trace_stack(struct trace_buffer *buffer, 908 unsigned long flags, 909 int skip, int pc, struct pt_regs *regs); 910 static inline void ftrace_trace_stack(struct trace_array *tr, 911 struct trace_buffer *buffer, 912 unsigned long flags, 913 int skip, int pc, struct pt_regs *regs); 914 915 #else 916 static inline void __ftrace_trace_stack(struct trace_buffer *buffer, 917 unsigned long flags, 918 int skip, int pc, struct pt_regs *regs) 919 { 920 } 921 static inline void ftrace_trace_stack(struct trace_array *tr, 922 struct trace_buffer *buffer, 923 unsigned long flags, 924 int skip, int pc, struct pt_regs *regs) 925 { 926 } 927 928 #endif 929 930 static __always_inline void 931 trace_event_setup(struct ring_buffer_event *event, 932 int type, unsigned long flags, int pc) 933 { 934 struct trace_entry *ent = ring_buffer_event_data(event); 935 936 tracing_generic_entry_update(ent, type, flags, pc); 937 } 938 939 static __always_inline struct ring_buffer_event * 940 __trace_buffer_lock_reserve(struct trace_buffer *buffer, 941 int type, 942 unsigned long len, 943 unsigned long flags, int pc) 944 { 945 struct ring_buffer_event *event; 946 947 event = ring_buffer_lock_reserve(buffer, len); 948 if (event != NULL) 949 trace_event_setup(event, type, flags, pc); 950 951 return event; 952 } 953 954 void tracer_tracing_on(struct trace_array *tr) 955 { 956 if (tr->array_buffer.buffer) 957 ring_buffer_record_on(tr->array_buffer.buffer); 958 /* 959 * This flag is looked at when buffers haven't been allocated 960 * yet, or by some tracers (like irqsoff), that just want to 961 * know if the ring buffer has been disabled, but it can handle 962 * races of where it gets disabled but we still do a record. 963 * As the check is in the fast path of the tracers, it is more 964 * important to be fast than accurate. 965 */ 966 tr->buffer_disabled = 0; 967 /* Make the flag seen by readers */ 968 smp_wmb(); 969 } 970 971 /** 972 * tracing_on - enable tracing buffers 973 * 974 * This function enables tracing buffers that may have been 975 * disabled with tracing_off. 976 */ 977 void tracing_on(void) 978 { 979 tracer_tracing_on(&global_trace); 980 } 981 EXPORT_SYMBOL_GPL(tracing_on); 982 983 984 static __always_inline void 985 __buffer_unlock_commit(struct trace_buffer *buffer, struct ring_buffer_event *event) 986 { 987 __this_cpu_write(trace_taskinfo_save, true); 988 989 /* If this is the temp buffer, we need to commit fully */ 990 if (this_cpu_read(trace_buffered_event) == event) { 991 /* Length is in event->array[0] */ 992 ring_buffer_write(buffer, event->array[0], &event->array[1]); 993 /* Release the temp buffer */ 994 this_cpu_dec(trace_buffered_event_cnt); 995 } else 996 ring_buffer_unlock_commit(buffer, event); 997 } 998 999 /** 1000 * __trace_puts - write a constant string into the trace buffer. 1001 * @ip: The address of the caller 1002 * @str: The constant string to write 1003 * @size: The size of the string. 1004 */ 1005 int __trace_puts(unsigned long ip, const char *str, int size) 1006 { 1007 struct ring_buffer_event *event; 1008 struct trace_buffer *buffer; 1009 struct print_entry *entry; 1010 unsigned long irq_flags; 1011 int alloc; 1012 int pc; 1013 1014 if (!(global_trace.trace_flags & TRACE_ITER_PRINTK)) 1015 return 0; 1016 1017 pc = preempt_count(); 1018 1019 if (unlikely(tracing_selftest_running || tracing_disabled)) 1020 return 0; 1021 1022 alloc = sizeof(*entry) + size + 2; /* possible \n added */ 1023 1024 local_save_flags(irq_flags); 1025 buffer = global_trace.array_buffer.buffer; 1026 ring_buffer_nest_start(buffer); 1027 event = __trace_buffer_lock_reserve(buffer, TRACE_PRINT, alloc, 1028 irq_flags, pc); 1029 if (!event) { 1030 size = 0; 1031 goto out; 1032 } 1033 1034 entry = ring_buffer_event_data(event); 1035 entry->ip = ip; 1036 1037 memcpy(&entry->buf, str, size); 1038 1039 /* Add a newline if necessary */ 1040 if (entry->buf[size - 1] != '\n') { 1041 entry->buf[size] = '\n'; 1042 entry->buf[size + 1] = '\0'; 1043 } else 1044 entry->buf[size] = '\0'; 1045 1046 __buffer_unlock_commit(buffer, event); 1047 ftrace_trace_stack(&global_trace, buffer, irq_flags, 4, pc, NULL); 1048 out: 1049 ring_buffer_nest_end(buffer); 1050 return size; 1051 } 1052 EXPORT_SYMBOL_GPL(__trace_puts); 1053 1054 /** 1055 * __trace_bputs - write the pointer to a constant string into trace buffer 1056 * @ip: The address of the caller 1057 * @str: The constant string to write to the buffer to 1058 */ 1059 int __trace_bputs(unsigned long ip, const char *str) 1060 { 1061 struct ring_buffer_event *event; 1062 struct trace_buffer *buffer; 1063 struct bputs_entry *entry; 1064 unsigned long irq_flags; 1065 int size = sizeof(struct bputs_entry); 1066 int ret = 0; 1067 int pc; 1068 1069 if (!(global_trace.trace_flags & TRACE_ITER_PRINTK)) 1070 return 0; 1071 1072 pc = preempt_count(); 1073 1074 if (unlikely(tracing_selftest_running || tracing_disabled)) 1075 return 0; 1076 1077 local_save_flags(irq_flags); 1078 buffer = global_trace.array_buffer.buffer; 1079 1080 ring_buffer_nest_start(buffer); 1081 event = __trace_buffer_lock_reserve(buffer, TRACE_BPUTS, size, 1082 irq_flags, pc); 1083 if (!event) 1084 goto out; 1085 1086 entry = ring_buffer_event_data(event); 1087 entry->ip = ip; 1088 entry->str = str; 1089 1090 __buffer_unlock_commit(buffer, event); 1091 ftrace_trace_stack(&global_trace, buffer, irq_flags, 4, pc, NULL); 1092 1093 ret = 1; 1094 out: 1095 ring_buffer_nest_end(buffer); 1096 return ret; 1097 } 1098 EXPORT_SYMBOL_GPL(__trace_bputs); 1099 1100 #ifdef CONFIG_TRACER_SNAPSHOT 1101 static void tracing_snapshot_instance_cond(struct trace_array *tr, 1102 void *cond_data) 1103 { 1104 struct tracer *tracer = tr->current_trace; 1105 unsigned long flags; 1106 1107 if (in_nmi()) { 1108 internal_trace_puts("*** SNAPSHOT CALLED FROM NMI CONTEXT ***\n"); 1109 internal_trace_puts("*** snapshot is being ignored ***\n"); 1110 return; 1111 } 1112 1113 if (!tr->allocated_snapshot) { 1114 internal_trace_puts("*** SNAPSHOT NOT ALLOCATED ***\n"); 1115 internal_trace_puts("*** stopping trace here! ***\n"); 1116 tracing_off(); 1117 return; 1118 } 1119 1120 /* Note, snapshot can not be used when the tracer uses it */ 1121 if (tracer->use_max_tr) { 1122 internal_trace_puts("*** LATENCY TRACER ACTIVE ***\n"); 1123 internal_trace_puts("*** Can not use snapshot (sorry) ***\n"); 1124 return; 1125 } 1126 1127 local_irq_save(flags); 1128 update_max_tr(tr, current, smp_processor_id(), cond_data); 1129 local_irq_restore(flags); 1130 } 1131 1132 void tracing_snapshot_instance(struct trace_array *tr) 1133 { 1134 tracing_snapshot_instance_cond(tr, NULL); 1135 } 1136 1137 /** 1138 * tracing_snapshot - take a snapshot of the current buffer. 1139 * 1140 * This causes a swap between the snapshot buffer and the current live 1141 * tracing buffer. You can use this to take snapshots of the live 1142 * trace when some condition is triggered, but continue to trace. 1143 * 1144 * Note, make sure to allocate the snapshot with either 1145 * a tracing_snapshot_alloc(), or by doing it manually 1146 * with: echo 1 > /sys/kernel/debug/tracing/snapshot 1147 * 1148 * If the snapshot buffer is not allocated, it will stop tracing. 1149 * Basically making a permanent snapshot. 1150 */ 1151 void tracing_snapshot(void) 1152 { 1153 struct trace_array *tr = &global_trace; 1154 1155 tracing_snapshot_instance(tr); 1156 } 1157 EXPORT_SYMBOL_GPL(tracing_snapshot); 1158 1159 /** 1160 * tracing_snapshot_cond - conditionally take a snapshot of the current buffer. 1161 * @tr: The tracing instance to snapshot 1162 * @cond_data: The data to be tested conditionally, and possibly saved 1163 * 1164 * This is the same as tracing_snapshot() except that the snapshot is 1165 * conditional - the snapshot will only happen if the 1166 * cond_snapshot.update() implementation receiving the cond_data 1167 * returns true, which means that the trace array's cond_snapshot 1168 * update() operation used the cond_data to determine whether the 1169 * snapshot should be taken, and if it was, presumably saved it along 1170 * with the snapshot. 1171 */ 1172 void tracing_snapshot_cond(struct trace_array *tr, void *cond_data) 1173 { 1174 tracing_snapshot_instance_cond(tr, cond_data); 1175 } 1176 EXPORT_SYMBOL_GPL(tracing_snapshot_cond); 1177 1178 /** 1179 * tracing_snapshot_cond_data - get the user data associated with a snapshot 1180 * @tr: The tracing instance 1181 * 1182 * When the user enables a conditional snapshot using 1183 * tracing_snapshot_cond_enable(), the user-defined cond_data is saved 1184 * with the snapshot. This accessor is used to retrieve it. 1185 * 1186 * Should not be called from cond_snapshot.update(), since it takes 1187 * the tr->max_lock lock, which the code calling 1188 * cond_snapshot.update() has already done. 1189 * 1190 * Returns the cond_data associated with the trace array's snapshot. 1191 */ 1192 void *tracing_cond_snapshot_data(struct trace_array *tr) 1193 { 1194 void *cond_data = NULL; 1195 1196 arch_spin_lock(&tr->max_lock); 1197 1198 if (tr->cond_snapshot) 1199 cond_data = tr->cond_snapshot->cond_data; 1200 1201 arch_spin_unlock(&tr->max_lock); 1202 1203 return cond_data; 1204 } 1205 EXPORT_SYMBOL_GPL(tracing_cond_snapshot_data); 1206 1207 static int resize_buffer_duplicate_size(struct array_buffer *trace_buf, 1208 struct array_buffer *size_buf, int cpu_id); 1209 static void set_buffer_entries(struct array_buffer *buf, unsigned long val); 1210 1211 int tracing_alloc_snapshot_instance(struct trace_array *tr) 1212 { 1213 int ret; 1214 1215 if (!tr->allocated_snapshot) { 1216 1217 /* allocate spare buffer */ 1218 ret = resize_buffer_duplicate_size(&tr->max_buffer, 1219 &tr->array_buffer, RING_BUFFER_ALL_CPUS); 1220 if (ret < 0) 1221 return ret; 1222 1223 tr->allocated_snapshot = true; 1224 } 1225 1226 return 0; 1227 } 1228 1229 static void free_snapshot(struct trace_array *tr) 1230 { 1231 /* 1232 * We don't free the ring buffer. instead, resize it because 1233 * The max_tr ring buffer has some state (e.g. ring->clock) and 1234 * we want preserve it. 1235 */ 1236 ring_buffer_resize(tr->max_buffer.buffer, 1, RING_BUFFER_ALL_CPUS); 1237 set_buffer_entries(&tr->max_buffer, 1); 1238 tracing_reset_online_cpus(&tr->max_buffer); 1239 tr->allocated_snapshot = false; 1240 } 1241 1242 /** 1243 * tracing_alloc_snapshot - allocate snapshot buffer. 1244 * 1245 * This only allocates the snapshot buffer if it isn't already 1246 * allocated - it doesn't also take a snapshot. 1247 * 1248 * This is meant to be used in cases where the snapshot buffer needs 1249 * to be set up for events that can't sleep but need to be able to 1250 * trigger a snapshot. 1251 */ 1252 int tracing_alloc_snapshot(void) 1253 { 1254 struct trace_array *tr = &global_trace; 1255 int ret; 1256 1257 ret = tracing_alloc_snapshot_instance(tr); 1258 WARN_ON(ret < 0); 1259 1260 return ret; 1261 } 1262 EXPORT_SYMBOL_GPL(tracing_alloc_snapshot); 1263 1264 /** 1265 * tracing_snapshot_alloc - allocate and take a snapshot of the current buffer. 1266 * 1267 * This is similar to tracing_snapshot(), but it will allocate the 1268 * snapshot buffer if it isn't already allocated. Use this only 1269 * where it is safe to sleep, as the allocation may sleep. 1270 * 1271 * This causes a swap between the snapshot buffer and the current live 1272 * tracing buffer. You can use this to take snapshots of the live 1273 * trace when some condition is triggered, but continue to trace. 1274 */ 1275 void tracing_snapshot_alloc(void) 1276 { 1277 int ret; 1278 1279 ret = tracing_alloc_snapshot(); 1280 if (ret < 0) 1281 return; 1282 1283 tracing_snapshot(); 1284 } 1285 EXPORT_SYMBOL_GPL(tracing_snapshot_alloc); 1286 1287 /** 1288 * tracing_snapshot_cond_enable - enable conditional snapshot for an instance 1289 * @tr: The tracing instance 1290 * @cond_data: User data to associate with the snapshot 1291 * @update: Implementation of the cond_snapshot update function 1292 * 1293 * Check whether the conditional snapshot for the given instance has 1294 * already been enabled, or if the current tracer is already using a 1295 * snapshot; if so, return -EBUSY, else create a cond_snapshot and 1296 * save the cond_data and update function inside. 1297 * 1298 * Returns 0 if successful, error otherwise. 1299 */ 1300 int tracing_snapshot_cond_enable(struct trace_array *tr, void *cond_data, 1301 cond_update_fn_t update) 1302 { 1303 struct cond_snapshot *cond_snapshot; 1304 int ret = 0; 1305 1306 cond_snapshot = kzalloc(sizeof(*cond_snapshot), GFP_KERNEL); 1307 if (!cond_snapshot) 1308 return -ENOMEM; 1309 1310 cond_snapshot->cond_data = cond_data; 1311 cond_snapshot->update = update; 1312 1313 mutex_lock(&trace_types_lock); 1314 1315 ret = tracing_alloc_snapshot_instance(tr); 1316 if (ret) 1317 goto fail_unlock; 1318 1319 if (tr->current_trace->use_max_tr) { 1320 ret = -EBUSY; 1321 goto fail_unlock; 1322 } 1323 1324 /* 1325 * The cond_snapshot can only change to NULL without the 1326 * trace_types_lock. We don't care if we race with it going 1327 * to NULL, but we want to make sure that it's not set to 1328 * something other than NULL when we get here, which we can 1329 * do safely with only holding the trace_types_lock and not 1330 * having to take the max_lock. 1331 */ 1332 if (tr->cond_snapshot) { 1333 ret = -EBUSY; 1334 goto fail_unlock; 1335 } 1336 1337 arch_spin_lock(&tr->max_lock); 1338 tr->cond_snapshot = cond_snapshot; 1339 arch_spin_unlock(&tr->max_lock); 1340 1341 mutex_unlock(&trace_types_lock); 1342 1343 return ret; 1344 1345 fail_unlock: 1346 mutex_unlock(&trace_types_lock); 1347 kfree(cond_snapshot); 1348 return ret; 1349 } 1350 EXPORT_SYMBOL_GPL(tracing_snapshot_cond_enable); 1351 1352 /** 1353 * tracing_snapshot_cond_disable - disable conditional snapshot for an instance 1354 * @tr: The tracing instance 1355 * 1356 * Check whether the conditional snapshot for the given instance is 1357 * enabled; if so, free the cond_snapshot associated with it, 1358 * otherwise return -EINVAL. 1359 * 1360 * Returns 0 if successful, error otherwise. 1361 */ 1362 int tracing_snapshot_cond_disable(struct trace_array *tr) 1363 { 1364 int ret = 0; 1365 1366 arch_spin_lock(&tr->max_lock); 1367 1368 if (!tr->cond_snapshot) 1369 ret = -EINVAL; 1370 else { 1371 kfree(tr->cond_snapshot); 1372 tr->cond_snapshot = NULL; 1373 } 1374 1375 arch_spin_unlock(&tr->max_lock); 1376 1377 return ret; 1378 } 1379 EXPORT_SYMBOL_GPL(tracing_snapshot_cond_disable); 1380 #else 1381 void tracing_snapshot(void) 1382 { 1383 WARN_ONCE(1, "Snapshot feature not enabled, but internal snapshot used"); 1384 } 1385 EXPORT_SYMBOL_GPL(tracing_snapshot); 1386 void tracing_snapshot_cond(struct trace_array *tr, void *cond_data) 1387 { 1388 WARN_ONCE(1, "Snapshot feature not enabled, but internal conditional snapshot used"); 1389 } 1390 EXPORT_SYMBOL_GPL(tracing_snapshot_cond); 1391 int tracing_alloc_snapshot(void) 1392 { 1393 WARN_ONCE(1, "Snapshot feature not enabled, but snapshot allocation used"); 1394 return -ENODEV; 1395 } 1396 EXPORT_SYMBOL_GPL(tracing_alloc_snapshot); 1397 void tracing_snapshot_alloc(void) 1398 { 1399 /* Give warning */ 1400 tracing_snapshot(); 1401 } 1402 EXPORT_SYMBOL_GPL(tracing_snapshot_alloc); 1403 void *tracing_cond_snapshot_data(struct trace_array *tr) 1404 { 1405 return NULL; 1406 } 1407 EXPORT_SYMBOL_GPL(tracing_cond_snapshot_data); 1408 int tracing_snapshot_cond_enable(struct trace_array *tr, void *cond_data, cond_update_fn_t update) 1409 { 1410 return -ENODEV; 1411 } 1412 EXPORT_SYMBOL_GPL(tracing_snapshot_cond_enable); 1413 int tracing_snapshot_cond_disable(struct trace_array *tr) 1414 { 1415 return false; 1416 } 1417 EXPORT_SYMBOL_GPL(tracing_snapshot_cond_disable); 1418 #endif /* CONFIG_TRACER_SNAPSHOT */ 1419 1420 void tracer_tracing_off(struct trace_array *tr) 1421 { 1422 if (tr->array_buffer.buffer) 1423 ring_buffer_record_off(tr->array_buffer.buffer); 1424 /* 1425 * This flag is looked at when buffers haven't been allocated 1426 * yet, or by some tracers (like irqsoff), that just want to 1427 * know if the ring buffer has been disabled, but it can handle 1428 * races of where it gets disabled but we still do a record. 1429 * As the check is in the fast path of the tracers, it is more 1430 * important to be fast than accurate. 1431 */ 1432 tr->buffer_disabled = 1; 1433 /* Make the flag seen by readers */ 1434 smp_wmb(); 1435 } 1436 1437 /** 1438 * tracing_off - turn off tracing buffers 1439 * 1440 * This function stops the tracing buffers from recording data. 1441 * It does not disable any overhead the tracers themselves may 1442 * be causing. This function simply causes all recording to 1443 * the ring buffers to fail. 1444 */ 1445 void tracing_off(void) 1446 { 1447 tracer_tracing_off(&global_trace); 1448 } 1449 EXPORT_SYMBOL_GPL(tracing_off); 1450 1451 void disable_trace_on_warning(void) 1452 { 1453 if (__disable_trace_on_warning) { 1454 trace_array_printk_buf(global_trace.array_buffer.buffer, _THIS_IP_, 1455 "Disabling tracing due to warning\n"); 1456 tracing_off(); 1457 } 1458 } 1459 1460 /** 1461 * tracer_tracing_is_on - show real state of ring buffer enabled 1462 * @tr : the trace array to know if ring buffer is enabled 1463 * 1464 * Shows real state of the ring buffer if it is enabled or not. 1465 */ 1466 bool tracer_tracing_is_on(struct trace_array *tr) 1467 { 1468 if (tr->array_buffer.buffer) 1469 return ring_buffer_record_is_on(tr->array_buffer.buffer); 1470 return !tr->buffer_disabled; 1471 } 1472 1473 /** 1474 * tracing_is_on - show state of ring buffers enabled 1475 */ 1476 int tracing_is_on(void) 1477 { 1478 return tracer_tracing_is_on(&global_trace); 1479 } 1480 EXPORT_SYMBOL_GPL(tracing_is_on); 1481 1482 static int __init set_buf_size(char *str) 1483 { 1484 unsigned long buf_size; 1485 1486 if (!str) 1487 return 0; 1488 buf_size = memparse(str, &str); 1489 /* nr_entries can not be zero */ 1490 if (buf_size == 0) 1491 return 0; 1492 trace_buf_size = buf_size; 1493 return 1; 1494 } 1495 __setup("trace_buf_size=", set_buf_size); 1496 1497 static int __init set_tracing_thresh(char *str) 1498 { 1499 unsigned long threshold; 1500 int ret; 1501 1502 if (!str) 1503 return 0; 1504 ret = kstrtoul(str, 0, &threshold); 1505 if (ret < 0) 1506 return 0; 1507 tracing_thresh = threshold * 1000; 1508 return 1; 1509 } 1510 __setup("tracing_thresh=", set_tracing_thresh); 1511 1512 unsigned long nsecs_to_usecs(unsigned long nsecs) 1513 { 1514 return nsecs / 1000; 1515 } 1516 1517 /* 1518 * TRACE_FLAGS is defined as a tuple matching bit masks with strings. 1519 * It uses C(a, b) where 'a' is the eval (enum) name and 'b' is the string that 1520 * matches it. By defining "C(a, b) b", TRACE_FLAGS becomes a list 1521 * of strings in the order that the evals (enum) were defined. 1522 */ 1523 #undef C 1524 #define C(a, b) b 1525 1526 /* These must match the bit postions in trace_iterator_flags */ 1527 static const char *trace_options[] = { 1528 TRACE_FLAGS 1529 NULL 1530 }; 1531 1532 static struct { 1533 u64 (*func)(void); 1534 const char *name; 1535 int in_ns; /* is this clock in nanoseconds? */ 1536 } trace_clocks[] = { 1537 { trace_clock_local, "local", 1 }, 1538 { trace_clock_global, "global", 1 }, 1539 { trace_clock_counter, "counter", 0 }, 1540 { trace_clock_jiffies, "uptime", 0 }, 1541 { trace_clock, "perf", 1 }, 1542 { ktime_get_mono_fast_ns, "mono", 1 }, 1543 { ktime_get_raw_fast_ns, "mono_raw", 1 }, 1544 { ktime_get_boot_fast_ns, "boot", 1 }, 1545 ARCH_TRACE_CLOCKS 1546 }; 1547 1548 bool trace_clock_in_ns(struct trace_array *tr) 1549 { 1550 if (trace_clocks[tr->clock_id].in_ns) 1551 return true; 1552 1553 return false; 1554 } 1555 1556 /* 1557 * trace_parser_get_init - gets the buffer for trace parser 1558 */ 1559 int trace_parser_get_init(struct trace_parser *parser, int size) 1560 { 1561 memset(parser, 0, sizeof(*parser)); 1562 1563 parser->buffer = kmalloc(size, GFP_KERNEL); 1564 if (!parser->buffer) 1565 return 1; 1566 1567 parser->size = size; 1568 return 0; 1569 } 1570 1571 /* 1572 * trace_parser_put - frees the buffer for trace parser 1573 */ 1574 void trace_parser_put(struct trace_parser *parser) 1575 { 1576 kfree(parser->buffer); 1577 parser->buffer = NULL; 1578 } 1579 1580 /* 1581 * trace_get_user - reads the user input string separated by space 1582 * (matched by isspace(ch)) 1583 * 1584 * For each string found the 'struct trace_parser' is updated, 1585 * and the function returns. 1586 * 1587 * Returns number of bytes read. 1588 * 1589 * See kernel/trace/trace.h for 'struct trace_parser' details. 1590 */ 1591 int trace_get_user(struct trace_parser *parser, const char __user *ubuf, 1592 size_t cnt, loff_t *ppos) 1593 { 1594 char ch; 1595 size_t read = 0; 1596 ssize_t ret; 1597 1598 if (!*ppos) 1599 trace_parser_clear(parser); 1600 1601 ret = get_user(ch, ubuf++); 1602 if (ret) 1603 goto out; 1604 1605 read++; 1606 cnt--; 1607 1608 /* 1609 * The parser is not finished with the last write, 1610 * continue reading the user input without skipping spaces. 1611 */ 1612 if (!parser->cont) { 1613 /* skip white space */ 1614 while (cnt && isspace(ch)) { 1615 ret = get_user(ch, ubuf++); 1616 if (ret) 1617 goto out; 1618 read++; 1619 cnt--; 1620 } 1621 1622 parser->idx = 0; 1623 1624 /* only spaces were written */ 1625 if (isspace(ch) || !ch) { 1626 *ppos += read; 1627 ret = read; 1628 goto out; 1629 } 1630 } 1631 1632 /* read the non-space input */ 1633 while (cnt && !isspace(ch) && ch) { 1634 if (parser->idx < parser->size - 1) 1635 parser->buffer[parser->idx++] = ch; 1636 else { 1637 ret = -EINVAL; 1638 goto out; 1639 } 1640 ret = get_user(ch, ubuf++); 1641 if (ret) 1642 goto out; 1643 read++; 1644 cnt--; 1645 } 1646 1647 /* We either got finished input or we have to wait for another call. */ 1648 if (isspace(ch) || !ch) { 1649 parser->buffer[parser->idx] = 0; 1650 parser->cont = false; 1651 } else if (parser->idx < parser->size - 1) { 1652 parser->cont = true; 1653 parser->buffer[parser->idx++] = ch; 1654 /* Make sure the parsed string always terminates with '\0'. */ 1655 parser->buffer[parser->idx] = 0; 1656 } else { 1657 ret = -EINVAL; 1658 goto out; 1659 } 1660 1661 *ppos += read; 1662 ret = read; 1663 1664 out: 1665 return ret; 1666 } 1667 1668 /* TODO add a seq_buf_to_buffer() */ 1669 static ssize_t trace_seq_to_buffer(struct trace_seq *s, void *buf, size_t cnt) 1670 { 1671 int len; 1672 1673 if (trace_seq_used(s) <= s->seq.readpos) 1674 return -EBUSY; 1675 1676 len = trace_seq_used(s) - s->seq.readpos; 1677 if (cnt > len) 1678 cnt = len; 1679 memcpy(buf, s->buffer + s->seq.readpos, cnt); 1680 1681 s->seq.readpos += cnt; 1682 return cnt; 1683 } 1684 1685 unsigned long __read_mostly tracing_thresh; 1686 static const struct file_operations tracing_max_lat_fops; 1687 1688 #if (defined(CONFIG_TRACER_MAX_TRACE) || defined(CONFIG_HWLAT_TRACER)) && \ 1689 defined(CONFIG_FSNOTIFY) 1690 1691 static struct workqueue_struct *fsnotify_wq; 1692 1693 static void latency_fsnotify_workfn(struct work_struct *work) 1694 { 1695 struct trace_array *tr = container_of(work, struct trace_array, 1696 fsnotify_work); 1697 fsnotify_inode(tr->d_max_latency->d_inode, FS_MODIFY); 1698 } 1699 1700 static void latency_fsnotify_workfn_irq(struct irq_work *iwork) 1701 { 1702 struct trace_array *tr = container_of(iwork, struct trace_array, 1703 fsnotify_irqwork); 1704 queue_work(fsnotify_wq, &tr->fsnotify_work); 1705 } 1706 1707 static void trace_create_maxlat_file(struct trace_array *tr, 1708 struct dentry *d_tracer) 1709 { 1710 INIT_WORK(&tr->fsnotify_work, latency_fsnotify_workfn); 1711 init_irq_work(&tr->fsnotify_irqwork, latency_fsnotify_workfn_irq); 1712 tr->d_max_latency = trace_create_file("tracing_max_latency", 0644, 1713 d_tracer, &tr->max_latency, 1714 &tracing_max_lat_fops); 1715 } 1716 1717 __init static int latency_fsnotify_init(void) 1718 { 1719 fsnotify_wq = alloc_workqueue("tr_max_lat_wq", 1720 WQ_UNBOUND | WQ_HIGHPRI, 0); 1721 if (!fsnotify_wq) { 1722 pr_err("Unable to allocate tr_max_lat_wq\n"); 1723 return -ENOMEM; 1724 } 1725 return 0; 1726 } 1727 1728 late_initcall_sync(latency_fsnotify_init); 1729 1730 void latency_fsnotify(struct trace_array *tr) 1731 { 1732 if (!fsnotify_wq) 1733 return; 1734 /* 1735 * We cannot call queue_work(&tr->fsnotify_work) from here because it's 1736 * possible that we are called from __schedule() or do_idle(), which 1737 * could cause a deadlock. 1738 */ 1739 irq_work_queue(&tr->fsnotify_irqwork); 1740 } 1741 1742 /* 1743 * (defined(CONFIG_TRACER_MAX_TRACE) || defined(CONFIG_HWLAT_TRACER)) && \ 1744 * defined(CONFIG_FSNOTIFY) 1745 */ 1746 #else 1747 1748 #define trace_create_maxlat_file(tr, d_tracer) \ 1749 trace_create_file("tracing_max_latency", 0644, d_tracer, \ 1750 &tr->max_latency, &tracing_max_lat_fops) 1751 1752 #endif 1753 1754 #ifdef CONFIG_TRACER_MAX_TRACE 1755 /* 1756 * Copy the new maximum trace into the separate maximum-trace 1757 * structure. (this way the maximum trace is permanently saved, 1758 * for later retrieval via /sys/kernel/tracing/tracing_max_latency) 1759 */ 1760 static void 1761 __update_max_tr(struct trace_array *tr, struct task_struct *tsk, int cpu) 1762 { 1763 struct array_buffer *trace_buf = &tr->array_buffer; 1764 struct array_buffer *max_buf = &tr->max_buffer; 1765 struct trace_array_cpu *data = per_cpu_ptr(trace_buf->data, cpu); 1766 struct trace_array_cpu *max_data = per_cpu_ptr(max_buf->data, cpu); 1767 1768 max_buf->cpu = cpu; 1769 max_buf->time_start = data->preempt_timestamp; 1770 1771 max_data->saved_latency = tr->max_latency; 1772 max_data->critical_start = data->critical_start; 1773 max_data->critical_end = data->critical_end; 1774 1775 strncpy(max_data->comm, tsk->comm, TASK_COMM_LEN); 1776 max_data->pid = tsk->pid; 1777 /* 1778 * If tsk == current, then use current_uid(), as that does not use 1779 * RCU. The irq tracer can be called out of RCU scope. 1780 */ 1781 if (tsk == current) 1782 max_data->uid = current_uid(); 1783 else 1784 max_data->uid = task_uid(tsk); 1785 1786 max_data->nice = tsk->static_prio - 20 - MAX_RT_PRIO; 1787 max_data->policy = tsk->policy; 1788 max_data->rt_priority = tsk->rt_priority; 1789 1790 /* record this tasks comm */ 1791 tracing_record_cmdline(tsk); 1792 latency_fsnotify(tr); 1793 } 1794 1795 /** 1796 * update_max_tr - snapshot all trace buffers from global_trace to max_tr 1797 * @tr: tracer 1798 * @tsk: the task with the latency 1799 * @cpu: The cpu that initiated the trace. 1800 * @cond_data: User data associated with a conditional snapshot 1801 * 1802 * Flip the buffers between the @tr and the max_tr and record information 1803 * about which task was the cause of this latency. 1804 */ 1805 void 1806 update_max_tr(struct trace_array *tr, struct task_struct *tsk, int cpu, 1807 void *cond_data) 1808 { 1809 if (tr->stop_count) 1810 return; 1811 1812 WARN_ON_ONCE(!irqs_disabled()); 1813 1814 if (!tr->allocated_snapshot) { 1815 /* Only the nop tracer should hit this when disabling */ 1816 WARN_ON_ONCE(tr->current_trace != &nop_trace); 1817 return; 1818 } 1819 1820 arch_spin_lock(&tr->max_lock); 1821 1822 /* Inherit the recordable setting from array_buffer */ 1823 if (ring_buffer_record_is_set_on(tr->array_buffer.buffer)) 1824 ring_buffer_record_on(tr->max_buffer.buffer); 1825 else 1826 ring_buffer_record_off(tr->max_buffer.buffer); 1827 1828 #ifdef CONFIG_TRACER_SNAPSHOT 1829 if (tr->cond_snapshot && !tr->cond_snapshot->update(tr, cond_data)) 1830 goto out_unlock; 1831 #endif 1832 swap(tr->array_buffer.buffer, tr->max_buffer.buffer); 1833 1834 __update_max_tr(tr, tsk, cpu); 1835 1836 out_unlock: 1837 arch_spin_unlock(&tr->max_lock); 1838 } 1839 1840 /** 1841 * update_max_tr_single - only copy one trace over, and reset the rest 1842 * @tr: tracer 1843 * @tsk: task with the latency 1844 * @cpu: the cpu of the buffer to copy. 1845 * 1846 * Flip the trace of a single CPU buffer between the @tr and the max_tr. 1847 */ 1848 void 1849 update_max_tr_single(struct trace_array *tr, struct task_struct *tsk, int cpu) 1850 { 1851 int ret; 1852 1853 if (tr->stop_count) 1854 return; 1855 1856 WARN_ON_ONCE(!irqs_disabled()); 1857 if (!tr->allocated_snapshot) { 1858 /* Only the nop tracer should hit this when disabling */ 1859 WARN_ON_ONCE(tr->current_trace != &nop_trace); 1860 return; 1861 } 1862 1863 arch_spin_lock(&tr->max_lock); 1864 1865 ret = ring_buffer_swap_cpu(tr->max_buffer.buffer, tr->array_buffer.buffer, cpu); 1866 1867 if (ret == -EBUSY) { 1868 /* 1869 * We failed to swap the buffer due to a commit taking 1870 * place on this CPU. We fail to record, but we reset 1871 * the max trace buffer (no one writes directly to it) 1872 * and flag that it failed. 1873 */ 1874 trace_array_printk_buf(tr->max_buffer.buffer, _THIS_IP_, 1875 "Failed to swap buffers due to commit in progress\n"); 1876 } 1877 1878 WARN_ON_ONCE(ret && ret != -EAGAIN && ret != -EBUSY); 1879 1880 __update_max_tr(tr, tsk, cpu); 1881 arch_spin_unlock(&tr->max_lock); 1882 } 1883 #endif /* CONFIG_TRACER_MAX_TRACE */ 1884 1885 static int wait_on_pipe(struct trace_iterator *iter, int full) 1886 { 1887 /* Iterators are static, they should be filled or empty */ 1888 if (trace_buffer_iter(iter, iter->cpu_file)) 1889 return 0; 1890 1891 return ring_buffer_wait(iter->array_buffer->buffer, iter->cpu_file, 1892 full); 1893 } 1894 1895 #ifdef CONFIG_FTRACE_STARTUP_TEST 1896 static bool selftests_can_run; 1897 1898 struct trace_selftests { 1899 struct list_head list; 1900 struct tracer *type; 1901 }; 1902 1903 static LIST_HEAD(postponed_selftests); 1904 1905 static int save_selftest(struct tracer *type) 1906 { 1907 struct trace_selftests *selftest; 1908 1909 selftest = kmalloc(sizeof(*selftest), GFP_KERNEL); 1910 if (!selftest) 1911 return -ENOMEM; 1912 1913 selftest->type = type; 1914 list_add(&selftest->list, &postponed_selftests); 1915 return 0; 1916 } 1917 1918 static int run_tracer_selftest(struct tracer *type) 1919 { 1920 struct trace_array *tr = &global_trace; 1921 struct tracer *saved_tracer = tr->current_trace; 1922 int ret; 1923 1924 if (!type->selftest || tracing_selftest_disabled) 1925 return 0; 1926 1927 /* 1928 * If a tracer registers early in boot up (before scheduling is 1929 * initialized and such), then do not run its selftests yet. 1930 * Instead, run it a little later in the boot process. 1931 */ 1932 if (!selftests_can_run) 1933 return save_selftest(type); 1934 1935 /* 1936 * Run a selftest on this tracer. 1937 * Here we reset the trace buffer, and set the current 1938 * tracer to be this tracer. The tracer can then run some 1939 * internal tracing to verify that everything is in order. 1940 * If we fail, we do not register this tracer. 1941 */ 1942 tracing_reset_online_cpus(&tr->array_buffer); 1943 1944 tr->current_trace = type; 1945 1946 #ifdef CONFIG_TRACER_MAX_TRACE 1947 if (type->use_max_tr) { 1948 /* If we expanded the buffers, make sure the max is expanded too */ 1949 if (ring_buffer_expanded) 1950 ring_buffer_resize(tr->max_buffer.buffer, trace_buf_size, 1951 RING_BUFFER_ALL_CPUS); 1952 tr->allocated_snapshot = true; 1953 } 1954 #endif 1955 1956 /* the test is responsible for initializing and enabling */ 1957 pr_info("Testing tracer %s: ", type->name); 1958 ret = type->selftest(type, tr); 1959 /* the test is responsible for resetting too */ 1960 tr->current_trace = saved_tracer; 1961 if (ret) { 1962 printk(KERN_CONT "FAILED!\n"); 1963 /* Add the warning after printing 'FAILED' */ 1964 WARN_ON(1); 1965 return -1; 1966 } 1967 /* Only reset on passing, to avoid touching corrupted buffers */ 1968 tracing_reset_online_cpus(&tr->array_buffer); 1969 1970 #ifdef CONFIG_TRACER_MAX_TRACE 1971 if (type->use_max_tr) { 1972 tr->allocated_snapshot = false; 1973 1974 /* Shrink the max buffer again */ 1975 if (ring_buffer_expanded) 1976 ring_buffer_resize(tr->max_buffer.buffer, 1, 1977 RING_BUFFER_ALL_CPUS); 1978 } 1979 #endif 1980 1981 printk(KERN_CONT "PASSED\n"); 1982 return 0; 1983 } 1984 1985 static __init int init_trace_selftests(void) 1986 { 1987 struct trace_selftests *p, *n; 1988 struct tracer *t, **last; 1989 int ret; 1990 1991 selftests_can_run = true; 1992 1993 mutex_lock(&trace_types_lock); 1994 1995 if (list_empty(&postponed_selftests)) 1996 goto out; 1997 1998 pr_info("Running postponed tracer tests:\n"); 1999 2000 tracing_selftest_running = true; 2001 list_for_each_entry_safe(p, n, &postponed_selftests, list) { 2002 /* This loop can take minutes when sanitizers are enabled, so 2003 * lets make sure we allow RCU processing. 2004 */ 2005 cond_resched(); 2006 ret = run_tracer_selftest(p->type); 2007 /* If the test fails, then warn and remove from available_tracers */ 2008 if (ret < 0) { 2009 WARN(1, "tracer: %s failed selftest, disabling\n", 2010 p->type->name); 2011 last = &trace_types; 2012 for (t = trace_types; t; t = t->next) { 2013 if (t == p->type) { 2014 *last = t->next; 2015 break; 2016 } 2017 last = &t->next; 2018 } 2019 } 2020 list_del(&p->list); 2021 kfree(p); 2022 } 2023 tracing_selftest_running = false; 2024 2025 out: 2026 mutex_unlock(&trace_types_lock); 2027 2028 return 0; 2029 } 2030 core_initcall(init_trace_selftests); 2031 #else 2032 static inline int run_tracer_selftest(struct tracer *type) 2033 { 2034 return 0; 2035 } 2036 #endif /* CONFIG_FTRACE_STARTUP_TEST */ 2037 2038 static void add_tracer_options(struct trace_array *tr, struct tracer *t); 2039 2040 static void __init apply_trace_boot_options(void); 2041 2042 /** 2043 * register_tracer - register a tracer with the ftrace system. 2044 * @type: the plugin for the tracer 2045 * 2046 * Register a new plugin tracer. 2047 */ 2048 int __init register_tracer(struct tracer *type) 2049 { 2050 struct tracer *t; 2051 int ret = 0; 2052 2053 if (!type->name) { 2054 pr_info("Tracer must have a name\n"); 2055 return -1; 2056 } 2057 2058 if (strlen(type->name) >= MAX_TRACER_SIZE) { 2059 pr_info("Tracer has a name longer than %d\n", MAX_TRACER_SIZE); 2060 return -1; 2061 } 2062 2063 if (security_locked_down(LOCKDOWN_TRACEFS)) { 2064 pr_warn("Can not register tracer %s due to lockdown\n", 2065 type->name); 2066 return -EPERM; 2067 } 2068 2069 mutex_lock(&trace_types_lock); 2070 2071 tracing_selftest_running = true; 2072 2073 for (t = trace_types; t; t = t->next) { 2074 if (strcmp(type->name, t->name) == 0) { 2075 /* already found */ 2076 pr_info("Tracer %s already registered\n", 2077 type->name); 2078 ret = -1; 2079 goto out; 2080 } 2081 } 2082 2083 if (!type->set_flag) 2084 type->set_flag = &dummy_set_flag; 2085 if (!type->flags) { 2086 /*allocate a dummy tracer_flags*/ 2087 type->flags = kmalloc(sizeof(*type->flags), GFP_KERNEL); 2088 if (!type->flags) { 2089 ret = -ENOMEM; 2090 goto out; 2091 } 2092 type->flags->val = 0; 2093 type->flags->opts = dummy_tracer_opt; 2094 } else 2095 if (!type->flags->opts) 2096 type->flags->opts = dummy_tracer_opt; 2097 2098 /* store the tracer for __set_tracer_option */ 2099 type->flags->trace = type; 2100 2101 ret = run_tracer_selftest(type); 2102 if (ret < 0) 2103 goto out; 2104 2105 type->next = trace_types; 2106 trace_types = type; 2107 add_tracer_options(&global_trace, type); 2108 2109 out: 2110 tracing_selftest_running = false; 2111 mutex_unlock(&trace_types_lock); 2112 2113 if (ret || !default_bootup_tracer) 2114 goto out_unlock; 2115 2116 if (strncmp(default_bootup_tracer, type->name, MAX_TRACER_SIZE)) 2117 goto out_unlock; 2118 2119 printk(KERN_INFO "Starting tracer '%s'\n", type->name); 2120 /* Do we want this tracer to start on bootup? */ 2121 tracing_set_tracer(&global_trace, type->name); 2122 default_bootup_tracer = NULL; 2123 2124 apply_trace_boot_options(); 2125 2126 /* disable other selftests, since this will break it. */ 2127 disable_tracing_selftest("running a tracer"); 2128 2129 out_unlock: 2130 return ret; 2131 } 2132 2133 static void tracing_reset_cpu(struct array_buffer *buf, int cpu) 2134 { 2135 struct trace_buffer *buffer = buf->buffer; 2136 2137 if (!buffer) 2138 return; 2139 2140 ring_buffer_record_disable(buffer); 2141 2142 /* Make sure all commits have finished */ 2143 synchronize_rcu(); 2144 ring_buffer_reset_cpu(buffer, cpu); 2145 2146 ring_buffer_record_enable(buffer); 2147 } 2148 2149 void tracing_reset_online_cpus(struct array_buffer *buf) 2150 { 2151 struct trace_buffer *buffer = buf->buffer; 2152 2153 if (!buffer) 2154 return; 2155 2156 ring_buffer_record_disable(buffer); 2157 2158 /* Make sure all commits have finished */ 2159 synchronize_rcu(); 2160 2161 buf->time_start = buffer_ftrace_now(buf, buf->cpu); 2162 2163 ring_buffer_reset_online_cpus(buffer); 2164 2165 ring_buffer_record_enable(buffer); 2166 } 2167 2168 /* Must have trace_types_lock held */ 2169 void tracing_reset_all_online_cpus(void) 2170 { 2171 struct trace_array *tr; 2172 2173 list_for_each_entry(tr, &ftrace_trace_arrays, list) { 2174 if (!tr->clear_trace) 2175 continue; 2176 tr->clear_trace = false; 2177 tracing_reset_online_cpus(&tr->array_buffer); 2178 #ifdef CONFIG_TRACER_MAX_TRACE 2179 tracing_reset_online_cpus(&tr->max_buffer); 2180 #endif 2181 } 2182 } 2183 2184 static int *tgid_map; 2185 2186 #define SAVED_CMDLINES_DEFAULT 128 2187 #define NO_CMDLINE_MAP UINT_MAX 2188 static arch_spinlock_t trace_cmdline_lock = __ARCH_SPIN_LOCK_UNLOCKED; 2189 struct saved_cmdlines_buffer { 2190 unsigned map_pid_to_cmdline[PID_MAX_DEFAULT+1]; 2191 unsigned *map_cmdline_to_pid; 2192 unsigned cmdline_num; 2193 int cmdline_idx; 2194 char *saved_cmdlines; 2195 }; 2196 static struct saved_cmdlines_buffer *savedcmd; 2197 2198 /* temporary disable recording */ 2199 static atomic_t trace_record_taskinfo_disabled __read_mostly; 2200 2201 static inline char *get_saved_cmdlines(int idx) 2202 { 2203 return &savedcmd->saved_cmdlines[idx * TASK_COMM_LEN]; 2204 } 2205 2206 static inline void set_cmdline(int idx, const char *cmdline) 2207 { 2208 strncpy(get_saved_cmdlines(idx), cmdline, TASK_COMM_LEN); 2209 } 2210 2211 static int allocate_cmdlines_buffer(unsigned int val, 2212 struct saved_cmdlines_buffer *s) 2213 { 2214 s->map_cmdline_to_pid = kmalloc_array(val, 2215 sizeof(*s->map_cmdline_to_pid), 2216 GFP_KERNEL); 2217 if (!s->map_cmdline_to_pid) 2218 return -ENOMEM; 2219 2220 s->saved_cmdlines = kmalloc_array(TASK_COMM_LEN, val, GFP_KERNEL); 2221 if (!s->saved_cmdlines) { 2222 kfree(s->map_cmdline_to_pid); 2223 return -ENOMEM; 2224 } 2225 2226 s->cmdline_idx = 0; 2227 s->cmdline_num = val; 2228 memset(&s->map_pid_to_cmdline, NO_CMDLINE_MAP, 2229 sizeof(s->map_pid_to_cmdline)); 2230 memset(s->map_cmdline_to_pid, NO_CMDLINE_MAP, 2231 val * sizeof(*s->map_cmdline_to_pid)); 2232 2233 return 0; 2234 } 2235 2236 static int trace_create_savedcmd(void) 2237 { 2238 int ret; 2239 2240 savedcmd = kmalloc(sizeof(*savedcmd), GFP_KERNEL); 2241 if (!savedcmd) 2242 return -ENOMEM; 2243 2244 ret = allocate_cmdlines_buffer(SAVED_CMDLINES_DEFAULT, savedcmd); 2245 if (ret < 0) { 2246 kfree(savedcmd); 2247 savedcmd = NULL; 2248 return -ENOMEM; 2249 } 2250 2251 return 0; 2252 } 2253 2254 int is_tracing_stopped(void) 2255 { 2256 return global_trace.stop_count; 2257 } 2258 2259 /** 2260 * tracing_start - quick start of the tracer 2261 * 2262 * If tracing is enabled but was stopped by tracing_stop, 2263 * this will start the tracer back up. 2264 */ 2265 void tracing_start(void) 2266 { 2267 struct trace_buffer *buffer; 2268 unsigned long flags; 2269 2270 if (tracing_disabled) 2271 return; 2272 2273 raw_spin_lock_irqsave(&global_trace.start_lock, flags); 2274 if (--global_trace.stop_count) { 2275 if (global_trace.stop_count < 0) { 2276 /* Someone screwed up their debugging */ 2277 WARN_ON_ONCE(1); 2278 global_trace.stop_count = 0; 2279 } 2280 goto out; 2281 } 2282 2283 /* Prevent the buffers from switching */ 2284 arch_spin_lock(&global_trace.max_lock); 2285 2286 buffer = global_trace.array_buffer.buffer; 2287 if (buffer) 2288 ring_buffer_record_enable(buffer); 2289 2290 #ifdef CONFIG_TRACER_MAX_TRACE 2291 buffer = global_trace.max_buffer.buffer; 2292 if (buffer) 2293 ring_buffer_record_enable(buffer); 2294 #endif 2295 2296 arch_spin_unlock(&global_trace.max_lock); 2297 2298 out: 2299 raw_spin_unlock_irqrestore(&global_trace.start_lock, flags); 2300 } 2301 2302 static void tracing_start_tr(struct trace_array *tr) 2303 { 2304 struct trace_buffer *buffer; 2305 unsigned long flags; 2306 2307 if (tracing_disabled) 2308 return; 2309 2310 /* If global, we need to also start the max tracer */ 2311 if (tr->flags & TRACE_ARRAY_FL_GLOBAL) 2312 return tracing_start(); 2313 2314 raw_spin_lock_irqsave(&tr->start_lock, flags); 2315 2316 if (--tr->stop_count) { 2317 if (tr->stop_count < 0) { 2318 /* Someone screwed up their debugging */ 2319 WARN_ON_ONCE(1); 2320 tr->stop_count = 0; 2321 } 2322 goto out; 2323 } 2324 2325 buffer = tr->array_buffer.buffer; 2326 if (buffer) 2327 ring_buffer_record_enable(buffer); 2328 2329 out: 2330 raw_spin_unlock_irqrestore(&tr->start_lock, flags); 2331 } 2332 2333 /** 2334 * tracing_stop - quick stop of the tracer 2335 * 2336 * Light weight way to stop tracing. Use in conjunction with 2337 * tracing_start. 2338 */ 2339 void tracing_stop(void) 2340 { 2341 struct trace_buffer *buffer; 2342 unsigned long flags; 2343 2344 raw_spin_lock_irqsave(&global_trace.start_lock, flags); 2345 if (global_trace.stop_count++) 2346 goto out; 2347 2348 /* Prevent the buffers from switching */ 2349 arch_spin_lock(&global_trace.max_lock); 2350 2351 buffer = global_trace.array_buffer.buffer; 2352 if (buffer) 2353 ring_buffer_record_disable(buffer); 2354 2355 #ifdef CONFIG_TRACER_MAX_TRACE 2356 buffer = global_trace.max_buffer.buffer; 2357 if (buffer) 2358 ring_buffer_record_disable(buffer); 2359 #endif 2360 2361 arch_spin_unlock(&global_trace.max_lock); 2362 2363 out: 2364 raw_spin_unlock_irqrestore(&global_trace.start_lock, flags); 2365 } 2366 2367 static void tracing_stop_tr(struct trace_array *tr) 2368 { 2369 struct trace_buffer *buffer; 2370 unsigned long flags; 2371 2372 /* If global, we need to also stop the max tracer */ 2373 if (tr->flags & TRACE_ARRAY_FL_GLOBAL) 2374 return tracing_stop(); 2375 2376 raw_spin_lock_irqsave(&tr->start_lock, flags); 2377 if (tr->stop_count++) 2378 goto out; 2379 2380 buffer = tr->array_buffer.buffer; 2381 if (buffer) 2382 ring_buffer_record_disable(buffer); 2383 2384 out: 2385 raw_spin_unlock_irqrestore(&tr->start_lock, flags); 2386 } 2387 2388 static int trace_save_cmdline(struct task_struct *tsk) 2389 { 2390 unsigned pid, idx; 2391 2392 /* treat recording of idle task as a success */ 2393 if (!tsk->pid) 2394 return 1; 2395 2396 if (unlikely(tsk->pid > PID_MAX_DEFAULT)) 2397 return 0; 2398 2399 /* 2400 * It's not the end of the world if we don't get 2401 * the lock, but we also don't want to spin 2402 * nor do we want to disable interrupts, 2403 * so if we miss here, then better luck next time. 2404 */ 2405 if (!arch_spin_trylock(&trace_cmdline_lock)) 2406 return 0; 2407 2408 idx = savedcmd->map_pid_to_cmdline[tsk->pid]; 2409 if (idx == NO_CMDLINE_MAP) { 2410 idx = (savedcmd->cmdline_idx + 1) % savedcmd->cmdline_num; 2411 2412 /* 2413 * Check whether the cmdline buffer at idx has a pid 2414 * mapped. We are going to overwrite that entry so we 2415 * need to clear the map_pid_to_cmdline. Otherwise we 2416 * would read the new comm for the old pid. 2417 */ 2418 pid = savedcmd->map_cmdline_to_pid[idx]; 2419 if (pid != NO_CMDLINE_MAP) 2420 savedcmd->map_pid_to_cmdline[pid] = NO_CMDLINE_MAP; 2421 2422 savedcmd->map_cmdline_to_pid[idx] = tsk->pid; 2423 savedcmd->map_pid_to_cmdline[tsk->pid] = idx; 2424 2425 savedcmd->cmdline_idx = idx; 2426 } 2427 2428 set_cmdline(idx, tsk->comm); 2429 2430 arch_spin_unlock(&trace_cmdline_lock); 2431 2432 return 1; 2433 } 2434 2435 static void __trace_find_cmdline(int pid, char comm[]) 2436 { 2437 unsigned map; 2438 2439 if (!pid) { 2440 strcpy(comm, "<idle>"); 2441 return; 2442 } 2443 2444 if (WARN_ON_ONCE(pid < 0)) { 2445 strcpy(comm, "<XXX>"); 2446 return; 2447 } 2448 2449 if (pid > PID_MAX_DEFAULT) { 2450 strcpy(comm, "<...>"); 2451 return; 2452 } 2453 2454 map = savedcmd->map_pid_to_cmdline[pid]; 2455 if (map != NO_CMDLINE_MAP) 2456 strlcpy(comm, get_saved_cmdlines(map), TASK_COMM_LEN); 2457 else 2458 strcpy(comm, "<...>"); 2459 } 2460 2461 void trace_find_cmdline(int pid, char comm[]) 2462 { 2463 preempt_disable(); 2464 arch_spin_lock(&trace_cmdline_lock); 2465 2466 __trace_find_cmdline(pid, comm); 2467 2468 arch_spin_unlock(&trace_cmdline_lock); 2469 preempt_enable(); 2470 } 2471 2472 int trace_find_tgid(int pid) 2473 { 2474 if (unlikely(!tgid_map || !pid || pid > PID_MAX_DEFAULT)) 2475 return 0; 2476 2477 return tgid_map[pid]; 2478 } 2479 2480 static int trace_save_tgid(struct task_struct *tsk) 2481 { 2482 /* treat recording of idle task as a success */ 2483 if (!tsk->pid) 2484 return 1; 2485 2486 if (unlikely(!tgid_map || tsk->pid > PID_MAX_DEFAULT)) 2487 return 0; 2488 2489 tgid_map[tsk->pid] = tsk->tgid; 2490 return 1; 2491 } 2492 2493 static bool tracing_record_taskinfo_skip(int flags) 2494 { 2495 if (unlikely(!(flags & (TRACE_RECORD_CMDLINE | TRACE_RECORD_TGID)))) 2496 return true; 2497 if (atomic_read(&trace_record_taskinfo_disabled) || !tracing_is_on()) 2498 return true; 2499 if (!__this_cpu_read(trace_taskinfo_save)) 2500 return true; 2501 return false; 2502 } 2503 2504 /** 2505 * tracing_record_taskinfo - record the task info of a task 2506 * 2507 * @task: task to record 2508 * @flags: TRACE_RECORD_CMDLINE for recording comm 2509 * TRACE_RECORD_TGID for recording tgid 2510 */ 2511 void tracing_record_taskinfo(struct task_struct *task, int flags) 2512 { 2513 bool done; 2514 2515 if (tracing_record_taskinfo_skip(flags)) 2516 return; 2517 2518 /* 2519 * Record as much task information as possible. If some fail, continue 2520 * to try to record the others. 2521 */ 2522 done = !(flags & TRACE_RECORD_CMDLINE) || trace_save_cmdline(task); 2523 done &= !(flags & TRACE_RECORD_TGID) || trace_save_tgid(task); 2524 2525 /* If recording any information failed, retry again soon. */ 2526 if (!done) 2527 return; 2528 2529 __this_cpu_write(trace_taskinfo_save, false); 2530 } 2531 2532 /** 2533 * tracing_record_taskinfo_sched_switch - record task info for sched_switch 2534 * 2535 * @prev: previous task during sched_switch 2536 * @next: next task during sched_switch 2537 * @flags: TRACE_RECORD_CMDLINE for recording comm 2538 * TRACE_RECORD_TGID for recording tgid 2539 */ 2540 void tracing_record_taskinfo_sched_switch(struct task_struct *prev, 2541 struct task_struct *next, int flags) 2542 { 2543 bool done; 2544 2545 if (tracing_record_taskinfo_skip(flags)) 2546 return; 2547 2548 /* 2549 * Record as much task information as possible. If some fail, continue 2550 * to try to record the others. 2551 */ 2552 done = !(flags & TRACE_RECORD_CMDLINE) || trace_save_cmdline(prev); 2553 done &= !(flags & TRACE_RECORD_CMDLINE) || trace_save_cmdline(next); 2554 done &= !(flags & TRACE_RECORD_TGID) || trace_save_tgid(prev); 2555 done &= !(flags & TRACE_RECORD_TGID) || trace_save_tgid(next); 2556 2557 /* If recording any information failed, retry again soon. */ 2558 if (!done) 2559 return; 2560 2561 __this_cpu_write(trace_taskinfo_save, false); 2562 } 2563 2564 /* Helpers to record a specific task information */ 2565 void tracing_record_cmdline(struct task_struct *task) 2566 { 2567 tracing_record_taskinfo(task, TRACE_RECORD_CMDLINE); 2568 } 2569 2570 void tracing_record_tgid(struct task_struct *task) 2571 { 2572 tracing_record_taskinfo(task, TRACE_RECORD_TGID); 2573 } 2574 2575 /* 2576 * Several functions return TRACE_TYPE_PARTIAL_LINE if the trace_seq 2577 * overflowed, and TRACE_TYPE_HANDLED otherwise. This helper function 2578 * simplifies those functions and keeps them in sync. 2579 */ 2580 enum print_line_t trace_handle_return(struct trace_seq *s) 2581 { 2582 return trace_seq_has_overflowed(s) ? 2583 TRACE_TYPE_PARTIAL_LINE : TRACE_TYPE_HANDLED; 2584 } 2585 EXPORT_SYMBOL_GPL(trace_handle_return); 2586 2587 void 2588 tracing_generic_entry_update(struct trace_entry *entry, unsigned short type, 2589 unsigned long flags, int pc) 2590 { 2591 struct task_struct *tsk = current; 2592 2593 entry->preempt_count = pc & 0xff; 2594 entry->pid = (tsk) ? tsk->pid : 0; 2595 entry->type = type; 2596 entry->flags = 2597 #ifdef CONFIG_TRACE_IRQFLAGS_SUPPORT 2598 (irqs_disabled_flags(flags) ? TRACE_FLAG_IRQS_OFF : 0) | 2599 #else 2600 TRACE_FLAG_IRQS_NOSUPPORT | 2601 #endif 2602 ((pc & NMI_MASK ) ? TRACE_FLAG_NMI : 0) | 2603 ((pc & HARDIRQ_MASK) ? TRACE_FLAG_HARDIRQ : 0) | 2604 ((pc & SOFTIRQ_OFFSET) ? TRACE_FLAG_SOFTIRQ : 0) | 2605 (tif_need_resched() ? TRACE_FLAG_NEED_RESCHED : 0) | 2606 (test_preempt_need_resched() ? TRACE_FLAG_PREEMPT_RESCHED : 0); 2607 } 2608 EXPORT_SYMBOL_GPL(tracing_generic_entry_update); 2609 2610 struct ring_buffer_event * 2611 trace_buffer_lock_reserve(struct trace_buffer *buffer, 2612 int type, 2613 unsigned long len, 2614 unsigned long flags, int pc) 2615 { 2616 return __trace_buffer_lock_reserve(buffer, type, len, flags, pc); 2617 } 2618 2619 DEFINE_PER_CPU(struct ring_buffer_event *, trace_buffered_event); 2620 DEFINE_PER_CPU(int, trace_buffered_event_cnt); 2621 static int trace_buffered_event_ref; 2622 2623 /** 2624 * trace_buffered_event_enable - enable buffering events 2625 * 2626 * When events are being filtered, it is quicker to use a temporary 2627 * buffer to write the event data into if there's a likely chance 2628 * that it will not be committed. The discard of the ring buffer 2629 * is not as fast as committing, and is much slower than copying 2630 * a commit. 2631 * 2632 * When an event is to be filtered, allocate per cpu buffers to 2633 * write the event data into, and if the event is filtered and discarded 2634 * it is simply dropped, otherwise, the entire data is to be committed 2635 * in one shot. 2636 */ 2637 void trace_buffered_event_enable(void) 2638 { 2639 struct ring_buffer_event *event; 2640 struct page *page; 2641 int cpu; 2642 2643 WARN_ON_ONCE(!mutex_is_locked(&event_mutex)); 2644 2645 if (trace_buffered_event_ref++) 2646 return; 2647 2648 for_each_tracing_cpu(cpu) { 2649 page = alloc_pages_node(cpu_to_node(cpu), 2650 GFP_KERNEL | __GFP_NORETRY, 0); 2651 if (!page) 2652 goto failed; 2653 2654 event = page_address(page); 2655 memset(event, 0, sizeof(*event)); 2656 2657 per_cpu(trace_buffered_event, cpu) = event; 2658 2659 preempt_disable(); 2660 if (cpu == smp_processor_id() && 2661 __this_cpu_read(trace_buffered_event) != 2662 per_cpu(trace_buffered_event, cpu)) 2663 WARN_ON_ONCE(1); 2664 preempt_enable(); 2665 } 2666 2667 return; 2668 failed: 2669 trace_buffered_event_disable(); 2670 } 2671 2672 static void enable_trace_buffered_event(void *data) 2673 { 2674 /* Probably not needed, but do it anyway */ 2675 smp_rmb(); 2676 this_cpu_dec(trace_buffered_event_cnt); 2677 } 2678 2679 static void disable_trace_buffered_event(void *data) 2680 { 2681 this_cpu_inc(trace_buffered_event_cnt); 2682 } 2683 2684 /** 2685 * trace_buffered_event_disable - disable buffering events 2686 * 2687 * When a filter is removed, it is faster to not use the buffered 2688 * events, and to commit directly into the ring buffer. Free up 2689 * the temp buffers when there are no more users. This requires 2690 * special synchronization with current events. 2691 */ 2692 void trace_buffered_event_disable(void) 2693 { 2694 int cpu; 2695 2696 WARN_ON_ONCE(!mutex_is_locked(&event_mutex)); 2697 2698 if (WARN_ON_ONCE(!trace_buffered_event_ref)) 2699 return; 2700 2701 if (--trace_buffered_event_ref) 2702 return; 2703 2704 preempt_disable(); 2705 /* For each CPU, set the buffer as used. */ 2706 smp_call_function_many(tracing_buffer_mask, 2707 disable_trace_buffered_event, NULL, 1); 2708 preempt_enable(); 2709 2710 /* Wait for all current users to finish */ 2711 synchronize_rcu(); 2712 2713 for_each_tracing_cpu(cpu) { 2714 free_page((unsigned long)per_cpu(trace_buffered_event, cpu)); 2715 per_cpu(trace_buffered_event, cpu) = NULL; 2716 } 2717 /* 2718 * Make sure trace_buffered_event is NULL before clearing 2719 * trace_buffered_event_cnt. 2720 */ 2721 smp_wmb(); 2722 2723 preempt_disable(); 2724 /* Do the work on each cpu */ 2725 smp_call_function_many(tracing_buffer_mask, 2726 enable_trace_buffered_event, NULL, 1); 2727 preempt_enable(); 2728 } 2729 2730 static struct trace_buffer *temp_buffer; 2731 2732 struct ring_buffer_event * 2733 trace_event_buffer_lock_reserve(struct trace_buffer **current_rb, 2734 struct trace_event_file *trace_file, 2735 int type, unsigned long len, 2736 unsigned long flags, int pc) 2737 { 2738 struct ring_buffer_event *entry; 2739 int val; 2740 2741 *current_rb = trace_file->tr->array_buffer.buffer; 2742 2743 if (!ring_buffer_time_stamp_abs(*current_rb) && (trace_file->flags & 2744 (EVENT_FILE_FL_SOFT_DISABLED | EVENT_FILE_FL_FILTERED)) && 2745 (entry = this_cpu_read(trace_buffered_event))) { 2746 /* Try to use the per cpu buffer first */ 2747 val = this_cpu_inc_return(trace_buffered_event_cnt); 2748 if (val == 1) { 2749 trace_event_setup(entry, type, flags, pc); 2750 entry->array[0] = len; 2751 return entry; 2752 } 2753 this_cpu_dec(trace_buffered_event_cnt); 2754 } 2755 2756 entry = __trace_buffer_lock_reserve(*current_rb, 2757 type, len, flags, pc); 2758 /* 2759 * If tracing is off, but we have triggers enabled 2760 * we still need to look at the event data. Use the temp_buffer 2761 * to store the trace event for the trigger to use. It's recursive 2762 * safe and will not be recorded anywhere. 2763 */ 2764 if (!entry && trace_file->flags & EVENT_FILE_FL_TRIGGER_COND) { 2765 *current_rb = temp_buffer; 2766 entry = __trace_buffer_lock_reserve(*current_rb, 2767 type, len, flags, pc); 2768 } 2769 return entry; 2770 } 2771 EXPORT_SYMBOL_GPL(trace_event_buffer_lock_reserve); 2772 2773 static DEFINE_SPINLOCK(tracepoint_iter_lock); 2774 static DEFINE_MUTEX(tracepoint_printk_mutex); 2775 2776 static void output_printk(struct trace_event_buffer *fbuffer) 2777 { 2778 struct trace_event_call *event_call; 2779 struct trace_event_file *file; 2780 struct trace_event *event; 2781 unsigned long flags; 2782 struct trace_iterator *iter = tracepoint_print_iter; 2783 2784 /* We should never get here if iter is NULL */ 2785 if (WARN_ON_ONCE(!iter)) 2786 return; 2787 2788 event_call = fbuffer->trace_file->event_call; 2789 if (!event_call || !event_call->event.funcs || 2790 !event_call->event.funcs->trace) 2791 return; 2792 2793 file = fbuffer->trace_file; 2794 if (test_bit(EVENT_FILE_FL_SOFT_DISABLED_BIT, &file->flags) || 2795 (unlikely(file->flags & EVENT_FILE_FL_FILTERED) && 2796 !filter_match_preds(file->filter, fbuffer->entry))) 2797 return; 2798 2799 event = &fbuffer->trace_file->event_call->event; 2800 2801 spin_lock_irqsave(&tracepoint_iter_lock, flags); 2802 trace_seq_init(&iter->seq); 2803 iter->ent = fbuffer->entry; 2804 event_call->event.funcs->trace(iter, 0, event); 2805 trace_seq_putc(&iter->seq, 0); 2806 printk("%s", iter->seq.buffer); 2807 2808 spin_unlock_irqrestore(&tracepoint_iter_lock, flags); 2809 } 2810 2811 int tracepoint_printk_sysctl(struct ctl_table *table, int write, 2812 void *buffer, size_t *lenp, 2813 loff_t *ppos) 2814 { 2815 int save_tracepoint_printk; 2816 int ret; 2817 2818 mutex_lock(&tracepoint_printk_mutex); 2819 save_tracepoint_printk = tracepoint_printk; 2820 2821 ret = proc_dointvec(table, write, buffer, lenp, ppos); 2822 2823 /* 2824 * This will force exiting early, as tracepoint_printk 2825 * is always zero when tracepoint_printk_iter is not allocated 2826 */ 2827 if (!tracepoint_print_iter) 2828 tracepoint_printk = 0; 2829 2830 if (save_tracepoint_printk == tracepoint_printk) 2831 goto out; 2832 2833 if (tracepoint_printk) 2834 static_key_enable(&tracepoint_printk_key.key); 2835 else 2836 static_key_disable(&tracepoint_printk_key.key); 2837 2838 out: 2839 mutex_unlock(&tracepoint_printk_mutex); 2840 2841 return ret; 2842 } 2843 2844 void trace_event_buffer_commit(struct trace_event_buffer *fbuffer) 2845 { 2846 if (static_key_false(&tracepoint_printk_key.key)) 2847 output_printk(fbuffer); 2848 2849 if (static_branch_unlikely(&trace_event_exports_enabled)) 2850 ftrace_exports(fbuffer->event, TRACE_EXPORT_EVENT); 2851 event_trigger_unlock_commit_regs(fbuffer->trace_file, fbuffer->buffer, 2852 fbuffer->event, fbuffer->entry, 2853 fbuffer->flags, fbuffer->pc, fbuffer->regs); 2854 } 2855 EXPORT_SYMBOL_GPL(trace_event_buffer_commit); 2856 2857 /* 2858 * Skip 3: 2859 * 2860 * trace_buffer_unlock_commit_regs() 2861 * trace_event_buffer_commit() 2862 * trace_event_raw_event_xxx() 2863 */ 2864 # define STACK_SKIP 3 2865 2866 void trace_buffer_unlock_commit_regs(struct trace_array *tr, 2867 struct trace_buffer *buffer, 2868 struct ring_buffer_event *event, 2869 unsigned long flags, int pc, 2870 struct pt_regs *regs) 2871 { 2872 __buffer_unlock_commit(buffer, event); 2873 2874 /* 2875 * If regs is not set, then skip the necessary functions. 2876 * Note, we can still get here via blktrace, wakeup tracer 2877 * and mmiotrace, but that's ok if they lose a function or 2878 * two. They are not that meaningful. 2879 */ 2880 ftrace_trace_stack(tr, buffer, flags, regs ? 0 : STACK_SKIP, pc, regs); 2881 ftrace_trace_userstack(tr, buffer, flags, pc); 2882 } 2883 2884 /* 2885 * Similar to trace_buffer_unlock_commit_regs() but do not dump stack. 2886 */ 2887 void 2888 trace_buffer_unlock_commit_nostack(struct trace_buffer *buffer, 2889 struct ring_buffer_event *event) 2890 { 2891 __buffer_unlock_commit(buffer, event); 2892 } 2893 2894 void 2895 trace_function(struct trace_array *tr, 2896 unsigned long ip, unsigned long parent_ip, unsigned long flags, 2897 int pc) 2898 { 2899 struct trace_event_call *call = &event_function; 2900 struct trace_buffer *buffer = tr->array_buffer.buffer; 2901 struct ring_buffer_event *event; 2902 struct ftrace_entry *entry; 2903 2904 event = __trace_buffer_lock_reserve(buffer, TRACE_FN, sizeof(*entry), 2905 flags, pc); 2906 if (!event) 2907 return; 2908 entry = ring_buffer_event_data(event); 2909 entry->ip = ip; 2910 entry->parent_ip = parent_ip; 2911 2912 if (!call_filter_check_discard(call, entry, buffer, event)) { 2913 if (static_branch_unlikely(&trace_function_exports_enabled)) 2914 ftrace_exports(event, TRACE_EXPORT_FUNCTION); 2915 __buffer_unlock_commit(buffer, event); 2916 } 2917 } 2918 2919 #ifdef CONFIG_STACKTRACE 2920 2921 /* Allow 4 levels of nesting: normal, softirq, irq, NMI */ 2922 #define FTRACE_KSTACK_NESTING 4 2923 2924 #define FTRACE_KSTACK_ENTRIES (PAGE_SIZE / FTRACE_KSTACK_NESTING) 2925 2926 struct ftrace_stack { 2927 unsigned long calls[FTRACE_KSTACK_ENTRIES]; 2928 }; 2929 2930 2931 struct ftrace_stacks { 2932 struct ftrace_stack stacks[FTRACE_KSTACK_NESTING]; 2933 }; 2934 2935 static DEFINE_PER_CPU(struct ftrace_stacks, ftrace_stacks); 2936 static DEFINE_PER_CPU(int, ftrace_stack_reserve); 2937 2938 static void __ftrace_trace_stack(struct trace_buffer *buffer, 2939 unsigned long flags, 2940 int skip, int pc, struct pt_regs *regs) 2941 { 2942 struct trace_event_call *call = &event_kernel_stack; 2943 struct ring_buffer_event *event; 2944 unsigned int size, nr_entries; 2945 struct ftrace_stack *fstack; 2946 struct stack_entry *entry; 2947 int stackidx; 2948 2949 /* 2950 * Add one, for this function and the call to save_stack_trace() 2951 * If regs is set, then these functions will not be in the way. 2952 */ 2953 #ifndef CONFIG_UNWINDER_ORC 2954 if (!regs) 2955 skip++; 2956 #endif 2957 2958 preempt_disable_notrace(); 2959 2960 stackidx = __this_cpu_inc_return(ftrace_stack_reserve) - 1; 2961 2962 /* This should never happen. If it does, yell once and skip */ 2963 if (WARN_ON_ONCE(stackidx >= FTRACE_KSTACK_NESTING)) 2964 goto out; 2965 2966 /* 2967 * The above __this_cpu_inc_return() is 'atomic' cpu local. An 2968 * interrupt will either see the value pre increment or post 2969 * increment. If the interrupt happens pre increment it will have 2970 * restored the counter when it returns. We just need a barrier to 2971 * keep gcc from moving things around. 2972 */ 2973 barrier(); 2974 2975 fstack = this_cpu_ptr(ftrace_stacks.stacks) + stackidx; 2976 size = ARRAY_SIZE(fstack->calls); 2977 2978 if (regs) { 2979 nr_entries = stack_trace_save_regs(regs, fstack->calls, 2980 size, skip); 2981 } else { 2982 nr_entries = stack_trace_save(fstack->calls, size, skip); 2983 } 2984 2985 size = nr_entries * sizeof(unsigned long); 2986 event = __trace_buffer_lock_reserve(buffer, TRACE_STACK, 2987 sizeof(*entry) + size, flags, pc); 2988 if (!event) 2989 goto out; 2990 entry = ring_buffer_event_data(event); 2991 2992 memcpy(&entry->caller, fstack->calls, size); 2993 entry->size = nr_entries; 2994 2995 if (!call_filter_check_discard(call, entry, buffer, event)) 2996 __buffer_unlock_commit(buffer, event); 2997 2998 out: 2999 /* Again, don't let gcc optimize things here */ 3000 barrier(); 3001 __this_cpu_dec(ftrace_stack_reserve); 3002 preempt_enable_notrace(); 3003 3004 } 3005 3006 static inline void ftrace_trace_stack(struct trace_array *tr, 3007 struct trace_buffer *buffer, 3008 unsigned long flags, 3009 int skip, int pc, struct pt_regs *regs) 3010 { 3011 if (!(tr->trace_flags & TRACE_ITER_STACKTRACE)) 3012 return; 3013 3014 __ftrace_trace_stack(buffer, flags, skip, pc, regs); 3015 } 3016 3017 void __trace_stack(struct trace_array *tr, unsigned long flags, int skip, 3018 int pc) 3019 { 3020 struct trace_buffer *buffer = tr->array_buffer.buffer; 3021 3022 if (rcu_is_watching()) { 3023 __ftrace_trace_stack(buffer, flags, skip, pc, NULL); 3024 return; 3025 } 3026 3027 /* 3028 * When an NMI triggers, RCU is enabled via rcu_nmi_enter(), 3029 * but if the above rcu_is_watching() failed, then the NMI 3030 * triggered someplace critical, and rcu_irq_enter() should 3031 * not be called from NMI. 3032 */ 3033 if (unlikely(in_nmi())) 3034 return; 3035 3036 rcu_irq_enter_irqson(); 3037 __ftrace_trace_stack(buffer, flags, skip, pc, NULL); 3038 rcu_irq_exit_irqson(); 3039 } 3040 3041 /** 3042 * trace_dump_stack - record a stack back trace in the trace buffer 3043 * @skip: Number of functions to skip (helper handlers) 3044 */ 3045 void trace_dump_stack(int skip) 3046 { 3047 unsigned long flags; 3048 3049 if (tracing_disabled || tracing_selftest_running) 3050 return; 3051 3052 local_save_flags(flags); 3053 3054 #ifndef CONFIG_UNWINDER_ORC 3055 /* Skip 1 to skip this function. */ 3056 skip++; 3057 #endif 3058 __ftrace_trace_stack(global_trace.array_buffer.buffer, 3059 flags, skip, preempt_count(), NULL); 3060 } 3061 EXPORT_SYMBOL_GPL(trace_dump_stack); 3062 3063 #ifdef CONFIG_USER_STACKTRACE_SUPPORT 3064 static DEFINE_PER_CPU(int, user_stack_count); 3065 3066 static void 3067 ftrace_trace_userstack(struct trace_array *tr, 3068 struct trace_buffer *buffer, unsigned long flags, int pc) 3069 { 3070 struct trace_event_call *call = &event_user_stack; 3071 struct ring_buffer_event *event; 3072 struct userstack_entry *entry; 3073 3074 if (!(tr->trace_flags & TRACE_ITER_USERSTACKTRACE)) 3075 return; 3076 3077 /* 3078 * NMIs can not handle page faults, even with fix ups. 3079 * The save user stack can (and often does) fault. 3080 */ 3081 if (unlikely(in_nmi())) 3082 return; 3083 3084 /* 3085 * prevent recursion, since the user stack tracing may 3086 * trigger other kernel events. 3087 */ 3088 preempt_disable(); 3089 if (__this_cpu_read(user_stack_count)) 3090 goto out; 3091 3092 __this_cpu_inc(user_stack_count); 3093 3094 event = __trace_buffer_lock_reserve(buffer, TRACE_USER_STACK, 3095 sizeof(*entry), flags, pc); 3096 if (!event) 3097 goto out_drop_count; 3098 entry = ring_buffer_event_data(event); 3099 3100 entry->tgid = current->tgid; 3101 memset(&entry->caller, 0, sizeof(entry->caller)); 3102 3103 stack_trace_save_user(entry->caller, FTRACE_STACK_ENTRIES); 3104 if (!call_filter_check_discard(call, entry, buffer, event)) 3105 __buffer_unlock_commit(buffer, event); 3106 3107 out_drop_count: 3108 __this_cpu_dec(user_stack_count); 3109 out: 3110 preempt_enable(); 3111 } 3112 #else /* CONFIG_USER_STACKTRACE_SUPPORT */ 3113 static void ftrace_trace_userstack(struct trace_array *tr, 3114 struct trace_buffer *buffer, 3115 unsigned long flags, int pc) 3116 { 3117 } 3118 #endif /* !CONFIG_USER_STACKTRACE_SUPPORT */ 3119 3120 #endif /* CONFIG_STACKTRACE */ 3121 3122 /* created for use with alloc_percpu */ 3123 struct trace_buffer_struct { 3124 int nesting; 3125 char buffer[4][TRACE_BUF_SIZE]; 3126 }; 3127 3128 static struct trace_buffer_struct *trace_percpu_buffer; 3129 3130 /* 3131 * This allows for lockless recording. If we're nested too deeply, then 3132 * this returns NULL. 3133 */ 3134 static char *get_trace_buf(void) 3135 { 3136 struct trace_buffer_struct *buffer = this_cpu_ptr(trace_percpu_buffer); 3137 3138 if (!buffer || buffer->nesting >= 4) 3139 return NULL; 3140 3141 buffer->nesting++; 3142 3143 /* Interrupts must see nesting incremented before we use the buffer */ 3144 barrier(); 3145 return &buffer->buffer[buffer->nesting - 1][0]; 3146 } 3147 3148 static void put_trace_buf(void) 3149 { 3150 /* Don't let the decrement of nesting leak before this */ 3151 barrier(); 3152 this_cpu_dec(trace_percpu_buffer->nesting); 3153 } 3154 3155 static int alloc_percpu_trace_buffer(void) 3156 { 3157 struct trace_buffer_struct *buffers; 3158 3159 if (trace_percpu_buffer) 3160 return 0; 3161 3162 buffers = alloc_percpu(struct trace_buffer_struct); 3163 if (MEM_FAIL(!buffers, "Could not allocate percpu trace_printk buffer")) 3164 return -ENOMEM; 3165 3166 trace_percpu_buffer = buffers; 3167 return 0; 3168 } 3169 3170 static int buffers_allocated; 3171 3172 void trace_printk_init_buffers(void) 3173 { 3174 if (buffers_allocated) 3175 return; 3176 3177 if (alloc_percpu_trace_buffer()) 3178 return; 3179 3180 /* trace_printk() is for debug use only. Don't use it in production. */ 3181 3182 pr_warn("\n"); 3183 pr_warn("**********************************************************\n"); 3184 pr_warn("** NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE **\n"); 3185 pr_warn("** **\n"); 3186 pr_warn("** trace_printk() being used. Allocating extra memory. **\n"); 3187 pr_warn("** **\n"); 3188 pr_warn("** This means that this is a DEBUG kernel and it is **\n"); 3189 pr_warn("** unsafe for production use. **\n"); 3190 pr_warn("** **\n"); 3191 pr_warn("** If you see this message and you are not debugging **\n"); 3192 pr_warn("** the kernel, report this immediately to your vendor! **\n"); 3193 pr_warn("** **\n"); 3194 pr_warn("** NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE **\n"); 3195 pr_warn("**********************************************************\n"); 3196 3197 /* Expand the buffers to set size */ 3198 tracing_update_buffers(); 3199 3200 buffers_allocated = 1; 3201 3202 /* 3203 * trace_printk_init_buffers() can be called by modules. 3204 * If that happens, then we need to start cmdline recording 3205 * directly here. If the global_trace.buffer is already 3206 * allocated here, then this was called by module code. 3207 */ 3208 if (global_trace.array_buffer.buffer) 3209 tracing_start_cmdline_record(); 3210 } 3211 EXPORT_SYMBOL_GPL(trace_printk_init_buffers); 3212 3213 void trace_printk_start_comm(void) 3214 { 3215 /* Start tracing comms if trace printk is set */ 3216 if (!buffers_allocated) 3217 return; 3218 tracing_start_cmdline_record(); 3219 } 3220 3221 static void trace_printk_start_stop_comm(int enabled) 3222 { 3223 if (!buffers_allocated) 3224 return; 3225 3226 if (enabled) 3227 tracing_start_cmdline_record(); 3228 else 3229 tracing_stop_cmdline_record(); 3230 } 3231 3232 /** 3233 * trace_vbprintk - write binary msg to tracing buffer 3234 * @ip: The address of the caller 3235 * @fmt: The string format to write to the buffer 3236 * @args: Arguments for @fmt 3237 */ 3238 int trace_vbprintk(unsigned long ip, const char *fmt, va_list args) 3239 { 3240 struct trace_event_call *call = &event_bprint; 3241 struct ring_buffer_event *event; 3242 struct trace_buffer *buffer; 3243 struct trace_array *tr = &global_trace; 3244 struct bprint_entry *entry; 3245 unsigned long flags; 3246 char *tbuffer; 3247 int len = 0, size, pc; 3248 3249 if (unlikely(tracing_selftest_running || tracing_disabled)) 3250 return 0; 3251 3252 /* Don't pollute graph traces with trace_vprintk internals */ 3253 pause_graph_tracing(); 3254 3255 pc = preempt_count(); 3256 preempt_disable_notrace(); 3257 3258 tbuffer = get_trace_buf(); 3259 if (!tbuffer) { 3260 len = 0; 3261 goto out_nobuffer; 3262 } 3263 3264 len = vbin_printf((u32 *)tbuffer, TRACE_BUF_SIZE/sizeof(int), fmt, args); 3265 3266 if (len > TRACE_BUF_SIZE/sizeof(int) || len < 0) 3267 goto out_put; 3268 3269 local_save_flags(flags); 3270 size = sizeof(*entry) + sizeof(u32) * len; 3271 buffer = tr->array_buffer.buffer; 3272 ring_buffer_nest_start(buffer); 3273 event = __trace_buffer_lock_reserve(buffer, TRACE_BPRINT, size, 3274 flags, pc); 3275 if (!event) 3276 goto out; 3277 entry = ring_buffer_event_data(event); 3278 entry->ip = ip; 3279 entry->fmt = fmt; 3280 3281 memcpy(entry->buf, tbuffer, sizeof(u32) * len); 3282 if (!call_filter_check_discard(call, entry, buffer, event)) { 3283 __buffer_unlock_commit(buffer, event); 3284 ftrace_trace_stack(tr, buffer, flags, 6, pc, NULL); 3285 } 3286 3287 out: 3288 ring_buffer_nest_end(buffer); 3289 out_put: 3290 put_trace_buf(); 3291 3292 out_nobuffer: 3293 preempt_enable_notrace(); 3294 unpause_graph_tracing(); 3295 3296 return len; 3297 } 3298 EXPORT_SYMBOL_GPL(trace_vbprintk); 3299 3300 __printf(3, 0) 3301 static int 3302 __trace_array_vprintk(struct trace_buffer *buffer, 3303 unsigned long ip, const char *fmt, va_list args) 3304 { 3305 struct trace_event_call *call = &event_print; 3306 struct ring_buffer_event *event; 3307 int len = 0, size, pc; 3308 struct print_entry *entry; 3309 unsigned long flags; 3310 char *tbuffer; 3311 3312 if (tracing_disabled || tracing_selftest_running) 3313 return 0; 3314 3315 /* Don't pollute graph traces with trace_vprintk internals */ 3316 pause_graph_tracing(); 3317 3318 pc = preempt_count(); 3319 preempt_disable_notrace(); 3320 3321 3322 tbuffer = get_trace_buf(); 3323 if (!tbuffer) { 3324 len = 0; 3325 goto out_nobuffer; 3326 } 3327 3328 len = vscnprintf(tbuffer, TRACE_BUF_SIZE, fmt, args); 3329 3330 local_save_flags(flags); 3331 size = sizeof(*entry) + len + 1; 3332 ring_buffer_nest_start(buffer); 3333 event = __trace_buffer_lock_reserve(buffer, TRACE_PRINT, size, 3334 flags, pc); 3335 if (!event) 3336 goto out; 3337 entry = ring_buffer_event_data(event); 3338 entry->ip = ip; 3339 3340 memcpy(&entry->buf, tbuffer, len + 1); 3341 if (!call_filter_check_discard(call, entry, buffer, event)) { 3342 __buffer_unlock_commit(buffer, event); 3343 ftrace_trace_stack(&global_trace, buffer, flags, 6, pc, NULL); 3344 } 3345 3346 out: 3347 ring_buffer_nest_end(buffer); 3348 put_trace_buf(); 3349 3350 out_nobuffer: 3351 preempt_enable_notrace(); 3352 unpause_graph_tracing(); 3353 3354 return len; 3355 } 3356 3357 __printf(3, 0) 3358 int trace_array_vprintk(struct trace_array *tr, 3359 unsigned long ip, const char *fmt, va_list args) 3360 { 3361 return __trace_array_vprintk(tr->array_buffer.buffer, ip, fmt, args); 3362 } 3363 3364 /** 3365 * trace_array_printk - Print a message to a specific instance 3366 * @tr: The instance trace_array descriptor 3367 * @ip: The instruction pointer that this is called from. 3368 * @fmt: The format to print (printf format) 3369 * 3370 * If a subsystem sets up its own instance, they have the right to 3371 * printk strings into their tracing instance buffer using this 3372 * function. Note, this function will not write into the top level 3373 * buffer (use trace_printk() for that), as writing into the top level 3374 * buffer should only have events that can be individually disabled. 3375 * trace_printk() is only used for debugging a kernel, and should not 3376 * be ever encorporated in normal use. 3377 * 3378 * trace_array_printk() can be used, as it will not add noise to the 3379 * top level tracing buffer. 3380 * 3381 * Note, trace_array_init_printk() must be called on @tr before this 3382 * can be used. 3383 */ 3384 __printf(3, 0) 3385 int trace_array_printk(struct trace_array *tr, 3386 unsigned long ip, const char *fmt, ...) 3387 { 3388 int ret; 3389 va_list ap; 3390 3391 if (!tr) 3392 return -ENOENT; 3393 3394 /* This is only allowed for created instances */ 3395 if (tr == &global_trace) 3396 return 0; 3397 3398 if (!(tr->trace_flags & TRACE_ITER_PRINTK)) 3399 return 0; 3400 3401 va_start(ap, fmt); 3402 ret = trace_array_vprintk(tr, ip, fmt, ap); 3403 va_end(ap); 3404 return ret; 3405 } 3406 EXPORT_SYMBOL_GPL(trace_array_printk); 3407 3408 /** 3409 * trace_array_init_printk - Initialize buffers for trace_array_printk() 3410 * @tr: The trace array to initialize the buffers for 3411 * 3412 * As trace_array_printk() only writes into instances, they are OK to 3413 * have in the kernel (unlike trace_printk()). This needs to be called 3414 * before trace_array_printk() can be used on a trace_array. 3415 */ 3416 int trace_array_init_printk(struct trace_array *tr) 3417 { 3418 if (!tr) 3419 return -ENOENT; 3420 3421 /* This is only allowed for created instances */ 3422 if (tr == &global_trace) 3423 return -EINVAL; 3424 3425 return alloc_percpu_trace_buffer(); 3426 } 3427 EXPORT_SYMBOL_GPL(trace_array_init_printk); 3428 3429 __printf(3, 4) 3430 int trace_array_printk_buf(struct trace_buffer *buffer, 3431 unsigned long ip, const char *fmt, ...) 3432 { 3433 int ret; 3434 va_list ap; 3435 3436 if (!(global_trace.trace_flags & TRACE_ITER_PRINTK)) 3437 return 0; 3438 3439 va_start(ap, fmt); 3440 ret = __trace_array_vprintk(buffer, ip, fmt, ap); 3441 va_end(ap); 3442 return ret; 3443 } 3444 3445 __printf(2, 0) 3446 int trace_vprintk(unsigned long ip, const char *fmt, va_list args) 3447 { 3448 return trace_array_vprintk(&global_trace, ip, fmt, args); 3449 } 3450 EXPORT_SYMBOL_GPL(trace_vprintk); 3451 3452 static void trace_iterator_increment(struct trace_iterator *iter) 3453 { 3454 struct ring_buffer_iter *buf_iter = trace_buffer_iter(iter, iter->cpu); 3455 3456 iter->idx++; 3457 if (buf_iter) 3458 ring_buffer_iter_advance(buf_iter); 3459 } 3460 3461 static struct trace_entry * 3462 peek_next_entry(struct trace_iterator *iter, int cpu, u64 *ts, 3463 unsigned long *lost_events) 3464 { 3465 struct ring_buffer_event *event; 3466 struct ring_buffer_iter *buf_iter = trace_buffer_iter(iter, cpu); 3467 3468 if (buf_iter) { 3469 event = ring_buffer_iter_peek(buf_iter, ts); 3470 if (lost_events) 3471 *lost_events = ring_buffer_iter_dropped(buf_iter) ? 3472 (unsigned long)-1 : 0; 3473 } else { 3474 event = ring_buffer_peek(iter->array_buffer->buffer, cpu, ts, 3475 lost_events); 3476 } 3477 3478 if (event) { 3479 iter->ent_size = ring_buffer_event_length(event); 3480 return ring_buffer_event_data(event); 3481 } 3482 iter->ent_size = 0; 3483 return NULL; 3484 } 3485 3486 static struct trace_entry * 3487 __find_next_entry(struct trace_iterator *iter, int *ent_cpu, 3488 unsigned long *missing_events, u64 *ent_ts) 3489 { 3490 struct trace_buffer *buffer = iter->array_buffer->buffer; 3491 struct trace_entry *ent, *next = NULL; 3492 unsigned long lost_events = 0, next_lost = 0; 3493 int cpu_file = iter->cpu_file; 3494 u64 next_ts = 0, ts; 3495 int next_cpu = -1; 3496 int next_size = 0; 3497 int cpu; 3498 3499 /* 3500 * If we are in a per_cpu trace file, don't bother by iterating over 3501 * all cpu and peek directly. 3502 */ 3503 if (cpu_file > RING_BUFFER_ALL_CPUS) { 3504 if (ring_buffer_empty_cpu(buffer, cpu_file)) 3505 return NULL; 3506 ent = peek_next_entry(iter, cpu_file, ent_ts, missing_events); 3507 if (ent_cpu) 3508 *ent_cpu = cpu_file; 3509 3510 return ent; 3511 } 3512 3513 for_each_tracing_cpu(cpu) { 3514 3515 if (ring_buffer_empty_cpu(buffer, cpu)) 3516 continue; 3517 3518 ent = peek_next_entry(iter, cpu, &ts, &lost_events); 3519 3520 /* 3521 * Pick the entry with the smallest timestamp: 3522 */ 3523 if (ent && (!next || ts < next_ts)) { 3524 next = ent; 3525 next_cpu = cpu; 3526 next_ts = ts; 3527 next_lost = lost_events; 3528 next_size = iter->ent_size; 3529 } 3530 } 3531 3532 iter->ent_size = next_size; 3533 3534 if (ent_cpu) 3535 *ent_cpu = next_cpu; 3536 3537 if (ent_ts) 3538 *ent_ts = next_ts; 3539 3540 if (missing_events) 3541 *missing_events = next_lost; 3542 3543 return next; 3544 } 3545 3546 #define STATIC_TEMP_BUF_SIZE 128 3547 static char static_temp_buf[STATIC_TEMP_BUF_SIZE] __aligned(4); 3548 3549 /* Find the next real entry, without updating the iterator itself */ 3550 struct trace_entry *trace_find_next_entry(struct trace_iterator *iter, 3551 int *ent_cpu, u64 *ent_ts) 3552 { 3553 /* __find_next_entry will reset ent_size */ 3554 int ent_size = iter->ent_size; 3555 struct trace_entry *entry; 3556 3557 /* 3558 * If called from ftrace_dump(), then the iter->temp buffer 3559 * will be the static_temp_buf and not created from kmalloc. 3560 * If the entry size is greater than the buffer, we can 3561 * not save it. Just return NULL in that case. This is only 3562 * used to add markers when two consecutive events' time 3563 * stamps have a large delta. See trace_print_lat_context() 3564 */ 3565 if (iter->temp == static_temp_buf && 3566 STATIC_TEMP_BUF_SIZE < ent_size) 3567 return NULL; 3568 3569 /* 3570 * The __find_next_entry() may call peek_next_entry(), which may 3571 * call ring_buffer_peek() that may make the contents of iter->ent 3572 * undefined. Need to copy iter->ent now. 3573 */ 3574 if (iter->ent && iter->ent != iter->temp) { 3575 if ((!iter->temp || iter->temp_size < iter->ent_size) && 3576 !WARN_ON_ONCE(iter->temp == static_temp_buf)) { 3577 void *temp; 3578 temp = kmalloc(iter->ent_size, GFP_KERNEL); 3579 if (!temp) 3580 return NULL; 3581 kfree(iter->temp); 3582 iter->temp = temp; 3583 iter->temp_size = iter->ent_size; 3584 } 3585 memcpy(iter->temp, iter->ent, iter->ent_size); 3586 iter->ent = iter->temp; 3587 } 3588 entry = __find_next_entry(iter, ent_cpu, NULL, ent_ts); 3589 /* Put back the original ent_size */ 3590 iter->ent_size = ent_size; 3591 3592 return entry; 3593 } 3594 3595 /* Find the next real entry, and increment the iterator to the next entry */ 3596 void *trace_find_next_entry_inc(struct trace_iterator *iter) 3597 { 3598 iter->ent = __find_next_entry(iter, &iter->cpu, 3599 &iter->lost_events, &iter->ts); 3600 3601 if (iter->ent) 3602 trace_iterator_increment(iter); 3603 3604 return iter->ent ? iter : NULL; 3605 } 3606 3607 static void trace_consume(struct trace_iterator *iter) 3608 { 3609 ring_buffer_consume(iter->array_buffer->buffer, iter->cpu, &iter->ts, 3610 &iter->lost_events); 3611 } 3612 3613 static void *s_next(struct seq_file *m, void *v, loff_t *pos) 3614 { 3615 struct trace_iterator *iter = m->private; 3616 int i = (int)*pos; 3617 void *ent; 3618 3619 WARN_ON_ONCE(iter->leftover); 3620 3621 (*pos)++; 3622 3623 /* can't go backwards */ 3624 if (iter->idx > i) 3625 return NULL; 3626 3627 if (iter->idx < 0) 3628 ent = trace_find_next_entry_inc(iter); 3629 else 3630 ent = iter; 3631 3632 while (ent && iter->idx < i) 3633 ent = trace_find_next_entry_inc(iter); 3634 3635 iter->pos = *pos; 3636 3637 return ent; 3638 } 3639 3640 void tracing_iter_reset(struct trace_iterator *iter, int cpu) 3641 { 3642 struct ring_buffer_iter *buf_iter; 3643 unsigned long entries = 0; 3644 u64 ts; 3645 3646 per_cpu_ptr(iter->array_buffer->data, cpu)->skipped_entries = 0; 3647 3648 buf_iter = trace_buffer_iter(iter, cpu); 3649 if (!buf_iter) 3650 return; 3651 3652 ring_buffer_iter_reset(buf_iter); 3653 3654 /* 3655 * We could have the case with the max latency tracers 3656 * that a reset never took place on a cpu. This is evident 3657 * by the timestamp being before the start of the buffer. 3658 */ 3659 while (ring_buffer_iter_peek(buf_iter, &ts)) { 3660 if (ts >= iter->array_buffer->time_start) 3661 break; 3662 entries++; 3663 ring_buffer_iter_advance(buf_iter); 3664 } 3665 3666 per_cpu_ptr(iter->array_buffer->data, cpu)->skipped_entries = entries; 3667 } 3668 3669 /* 3670 * The current tracer is copied to avoid a global locking 3671 * all around. 3672 */ 3673 static void *s_start(struct seq_file *m, loff_t *pos) 3674 { 3675 struct trace_iterator *iter = m->private; 3676 struct trace_array *tr = iter->tr; 3677 int cpu_file = iter->cpu_file; 3678 void *p = NULL; 3679 loff_t l = 0; 3680 int cpu; 3681 3682 /* 3683 * copy the tracer to avoid using a global lock all around. 3684 * iter->trace is a copy of current_trace, the pointer to the 3685 * name may be used instead of a strcmp(), as iter->trace->name 3686 * will point to the same string as current_trace->name. 3687 */ 3688 mutex_lock(&trace_types_lock); 3689 if (unlikely(tr->current_trace && iter->trace->name != tr->current_trace->name)) 3690 *iter->trace = *tr->current_trace; 3691 mutex_unlock(&trace_types_lock); 3692 3693 #ifdef CONFIG_TRACER_MAX_TRACE 3694 if (iter->snapshot && iter->trace->use_max_tr) 3695 return ERR_PTR(-EBUSY); 3696 #endif 3697 3698 if (!iter->snapshot) 3699 atomic_inc(&trace_record_taskinfo_disabled); 3700 3701 if (*pos != iter->pos) { 3702 iter->ent = NULL; 3703 iter->cpu = 0; 3704 iter->idx = -1; 3705 3706 if (cpu_file == RING_BUFFER_ALL_CPUS) { 3707 for_each_tracing_cpu(cpu) 3708 tracing_iter_reset(iter, cpu); 3709 } else 3710 tracing_iter_reset(iter, cpu_file); 3711 3712 iter->leftover = 0; 3713 for (p = iter; p && l < *pos; p = s_next(m, p, &l)) 3714 ; 3715 3716 } else { 3717 /* 3718 * If we overflowed the seq_file before, then we want 3719 * to just reuse the trace_seq buffer again. 3720 */ 3721 if (iter->leftover) 3722 p = iter; 3723 else { 3724 l = *pos - 1; 3725 p = s_next(m, p, &l); 3726 } 3727 } 3728 3729 trace_event_read_lock(); 3730 trace_access_lock(cpu_file); 3731 return p; 3732 } 3733 3734 static void s_stop(struct seq_file *m, void *p) 3735 { 3736 struct trace_iterator *iter = m->private; 3737 3738 #ifdef CONFIG_TRACER_MAX_TRACE 3739 if (iter->snapshot && iter->trace->use_max_tr) 3740 return; 3741 #endif 3742 3743 if (!iter->snapshot) 3744 atomic_dec(&trace_record_taskinfo_disabled); 3745 3746 trace_access_unlock(iter->cpu_file); 3747 trace_event_read_unlock(); 3748 } 3749 3750 static void 3751 get_total_entries_cpu(struct array_buffer *buf, unsigned long *total, 3752 unsigned long *entries, int cpu) 3753 { 3754 unsigned long count; 3755 3756 count = ring_buffer_entries_cpu(buf->buffer, cpu); 3757 /* 3758 * If this buffer has skipped entries, then we hold all 3759 * entries for the trace and we need to ignore the 3760 * ones before the time stamp. 3761 */ 3762 if (per_cpu_ptr(buf->data, cpu)->skipped_entries) { 3763 count -= per_cpu_ptr(buf->data, cpu)->skipped_entries; 3764 /* total is the same as the entries */ 3765 *total = count; 3766 } else 3767 *total = count + 3768 ring_buffer_overrun_cpu(buf->buffer, cpu); 3769 *entries = count; 3770 } 3771 3772 static void 3773 get_total_entries(struct array_buffer *buf, 3774 unsigned long *total, unsigned long *entries) 3775 { 3776 unsigned long t, e; 3777 int cpu; 3778 3779 *total = 0; 3780 *entries = 0; 3781 3782 for_each_tracing_cpu(cpu) { 3783 get_total_entries_cpu(buf, &t, &e, cpu); 3784 *total += t; 3785 *entries += e; 3786 } 3787 } 3788 3789 unsigned long trace_total_entries_cpu(struct trace_array *tr, int cpu) 3790 { 3791 unsigned long total, entries; 3792 3793 if (!tr) 3794 tr = &global_trace; 3795 3796 get_total_entries_cpu(&tr->array_buffer, &total, &entries, cpu); 3797 3798 return entries; 3799 } 3800 3801 unsigned long trace_total_entries(struct trace_array *tr) 3802 { 3803 unsigned long total, entries; 3804 3805 if (!tr) 3806 tr = &global_trace; 3807 3808 get_total_entries(&tr->array_buffer, &total, &entries); 3809 3810 return entries; 3811 } 3812 3813 static void print_lat_help_header(struct seq_file *m) 3814 { 3815 seq_puts(m, "# _------=> CPU# \n" 3816 "# / _-----=> irqs-off \n" 3817 "# | / _----=> need-resched \n" 3818 "# || / _---=> hardirq/softirq \n" 3819 "# ||| / _--=> preempt-depth \n" 3820 "# |||| / delay \n" 3821 "# cmd pid ||||| time | caller \n" 3822 "# \\ / ||||| \\ | / \n"); 3823 } 3824 3825 static void print_event_info(struct array_buffer *buf, struct seq_file *m) 3826 { 3827 unsigned long total; 3828 unsigned long entries; 3829 3830 get_total_entries(buf, &total, &entries); 3831 seq_printf(m, "# entries-in-buffer/entries-written: %lu/%lu #P:%d\n", 3832 entries, total, num_online_cpus()); 3833 seq_puts(m, "#\n"); 3834 } 3835 3836 static void print_func_help_header(struct array_buffer *buf, struct seq_file *m, 3837 unsigned int flags) 3838 { 3839 bool tgid = flags & TRACE_ITER_RECORD_TGID; 3840 3841 print_event_info(buf, m); 3842 3843 seq_printf(m, "# TASK-PID %s CPU# TIMESTAMP FUNCTION\n", tgid ? " TGID " : ""); 3844 seq_printf(m, "# | | %s | | |\n", tgid ? " | " : ""); 3845 } 3846 3847 static void print_func_help_header_irq(struct array_buffer *buf, struct seq_file *m, 3848 unsigned int flags) 3849 { 3850 bool tgid = flags & TRACE_ITER_RECORD_TGID; 3851 const char *space = " "; 3852 int prec = tgid ? 12 : 2; 3853 3854 print_event_info(buf, m); 3855 3856 seq_printf(m, "# %.*s _-----=> irqs-off\n", prec, space); 3857 seq_printf(m, "# %.*s / _----=> need-resched\n", prec, space); 3858 seq_printf(m, "# %.*s| / _---=> hardirq/softirq\n", prec, space); 3859 seq_printf(m, "# %.*s|| / _--=> preempt-depth\n", prec, space); 3860 seq_printf(m, "# %.*s||| / delay\n", prec, space); 3861 seq_printf(m, "# TASK-PID %.*s CPU# |||| TIMESTAMP FUNCTION\n", prec, " TGID "); 3862 seq_printf(m, "# | | %.*s | |||| | |\n", prec, " | "); 3863 } 3864 3865 void 3866 print_trace_header(struct seq_file *m, struct trace_iterator *iter) 3867 { 3868 unsigned long sym_flags = (global_trace.trace_flags & TRACE_ITER_SYM_MASK); 3869 struct array_buffer *buf = iter->array_buffer; 3870 struct trace_array_cpu *data = per_cpu_ptr(buf->data, buf->cpu); 3871 struct tracer *type = iter->trace; 3872 unsigned long entries; 3873 unsigned long total; 3874 const char *name = "preemption"; 3875 3876 name = type->name; 3877 3878 get_total_entries(buf, &total, &entries); 3879 3880 seq_printf(m, "# %s latency trace v1.1.5 on %s\n", 3881 name, UTS_RELEASE); 3882 seq_puts(m, "# -----------------------------------" 3883 "---------------------------------\n"); 3884 seq_printf(m, "# latency: %lu us, #%lu/%lu, CPU#%d |" 3885 " (M:%s VP:%d, KP:%d, SP:%d HP:%d", 3886 nsecs_to_usecs(data->saved_latency), 3887 entries, 3888 total, 3889 buf->cpu, 3890 #if defined(CONFIG_PREEMPT_NONE) 3891 "server", 3892 #elif defined(CONFIG_PREEMPT_VOLUNTARY) 3893 "desktop", 3894 #elif defined(CONFIG_PREEMPT) 3895 "preempt", 3896 #elif defined(CONFIG_PREEMPT_RT) 3897 "preempt_rt", 3898 #else 3899 "unknown", 3900 #endif 3901 /* These are reserved for later use */ 3902 0, 0, 0, 0); 3903 #ifdef CONFIG_SMP 3904 seq_printf(m, " #P:%d)\n", num_online_cpus()); 3905 #else 3906 seq_puts(m, ")\n"); 3907 #endif 3908 seq_puts(m, "# -----------------\n"); 3909 seq_printf(m, "# | task: %.16s-%d " 3910 "(uid:%d nice:%ld policy:%ld rt_prio:%ld)\n", 3911 data->comm, data->pid, 3912 from_kuid_munged(seq_user_ns(m), data->uid), data->nice, 3913 data->policy, data->rt_priority); 3914 seq_puts(m, "# -----------------\n"); 3915 3916 if (data->critical_start) { 3917 seq_puts(m, "# => started at: "); 3918 seq_print_ip_sym(&iter->seq, data->critical_start, sym_flags); 3919 trace_print_seq(m, &iter->seq); 3920 seq_puts(m, "\n# => ended at: "); 3921 seq_print_ip_sym(&iter->seq, data->critical_end, sym_flags); 3922 trace_print_seq(m, &iter->seq); 3923 seq_puts(m, "\n#\n"); 3924 } 3925 3926 seq_puts(m, "#\n"); 3927 } 3928 3929 static void test_cpu_buff_start(struct trace_iterator *iter) 3930 { 3931 struct trace_seq *s = &iter->seq; 3932 struct trace_array *tr = iter->tr; 3933 3934 if (!(tr->trace_flags & TRACE_ITER_ANNOTATE)) 3935 return; 3936 3937 if (!(iter->iter_flags & TRACE_FILE_ANNOTATE)) 3938 return; 3939 3940 if (cpumask_available(iter->started) && 3941 cpumask_test_cpu(iter->cpu, iter->started)) 3942 return; 3943 3944 if (per_cpu_ptr(iter->array_buffer->data, iter->cpu)->skipped_entries) 3945 return; 3946 3947 if (cpumask_available(iter->started)) 3948 cpumask_set_cpu(iter->cpu, iter->started); 3949 3950 /* Don't print started cpu buffer for the first entry of the trace */ 3951 if (iter->idx > 1) 3952 trace_seq_printf(s, "##### CPU %u buffer started ####\n", 3953 iter->cpu); 3954 } 3955 3956 static enum print_line_t print_trace_fmt(struct trace_iterator *iter) 3957 { 3958 struct trace_array *tr = iter->tr; 3959 struct trace_seq *s = &iter->seq; 3960 unsigned long sym_flags = (tr->trace_flags & TRACE_ITER_SYM_MASK); 3961 struct trace_entry *entry; 3962 struct trace_event *event; 3963 3964 entry = iter->ent; 3965 3966 test_cpu_buff_start(iter); 3967 3968 event = ftrace_find_event(entry->type); 3969 3970 if (tr->trace_flags & TRACE_ITER_CONTEXT_INFO) { 3971 if (iter->iter_flags & TRACE_FILE_LAT_FMT) 3972 trace_print_lat_context(iter); 3973 else 3974 trace_print_context(iter); 3975 } 3976 3977 if (trace_seq_has_overflowed(s)) 3978 return TRACE_TYPE_PARTIAL_LINE; 3979 3980 if (event) 3981 return event->funcs->trace(iter, sym_flags, event); 3982 3983 trace_seq_printf(s, "Unknown type %d\n", entry->type); 3984 3985 return trace_handle_return(s); 3986 } 3987 3988 static enum print_line_t print_raw_fmt(struct trace_iterator *iter) 3989 { 3990 struct trace_array *tr = iter->tr; 3991 struct trace_seq *s = &iter->seq; 3992 struct trace_entry *entry; 3993 struct trace_event *event; 3994 3995 entry = iter->ent; 3996 3997 if (tr->trace_flags & TRACE_ITER_CONTEXT_INFO) 3998 trace_seq_printf(s, "%d %d %llu ", 3999 entry->pid, iter->cpu, iter->ts); 4000 4001 if (trace_seq_has_overflowed(s)) 4002 return TRACE_TYPE_PARTIAL_LINE; 4003 4004 event = ftrace_find_event(entry->type); 4005 if (event) 4006 return event->funcs->raw(iter, 0, event); 4007 4008 trace_seq_printf(s, "%d ?\n", entry->type); 4009 4010 return trace_handle_return(s); 4011 } 4012 4013 static enum print_line_t print_hex_fmt(struct trace_iterator *iter) 4014 { 4015 struct trace_array *tr = iter->tr; 4016 struct trace_seq *s = &iter->seq; 4017 unsigned char newline = '\n'; 4018 struct trace_entry *entry; 4019 struct trace_event *event; 4020 4021 entry = iter->ent; 4022 4023 if (tr->trace_flags & TRACE_ITER_CONTEXT_INFO) { 4024 SEQ_PUT_HEX_FIELD(s, entry->pid); 4025 SEQ_PUT_HEX_FIELD(s, iter->cpu); 4026 SEQ_PUT_HEX_FIELD(s, iter->ts); 4027 if (trace_seq_has_overflowed(s)) 4028 return TRACE_TYPE_PARTIAL_LINE; 4029 } 4030 4031 event = ftrace_find_event(entry->type); 4032 if (event) { 4033 enum print_line_t ret = event->funcs->hex(iter, 0, event); 4034 if (ret != TRACE_TYPE_HANDLED) 4035 return ret; 4036 } 4037 4038 SEQ_PUT_FIELD(s, newline); 4039 4040 return trace_handle_return(s); 4041 } 4042 4043 static enum print_line_t print_bin_fmt(struct trace_iterator *iter) 4044 { 4045 struct trace_array *tr = iter->tr; 4046 struct trace_seq *s = &iter->seq; 4047 struct trace_entry *entry; 4048 struct trace_event *event; 4049 4050 entry = iter->ent; 4051 4052 if (tr->trace_flags & TRACE_ITER_CONTEXT_INFO) { 4053 SEQ_PUT_FIELD(s, entry->pid); 4054 SEQ_PUT_FIELD(s, iter->cpu); 4055 SEQ_PUT_FIELD(s, iter->ts); 4056 if (trace_seq_has_overflowed(s)) 4057 return TRACE_TYPE_PARTIAL_LINE; 4058 } 4059 4060 event = ftrace_find_event(entry->type); 4061 return event ? event->funcs->binary(iter, 0, event) : 4062 TRACE_TYPE_HANDLED; 4063 } 4064 4065 int trace_empty(struct trace_iterator *iter) 4066 { 4067 struct ring_buffer_iter *buf_iter; 4068 int cpu; 4069 4070 /* If we are looking at one CPU buffer, only check that one */ 4071 if (iter->cpu_file != RING_BUFFER_ALL_CPUS) { 4072 cpu = iter->cpu_file; 4073 buf_iter = trace_buffer_iter(iter, cpu); 4074 if (buf_iter) { 4075 if (!ring_buffer_iter_empty(buf_iter)) 4076 return 0; 4077 } else { 4078 if (!ring_buffer_empty_cpu(iter->array_buffer->buffer, cpu)) 4079 return 0; 4080 } 4081 return 1; 4082 } 4083 4084 for_each_tracing_cpu(cpu) { 4085 buf_iter = trace_buffer_iter(iter, cpu); 4086 if (buf_iter) { 4087 if (!ring_buffer_iter_empty(buf_iter)) 4088 return 0; 4089 } else { 4090 if (!ring_buffer_empty_cpu(iter->array_buffer->buffer, cpu)) 4091 return 0; 4092 } 4093 } 4094 4095 return 1; 4096 } 4097 4098 /* Called with trace_event_read_lock() held. */ 4099 enum print_line_t print_trace_line(struct trace_iterator *iter) 4100 { 4101 struct trace_array *tr = iter->tr; 4102 unsigned long trace_flags = tr->trace_flags; 4103 enum print_line_t ret; 4104 4105 if (iter->lost_events) { 4106 if (iter->lost_events == (unsigned long)-1) 4107 trace_seq_printf(&iter->seq, "CPU:%d [LOST EVENTS]\n", 4108 iter->cpu); 4109 else 4110 trace_seq_printf(&iter->seq, "CPU:%d [LOST %lu EVENTS]\n", 4111 iter->cpu, iter->lost_events); 4112 if (trace_seq_has_overflowed(&iter->seq)) 4113 return TRACE_TYPE_PARTIAL_LINE; 4114 } 4115 4116 if (iter->trace && iter->trace->print_line) { 4117 ret = iter->trace->print_line(iter); 4118 if (ret != TRACE_TYPE_UNHANDLED) 4119 return ret; 4120 } 4121 4122 if (iter->ent->type == TRACE_BPUTS && 4123 trace_flags & TRACE_ITER_PRINTK && 4124 trace_flags & TRACE_ITER_PRINTK_MSGONLY) 4125 return trace_print_bputs_msg_only(iter); 4126 4127 if (iter->ent->type == TRACE_BPRINT && 4128 trace_flags & TRACE_ITER_PRINTK && 4129 trace_flags & TRACE_ITER_PRINTK_MSGONLY) 4130 return trace_print_bprintk_msg_only(iter); 4131 4132 if (iter->ent->type == TRACE_PRINT && 4133 trace_flags & TRACE_ITER_PRINTK && 4134 trace_flags & TRACE_ITER_PRINTK_MSGONLY) 4135 return trace_print_printk_msg_only(iter); 4136 4137 if (trace_flags & TRACE_ITER_BIN) 4138 return print_bin_fmt(iter); 4139 4140 if (trace_flags & TRACE_ITER_HEX) 4141 return print_hex_fmt(iter); 4142 4143 if (trace_flags & TRACE_ITER_RAW) 4144 return print_raw_fmt(iter); 4145 4146 return print_trace_fmt(iter); 4147 } 4148 4149 void trace_latency_header(struct seq_file *m) 4150 { 4151 struct trace_iterator *iter = m->private; 4152 struct trace_array *tr = iter->tr; 4153 4154 /* print nothing if the buffers are empty */ 4155 if (trace_empty(iter)) 4156 return; 4157 4158 if (iter->iter_flags & TRACE_FILE_LAT_FMT) 4159 print_trace_header(m, iter); 4160 4161 if (!(tr->trace_flags & TRACE_ITER_VERBOSE)) 4162 print_lat_help_header(m); 4163 } 4164 4165 void trace_default_header(struct seq_file *m) 4166 { 4167 struct trace_iterator *iter = m->private; 4168 struct trace_array *tr = iter->tr; 4169 unsigned long trace_flags = tr->trace_flags; 4170 4171 if (!(trace_flags & TRACE_ITER_CONTEXT_INFO)) 4172 return; 4173 4174 if (iter->iter_flags & TRACE_FILE_LAT_FMT) { 4175 /* print nothing if the buffers are empty */ 4176 if (trace_empty(iter)) 4177 return; 4178 print_trace_header(m, iter); 4179 if (!(trace_flags & TRACE_ITER_VERBOSE)) 4180 print_lat_help_header(m); 4181 } else { 4182 if (!(trace_flags & TRACE_ITER_VERBOSE)) { 4183 if (trace_flags & TRACE_ITER_IRQ_INFO) 4184 print_func_help_header_irq(iter->array_buffer, 4185 m, trace_flags); 4186 else 4187 print_func_help_header(iter->array_buffer, m, 4188 trace_flags); 4189 } 4190 } 4191 } 4192 4193 static void test_ftrace_alive(struct seq_file *m) 4194 { 4195 if (!ftrace_is_dead()) 4196 return; 4197 seq_puts(m, "# WARNING: FUNCTION TRACING IS CORRUPTED\n" 4198 "# MAY BE MISSING FUNCTION EVENTS\n"); 4199 } 4200 4201 #ifdef CONFIG_TRACER_MAX_TRACE 4202 static void show_snapshot_main_help(struct seq_file *m) 4203 { 4204 seq_puts(m, "# echo 0 > snapshot : Clears and frees snapshot buffer\n" 4205 "# echo 1 > snapshot : Allocates snapshot buffer, if not already allocated.\n" 4206 "# Takes a snapshot of the main buffer.\n" 4207 "# echo 2 > snapshot : Clears snapshot buffer (but does not allocate or free)\n" 4208 "# (Doesn't have to be '2' works with any number that\n" 4209 "# is not a '0' or '1')\n"); 4210 } 4211 4212 static void show_snapshot_percpu_help(struct seq_file *m) 4213 { 4214 seq_puts(m, "# echo 0 > snapshot : Invalid for per_cpu snapshot file.\n"); 4215 #ifdef CONFIG_RING_BUFFER_ALLOW_SWAP 4216 seq_puts(m, "# echo 1 > snapshot : Allocates snapshot buffer, if not already allocated.\n" 4217 "# Takes a snapshot of the main buffer for this cpu.\n"); 4218 #else 4219 seq_puts(m, "# echo 1 > snapshot : Not supported with this kernel.\n" 4220 "# Must use main snapshot file to allocate.\n"); 4221 #endif 4222 seq_puts(m, "# echo 2 > snapshot : Clears this cpu's snapshot buffer (but does not allocate)\n" 4223 "# (Doesn't have to be '2' works with any number that\n" 4224 "# is not a '0' or '1')\n"); 4225 } 4226 4227 static void print_snapshot_help(struct seq_file *m, struct trace_iterator *iter) 4228 { 4229 if (iter->tr->allocated_snapshot) 4230 seq_puts(m, "#\n# * Snapshot is allocated *\n#\n"); 4231 else 4232 seq_puts(m, "#\n# * Snapshot is freed *\n#\n"); 4233 4234 seq_puts(m, "# Snapshot commands:\n"); 4235 if (iter->cpu_file == RING_BUFFER_ALL_CPUS) 4236 show_snapshot_main_help(m); 4237 else 4238 show_snapshot_percpu_help(m); 4239 } 4240 #else 4241 /* Should never be called */ 4242 static inline void print_snapshot_help(struct seq_file *m, struct trace_iterator *iter) { } 4243 #endif 4244 4245 static int s_show(struct seq_file *m, void *v) 4246 { 4247 struct trace_iterator *iter = v; 4248 int ret; 4249 4250 if (iter->ent == NULL) { 4251 if (iter->tr) { 4252 seq_printf(m, "# tracer: %s\n", iter->trace->name); 4253 seq_puts(m, "#\n"); 4254 test_ftrace_alive(m); 4255 } 4256 if (iter->snapshot && trace_empty(iter)) 4257 print_snapshot_help(m, iter); 4258 else if (iter->trace && iter->trace->print_header) 4259 iter->trace->print_header(m); 4260 else 4261 trace_default_header(m); 4262 4263 } else if (iter->leftover) { 4264 /* 4265 * If we filled the seq_file buffer earlier, we 4266 * want to just show it now. 4267 */ 4268 ret = trace_print_seq(m, &iter->seq); 4269 4270 /* ret should this time be zero, but you never know */ 4271 iter->leftover = ret; 4272 4273 } else { 4274 print_trace_line(iter); 4275 ret = trace_print_seq(m, &iter->seq); 4276 /* 4277 * If we overflow the seq_file buffer, then it will 4278 * ask us for this data again at start up. 4279 * Use that instead. 4280 * ret is 0 if seq_file write succeeded. 4281 * -1 otherwise. 4282 */ 4283 iter->leftover = ret; 4284 } 4285 4286 return 0; 4287 } 4288 4289 /* 4290 * Should be used after trace_array_get(), trace_types_lock 4291 * ensures that i_cdev was already initialized. 4292 */ 4293 static inline int tracing_get_cpu(struct inode *inode) 4294 { 4295 if (inode->i_cdev) /* See trace_create_cpu_file() */ 4296 return (long)inode->i_cdev - 1; 4297 return RING_BUFFER_ALL_CPUS; 4298 } 4299 4300 static const struct seq_operations tracer_seq_ops = { 4301 .start = s_start, 4302 .next = s_next, 4303 .stop = s_stop, 4304 .show = s_show, 4305 }; 4306 4307 static struct trace_iterator * 4308 __tracing_open(struct inode *inode, struct file *file, bool snapshot) 4309 { 4310 struct trace_array *tr = inode->i_private; 4311 struct trace_iterator *iter; 4312 int cpu; 4313 4314 if (tracing_disabled) 4315 return ERR_PTR(-ENODEV); 4316 4317 iter = __seq_open_private(file, &tracer_seq_ops, sizeof(*iter)); 4318 if (!iter) 4319 return ERR_PTR(-ENOMEM); 4320 4321 iter->buffer_iter = kcalloc(nr_cpu_ids, sizeof(*iter->buffer_iter), 4322 GFP_KERNEL); 4323 if (!iter->buffer_iter) 4324 goto release; 4325 4326 /* 4327 * trace_find_next_entry() may need to save off iter->ent. 4328 * It will place it into the iter->temp buffer. As most 4329 * events are less than 128, allocate a buffer of that size. 4330 * If one is greater, then trace_find_next_entry() will 4331 * allocate a new buffer to adjust for the bigger iter->ent. 4332 * It's not critical if it fails to get allocated here. 4333 */ 4334 iter->temp = kmalloc(128, GFP_KERNEL); 4335 if (iter->temp) 4336 iter->temp_size = 128; 4337 4338 /* 4339 * We make a copy of the current tracer to avoid concurrent 4340 * changes on it while we are reading. 4341 */ 4342 mutex_lock(&trace_types_lock); 4343 iter->trace = kzalloc(sizeof(*iter->trace), GFP_KERNEL); 4344 if (!iter->trace) 4345 goto fail; 4346 4347 *iter->trace = *tr->current_trace; 4348 4349 if (!zalloc_cpumask_var(&iter->started, GFP_KERNEL)) 4350 goto fail; 4351 4352 iter->tr = tr; 4353 4354 #ifdef CONFIG_TRACER_MAX_TRACE 4355 /* Currently only the top directory has a snapshot */ 4356 if (tr->current_trace->print_max || snapshot) 4357 iter->array_buffer = &tr->max_buffer; 4358 else 4359 #endif 4360 iter->array_buffer = &tr->array_buffer; 4361 iter->snapshot = snapshot; 4362 iter->pos = -1; 4363 iter->cpu_file = tracing_get_cpu(inode); 4364 mutex_init(&iter->mutex); 4365 4366 /* Notify the tracer early; before we stop tracing. */ 4367 if (iter->trace->open) 4368 iter->trace->open(iter); 4369 4370 /* Annotate start of buffers if we had overruns */ 4371 if (ring_buffer_overruns(iter->array_buffer->buffer)) 4372 iter->iter_flags |= TRACE_FILE_ANNOTATE; 4373 4374 /* Output in nanoseconds only if we are using a clock in nanoseconds. */ 4375 if (trace_clocks[tr->clock_id].in_ns) 4376 iter->iter_flags |= TRACE_FILE_TIME_IN_NS; 4377 4378 /* 4379 * If pause-on-trace is enabled, then stop the trace while 4380 * dumping, unless this is the "snapshot" file 4381 */ 4382 if (!iter->snapshot && (tr->trace_flags & TRACE_ITER_PAUSE_ON_TRACE)) 4383 tracing_stop_tr(tr); 4384 4385 if (iter->cpu_file == RING_BUFFER_ALL_CPUS) { 4386 for_each_tracing_cpu(cpu) { 4387 iter->buffer_iter[cpu] = 4388 ring_buffer_read_prepare(iter->array_buffer->buffer, 4389 cpu, GFP_KERNEL); 4390 } 4391 ring_buffer_read_prepare_sync(); 4392 for_each_tracing_cpu(cpu) { 4393 ring_buffer_read_start(iter->buffer_iter[cpu]); 4394 tracing_iter_reset(iter, cpu); 4395 } 4396 } else { 4397 cpu = iter->cpu_file; 4398 iter->buffer_iter[cpu] = 4399 ring_buffer_read_prepare(iter->array_buffer->buffer, 4400 cpu, GFP_KERNEL); 4401 ring_buffer_read_prepare_sync(); 4402 ring_buffer_read_start(iter->buffer_iter[cpu]); 4403 tracing_iter_reset(iter, cpu); 4404 } 4405 4406 mutex_unlock(&trace_types_lock); 4407 4408 return iter; 4409 4410 fail: 4411 mutex_unlock(&trace_types_lock); 4412 kfree(iter->trace); 4413 kfree(iter->temp); 4414 kfree(iter->buffer_iter); 4415 release: 4416 seq_release_private(inode, file); 4417 return ERR_PTR(-ENOMEM); 4418 } 4419 4420 int tracing_open_generic(struct inode *inode, struct file *filp) 4421 { 4422 int ret; 4423 4424 ret = tracing_check_open_get_tr(NULL); 4425 if (ret) 4426 return ret; 4427 4428 filp->private_data = inode->i_private; 4429 return 0; 4430 } 4431 4432 bool tracing_is_disabled(void) 4433 { 4434 return (tracing_disabled) ? true: false; 4435 } 4436 4437 /* 4438 * Open and update trace_array ref count. 4439 * Must have the current trace_array passed to it. 4440 */ 4441 int tracing_open_generic_tr(struct inode *inode, struct file *filp) 4442 { 4443 struct trace_array *tr = inode->i_private; 4444 int ret; 4445 4446 ret = tracing_check_open_get_tr(tr); 4447 if (ret) 4448 return ret; 4449 4450 filp->private_data = inode->i_private; 4451 4452 return 0; 4453 } 4454 4455 static int tracing_release(struct inode *inode, struct file *file) 4456 { 4457 struct trace_array *tr = inode->i_private; 4458 struct seq_file *m = file->private_data; 4459 struct trace_iterator *iter; 4460 int cpu; 4461 4462 if (!(file->f_mode & FMODE_READ)) { 4463 trace_array_put(tr); 4464 return 0; 4465 } 4466 4467 /* Writes do not use seq_file */ 4468 iter = m->private; 4469 mutex_lock(&trace_types_lock); 4470 4471 for_each_tracing_cpu(cpu) { 4472 if (iter->buffer_iter[cpu]) 4473 ring_buffer_read_finish(iter->buffer_iter[cpu]); 4474 } 4475 4476 if (iter->trace && iter->trace->close) 4477 iter->trace->close(iter); 4478 4479 if (!iter->snapshot && tr->stop_count) 4480 /* reenable tracing if it was previously enabled */ 4481 tracing_start_tr(tr); 4482 4483 __trace_array_put(tr); 4484 4485 mutex_unlock(&trace_types_lock); 4486 4487 mutex_destroy(&iter->mutex); 4488 free_cpumask_var(iter->started); 4489 kfree(iter->temp); 4490 kfree(iter->trace); 4491 kfree(iter->buffer_iter); 4492 seq_release_private(inode, file); 4493 4494 return 0; 4495 } 4496 4497 static int tracing_release_generic_tr(struct inode *inode, struct file *file) 4498 { 4499 struct trace_array *tr = inode->i_private; 4500 4501 trace_array_put(tr); 4502 return 0; 4503 } 4504 4505 static int tracing_single_release_tr(struct inode *inode, struct file *file) 4506 { 4507 struct trace_array *tr = inode->i_private; 4508 4509 trace_array_put(tr); 4510 4511 return single_release(inode, file); 4512 } 4513 4514 static int tracing_open(struct inode *inode, struct file *file) 4515 { 4516 struct trace_array *tr = inode->i_private; 4517 struct trace_iterator *iter; 4518 int ret; 4519 4520 ret = tracing_check_open_get_tr(tr); 4521 if (ret) 4522 return ret; 4523 4524 /* If this file was open for write, then erase contents */ 4525 if ((file->f_mode & FMODE_WRITE) && (file->f_flags & O_TRUNC)) { 4526 int cpu = tracing_get_cpu(inode); 4527 struct array_buffer *trace_buf = &tr->array_buffer; 4528 4529 #ifdef CONFIG_TRACER_MAX_TRACE 4530 if (tr->current_trace->print_max) 4531 trace_buf = &tr->max_buffer; 4532 #endif 4533 4534 if (cpu == RING_BUFFER_ALL_CPUS) 4535 tracing_reset_online_cpus(trace_buf); 4536 else 4537 tracing_reset_cpu(trace_buf, cpu); 4538 } 4539 4540 if (file->f_mode & FMODE_READ) { 4541 iter = __tracing_open(inode, file, false); 4542 if (IS_ERR(iter)) 4543 ret = PTR_ERR(iter); 4544 else if (tr->trace_flags & TRACE_ITER_LATENCY_FMT) 4545 iter->iter_flags |= TRACE_FILE_LAT_FMT; 4546 } 4547 4548 if (ret < 0) 4549 trace_array_put(tr); 4550 4551 return ret; 4552 } 4553 4554 /* 4555 * Some tracers are not suitable for instance buffers. 4556 * A tracer is always available for the global array (toplevel) 4557 * or if it explicitly states that it is. 4558 */ 4559 static bool 4560 trace_ok_for_array(struct tracer *t, struct trace_array *tr) 4561 { 4562 return (tr->flags & TRACE_ARRAY_FL_GLOBAL) || t->allow_instances; 4563 } 4564 4565 /* Find the next tracer that this trace array may use */ 4566 static struct tracer * 4567 get_tracer_for_array(struct trace_array *tr, struct tracer *t) 4568 { 4569 while (t && !trace_ok_for_array(t, tr)) 4570 t = t->next; 4571 4572 return t; 4573 } 4574 4575 static void * 4576 t_next(struct seq_file *m, void *v, loff_t *pos) 4577 { 4578 struct trace_array *tr = m->private; 4579 struct tracer *t = v; 4580 4581 (*pos)++; 4582 4583 if (t) 4584 t = get_tracer_for_array(tr, t->next); 4585 4586 return t; 4587 } 4588 4589 static void *t_start(struct seq_file *m, loff_t *pos) 4590 { 4591 struct trace_array *tr = m->private; 4592 struct tracer *t; 4593 loff_t l = 0; 4594 4595 mutex_lock(&trace_types_lock); 4596 4597 t = get_tracer_for_array(tr, trace_types); 4598 for (; t && l < *pos; t = t_next(m, t, &l)) 4599 ; 4600 4601 return t; 4602 } 4603 4604 static void t_stop(struct seq_file *m, void *p) 4605 { 4606 mutex_unlock(&trace_types_lock); 4607 } 4608 4609 static int t_show(struct seq_file *m, void *v) 4610 { 4611 struct tracer *t = v; 4612 4613 if (!t) 4614 return 0; 4615 4616 seq_puts(m, t->name); 4617 if (t->next) 4618 seq_putc(m, ' '); 4619 else 4620 seq_putc(m, '\n'); 4621 4622 return 0; 4623 } 4624 4625 static const struct seq_operations show_traces_seq_ops = { 4626 .start = t_start, 4627 .next = t_next, 4628 .stop = t_stop, 4629 .show = t_show, 4630 }; 4631 4632 static int show_traces_open(struct inode *inode, struct file *file) 4633 { 4634 struct trace_array *tr = inode->i_private; 4635 struct seq_file *m; 4636 int ret; 4637 4638 ret = tracing_check_open_get_tr(tr); 4639 if (ret) 4640 return ret; 4641 4642 ret = seq_open(file, &show_traces_seq_ops); 4643 if (ret) { 4644 trace_array_put(tr); 4645 return ret; 4646 } 4647 4648 m = file->private_data; 4649 m->private = tr; 4650 4651 return 0; 4652 } 4653 4654 static int show_traces_release(struct inode *inode, struct file *file) 4655 { 4656 struct trace_array *tr = inode->i_private; 4657 4658 trace_array_put(tr); 4659 return seq_release(inode, file); 4660 } 4661 4662 static ssize_t 4663 tracing_write_stub(struct file *filp, const char __user *ubuf, 4664 size_t count, loff_t *ppos) 4665 { 4666 return count; 4667 } 4668 4669 loff_t tracing_lseek(struct file *file, loff_t offset, int whence) 4670 { 4671 int ret; 4672 4673 if (file->f_mode & FMODE_READ) 4674 ret = seq_lseek(file, offset, whence); 4675 else 4676 file->f_pos = ret = 0; 4677 4678 return ret; 4679 } 4680 4681 static const struct file_operations tracing_fops = { 4682 .open = tracing_open, 4683 .read = seq_read, 4684 .write = tracing_write_stub, 4685 .llseek = tracing_lseek, 4686 .release = tracing_release, 4687 }; 4688 4689 static const struct file_operations show_traces_fops = { 4690 .open = show_traces_open, 4691 .read = seq_read, 4692 .llseek = seq_lseek, 4693 .release = show_traces_release, 4694 }; 4695 4696 static ssize_t 4697 tracing_cpumask_read(struct file *filp, char __user *ubuf, 4698 size_t count, loff_t *ppos) 4699 { 4700 struct trace_array *tr = file_inode(filp)->i_private; 4701 char *mask_str; 4702 int len; 4703 4704 len = snprintf(NULL, 0, "%*pb\n", 4705 cpumask_pr_args(tr->tracing_cpumask)) + 1; 4706 mask_str = kmalloc(len, GFP_KERNEL); 4707 if (!mask_str) 4708 return -ENOMEM; 4709 4710 len = snprintf(mask_str, len, "%*pb\n", 4711 cpumask_pr_args(tr->tracing_cpumask)); 4712 if (len >= count) { 4713 count = -EINVAL; 4714 goto out_err; 4715 } 4716 count = simple_read_from_buffer(ubuf, count, ppos, mask_str, len); 4717 4718 out_err: 4719 kfree(mask_str); 4720 4721 return count; 4722 } 4723 4724 int tracing_set_cpumask(struct trace_array *tr, 4725 cpumask_var_t tracing_cpumask_new) 4726 { 4727 int cpu; 4728 4729 if (!tr) 4730 return -EINVAL; 4731 4732 local_irq_disable(); 4733 arch_spin_lock(&tr->max_lock); 4734 for_each_tracing_cpu(cpu) { 4735 /* 4736 * Increase/decrease the disabled counter if we are 4737 * about to flip a bit in the cpumask: 4738 */ 4739 if (cpumask_test_cpu(cpu, tr->tracing_cpumask) && 4740 !cpumask_test_cpu(cpu, tracing_cpumask_new)) { 4741 atomic_inc(&per_cpu_ptr(tr->array_buffer.data, cpu)->disabled); 4742 ring_buffer_record_disable_cpu(tr->array_buffer.buffer, cpu); 4743 } 4744 if (!cpumask_test_cpu(cpu, tr->tracing_cpumask) && 4745 cpumask_test_cpu(cpu, tracing_cpumask_new)) { 4746 atomic_dec(&per_cpu_ptr(tr->array_buffer.data, cpu)->disabled); 4747 ring_buffer_record_enable_cpu(tr->array_buffer.buffer, cpu); 4748 } 4749 } 4750 arch_spin_unlock(&tr->max_lock); 4751 local_irq_enable(); 4752 4753 cpumask_copy(tr->tracing_cpumask, tracing_cpumask_new); 4754 4755 return 0; 4756 } 4757 4758 static ssize_t 4759 tracing_cpumask_write(struct file *filp, const char __user *ubuf, 4760 size_t count, loff_t *ppos) 4761 { 4762 struct trace_array *tr = file_inode(filp)->i_private; 4763 cpumask_var_t tracing_cpumask_new; 4764 int err; 4765 4766 if (!alloc_cpumask_var(&tracing_cpumask_new, GFP_KERNEL)) 4767 return -ENOMEM; 4768 4769 err = cpumask_parse_user(ubuf, count, tracing_cpumask_new); 4770 if (err) 4771 goto err_free; 4772 4773 err = tracing_set_cpumask(tr, tracing_cpumask_new); 4774 if (err) 4775 goto err_free; 4776 4777 free_cpumask_var(tracing_cpumask_new); 4778 4779 return count; 4780 4781 err_free: 4782 free_cpumask_var(tracing_cpumask_new); 4783 4784 return err; 4785 } 4786 4787 static const struct file_operations tracing_cpumask_fops = { 4788 .open = tracing_open_generic_tr, 4789 .read = tracing_cpumask_read, 4790 .write = tracing_cpumask_write, 4791 .release = tracing_release_generic_tr, 4792 .llseek = generic_file_llseek, 4793 }; 4794 4795 static int tracing_trace_options_show(struct seq_file *m, void *v) 4796 { 4797 struct tracer_opt *trace_opts; 4798 struct trace_array *tr = m->private; 4799 u32 tracer_flags; 4800 int i; 4801 4802 mutex_lock(&trace_types_lock); 4803 tracer_flags = tr->current_trace->flags->val; 4804 trace_opts = tr->current_trace->flags->opts; 4805 4806 for (i = 0; trace_options[i]; i++) { 4807 if (tr->trace_flags & (1 << i)) 4808 seq_printf(m, "%s\n", trace_options[i]); 4809 else 4810 seq_printf(m, "no%s\n", trace_options[i]); 4811 } 4812 4813 for (i = 0; trace_opts[i].name; i++) { 4814 if (tracer_flags & trace_opts[i].bit) 4815 seq_printf(m, "%s\n", trace_opts[i].name); 4816 else 4817 seq_printf(m, "no%s\n", trace_opts[i].name); 4818 } 4819 mutex_unlock(&trace_types_lock); 4820 4821 return 0; 4822 } 4823 4824 static int __set_tracer_option(struct trace_array *tr, 4825 struct tracer_flags *tracer_flags, 4826 struct tracer_opt *opts, int neg) 4827 { 4828 struct tracer *trace = tracer_flags->trace; 4829 int ret; 4830 4831 ret = trace->set_flag(tr, tracer_flags->val, opts->bit, !neg); 4832 if (ret) 4833 return ret; 4834 4835 if (neg) 4836 tracer_flags->val &= ~opts->bit; 4837 else 4838 tracer_flags->val |= opts->bit; 4839 return 0; 4840 } 4841 4842 /* Try to assign a tracer specific option */ 4843 static int set_tracer_option(struct trace_array *tr, char *cmp, int neg) 4844 { 4845 struct tracer *trace = tr->current_trace; 4846 struct tracer_flags *tracer_flags = trace->flags; 4847 struct tracer_opt *opts = NULL; 4848 int i; 4849 4850 for (i = 0; tracer_flags->opts[i].name; i++) { 4851 opts = &tracer_flags->opts[i]; 4852 4853 if (strcmp(cmp, opts->name) == 0) 4854 return __set_tracer_option(tr, trace->flags, opts, neg); 4855 } 4856 4857 return -EINVAL; 4858 } 4859 4860 /* Some tracers require overwrite to stay enabled */ 4861 int trace_keep_overwrite(struct tracer *tracer, u32 mask, int set) 4862 { 4863 if (tracer->enabled && (mask & TRACE_ITER_OVERWRITE) && !set) 4864 return -1; 4865 4866 return 0; 4867 } 4868 4869 int set_tracer_flag(struct trace_array *tr, unsigned int mask, int enabled) 4870 { 4871 if ((mask == TRACE_ITER_RECORD_TGID) || 4872 (mask == TRACE_ITER_RECORD_CMD)) 4873 lockdep_assert_held(&event_mutex); 4874 4875 /* do nothing if flag is already set */ 4876 if (!!(tr->trace_flags & mask) == !!enabled) 4877 return 0; 4878 4879 /* Give the tracer a chance to approve the change */ 4880 if (tr->current_trace->flag_changed) 4881 if (tr->current_trace->flag_changed(tr, mask, !!enabled)) 4882 return -EINVAL; 4883 4884 if (enabled) 4885 tr->trace_flags |= mask; 4886 else 4887 tr->trace_flags &= ~mask; 4888 4889 if (mask == TRACE_ITER_RECORD_CMD) 4890 trace_event_enable_cmd_record(enabled); 4891 4892 if (mask == TRACE_ITER_RECORD_TGID) { 4893 if (!tgid_map) 4894 tgid_map = kvcalloc(PID_MAX_DEFAULT + 1, 4895 sizeof(*tgid_map), 4896 GFP_KERNEL); 4897 if (!tgid_map) { 4898 tr->trace_flags &= ~TRACE_ITER_RECORD_TGID; 4899 return -ENOMEM; 4900 } 4901 4902 trace_event_enable_tgid_record(enabled); 4903 } 4904 4905 if (mask == TRACE_ITER_EVENT_FORK) 4906 trace_event_follow_fork(tr, enabled); 4907 4908 if (mask == TRACE_ITER_FUNC_FORK) 4909 ftrace_pid_follow_fork(tr, enabled); 4910 4911 if (mask == TRACE_ITER_OVERWRITE) { 4912 ring_buffer_change_overwrite(tr->array_buffer.buffer, enabled); 4913 #ifdef CONFIG_TRACER_MAX_TRACE 4914 ring_buffer_change_overwrite(tr->max_buffer.buffer, enabled); 4915 #endif 4916 } 4917 4918 if (mask == TRACE_ITER_PRINTK) { 4919 trace_printk_start_stop_comm(enabled); 4920 trace_printk_control(enabled); 4921 } 4922 4923 return 0; 4924 } 4925 4926 int trace_set_options(struct trace_array *tr, char *option) 4927 { 4928 char *cmp; 4929 int neg = 0; 4930 int ret; 4931 size_t orig_len = strlen(option); 4932 int len; 4933 4934 cmp = strstrip(option); 4935 4936 len = str_has_prefix(cmp, "no"); 4937 if (len) 4938 neg = 1; 4939 4940 cmp += len; 4941 4942 mutex_lock(&event_mutex); 4943 mutex_lock(&trace_types_lock); 4944 4945 ret = match_string(trace_options, -1, cmp); 4946 /* If no option could be set, test the specific tracer options */ 4947 if (ret < 0) 4948 ret = set_tracer_option(tr, cmp, neg); 4949 else 4950 ret = set_tracer_flag(tr, 1 << ret, !neg); 4951 4952 mutex_unlock(&trace_types_lock); 4953 mutex_unlock(&event_mutex); 4954 4955 /* 4956 * If the first trailing whitespace is replaced with '\0' by strstrip, 4957 * turn it back into a space. 4958 */ 4959 if (orig_len > strlen(option)) 4960 option[strlen(option)] = ' '; 4961 4962 return ret; 4963 } 4964 4965 static void __init apply_trace_boot_options(void) 4966 { 4967 char *buf = trace_boot_options_buf; 4968 char *option; 4969 4970 while (true) { 4971 option = strsep(&buf, ","); 4972 4973 if (!option) 4974 break; 4975 4976 if (*option) 4977 trace_set_options(&global_trace, option); 4978 4979 /* Put back the comma to allow this to be called again */ 4980 if (buf) 4981 *(buf - 1) = ','; 4982 } 4983 } 4984 4985 static ssize_t 4986 tracing_trace_options_write(struct file *filp, const char __user *ubuf, 4987 size_t cnt, loff_t *ppos) 4988 { 4989 struct seq_file *m = filp->private_data; 4990 struct trace_array *tr = m->private; 4991 char buf[64]; 4992 int ret; 4993 4994 if (cnt >= sizeof(buf)) 4995 return -EINVAL; 4996 4997 if (copy_from_user(buf, ubuf, cnt)) 4998 return -EFAULT; 4999 5000 buf[cnt] = 0; 5001 5002 ret = trace_set_options(tr, buf); 5003 if (ret < 0) 5004 return ret; 5005 5006 *ppos += cnt; 5007 5008 return cnt; 5009 } 5010 5011 static int tracing_trace_options_open(struct inode *inode, struct file *file) 5012 { 5013 struct trace_array *tr = inode->i_private; 5014 int ret; 5015 5016 ret = tracing_check_open_get_tr(tr); 5017 if (ret) 5018 return ret; 5019 5020 ret = single_open(file, tracing_trace_options_show, inode->i_private); 5021 if (ret < 0) 5022 trace_array_put(tr); 5023 5024 return ret; 5025 } 5026 5027 static const struct file_operations tracing_iter_fops = { 5028 .open = tracing_trace_options_open, 5029 .read = seq_read, 5030 .llseek = seq_lseek, 5031 .release = tracing_single_release_tr, 5032 .write = tracing_trace_options_write, 5033 }; 5034 5035 static const char readme_msg[] = 5036 "tracing mini-HOWTO:\n\n" 5037 "# echo 0 > tracing_on : quick way to disable tracing\n" 5038 "# echo 1 > tracing_on : quick way to re-enable tracing\n\n" 5039 " Important files:\n" 5040 " trace\t\t\t- The static contents of the buffer\n" 5041 "\t\t\t To clear the buffer write into this file: echo > trace\n" 5042 " trace_pipe\t\t- A consuming read to see the contents of the buffer\n" 5043 " current_tracer\t- function and latency tracers\n" 5044 " available_tracers\t- list of configured tracers for current_tracer\n" 5045 " error_log\t- error log for failed commands (that support it)\n" 5046 " buffer_size_kb\t- view and modify size of per cpu buffer\n" 5047 " buffer_total_size_kb - view total size of all cpu buffers\n\n" 5048 " trace_clock\t\t-change the clock used to order events\n" 5049 " local: Per cpu clock but may not be synced across CPUs\n" 5050 " global: Synced across CPUs but slows tracing down.\n" 5051 " counter: Not a clock, but just an increment\n" 5052 " uptime: Jiffy counter from time of boot\n" 5053 " perf: Same clock that perf events use\n" 5054 #ifdef CONFIG_X86_64 5055 " x86-tsc: TSC cycle counter\n" 5056 #endif 5057 "\n timestamp_mode\t-view the mode used to timestamp events\n" 5058 " delta: Delta difference against a buffer-wide timestamp\n" 5059 " absolute: Absolute (standalone) timestamp\n" 5060 "\n trace_marker\t\t- Writes into this file writes into the kernel buffer\n" 5061 "\n trace_marker_raw\t\t- Writes into this file writes binary data into the kernel buffer\n" 5062 " tracing_cpumask\t- Limit which CPUs to trace\n" 5063 " instances\t\t- Make sub-buffers with: mkdir instances/foo\n" 5064 "\t\t\t Remove sub-buffer with rmdir\n" 5065 " trace_options\t\t- Set format or modify how tracing happens\n" 5066 "\t\t\t Disable an option by prefixing 'no' to the\n" 5067 "\t\t\t option name\n" 5068 " saved_cmdlines_size\t- echo command number in here to store comm-pid list\n" 5069 #ifdef CONFIG_DYNAMIC_FTRACE 5070 "\n available_filter_functions - list of functions that can be filtered on\n" 5071 " set_ftrace_filter\t- echo function name in here to only trace these\n" 5072 "\t\t\t functions\n" 5073 "\t accepts: func_full_name or glob-matching-pattern\n" 5074 "\t modules: Can select a group via module\n" 5075 "\t Format: :mod:<module-name>\n" 5076 "\t example: echo :mod:ext3 > set_ftrace_filter\n" 5077 "\t triggers: a command to perform when function is hit\n" 5078 "\t Format: <function>:<trigger>[:count]\n" 5079 "\t trigger: traceon, traceoff\n" 5080 "\t\t enable_event:<system>:<event>\n" 5081 "\t\t disable_event:<system>:<event>\n" 5082 #ifdef CONFIG_STACKTRACE 5083 "\t\t stacktrace\n" 5084 #endif 5085 #ifdef CONFIG_TRACER_SNAPSHOT 5086 "\t\t snapshot\n" 5087 #endif 5088 "\t\t dump\n" 5089 "\t\t cpudump\n" 5090 "\t example: echo do_fault:traceoff > set_ftrace_filter\n" 5091 "\t echo do_trap:traceoff:3 > set_ftrace_filter\n" 5092 "\t The first one will disable tracing every time do_fault is hit\n" 5093 "\t The second will disable tracing at most 3 times when do_trap is hit\n" 5094 "\t The first time do trap is hit and it disables tracing, the\n" 5095 "\t counter will decrement to 2. If tracing is already disabled,\n" 5096 "\t the counter will not decrement. It only decrements when the\n" 5097 "\t trigger did work\n" 5098 "\t To remove trigger without count:\n" 5099 "\t echo '!<function>:<trigger> > set_ftrace_filter\n" 5100 "\t To remove trigger with a count:\n" 5101 "\t echo '!<function>:<trigger>:0 > set_ftrace_filter\n" 5102 " set_ftrace_notrace\t- echo function name in here to never trace.\n" 5103 "\t accepts: func_full_name, *func_end, func_begin*, *func_middle*\n" 5104 "\t modules: Can select a group via module command :mod:\n" 5105 "\t Does not accept triggers\n" 5106 #endif /* CONFIG_DYNAMIC_FTRACE */ 5107 #ifdef CONFIG_FUNCTION_TRACER 5108 " set_ftrace_pid\t- Write pid(s) to only function trace those pids\n" 5109 "\t\t (function)\n" 5110 " set_ftrace_notrace_pid\t- Write pid(s) to not function trace those pids\n" 5111 "\t\t (function)\n" 5112 #endif 5113 #ifdef CONFIG_FUNCTION_GRAPH_TRACER 5114 " set_graph_function\t- Trace the nested calls of a function (function_graph)\n" 5115 " set_graph_notrace\t- Do not trace the nested calls of a function (function_graph)\n" 5116 " max_graph_depth\t- Trace a limited depth of nested calls (0 is unlimited)\n" 5117 #endif 5118 #ifdef CONFIG_TRACER_SNAPSHOT 5119 "\n snapshot\t\t- Like 'trace' but shows the content of the static\n" 5120 "\t\t\t snapshot buffer. Read the contents for more\n" 5121 "\t\t\t information\n" 5122 #endif 5123 #ifdef CONFIG_STACK_TRACER 5124 " stack_trace\t\t- Shows the max stack trace when active\n" 5125 " stack_max_size\t- Shows current max stack size that was traced\n" 5126 "\t\t\t Write into this file to reset the max size (trigger a\n" 5127 "\t\t\t new trace)\n" 5128 #ifdef CONFIG_DYNAMIC_FTRACE 5129 " stack_trace_filter\t- Like set_ftrace_filter but limits what stack_trace\n" 5130 "\t\t\t traces\n" 5131 #endif 5132 #endif /* CONFIG_STACK_TRACER */ 5133 #ifdef CONFIG_DYNAMIC_EVENTS 5134 " dynamic_events\t\t- Create/append/remove/show the generic dynamic events\n" 5135 "\t\t\t Write into this file to define/undefine new trace events.\n" 5136 #endif 5137 #ifdef CONFIG_KPROBE_EVENTS 5138 " kprobe_events\t\t- Create/append/remove/show the kernel dynamic events\n" 5139 "\t\t\t Write into this file to define/undefine new trace events.\n" 5140 #endif 5141 #ifdef CONFIG_UPROBE_EVENTS 5142 " uprobe_events\t\t- Create/append/remove/show the userspace dynamic events\n" 5143 "\t\t\t Write into this file to define/undefine new trace events.\n" 5144 #endif 5145 #if defined(CONFIG_KPROBE_EVENTS) || defined(CONFIG_UPROBE_EVENTS) 5146 "\t accepts: event-definitions (one definition per line)\n" 5147 "\t Format: p[:[<group>/]<event>] <place> [<args>]\n" 5148 "\t r[maxactive][:[<group>/]<event>] <place> [<args>]\n" 5149 #ifdef CONFIG_HIST_TRIGGERS 5150 "\t s:[synthetic/]<event> <field> [<field>]\n" 5151 #endif 5152 "\t -:[<group>/]<event>\n" 5153 #ifdef CONFIG_KPROBE_EVENTS 5154 "\t place: [<module>:]<symbol>[+<offset>]|<memaddr>\n" 5155 "place (kretprobe): [<module>:]<symbol>[+<offset>]%return|<memaddr>\n" 5156 #endif 5157 #ifdef CONFIG_UPROBE_EVENTS 5158 " place (uprobe): <path>:<offset>[%return][(ref_ctr_offset)]\n" 5159 #endif 5160 "\t args: <name>=fetcharg[:type]\n" 5161 "\t fetcharg: %<register>, @<address>, @<symbol>[+|-<offset>],\n" 5162 #ifdef CONFIG_HAVE_FUNCTION_ARG_ACCESS_API 5163 "\t $stack<index>, $stack, $retval, $comm, $arg<N>,\n" 5164 #else 5165 "\t $stack<index>, $stack, $retval, $comm,\n" 5166 #endif 5167 "\t +|-[u]<offset>(<fetcharg>), \\imm-value, \\\"imm-string\"\n" 5168 "\t type: s8/16/32/64, u8/16/32/64, x8/16/32/64, string, symbol,\n" 5169 "\t b<bit-width>@<bit-offset>/<container-size>, ustring,\n" 5170 "\t <type>\\[<array-size>\\]\n" 5171 #ifdef CONFIG_HIST_TRIGGERS 5172 "\t field: <stype> <name>;\n" 5173 "\t stype: u8/u16/u32/u64, s8/s16/s32/s64, pid_t,\n" 5174 "\t [unsigned] char/int/long\n" 5175 #endif 5176 #endif 5177 " events/\t\t- Directory containing all trace event subsystems:\n" 5178 " enable\t\t- Write 0/1 to enable/disable tracing of all events\n" 5179 " events/<system>/\t- Directory containing all trace events for <system>:\n" 5180 " enable\t\t- Write 0/1 to enable/disable tracing of all <system>\n" 5181 "\t\t\t events\n" 5182 " filter\t\t- If set, only events passing filter are traced\n" 5183 " events/<system>/<event>/\t- Directory containing control files for\n" 5184 "\t\t\t <event>:\n" 5185 " enable\t\t- Write 0/1 to enable/disable tracing of <event>\n" 5186 " filter\t\t- If set, only events passing filter are traced\n" 5187 " trigger\t\t- If set, a command to perform when event is hit\n" 5188 "\t Format: <trigger>[:count][if <filter>]\n" 5189 "\t trigger: traceon, traceoff\n" 5190 "\t enable_event:<system>:<event>\n" 5191 "\t disable_event:<system>:<event>\n" 5192 #ifdef CONFIG_HIST_TRIGGERS 5193 "\t enable_hist:<system>:<event>\n" 5194 "\t disable_hist:<system>:<event>\n" 5195 #endif 5196 #ifdef CONFIG_STACKTRACE 5197 "\t\t stacktrace\n" 5198 #endif 5199 #ifdef CONFIG_TRACER_SNAPSHOT 5200 "\t\t snapshot\n" 5201 #endif 5202 #ifdef CONFIG_HIST_TRIGGERS 5203 "\t\t hist (see below)\n" 5204 #endif 5205 "\t example: echo traceoff > events/block/block_unplug/trigger\n" 5206 "\t echo traceoff:3 > events/block/block_unplug/trigger\n" 5207 "\t echo 'enable_event:kmem:kmalloc:3 if nr_rq > 1' > \\\n" 5208 "\t events/block/block_unplug/trigger\n" 5209 "\t The first disables tracing every time block_unplug is hit.\n" 5210 "\t The second disables tracing the first 3 times block_unplug is hit.\n" 5211 "\t The third enables the kmalloc event the first 3 times block_unplug\n" 5212 "\t is hit and has value of greater than 1 for the 'nr_rq' event field.\n" 5213 "\t Like function triggers, the counter is only decremented if it\n" 5214 "\t enabled or disabled tracing.\n" 5215 "\t To remove a trigger without a count:\n" 5216 "\t echo '!<trigger> > <system>/<event>/trigger\n" 5217 "\t To remove a trigger with a count:\n" 5218 "\t echo '!<trigger>:0 > <system>/<event>/trigger\n" 5219 "\t Filters can be ignored when removing a trigger.\n" 5220 #ifdef CONFIG_HIST_TRIGGERS 5221 " hist trigger\t- If set, event hits are aggregated into a hash table\n" 5222 "\t Format: hist:keys=<field1[,field2,...]>\n" 5223 "\t [:values=<field1[,field2,...]>]\n" 5224 "\t [:sort=<field1[,field2,...]>]\n" 5225 "\t [:size=#entries]\n" 5226 "\t [:pause][:continue][:clear]\n" 5227 "\t [:name=histname1]\n" 5228 "\t [:<handler>.<action>]\n" 5229 "\t [if <filter>]\n\n" 5230 "\t When a matching event is hit, an entry is added to a hash\n" 5231 "\t table using the key(s) and value(s) named, and the value of a\n" 5232 "\t sum called 'hitcount' is incremented. Keys and values\n" 5233 "\t correspond to fields in the event's format description. Keys\n" 5234 "\t can be any field, or the special string 'stacktrace'.\n" 5235 "\t Compound keys consisting of up to two fields can be specified\n" 5236 "\t by the 'keys' keyword. Values must correspond to numeric\n" 5237 "\t fields. Sort keys consisting of up to two fields can be\n" 5238 "\t specified using the 'sort' keyword. The sort direction can\n" 5239 "\t be modified by appending '.descending' or '.ascending' to a\n" 5240 "\t sort field. The 'size' parameter can be used to specify more\n" 5241 "\t or fewer than the default 2048 entries for the hashtable size.\n" 5242 "\t If a hist trigger is given a name using the 'name' parameter,\n" 5243 "\t its histogram data will be shared with other triggers of the\n" 5244 "\t same name, and trigger hits will update this common data.\n\n" 5245 "\t Reading the 'hist' file for the event will dump the hash\n" 5246 "\t table in its entirety to stdout. If there are multiple hist\n" 5247 "\t triggers attached to an event, there will be a table for each\n" 5248 "\t trigger in the output. The table displayed for a named\n" 5249 "\t trigger will be the same as any other instance having the\n" 5250 "\t same name. The default format used to display a given field\n" 5251 "\t can be modified by appending any of the following modifiers\n" 5252 "\t to the field name, as applicable:\n\n" 5253 "\t .hex display a number as a hex value\n" 5254 "\t .sym display an address as a symbol\n" 5255 "\t .sym-offset display an address as a symbol and offset\n" 5256 "\t .execname display a common_pid as a program name\n" 5257 "\t .syscall display a syscall id as a syscall name\n" 5258 "\t .log2 display log2 value rather than raw number\n" 5259 "\t .usecs display a common_timestamp in microseconds\n\n" 5260 "\t The 'pause' parameter can be used to pause an existing hist\n" 5261 "\t trigger or to start a hist trigger but not log any events\n" 5262 "\t until told to do so. 'continue' can be used to start or\n" 5263 "\t restart a paused hist trigger.\n\n" 5264 "\t The 'clear' parameter will clear the contents of a running\n" 5265 "\t hist trigger and leave its current paused/active state\n" 5266 "\t unchanged.\n\n" 5267 "\t The enable_hist and disable_hist triggers can be used to\n" 5268 "\t have one event conditionally start and stop another event's\n" 5269 "\t already-attached hist trigger. The syntax is analogous to\n" 5270 "\t the enable_event and disable_event triggers.\n\n" 5271 "\t Hist trigger handlers and actions are executed whenever a\n" 5272 "\t a histogram entry is added or updated. They take the form:\n\n" 5273 "\t <handler>.<action>\n\n" 5274 "\t The available handlers are:\n\n" 5275 "\t onmatch(matching.event) - invoke on addition or update\n" 5276 "\t onmax(var) - invoke if var exceeds current max\n" 5277 "\t onchange(var) - invoke action if var changes\n\n" 5278 "\t The available actions are:\n\n" 5279 "\t trace(<synthetic_event>,param list) - generate synthetic event\n" 5280 "\t save(field,...) - save current event fields\n" 5281 #ifdef CONFIG_TRACER_SNAPSHOT 5282 "\t snapshot() - snapshot the trace buffer\n\n" 5283 #endif 5284 #ifdef CONFIG_SYNTH_EVENTS 5285 " events/synthetic_events\t- Create/append/remove/show synthetic events\n" 5286 "\t Write into this file to define/undefine new synthetic events.\n" 5287 "\t example: echo 'myevent u64 lat; char name[]' >> synthetic_events\n" 5288 #endif 5289 #endif 5290 ; 5291 5292 static ssize_t 5293 tracing_readme_read(struct file *filp, char __user *ubuf, 5294 size_t cnt, loff_t *ppos) 5295 { 5296 return simple_read_from_buffer(ubuf, cnt, ppos, 5297 readme_msg, strlen(readme_msg)); 5298 } 5299 5300 static const struct file_operations tracing_readme_fops = { 5301 .open = tracing_open_generic, 5302 .read = tracing_readme_read, 5303 .llseek = generic_file_llseek, 5304 }; 5305 5306 static void *saved_tgids_next(struct seq_file *m, void *v, loff_t *pos) 5307 { 5308 int *ptr = v; 5309 5310 if (*pos || m->count) 5311 ptr++; 5312 5313 (*pos)++; 5314 5315 for (; ptr <= &tgid_map[PID_MAX_DEFAULT]; ptr++) { 5316 if (trace_find_tgid(*ptr)) 5317 return ptr; 5318 } 5319 5320 return NULL; 5321 } 5322 5323 static void *saved_tgids_start(struct seq_file *m, loff_t *pos) 5324 { 5325 void *v; 5326 loff_t l = 0; 5327 5328 if (!tgid_map) 5329 return NULL; 5330 5331 v = &tgid_map[0]; 5332 while (l <= *pos) { 5333 v = saved_tgids_next(m, v, &l); 5334 if (!v) 5335 return NULL; 5336 } 5337 5338 return v; 5339 } 5340 5341 static void saved_tgids_stop(struct seq_file *m, void *v) 5342 { 5343 } 5344 5345 static int saved_tgids_show(struct seq_file *m, void *v) 5346 { 5347 int pid = (int *)v - tgid_map; 5348 5349 seq_printf(m, "%d %d\n", pid, trace_find_tgid(pid)); 5350 return 0; 5351 } 5352 5353 static const struct seq_operations tracing_saved_tgids_seq_ops = { 5354 .start = saved_tgids_start, 5355 .stop = saved_tgids_stop, 5356 .next = saved_tgids_next, 5357 .show = saved_tgids_show, 5358 }; 5359 5360 static int tracing_saved_tgids_open(struct inode *inode, struct file *filp) 5361 { 5362 int ret; 5363 5364 ret = tracing_check_open_get_tr(NULL); 5365 if (ret) 5366 return ret; 5367 5368 return seq_open(filp, &tracing_saved_tgids_seq_ops); 5369 } 5370 5371 5372 static const struct file_operations tracing_saved_tgids_fops = { 5373 .open = tracing_saved_tgids_open, 5374 .read = seq_read, 5375 .llseek = seq_lseek, 5376 .release = seq_release, 5377 }; 5378 5379 static void *saved_cmdlines_next(struct seq_file *m, void *v, loff_t *pos) 5380 { 5381 unsigned int *ptr = v; 5382 5383 if (*pos || m->count) 5384 ptr++; 5385 5386 (*pos)++; 5387 5388 for (; ptr < &savedcmd->map_cmdline_to_pid[savedcmd->cmdline_num]; 5389 ptr++) { 5390 if (*ptr == -1 || *ptr == NO_CMDLINE_MAP) 5391 continue; 5392 5393 return ptr; 5394 } 5395 5396 return NULL; 5397 } 5398 5399 static void *saved_cmdlines_start(struct seq_file *m, loff_t *pos) 5400 { 5401 void *v; 5402 loff_t l = 0; 5403 5404 preempt_disable(); 5405 arch_spin_lock(&trace_cmdline_lock); 5406 5407 v = &savedcmd->map_cmdline_to_pid[0]; 5408 while (l <= *pos) { 5409 v = saved_cmdlines_next(m, v, &l); 5410 if (!v) 5411 return NULL; 5412 } 5413 5414 return v; 5415 } 5416 5417 static void saved_cmdlines_stop(struct seq_file *m, void *v) 5418 { 5419 arch_spin_unlock(&trace_cmdline_lock); 5420 preempt_enable(); 5421 } 5422 5423 static int saved_cmdlines_show(struct seq_file *m, void *v) 5424 { 5425 char buf[TASK_COMM_LEN]; 5426 unsigned int *pid = v; 5427 5428 __trace_find_cmdline(*pid, buf); 5429 seq_printf(m, "%d %s\n", *pid, buf); 5430 return 0; 5431 } 5432 5433 static const struct seq_operations tracing_saved_cmdlines_seq_ops = { 5434 .start = saved_cmdlines_start, 5435 .next = saved_cmdlines_next, 5436 .stop = saved_cmdlines_stop, 5437 .show = saved_cmdlines_show, 5438 }; 5439 5440 static int tracing_saved_cmdlines_open(struct inode *inode, struct file *filp) 5441 { 5442 int ret; 5443 5444 ret = tracing_check_open_get_tr(NULL); 5445 if (ret) 5446 return ret; 5447 5448 return seq_open(filp, &tracing_saved_cmdlines_seq_ops); 5449 } 5450 5451 static const struct file_operations tracing_saved_cmdlines_fops = { 5452 .open = tracing_saved_cmdlines_open, 5453 .read = seq_read, 5454 .llseek = seq_lseek, 5455 .release = seq_release, 5456 }; 5457 5458 static ssize_t 5459 tracing_saved_cmdlines_size_read(struct file *filp, char __user *ubuf, 5460 size_t cnt, loff_t *ppos) 5461 { 5462 char buf[64]; 5463 int r; 5464 5465 arch_spin_lock(&trace_cmdline_lock); 5466 r = scnprintf(buf, sizeof(buf), "%u\n", savedcmd->cmdline_num); 5467 arch_spin_unlock(&trace_cmdline_lock); 5468 5469 return simple_read_from_buffer(ubuf, cnt, ppos, buf, r); 5470 } 5471 5472 static void free_saved_cmdlines_buffer(struct saved_cmdlines_buffer *s) 5473 { 5474 kfree(s->saved_cmdlines); 5475 kfree(s->map_cmdline_to_pid); 5476 kfree(s); 5477 } 5478 5479 static int tracing_resize_saved_cmdlines(unsigned int val) 5480 { 5481 struct saved_cmdlines_buffer *s, *savedcmd_temp; 5482 5483 s = kmalloc(sizeof(*s), GFP_KERNEL); 5484 if (!s) 5485 return -ENOMEM; 5486 5487 if (allocate_cmdlines_buffer(val, s) < 0) { 5488 kfree(s); 5489 return -ENOMEM; 5490 } 5491 5492 arch_spin_lock(&trace_cmdline_lock); 5493 savedcmd_temp = savedcmd; 5494 savedcmd = s; 5495 arch_spin_unlock(&trace_cmdline_lock); 5496 free_saved_cmdlines_buffer(savedcmd_temp); 5497 5498 return 0; 5499 } 5500 5501 static ssize_t 5502 tracing_saved_cmdlines_size_write(struct file *filp, const char __user *ubuf, 5503 size_t cnt, loff_t *ppos) 5504 { 5505 unsigned long val; 5506 int ret; 5507 5508 ret = kstrtoul_from_user(ubuf, cnt, 10, &val); 5509 if (ret) 5510 return ret; 5511 5512 /* must have at least 1 entry or less than PID_MAX_DEFAULT */ 5513 if (!val || val > PID_MAX_DEFAULT) 5514 return -EINVAL; 5515 5516 ret = tracing_resize_saved_cmdlines((unsigned int)val); 5517 if (ret < 0) 5518 return ret; 5519 5520 *ppos += cnt; 5521 5522 return cnt; 5523 } 5524 5525 static const struct file_operations tracing_saved_cmdlines_size_fops = { 5526 .open = tracing_open_generic, 5527 .read = tracing_saved_cmdlines_size_read, 5528 .write = tracing_saved_cmdlines_size_write, 5529 }; 5530 5531 #ifdef CONFIG_TRACE_EVAL_MAP_FILE 5532 static union trace_eval_map_item * 5533 update_eval_map(union trace_eval_map_item *ptr) 5534 { 5535 if (!ptr->map.eval_string) { 5536 if (ptr->tail.next) { 5537 ptr = ptr->tail.next; 5538 /* Set ptr to the next real item (skip head) */ 5539 ptr++; 5540 } else 5541 return NULL; 5542 } 5543 return ptr; 5544 } 5545 5546 static void *eval_map_next(struct seq_file *m, void *v, loff_t *pos) 5547 { 5548 union trace_eval_map_item *ptr = v; 5549 5550 /* 5551 * Paranoid! If ptr points to end, we don't want to increment past it. 5552 * This really should never happen. 5553 */ 5554 (*pos)++; 5555 ptr = update_eval_map(ptr); 5556 if (WARN_ON_ONCE(!ptr)) 5557 return NULL; 5558 5559 ptr++; 5560 ptr = update_eval_map(ptr); 5561 5562 return ptr; 5563 } 5564 5565 static void *eval_map_start(struct seq_file *m, loff_t *pos) 5566 { 5567 union trace_eval_map_item *v; 5568 loff_t l = 0; 5569 5570 mutex_lock(&trace_eval_mutex); 5571 5572 v = trace_eval_maps; 5573 if (v) 5574 v++; 5575 5576 while (v && l < *pos) { 5577 v = eval_map_next(m, v, &l); 5578 } 5579 5580 return v; 5581 } 5582 5583 static void eval_map_stop(struct seq_file *m, void *v) 5584 { 5585 mutex_unlock(&trace_eval_mutex); 5586 } 5587 5588 static int eval_map_show(struct seq_file *m, void *v) 5589 { 5590 union trace_eval_map_item *ptr = v; 5591 5592 seq_printf(m, "%s %ld (%s)\n", 5593 ptr->map.eval_string, ptr->map.eval_value, 5594 ptr->map.system); 5595 5596 return 0; 5597 } 5598 5599 static const struct seq_operations tracing_eval_map_seq_ops = { 5600 .start = eval_map_start, 5601 .next = eval_map_next, 5602 .stop = eval_map_stop, 5603 .show = eval_map_show, 5604 }; 5605 5606 static int tracing_eval_map_open(struct inode *inode, struct file *filp) 5607 { 5608 int ret; 5609 5610 ret = tracing_check_open_get_tr(NULL); 5611 if (ret) 5612 return ret; 5613 5614 return seq_open(filp, &tracing_eval_map_seq_ops); 5615 } 5616 5617 static const struct file_operations tracing_eval_map_fops = { 5618 .open = tracing_eval_map_open, 5619 .read = seq_read, 5620 .llseek = seq_lseek, 5621 .release = seq_release, 5622 }; 5623 5624 static inline union trace_eval_map_item * 5625 trace_eval_jmp_to_tail(union trace_eval_map_item *ptr) 5626 { 5627 /* Return tail of array given the head */ 5628 return ptr + ptr->head.length + 1; 5629 } 5630 5631 static void 5632 trace_insert_eval_map_file(struct module *mod, struct trace_eval_map **start, 5633 int len) 5634 { 5635 struct trace_eval_map **stop; 5636 struct trace_eval_map **map; 5637 union trace_eval_map_item *map_array; 5638 union trace_eval_map_item *ptr; 5639 5640 stop = start + len; 5641 5642 /* 5643 * The trace_eval_maps contains the map plus a head and tail item, 5644 * where the head holds the module and length of array, and the 5645 * tail holds a pointer to the next list. 5646 */ 5647 map_array = kmalloc_array(len + 2, sizeof(*map_array), GFP_KERNEL); 5648 if (!map_array) { 5649 pr_warn("Unable to allocate trace eval mapping\n"); 5650 return; 5651 } 5652 5653 mutex_lock(&trace_eval_mutex); 5654 5655 if (!trace_eval_maps) 5656 trace_eval_maps = map_array; 5657 else { 5658 ptr = trace_eval_maps; 5659 for (;;) { 5660 ptr = trace_eval_jmp_to_tail(ptr); 5661 if (!ptr->tail.next) 5662 break; 5663 ptr = ptr->tail.next; 5664 5665 } 5666 ptr->tail.next = map_array; 5667 } 5668 map_array->head.mod = mod; 5669 map_array->head.length = len; 5670 map_array++; 5671 5672 for (map = start; (unsigned long)map < (unsigned long)stop; map++) { 5673 map_array->map = **map; 5674 map_array++; 5675 } 5676 memset(map_array, 0, sizeof(*map_array)); 5677 5678 mutex_unlock(&trace_eval_mutex); 5679 } 5680 5681 static void trace_create_eval_file(struct dentry *d_tracer) 5682 { 5683 trace_create_file("eval_map", 0444, d_tracer, 5684 NULL, &tracing_eval_map_fops); 5685 } 5686 5687 #else /* CONFIG_TRACE_EVAL_MAP_FILE */ 5688 static inline void trace_create_eval_file(struct dentry *d_tracer) { } 5689 static inline void trace_insert_eval_map_file(struct module *mod, 5690 struct trace_eval_map **start, int len) { } 5691 #endif /* !CONFIG_TRACE_EVAL_MAP_FILE */ 5692 5693 static void trace_insert_eval_map(struct module *mod, 5694 struct trace_eval_map **start, int len) 5695 { 5696 struct trace_eval_map **map; 5697 5698 if (len <= 0) 5699 return; 5700 5701 map = start; 5702 5703 trace_event_eval_update(map, len); 5704 5705 trace_insert_eval_map_file(mod, start, len); 5706 } 5707 5708 static ssize_t 5709 tracing_set_trace_read(struct file *filp, char __user *ubuf, 5710 size_t cnt, loff_t *ppos) 5711 { 5712 struct trace_array *tr = filp->private_data; 5713 char buf[MAX_TRACER_SIZE+2]; 5714 int r; 5715 5716 mutex_lock(&trace_types_lock); 5717 r = sprintf(buf, "%s\n", tr->current_trace->name); 5718 mutex_unlock(&trace_types_lock); 5719 5720 return simple_read_from_buffer(ubuf, cnt, ppos, buf, r); 5721 } 5722 5723 int tracer_init(struct tracer *t, struct trace_array *tr) 5724 { 5725 tracing_reset_online_cpus(&tr->array_buffer); 5726 return t->init(tr); 5727 } 5728 5729 static void set_buffer_entries(struct array_buffer *buf, unsigned long val) 5730 { 5731 int cpu; 5732 5733 for_each_tracing_cpu(cpu) 5734 per_cpu_ptr(buf->data, cpu)->entries = val; 5735 } 5736 5737 #ifdef CONFIG_TRACER_MAX_TRACE 5738 /* resize @tr's buffer to the size of @size_tr's entries */ 5739 static int resize_buffer_duplicate_size(struct array_buffer *trace_buf, 5740 struct array_buffer *size_buf, int cpu_id) 5741 { 5742 int cpu, ret = 0; 5743 5744 if (cpu_id == RING_BUFFER_ALL_CPUS) { 5745 for_each_tracing_cpu(cpu) { 5746 ret = ring_buffer_resize(trace_buf->buffer, 5747 per_cpu_ptr(size_buf->data, cpu)->entries, cpu); 5748 if (ret < 0) 5749 break; 5750 per_cpu_ptr(trace_buf->data, cpu)->entries = 5751 per_cpu_ptr(size_buf->data, cpu)->entries; 5752 } 5753 } else { 5754 ret = ring_buffer_resize(trace_buf->buffer, 5755 per_cpu_ptr(size_buf->data, cpu_id)->entries, cpu_id); 5756 if (ret == 0) 5757 per_cpu_ptr(trace_buf->data, cpu_id)->entries = 5758 per_cpu_ptr(size_buf->data, cpu_id)->entries; 5759 } 5760 5761 return ret; 5762 } 5763 #endif /* CONFIG_TRACER_MAX_TRACE */ 5764 5765 static int __tracing_resize_ring_buffer(struct trace_array *tr, 5766 unsigned long size, int cpu) 5767 { 5768 int ret; 5769 5770 /* 5771 * If kernel or user changes the size of the ring buffer 5772 * we use the size that was given, and we can forget about 5773 * expanding it later. 5774 */ 5775 ring_buffer_expanded = true; 5776 5777 /* May be called before buffers are initialized */ 5778 if (!tr->array_buffer.buffer) 5779 return 0; 5780 5781 ret = ring_buffer_resize(tr->array_buffer.buffer, size, cpu); 5782 if (ret < 0) 5783 return ret; 5784 5785 #ifdef CONFIG_TRACER_MAX_TRACE 5786 if (!(tr->flags & TRACE_ARRAY_FL_GLOBAL) || 5787 !tr->current_trace->use_max_tr) 5788 goto out; 5789 5790 ret = ring_buffer_resize(tr->max_buffer.buffer, size, cpu); 5791 if (ret < 0) { 5792 int r = resize_buffer_duplicate_size(&tr->array_buffer, 5793 &tr->array_buffer, cpu); 5794 if (r < 0) { 5795 /* 5796 * AARGH! We are left with different 5797 * size max buffer!!!! 5798 * The max buffer is our "snapshot" buffer. 5799 * When a tracer needs a snapshot (one of the 5800 * latency tracers), it swaps the max buffer 5801 * with the saved snap shot. We succeeded to 5802 * update the size of the main buffer, but failed to 5803 * update the size of the max buffer. But when we tried 5804 * to reset the main buffer to the original size, we 5805 * failed there too. This is very unlikely to 5806 * happen, but if it does, warn and kill all 5807 * tracing. 5808 */ 5809 WARN_ON(1); 5810 tracing_disabled = 1; 5811 } 5812 return ret; 5813 } 5814 5815 if (cpu == RING_BUFFER_ALL_CPUS) 5816 set_buffer_entries(&tr->max_buffer, size); 5817 else 5818 per_cpu_ptr(tr->max_buffer.data, cpu)->entries = size; 5819 5820 out: 5821 #endif /* CONFIG_TRACER_MAX_TRACE */ 5822 5823 if (cpu == RING_BUFFER_ALL_CPUS) 5824 set_buffer_entries(&tr->array_buffer, size); 5825 else 5826 per_cpu_ptr(tr->array_buffer.data, cpu)->entries = size; 5827 5828 return ret; 5829 } 5830 5831 ssize_t tracing_resize_ring_buffer(struct trace_array *tr, 5832 unsigned long size, int cpu_id) 5833 { 5834 int ret = size; 5835 5836 mutex_lock(&trace_types_lock); 5837 5838 if (cpu_id != RING_BUFFER_ALL_CPUS) { 5839 /* make sure, this cpu is enabled in the mask */ 5840 if (!cpumask_test_cpu(cpu_id, tracing_buffer_mask)) { 5841 ret = -EINVAL; 5842 goto out; 5843 } 5844 } 5845 5846 ret = __tracing_resize_ring_buffer(tr, size, cpu_id); 5847 if (ret < 0) 5848 ret = -ENOMEM; 5849 5850 out: 5851 mutex_unlock(&trace_types_lock); 5852 5853 return ret; 5854 } 5855 5856 5857 /** 5858 * tracing_update_buffers - used by tracing facility to expand ring buffers 5859 * 5860 * To save on memory when the tracing is never used on a system with it 5861 * configured in. The ring buffers are set to a minimum size. But once 5862 * a user starts to use the tracing facility, then they need to grow 5863 * to their default size. 5864 * 5865 * This function is to be called when a tracer is about to be used. 5866 */ 5867 int tracing_update_buffers(void) 5868 { 5869 int ret = 0; 5870 5871 mutex_lock(&trace_types_lock); 5872 if (!ring_buffer_expanded) 5873 ret = __tracing_resize_ring_buffer(&global_trace, trace_buf_size, 5874 RING_BUFFER_ALL_CPUS); 5875 mutex_unlock(&trace_types_lock); 5876 5877 return ret; 5878 } 5879 5880 struct trace_option_dentry; 5881 5882 static void 5883 create_trace_option_files(struct trace_array *tr, struct tracer *tracer); 5884 5885 /* 5886 * Used to clear out the tracer before deletion of an instance. 5887 * Must have trace_types_lock held. 5888 */ 5889 static void tracing_set_nop(struct trace_array *tr) 5890 { 5891 if (tr->current_trace == &nop_trace) 5892 return; 5893 5894 tr->current_trace->enabled--; 5895 5896 if (tr->current_trace->reset) 5897 tr->current_trace->reset(tr); 5898 5899 tr->current_trace = &nop_trace; 5900 } 5901 5902 static void add_tracer_options(struct trace_array *tr, struct tracer *t) 5903 { 5904 /* Only enable if the directory has been created already. */ 5905 if (!tr->dir) 5906 return; 5907 5908 create_trace_option_files(tr, t); 5909 } 5910 5911 int tracing_set_tracer(struct trace_array *tr, const char *buf) 5912 { 5913 struct tracer *t; 5914 #ifdef CONFIG_TRACER_MAX_TRACE 5915 bool had_max_tr; 5916 #endif 5917 int ret = 0; 5918 5919 mutex_lock(&trace_types_lock); 5920 5921 if (!ring_buffer_expanded) { 5922 ret = __tracing_resize_ring_buffer(tr, trace_buf_size, 5923 RING_BUFFER_ALL_CPUS); 5924 if (ret < 0) 5925 goto out; 5926 ret = 0; 5927 } 5928 5929 for (t = trace_types; t; t = t->next) { 5930 if (strcmp(t->name, buf) == 0) 5931 break; 5932 } 5933 if (!t) { 5934 ret = -EINVAL; 5935 goto out; 5936 } 5937 if (t == tr->current_trace) 5938 goto out; 5939 5940 #ifdef CONFIG_TRACER_SNAPSHOT 5941 if (t->use_max_tr) { 5942 arch_spin_lock(&tr->max_lock); 5943 if (tr->cond_snapshot) 5944 ret = -EBUSY; 5945 arch_spin_unlock(&tr->max_lock); 5946 if (ret) 5947 goto out; 5948 } 5949 #endif 5950 /* Some tracers won't work on kernel command line */ 5951 if (system_state < SYSTEM_RUNNING && t->noboot) { 5952 pr_warn("Tracer '%s' is not allowed on command line, ignored\n", 5953 t->name); 5954 goto out; 5955 } 5956 5957 /* Some tracers are only allowed for the top level buffer */ 5958 if (!trace_ok_for_array(t, tr)) { 5959 ret = -EINVAL; 5960 goto out; 5961 } 5962 5963 /* If trace pipe files are being read, we can't change the tracer */ 5964 if (tr->trace_ref) { 5965 ret = -EBUSY; 5966 goto out; 5967 } 5968 5969 trace_branch_disable(); 5970 5971 tr->current_trace->enabled--; 5972 5973 if (tr->current_trace->reset) 5974 tr->current_trace->reset(tr); 5975 5976 /* Current trace needs to be nop_trace before synchronize_rcu */ 5977 tr->current_trace = &nop_trace; 5978 5979 #ifdef CONFIG_TRACER_MAX_TRACE 5980 had_max_tr = tr->allocated_snapshot; 5981 5982 if (had_max_tr && !t->use_max_tr) { 5983 /* 5984 * We need to make sure that the update_max_tr sees that 5985 * current_trace changed to nop_trace to keep it from 5986 * swapping the buffers after we resize it. 5987 * The update_max_tr is called from interrupts disabled 5988 * so a synchronized_sched() is sufficient. 5989 */ 5990 synchronize_rcu(); 5991 free_snapshot(tr); 5992 } 5993 #endif 5994 5995 #ifdef CONFIG_TRACER_MAX_TRACE 5996 if (t->use_max_tr && !had_max_tr) { 5997 ret = tracing_alloc_snapshot_instance(tr); 5998 if (ret < 0) 5999 goto out; 6000 } 6001 #endif 6002 6003 if (t->init) { 6004 ret = tracer_init(t, tr); 6005 if (ret) 6006 goto out; 6007 } 6008 6009 tr->current_trace = t; 6010 tr->current_trace->enabled++; 6011 trace_branch_enable(tr); 6012 out: 6013 mutex_unlock(&trace_types_lock); 6014 6015 return ret; 6016 } 6017 6018 static ssize_t 6019 tracing_set_trace_write(struct file *filp, const char __user *ubuf, 6020 size_t cnt, loff_t *ppos) 6021 { 6022 struct trace_array *tr = filp->private_data; 6023 char buf[MAX_TRACER_SIZE+1]; 6024 int i; 6025 size_t ret; 6026 int err; 6027 6028 ret = cnt; 6029 6030 if (cnt > MAX_TRACER_SIZE) 6031 cnt = MAX_TRACER_SIZE; 6032 6033 if (copy_from_user(buf, ubuf, cnt)) 6034 return -EFAULT; 6035 6036 buf[cnt] = 0; 6037 6038 /* strip ending whitespace. */ 6039 for (i = cnt - 1; i > 0 && isspace(buf[i]); i--) 6040 buf[i] = 0; 6041 6042 err = tracing_set_tracer(tr, buf); 6043 if (err) 6044 return err; 6045 6046 *ppos += ret; 6047 6048 return ret; 6049 } 6050 6051 static ssize_t 6052 tracing_nsecs_read(unsigned long *ptr, char __user *ubuf, 6053 size_t cnt, loff_t *ppos) 6054 { 6055 char buf[64]; 6056 int r; 6057 6058 r = snprintf(buf, sizeof(buf), "%ld\n", 6059 *ptr == (unsigned long)-1 ? -1 : nsecs_to_usecs(*ptr)); 6060 if (r > sizeof(buf)) 6061 r = sizeof(buf); 6062 return simple_read_from_buffer(ubuf, cnt, ppos, buf, r); 6063 } 6064 6065 static ssize_t 6066 tracing_nsecs_write(unsigned long *ptr, const char __user *ubuf, 6067 size_t cnt, loff_t *ppos) 6068 { 6069 unsigned long val; 6070 int ret; 6071 6072 ret = kstrtoul_from_user(ubuf, cnt, 10, &val); 6073 if (ret) 6074 return ret; 6075 6076 *ptr = val * 1000; 6077 6078 return cnt; 6079 } 6080 6081 static ssize_t 6082 tracing_thresh_read(struct file *filp, char __user *ubuf, 6083 size_t cnt, loff_t *ppos) 6084 { 6085 return tracing_nsecs_read(&tracing_thresh, ubuf, cnt, ppos); 6086 } 6087 6088 static ssize_t 6089 tracing_thresh_write(struct file *filp, const char __user *ubuf, 6090 size_t cnt, loff_t *ppos) 6091 { 6092 struct trace_array *tr = filp->private_data; 6093 int ret; 6094 6095 mutex_lock(&trace_types_lock); 6096 ret = tracing_nsecs_write(&tracing_thresh, ubuf, cnt, ppos); 6097 if (ret < 0) 6098 goto out; 6099 6100 if (tr->current_trace->update_thresh) { 6101 ret = tr->current_trace->update_thresh(tr); 6102 if (ret < 0) 6103 goto out; 6104 } 6105 6106 ret = cnt; 6107 out: 6108 mutex_unlock(&trace_types_lock); 6109 6110 return ret; 6111 } 6112 6113 #if defined(CONFIG_TRACER_MAX_TRACE) || defined(CONFIG_HWLAT_TRACER) 6114 6115 static ssize_t 6116 tracing_max_lat_read(struct file *filp, char __user *ubuf, 6117 size_t cnt, loff_t *ppos) 6118 { 6119 return tracing_nsecs_read(filp->private_data, ubuf, cnt, ppos); 6120 } 6121 6122 static ssize_t 6123 tracing_max_lat_write(struct file *filp, const char __user *ubuf, 6124 size_t cnt, loff_t *ppos) 6125 { 6126 return tracing_nsecs_write(filp->private_data, ubuf, cnt, ppos); 6127 } 6128 6129 #endif 6130 6131 static int tracing_open_pipe(struct inode *inode, struct file *filp) 6132 { 6133 struct trace_array *tr = inode->i_private; 6134 struct trace_iterator *iter; 6135 int ret; 6136 6137 ret = tracing_check_open_get_tr(tr); 6138 if (ret) 6139 return ret; 6140 6141 mutex_lock(&trace_types_lock); 6142 6143 /* create a buffer to store the information to pass to userspace */ 6144 iter = kzalloc(sizeof(*iter), GFP_KERNEL); 6145 if (!iter) { 6146 ret = -ENOMEM; 6147 __trace_array_put(tr); 6148 goto out; 6149 } 6150 6151 trace_seq_init(&iter->seq); 6152 iter->trace = tr->current_trace; 6153 6154 if (!alloc_cpumask_var(&iter->started, GFP_KERNEL)) { 6155 ret = -ENOMEM; 6156 goto fail; 6157 } 6158 6159 /* trace pipe does not show start of buffer */ 6160 cpumask_setall(iter->started); 6161 6162 if (tr->trace_flags & TRACE_ITER_LATENCY_FMT) 6163 iter->iter_flags |= TRACE_FILE_LAT_FMT; 6164 6165 /* Output in nanoseconds only if we are using a clock in nanoseconds. */ 6166 if (trace_clocks[tr->clock_id].in_ns) 6167 iter->iter_flags |= TRACE_FILE_TIME_IN_NS; 6168 6169 iter->tr = tr; 6170 iter->array_buffer = &tr->array_buffer; 6171 iter->cpu_file = tracing_get_cpu(inode); 6172 mutex_init(&iter->mutex); 6173 filp->private_data = iter; 6174 6175 if (iter->trace->pipe_open) 6176 iter->trace->pipe_open(iter); 6177 6178 nonseekable_open(inode, filp); 6179 6180 tr->trace_ref++; 6181 out: 6182 mutex_unlock(&trace_types_lock); 6183 return ret; 6184 6185 fail: 6186 kfree(iter); 6187 __trace_array_put(tr); 6188 mutex_unlock(&trace_types_lock); 6189 return ret; 6190 } 6191 6192 static int tracing_release_pipe(struct inode *inode, struct file *file) 6193 { 6194 struct trace_iterator *iter = file->private_data; 6195 struct trace_array *tr = inode->i_private; 6196 6197 mutex_lock(&trace_types_lock); 6198 6199 tr->trace_ref--; 6200 6201 if (iter->trace->pipe_close) 6202 iter->trace->pipe_close(iter); 6203 6204 mutex_unlock(&trace_types_lock); 6205 6206 free_cpumask_var(iter->started); 6207 mutex_destroy(&iter->mutex); 6208 kfree(iter); 6209 6210 trace_array_put(tr); 6211 6212 return 0; 6213 } 6214 6215 static __poll_t 6216 trace_poll(struct trace_iterator *iter, struct file *filp, poll_table *poll_table) 6217 { 6218 struct trace_array *tr = iter->tr; 6219 6220 /* Iterators are static, they should be filled or empty */ 6221 if (trace_buffer_iter(iter, iter->cpu_file)) 6222 return EPOLLIN | EPOLLRDNORM; 6223 6224 if (tr->trace_flags & TRACE_ITER_BLOCK) 6225 /* 6226 * Always select as readable when in blocking mode 6227 */ 6228 return EPOLLIN | EPOLLRDNORM; 6229 else 6230 return ring_buffer_poll_wait(iter->array_buffer->buffer, iter->cpu_file, 6231 filp, poll_table); 6232 } 6233 6234 static __poll_t 6235 tracing_poll_pipe(struct file *filp, poll_table *poll_table) 6236 { 6237 struct trace_iterator *iter = filp->private_data; 6238 6239 return trace_poll(iter, filp, poll_table); 6240 } 6241 6242 /* Must be called with iter->mutex held. */ 6243 static int tracing_wait_pipe(struct file *filp) 6244 { 6245 struct trace_iterator *iter = filp->private_data; 6246 int ret; 6247 6248 while (trace_empty(iter)) { 6249 6250 if ((filp->f_flags & O_NONBLOCK)) { 6251 return -EAGAIN; 6252 } 6253 6254 /* 6255 * We block until we read something and tracing is disabled. 6256 * We still block if tracing is disabled, but we have never 6257 * read anything. This allows a user to cat this file, and 6258 * then enable tracing. But after we have read something, 6259 * we give an EOF when tracing is again disabled. 6260 * 6261 * iter->pos will be 0 if we haven't read anything. 6262 */ 6263 if (!tracer_tracing_is_on(iter->tr) && iter->pos) 6264 break; 6265 6266 mutex_unlock(&iter->mutex); 6267 6268 ret = wait_on_pipe(iter, 0); 6269 6270 mutex_lock(&iter->mutex); 6271 6272 if (ret) 6273 return ret; 6274 } 6275 6276 return 1; 6277 } 6278 6279 /* 6280 * Consumer reader. 6281 */ 6282 static ssize_t 6283 tracing_read_pipe(struct file *filp, char __user *ubuf, 6284 size_t cnt, loff_t *ppos) 6285 { 6286 struct trace_iterator *iter = filp->private_data; 6287 ssize_t sret; 6288 6289 /* 6290 * Avoid more than one consumer on a single file descriptor 6291 * This is just a matter of traces coherency, the ring buffer itself 6292 * is protected. 6293 */ 6294 mutex_lock(&iter->mutex); 6295 6296 /* return any leftover data */ 6297 sret = trace_seq_to_user(&iter->seq, ubuf, cnt); 6298 if (sret != -EBUSY) 6299 goto out; 6300 6301 trace_seq_init(&iter->seq); 6302 6303 if (iter->trace->read) { 6304 sret = iter->trace->read(iter, filp, ubuf, cnt, ppos); 6305 if (sret) 6306 goto out; 6307 } 6308 6309 waitagain: 6310 sret = tracing_wait_pipe(filp); 6311 if (sret <= 0) 6312 goto out; 6313 6314 /* stop when tracing is finished */ 6315 if (trace_empty(iter)) { 6316 sret = 0; 6317 goto out; 6318 } 6319 6320 if (cnt >= PAGE_SIZE) 6321 cnt = PAGE_SIZE - 1; 6322 6323 /* reset all but tr, trace, and overruns */ 6324 memset(&iter->seq, 0, 6325 sizeof(struct trace_iterator) - 6326 offsetof(struct trace_iterator, seq)); 6327 cpumask_clear(iter->started); 6328 trace_seq_init(&iter->seq); 6329 iter->pos = -1; 6330 6331 trace_event_read_lock(); 6332 trace_access_lock(iter->cpu_file); 6333 while (trace_find_next_entry_inc(iter) != NULL) { 6334 enum print_line_t ret; 6335 int save_len = iter->seq.seq.len; 6336 6337 ret = print_trace_line(iter); 6338 if (ret == TRACE_TYPE_PARTIAL_LINE) { 6339 /* don't print partial lines */ 6340 iter->seq.seq.len = save_len; 6341 break; 6342 } 6343 if (ret != TRACE_TYPE_NO_CONSUME) 6344 trace_consume(iter); 6345 6346 if (trace_seq_used(&iter->seq) >= cnt) 6347 break; 6348 6349 /* 6350 * Setting the full flag means we reached the trace_seq buffer 6351 * size and we should leave by partial output condition above. 6352 * One of the trace_seq_* functions is not used properly. 6353 */ 6354 WARN_ONCE(iter->seq.full, "full flag set for trace type %d", 6355 iter->ent->type); 6356 } 6357 trace_access_unlock(iter->cpu_file); 6358 trace_event_read_unlock(); 6359 6360 /* Now copy what we have to the user */ 6361 sret = trace_seq_to_user(&iter->seq, ubuf, cnt); 6362 if (iter->seq.seq.readpos >= trace_seq_used(&iter->seq)) 6363 trace_seq_init(&iter->seq); 6364 6365 /* 6366 * If there was nothing to send to user, in spite of consuming trace 6367 * entries, go back to wait for more entries. 6368 */ 6369 if (sret == -EBUSY) 6370 goto waitagain; 6371 6372 out: 6373 mutex_unlock(&iter->mutex); 6374 6375 return sret; 6376 } 6377 6378 static void tracing_spd_release_pipe(struct splice_pipe_desc *spd, 6379 unsigned int idx) 6380 { 6381 __free_page(spd->pages[idx]); 6382 } 6383 6384 static size_t 6385 tracing_fill_pipe_page(size_t rem, struct trace_iterator *iter) 6386 { 6387 size_t count; 6388 int save_len; 6389 int ret; 6390 6391 /* Seq buffer is page-sized, exactly what we need. */ 6392 for (;;) { 6393 save_len = iter->seq.seq.len; 6394 ret = print_trace_line(iter); 6395 6396 if (trace_seq_has_overflowed(&iter->seq)) { 6397 iter->seq.seq.len = save_len; 6398 break; 6399 } 6400 6401 /* 6402 * This should not be hit, because it should only 6403 * be set if the iter->seq overflowed. But check it 6404 * anyway to be safe. 6405 */ 6406 if (ret == TRACE_TYPE_PARTIAL_LINE) { 6407 iter->seq.seq.len = save_len; 6408 break; 6409 } 6410 6411 count = trace_seq_used(&iter->seq) - save_len; 6412 if (rem < count) { 6413 rem = 0; 6414 iter->seq.seq.len = save_len; 6415 break; 6416 } 6417 6418 if (ret != TRACE_TYPE_NO_CONSUME) 6419 trace_consume(iter); 6420 rem -= count; 6421 if (!trace_find_next_entry_inc(iter)) { 6422 rem = 0; 6423 iter->ent = NULL; 6424 break; 6425 } 6426 } 6427 6428 return rem; 6429 } 6430 6431 static ssize_t tracing_splice_read_pipe(struct file *filp, 6432 loff_t *ppos, 6433 struct pipe_inode_info *pipe, 6434 size_t len, 6435 unsigned int flags) 6436 { 6437 struct page *pages_def[PIPE_DEF_BUFFERS]; 6438 struct partial_page partial_def[PIPE_DEF_BUFFERS]; 6439 struct trace_iterator *iter = filp->private_data; 6440 struct splice_pipe_desc spd = { 6441 .pages = pages_def, 6442 .partial = partial_def, 6443 .nr_pages = 0, /* This gets updated below. */ 6444 .nr_pages_max = PIPE_DEF_BUFFERS, 6445 .ops = &default_pipe_buf_ops, 6446 .spd_release = tracing_spd_release_pipe, 6447 }; 6448 ssize_t ret; 6449 size_t rem; 6450 unsigned int i; 6451 6452 if (splice_grow_spd(pipe, &spd)) 6453 return -ENOMEM; 6454 6455 mutex_lock(&iter->mutex); 6456 6457 if (iter->trace->splice_read) { 6458 ret = iter->trace->splice_read(iter, filp, 6459 ppos, pipe, len, flags); 6460 if (ret) 6461 goto out_err; 6462 } 6463 6464 ret = tracing_wait_pipe(filp); 6465 if (ret <= 0) 6466 goto out_err; 6467 6468 if (!iter->ent && !trace_find_next_entry_inc(iter)) { 6469 ret = -EFAULT; 6470 goto out_err; 6471 } 6472 6473 trace_event_read_lock(); 6474 trace_access_lock(iter->cpu_file); 6475 6476 /* Fill as many pages as possible. */ 6477 for (i = 0, rem = len; i < spd.nr_pages_max && rem; i++) { 6478 spd.pages[i] = alloc_page(GFP_KERNEL); 6479 if (!spd.pages[i]) 6480 break; 6481 6482 rem = tracing_fill_pipe_page(rem, iter); 6483 6484 /* Copy the data into the page, so we can start over. */ 6485 ret = trace_seq_to_buffer(&iter->seq, 6486 page_address(spd.pages[i]), 6487 trace_seq_used(&iter->seq)); 6488 if (ret < 0) { 6489 __free_page(spd.pages[i]); 6490 break; 6491 } 6492 spd.partial[i].offset = 0; 6493 spd.partial[i].len = trace_seq_used(&iter->seq); 6494 6495 trace_seq_init(&iter->seq); 6496 } 6497 6498 trace_access_unlock(iter->cpu_file); 6499 trace_event_read_unlock(); 6500 mutex_unlock(&iter->mutex); 6501 6502 spd.nr_pages = i; 6503 6504 if (i) 6505 ret = splice_to_pipe(pipe, &spd); 6506 else 6507 ret = 0; 6508 out: 6509 splice_shrink_spd(&spd); 6510 return ret; 6511 6512 out_err: 6513 mutex_unlock(&iter->mutex); 6514 goto out; 6515 } 6516 6517 static ssize_t 6518 tracing_entries_read(struct file *filp, char __user *ubuf, 6519 size_t cnt, loff_t *ppos) 6520 { 6521 struct inode *inode = file_inode(filp); 6522 struct trace_array *tr = inode->i_private; 6523 int cpu = tracing_get_cpu(inode); 6524 char buf[64]; 6525 int r = 0; 6526 ssize_t ret; 6527 6528 mutex_lock(&trace_types_lock); 6529 6530 if (cpu == RING_BUFFER_ALL_CPUS) { 6531 int cpu, buf_size_same; 6532 unsigned long size; 6533 6534 size = 0; 6535 buf_size_same = 1; 6536 /* check if all cpu sizes are same */ 6537 for_each_tracing_cpu(cpu) { 6538 /* fill in the size from first enabled cpu */ 6539 if (size == 0) 6540 size = per_cpu_ptr(tr->array_buffer.data, cpu)->entries; 6541 if (size != per_cpu_ptr(tr->array_buffer.data, cpu)->entries) { 6542 buf_size_same = 0; 6543 break; 6544 } 6545 } 6546 6547 if (buf_size_same) { 6548 if (!ring_buffer_expanded) 6549 r = sprintf(buf, "%lu (expanded: %lu)\n", 6550 size >> 10, 6551 trace_buf_size >> 10); 6552 else 6553 r = sprintf(buf, "%lu\n", size >> 10); 6554 } else 6555 r = sprintf(buf, "X\n"); 6556 } else 6557 r = sprintf(buf, "%lu\n", per_cpu_ptr(tr->array_buffer.data, cpu)->entries >> 10); 6558 6559 mutex_unlock(&trace_types_lock); 6560 6561 ret = simple_read_from_buffer(ubuf, cnt, ppos, buf, r); 6562 return ret; 6563 } 6564 6565 static ssize_t 6566 tracing_entries_write(struct file *filp, const char __user *ubuf, 6567 size_t cnt, loff_t *ppos) 6568 { 6569 struct inode *inode = file_inode(filp); 6570 struct trace_array *tr = inode->i_private; 6571 unsigned long val; 6572 int ret; 6573 6574 ret = kstrtoul_from_user(ubuf, cnt, 10, &val); 6575 if (ret) 6576 return ret; 6577 6578 /* must have at least 1 entry */ 6579 if (!val) 6580 return -EINVAL; 6581 6582 /* value is in KB */ 6583 val <<= 10; 6584 ret = tracing_resize_ring_buffer(tr, val, tracing_get_cpu(inode)); 6585 if (ret < 0) 6586 return ret; 6587 6588 *ppos += cnt; 6589 6590 return cnt; 6591 } 6592 6593 static ssize_t 6594 tracing_total_entries_read(struct file *filp, char __user *ubuf, 6595 size_t cnt, loff_t *ppos) 6596 { 6597 struct trace_array *tr = filp->private_data; 6598 char buf[64]; 6599 int r, cpu; 6600 unsigned long size = 0, expanded_size = 0; 6601 6602 mutex_lock(&trace_types_lock); 6603 for_each_tracing_cpu(cpu) { 6604 size += per_cpu_ptr(tr->array_buffer.data, cpu)->entries >> 10; 6605 if (!ring_buffer_expanded) 6606 expanded_size += trace_buf_size >> 10; 6607 } 6608 if (ring_buffer_expanded) 6609 r = sprintf(buf, "%lu\n", size); 6610 else 6611 r = sprintf(buf, "%lu (expanded: %lu)\n", size, expanded_size); 6612 mutex_unlock(&trace_types_lock); 6613 6614 return simple_read_from_buffer(ubuf, cnt, ppos, buf, r); 6615 } 6616 6617 static ssize_t 6618 tracing_free_buffer_write(struct file *filp, const char __user *ubuf, 6619 size_t cnt, loff_t *ppos) 6620 { 6621 /* 6622 * There is no need to read what the user has written, this function 6623 * is just to make sure that there is no error when "echo" is used 6624 */ 6625 6626 *ppos += cnt; 6627 6628 return cnt; 6629 } 6630 6631 static int 6632 tracing_free_buffer_release(struct inode *inode, struct file *filp) 6633 { 6634 struct trace_array *tr = inode->i_private; 6635 6636 /* disable tracing ? */ 6637 if (tr->trace_flags & TRACE_ITER_STOP_ON_FREE) 6638 tracer_tracing_off(tr); 6639 /* resize the ring buffer to 0 */ 6640 tracing_resize_ring_buffer(tr, 0, RING_BUFFER_ALL_CPUS); 6641 6642 trace_array_put(tr); 6643 6644 return 0; 6645 } 6646 6647 static ssize_t 6648 tracing_mark_write(struct file *filp, const char __user *ubuf, 6649 size_t cnt, loff_t *fpos) 6650 { 6651 struct trace_array *tr = filp->private_data; 6652 struct ring_buffer_event *event; 6653 enum event_trigger_type tt = ETT_NONE; 6654 struct trace_buffer *buffer; 6655 struct print_entry *entry; 6656 unsigned long irq_flags; 6657 ssize_t written; 6658 int size; 6659 int len; 6660 6661 /* Used in tracing_mark_raw_write() as well */ 6662 #define FAULTED_STR "<faulted>" 6663 #define FAULTED_SIZE (sizeof(FAULTED_STR) - 1) /* '\0' is already accounted for */ 6664 6665 if (tracing_disabled) 6666 return -EINVAL; 6667 6668 if (!(tr->trace_flags & TRACE_ITER_MARKERS)) 6669 return -EINVAL; 6670 6671 if (cnt > TRACE_BUF_SIZE) 6672 cnt = TRACE_BUF_SIZE; 6673 6674 BUILD_BUG_ON(TRACE_BUF_SIZE >= PAGE_SIZE); 6675 6676 local_save_flags(irq_flags); 6677 size = sizeof(*entry) + cnt + 2; /* add '\0' and possible '\n' */ 6678 6679 /* If less than "<faulted>", then make sure we can still add that */ 6680 if (cnt < FAULTED_SIZE) 6681 size += FAULTED_SIZE - cnt; 6682 6683 buffer = tr->array_buffer.buffer; 6684 event = __trace_buffer_lock_reserve(buffer, TRACE_PRINT, size, 6685 irq_flags, preempt_count()); 6686 if (unlikely(!event)) 6687 /* Ring buffer disabled, return as if not open for write */ 6688 return -EBADF; 6689 6690 entry = ring_buffer_event_data(event); 6691 entry->ip = _THIS_IP_; 6692 6693 len = __copy_from_user_inatomic(&entry->buf, ubuf, cnt); 6694 if (len) { 6695 memcpy(&entry->buf, FAULTED_STR, FAULTED_SIZE); 6696 cnt = FAULTED_SIZE; 6697 written = -EFAULT; 6698 } else 6699 written = cnt; 6700 6701 if (tr->trace_marker_file && !list_empty(&tr->trace_marker_file->triggers)) { 6702 /* do not add \n before testing triggers, but add \0 */ 6703 entry->buf[cnt] = '\0'; 6704 tt = event_triggers_call(tr->trace_marker_file, entry, event); 6705 } 6706 6707 if (entry->buf[cnt - 1] != '\n') { 6708 entry->buf[cnt] = '\n'; 6709 entry->buf[cnt + 1] = '\0'; 6710 } else 6711 entry->buf[cnt] = '\0'; 6712 6713 if (static_branch_unlikely(&trace_marker_exports_enabled)) 6714 ftrace_exports(event, TRACE_EXPORT_MARKER); 6715 __buffer_unlock_commit(buffer, event); 6716 6717 if (tt) 6718 event_triggers_post_call(tr->trace_marker_file, tt); 6719 6720 if (written > 0) 6721 *fpos += written; 6722 6723 return written; 6724 } 6725 6726 /* Limit it for now to 3K (including tag) */ 6727 #define RAW_DATA_MAX_SIZE (1024*3) 6728 6729 static ssize_t 6730 tracing_mark_raw_write(struct file *filp, const char __user *ubuf, 6731 size_t cnt, loff_t *fpos) 6732 { 6733 struct trace_array *tr = filp->private_data; 6734 struct ring_buffer_event *event; 6735 struct trace_buffer *buffer; 6736 struct raw_data_entry *entry; 6737 unsigned long irq_flags; 6738 ssize_t written; 6739 int size; 6740 int len; 6741 6742 #define FAULT_SIZE_ID (FAULTED_SIZE + sizeof(int)) 6743 6744 if (tracing_disabled) 6745 return -EINVAL; 6746 6747 if (!(tr->trace_flags & TRACE_ITER_MARKERS)) 6748 return -EINVAL; 6749 6750 /* The marker must at least have a tag id */ 6751 if (cnt < sizeof(unsigned int) || cnt > RAW_DATA_MAX_SIZE) 6752 return -EINVAL; 6753 6754 if (cnt > TRACE_BUF_SIZE) 6755 cnt = TRACE_BUF_SIZE; 6756 6757 BUILD_BUG_ON(TRACE_BUF_SIZE >= PAGE_SIZE); 6758 6759 local_save_flags(irq_flags); 6760 size = sizeof(*entry) + cnt; 6761 if (cnt < FAULT_SIZE_ID) 6762 size += FAULT_SIZE_ID - cnt; 6763 6764 buffer = tr->array_buffer.buffer; 6765 event = __trace_buffer_lock_reserve(buffer, TRACE_RAW_DATA, size, 6766 irq_flags, preempt_count()); 6767 if (!event) 6768 /* Ring buffer disabled, return as if not open for write */ 6769 return -EBADF; 6770 6771 entry = ring_buffer_event_data(event); 6772 6773 len = __copy_from_user_inatomic(&entry->id, ubuf, cnt); 6774 if (len) { 6775 entry->id = -1; 6776 memcpy(&entry->buf, FAULTED_STR, FAULTED_SIZE); 6777 written = -EFAULT; 6778 } else 6779 written = cnt; 6780 6781 __buffer_unlock_commit(buffer, event); 6782 6783 if (written > 0) 6784 *fpos += written; 6785 6786 return written; 6787 } 6788 6789 static int tracing_clock_show(struct seq_file *m, void *v) 6790 { 6791 struct trace_array *tr = m->private; 6792 int i; 6793 6794 for (i = 0; i < ARRAY_SIZE(trace_clocks); i++) 6795 seq_printf(m, 6796 "%s%s%s%s", i ? " " : "", 6797 i == tr->clock_id ? "[" : "", trace_clocks[i].name, 6798 i == tr->clock_id ? "]" : ""); 6799 seq_putc(m, '\n'); 6800 6801 return 0; 6802 } 6803 6804 int tracing_set_clock(struct trace_array *tr, const char *clockstr) 6805 { 6806 int i; 6807 6808 for (i = 0; i < ARRAY_SIZE(trace_clocks); i++) { 6809 if (strcmp(trace_clocks[i].name, clockstr) == 0) 6810 break; 6811 } 6812 if (i == ARRAY_SIZE(trace_clocks)) 6813 return -EINVAL; 6814 6815 mutex_lock(&trace_types_lock); 6816 6817 tr->clock_id = i; 6818 6819 ring_buffer_set_clock(tr->array_buffer.buffer, trace_clocks[i].func); 6820 6821 /* 6822 * New clock may not be consistent with the previous clock. 6823 * Reset the buffer so that it doesn't have incomparable timestamps. 6824 */ 6825 tracing_reset_online_cpus(&tr->array_buffer); 6826 6827 #ifdef CONFIG_TRACER_MAX_TRACE 6828 if (tr->max_buffer.buffer) 6829 ring_buffer_set_clock(tr->max_buffer.buffer, trace_clocks[i].func); 6830 tracing_reset_online_cpus(&tr->max_buffer); 6831 #endif 6832 6833 mutex_unlock(&trace_types_lock); 6834 6835 return 0; 6836 } 6837 6838 static ssize_t tracing_clock_write(struct file *filp, const char __user *ubuf, 6839 size_t cnt, loff_t *fpos) 6840 { 6841 struct seq_file *m = filp->private_data; 6842 struct trace_array *tr = m->private; 6843 char buf[64]; 6844 const char *clockstr; 6845 int ret; 6846 6847 if (cnt >= sizeof(buf)) 6848 return -EINVAL; 6849 6850 if (copy_from_user(buf, ubuf, cnt)) 6851 return -EFAULT; 6852 6853 buf[cnt] = 0; 6854 6855 clockstr = strstrip(buf); 6856 6857 ret = tracing_set_clock(tr, clockstr); 6858 if (ret) 6859 return ret; 6860 6861 *fpos += cnt; 6862 6863 return cnt; 6864 } 6865 6866 static int tracing_clock_open(struct inode *inode, struct file *file) 6867 { 6868 struct trace_array *tr = inode->i_private; 6869 int ret; 6870 6871 ret = tracing_check_open_get_tr(tr); 6872 if (ret) 6873 return ret; 6874 6875 ret = single_open(file, tracing_clock_show, inode->i_private); 6876 if (ret < 0) 6877 trace_array_put(tr); 6878 6879 return ret; 6880 } 6881 6882 static int tracing_time_stamp_mode_show(struct seq_file *m, void *v) 6883 { 6884 struct trace_array *tr = m->private; 6885 6886 mutex_lock(&trace_types_lock); 6887 6888 if (ring_buffer_time_stamp_abs(tr->array_buffer.buffer)) 6889 seq_puts(m, "delta [absolute]\n"); 6890 else 6891 seq_puts(m, "[delta] absolute\n"); 6892 6893 mutex_unlock(&trace_types_lock); 6894 6895 return 0; 6896 } 6897 6898 static int tracing_time_stamp_mode_open(struct inode *inode, struct file *file) 6899 { 6900 struct trace_array *tr = inode->i_private; 6901 int ret; 6902 6903 ret = tracing_check_open_get_tr(tr); 6904 if (ret) 6905 return ret; 6906 6907 ret = single_open(file, tracing_time_stamp_mode_show, inode->i_private); 6908 if (ret < 0) 6909 trace_array_put(tr); 6910 6911 return ret; 6912 } 6913 6914 int tracing_set_time_stamp_abs(struct trace_array *tr, bool abs) 6915 { 6916 int ret = 0; 6917 6918 mutex_lock(&trace_types_lock); 6919 6920 if (abs && tr->time_stamp_abs_ref++) 6921 goto out; 6922 6923 if (!abs) { 6924 if (WARN_ON_ONCE(!tr->time_stamp_abs_ref)) { 6925 ret = -EINVAL; 6926 goto out; 6927 } 6928 6929 if (--tr->time_stamp_abs_ref) 6930 goto out; 6931 } 6932 6933 ring_buffer_set_time_stamp_abs(tr->array_buffer.buffer, abs); 6934 6935 #ifdef CONFIG_TRACER_MAX_TRACE 6936 if (tr->max_buffer.buffer) 6937 ring_buffer_set_time_stamp_abs(tr->max_buffer.buffer, abs); 6938 #endif 6939 out: 6940 mutex_unlock(&trace_types_lock); 6941 6942 return ret; 6943 } 6944 6945 struct ftrace_buffer_info { 6946 struct trace_iterator iter; 6947 void *spare; 6948 unsigned int spare_cpu; 6949 unsigned int read; 6950 }; 6951 6952 #ifdef CONFIG_TRACER_SNAPSHOT 6953 static int tracing_snapshot_open(struct inode *inode, struct file *file) 6954 { 6955 struct trace_array *tr = inode->i_private; 6956 struct trace_iterator *iter; 6957 struct seq_file *m; 6958 int ret; 6959 6960 ret = tracing_check_open_get_tr(tr); 6961 if (ret) 6962 return ret; 6963 6964 if (file->f_mode & FMODE_READ) { 6965 iter = __tracing_open(inode, file, true); 6966 if (IS_ERR(iter)) 6967 ret = PTR_ERR(iter); 6968 } else { 6969 /* Writes still need the seq_file to hold the private data */ 6970 ret = -ENOMEM; 6971 m = kzalloc(sizeof(*m), GFP_KERNEL); 6972 if (!m) 6973 goto out; 6974 iter = kzalloc(sizeof(*iter), GFP_KERNEL); 6975 if (!iter) { 6976 kfree(m); 6977 goto out; 6978 } 6979 ret = 0; 6980 6981 iter->tr = tr; 6982 iter->array_buffer = &tr->max_buffer; 6983 iter->cpu_file = tracing_get_cpu(inode); 6984 m->private = iter; 6985 file->private_data = m; 6986 } 6987 out: 6988 if (ret < 0) 6989 trace_array_put(tr); 6990 6991 return ret; 6992 } 6993 6994 static ssize_t 6995 tracing_snapshot_write(struct file *filp, const char __user *ubuf, size_t cnt, 6996 loff_t *ppos) 6997 { 6998 struct seq_file *m = filp->private_data; 6999 struct trace_iterator *iter = m->private; 7000 struct trace_array *tr = iter->tr; 7001 unsigned long val; 7002 int ret; 7003 7004 ret = tracing_update_buffers(); 7005 if (ret < 0) 7006 return ret; 7007 7008 ret = kstrtoul_from_user(ubuf, cnt, 10, &val); 7009 if (ret) 7010 return ret; 7011 7012 mutex_lock(&trace_types_lock); 7013 7014 if (tr->current_trace->use_max_tr) { 7015 ret = -EBUSY; 7016 goto out; 7017 } 7018 7019 arch_spin_lock(&tr->max_lock); 7020 if (tr->cond_snapshot) 7021 ret = -EBUSY; 7022 arch_spin_unlock(&tr->max_lock); 7023 if (ret) 7024 goto out; 7025 7026 switch (val) { 7027 case 0: 7028 if (iter->cpu_file != RING_BUFFER_ALL_CPUS) { 7029 ret = -EINVAL; 7030 break; 7031 } 7032 if (tr->allocated_snapshot) 7033 free_snapshot(tr); 7034 break; 7035 case 1: 7036 /* Only allow per-cpu swap if the ring buffer supports it */ 7037 #ifndef CONFIG_RING_BUFFER_ALLOW_SWAP 7038 if (iter->cpu_file != RING_BUFFER_ALL_CPUS) { 7039 ret = -EINVAL; 7040 break; 7041 } 7042 #endif 7043 if (tr->allocated_snapshot) 7044 ret = resize_buffer_duplicate_size(&tr->max_buffer, 7045 &tr->array_buffer, iter->cpu_file); 7046 else 7047 ret = tracing_alloc_snapshot_instance(tr); 7048 if (ret < 0) 7049 break; 7050 local_irq_disable(); 7051 /* Now, we're going to swap */ 7052 if (iter->cpu_file == RING_BUFFER_ALL_CPUS) 7053 update_max_tr(tr, current, smp_processor_id(), NULL); 7054 else 7055 update_max_tr_single(tr, current, iter->cpu_file); 7056 local_irq_enable(); 7057 break; 7058 default: 7059 if (tr->allocated_snapshot) { 7060 if (iter->cpu_file == RING_BUFFER_ALL_CPUS) 7061 tracing_reset_online_cpus(&tr->max_buffer); 7062 else 7063 tracing_reset_cpu(&tr->max_buffer, iter->cpu_file); 7064 } 7065 break; 7066 } 7067 7068 if (ret >= 0) { 7069 *ppos += cnt; 7070 ret = cnt; 7071 } 7072 out: 7073 mutex_unlock(&trace_types_lock); 7074 return ret; 7075 } 7076 7077 static int tracing_snapshot_release(struct inode *inode, struct file *file) 7078 { 7079 struct seq_file *m = file->private_data; 7080 int ret; 7081 7082 ret = tracing_release(inode, file); 7083 7084 if (file->f_mode & FMODE_READ) 7085 return ret; 7086 7087 /* If write only, the seq_file is just a stub */ 7088 if (m) 7089 kfree(m->private); 7090 kfree(m); 7091 7092 return 0; 7093 } 7094 7095 static int tracing_buffers_open(struct inode *inode, struct file *filp); 7096 static ssize_t tracing_buffers_read(struct file *filp, char __user *ubuf, 7097 size_t count, loff_t *ppos); 7098 static int tracing_buffers_release(struct inode *inode, struct file *file); 7099 static ssize_t tracing_buffers_splice_read(struct file *file, loff_t *ppos, 7100 struct pipe_inode_info *pipe, size_t len, unsigned int flags); 7101 7102 static int snapshot_raw_open(struct inode *inode, struct file *filp) 7103 { 7104 struct ftrace_buffer_info *info; 7105 int ret; 7106 7107 /* The following checks for tracefs lockdown */ 7108 ret = tracing_buffers_open(inode, filp); 7109 if (ret < 0) 7110 return ret; 7111 7112 info = filp->private_data; 7113 7114 if (info->iter.trace->use_max_tr) { 7115 tracing_buffers_release(inode, filp); 7116 return -EBUSY; 7117 } 7118 7119 info->iter.snapshot = true; 7120 info->iter.array_buffer = &info->iter.tr->max_buffer; 7121 7122 return ret; 7123 } 7124 7125 #endif /* CONFIG_TRACER_SNAPSHOT */ 7126 7127 7128 static const struct file_operations tracing_thresh_fops = { 7129 .open = tracing_open_generic, 7130 .read = tracing_thresh_read, 7131 .write = tracing_thresh_write, 7132 .llseek = generic_file_llseek, 7133 }; 7134 7135 #if defined(CONFIG_TRACER_MAX_TRACE) || defined(CONFIG_HWLAT_TRACER) 7136 static const struct file_operations tracing_max_lat_fops = { 7137 .open = tracing_open_generic, 7138 .read = tracing_max_lat_read, 7139 .write = tracing_max_lat_write, 7140 .llseek = generic_file_llseek, 7141 }; 7142 #endif 7143 7144 static const struct file_operations set_tracer_fops = { 7145 .open = tracing_open_generic, 7146 .read = tracing_set_trace_read, 7147 .write = tracing_set_trace_write, 7148 .llseek = generic_file_llseek, 7149 }; 7150 7151 static const struct file_operations tracing_pipe_fops = { 7152 .open = tracing_open_pipe, 7153 .poll = tracing_poll_pipe, 7154 .read = tracing_read_pipe, 7155 .splice_read = tracing_splice_read_pipe, 7156 .release = tracing_release_pipe, 7157 .llseek = no_llseek, 7158 }; 7159 7160 static const struct file_operations tracing_entries_fops = { 7161 .open = tracing_open_generic_tr, 7162 .read = tracing_entries_read, 7163 .write = tracing_entries_write, 7164 .llseek = generic_file_llseek, 7165 .release = tracing_release_generic_tr, 7166 }; 7167 7168 static const struct file_operations tracing_total_entries_fops = { 7169 .open = tracing_open_generic_tr, 7170 .read = tracing_total_entries_read, 7171 .llseek = generic_file_llseek, 7172 .release = tracing_release_generic_tr, 7173 }; 7174 7175 static const struct file_operations tracing_free_buffer_fops = { 7176 .open = tracing_open_generic_tr, 7177 .write = tracing_free_buffer_write, 7178 .release = tracing_free_buffer_release, 7179 }; 7180 7181 static const struct file_operations tracing_mark_fops = { 7182 .open = tracing_open_generic_tr, 7183 .write = tracing_mark_write, 7184 .llseek = generic_file_llseek, 7185 .release = tracing_release_generic_tr, 7186 }; 7187 7188 static const struct file_operations tracing_mark_raw_fops = { 7189 .open = tracing_open_generic_tr, 7190 .write = tracing_mark_raw_write, 7191 .llseek = generic_file_llseek, 7192 .release = tracing_release_generic_tr, 7193 }; 7194 7195 static const struct file_operations trace_clock_fops = { 7196 .open = tracing_clock_open, 7197 .read = seq_read, 7198 .llseek = seq_lseek, 7199 .release = tracing_single_release_tr, 7200 .write = tracing_clock_write, 7201 }; 7202 7203 static const struct file_operations trace_time_stamp_mode_fops = { 7204 .open = tracing_time_stamp_mode_open, 7205 .read = seq_read, 7206 .llseek = seq_lseek, 7207 .release = tracing_single_release_tr, 7208 }; 7209 7210 #ifdef CONFIG_TRACER_SNAPSHOT 7211 static const struct file_operations snapshot_fops = { 7212 .open = tracing_snapshot_open, 7213 .read = seq_read, 7214 .write = tracing_snapshot_write, 7215 .llseek = tracing_lseek, 7216 .release = tracing_snapshot_release, 7217 }; 7218 7219 static const struct file_operations snapshot_raw_fops = { 7220 .open = snapshot_raw_open, 7221 .read = tracing_buffers_read, 7222 .release = tracing_buffers_release, 7223 .splice_read = tracing_buffers_splice_read, 7224 .llseek = no_llseek, 7225 }; 7226 7227 #endif /* CONFIG_TRACER_SNAPSHOT */ 7228 7229 #define TRACING_LOG_ERRS_MAX 8 7230 #define TRACING_LOG_LOC_MAX 128 7231 7232 #define CMD_PREFIX " Command: " 7233 7234 struct err_info { 7235 const char **errs; /* ptr to loc-specific array of err strings */ 7236 u8 type; /* index into errs -> specific err string */ 7237 u8 pos; /* MAX_FILTER_STR_VAL = 256 */ 7238 u64 ts; 7239 }; 7240 7241 struct tracing_log_err { 7242 struct list_head list; 7243 struct err_info info; 7244 char loc[TRACING_LOG_LOC_MAX]; /* err location */ 7245 char cmd[MAX_FILTER_STR_VAL]; /* what caused err */ 7246 }; 7247 7248 static DEFINE_MUTEX(tracing_err_log_lock); 7249 7250 static struct tracing_log_err *get_tracing_log_err(struct trace_array *tr) 7251 { 7252 struct tracing_log_err *err; 7253 7254 if (tr->n_err_log_entries < TRACING_LOG_ERRS_MAX) { 7255 err = kzalloc(sizeof(*err), GFP_KERNEL); 7256 if (!err) 7257 err = ERR_PTR(-ENOMEM); 7258 tr->n_err_log_entries++; 7259 7260 return err; 7261 } 7262 7263 err = list_first_entry(&tr->err_log, struct tracing_log_err, list); 7264 list_del(&err->list); 7265 7266 return err; 7267 } 7268 7269 /** 7270 * err_pos - find the position of a string within a command for error careting 7271 * @cmd: The tracing command that caused the error 7272 * @str: The string to position the caret at within @cmd 7273 * 7274 * Finds the position of the first occurence of @str within @cmd. The 7275 * return value can be passed to tracing_log_err() for caret placement 7276 * within @cmd. 7277 * 7278 * Returns the index within @cmd of the first occurence of @str or 0 7279 * if @str was not found. 7280 */ 7281 unsigned int err_pos(char *cmd, const char *str) 7282 { 7283 char *found; 7284 7285 if (WARN_ON(!strlen(cmd))) 7286 return 0; 7287 7288 found = strstr(cmd, str); 7289 if (found) 7290 return found - cmd; 7291 7292 return 0; 7293 } 7294 7295 /** 7296 * tracing_log_err - write an error to the tracing error log 7297 * @tr: The associated trace array for the error (NULL for top level array) 7298 * @loc: A string describing where the error occurred 7299 * @cmd: The tracing command that caused the error 7300 * @errs: The array of loc-specific static error strings 7301 * @type: The index into errs[], which produces the specific static err string 7302 * @pos: The position the caret should be placed in the cmd 7303 * 7304 * Writes an error into tracing/error_log of the form: 7305 * 7306 * <loc>: error: <text> 7307 * Command: <cmd> 7308 * ^ 7309 * 7310 * tracing/error_log is a small log file containing the last 7311 * TRACING_LOG_ERRS_MAX errors (8). Memory for errors isn't allocated 7312 * unless there has been a tracing error, and the error log can be 7313 * cleared and have its memory freed by writing the empty string in 7314 * truncation mode to it i.e. echo > tracing/error_log. 7315 * 7316 * NOTE: the @errs array along with the @type param are used to 7317 * produce a static error string - this string is not copied and saved 7318 * when the error is logged - only a pointer to it is saved. See 7319 * existing callers for examples of how static strings are typically 7320 * defined for use with tracing_log_err(). 7321 */ 7322 void tracing_log_err(struct trace_array *tr, 7323 const char *loc, const char *cmd, 7324 const char **errs, u8 type, u8 pos) 7325 { 7326 struct tracing_log_err *err; 7327 7328 if (!tr) 7329 tr = &global_trace; 7330 7331 mutex_lock(&tracing_err_log_lock); 7332 err = get_tracing_log_err(tr); 7333 if (PTR_ERR(err) == -ENOMEM) { 7334 mutex_unlock(&tracing_err_log_lock); 7335 return; 7336 } 7337 7338 snprintf(err->loc, TRACING_LOG_LOC_MAX, "%s: error: ", loc); 7339 snprintf(err->cmd, MAX_FILTER_STR_VAL,"\n" CMD_PREFIX "%s\n", cmd); 7340 7341 err->info.errs = errs; 7342 err->info.type = type; 7343 err->info.pos = pos; 7344 err->info.ts = local_clock(); 7345 7346 list_add_tail(&err->list, &tr->err_log); 7347 mutex_unlock(&tracing_err_log_lock); 7348 } 7349 7350 static void clear_tracing_err_log(struct trace_array *tr) 7351 { 7352 struct tracing_log_err *err, *next; 7353 7354 mutex_lock(&tracing_err_log_lock); 7355 list_for_each_entry_safe(err, next, &tr->err_log, list) { 7356 list_del(&err->list); 7357 kfree(err); 7358 } 7359 7360 tr->n_err_log_entries = 0; 7361 mutex_unlock(&tracing_err_log_lock); 7362 } 7363 7364 static void *tracing_err_log_seq_start(struct seq_file *m, loff_t *pos) 7365 { 7366 struct trace_array *tr = m->private; 7367 7368 mutex_lock(&tracing_err_log_lock); 7369 7370 return seq_list_start(&tr->err_log, *pos); 7371 } 7372 7373 static void *tracing_err_log_seq_next(struct seq_file *m, void *v, loff_t *pos) 7374 { 7375 struct trace_array *tr = m->private; 7376 7377 return seq_list_next(v, &tr->err_log, pos); 7378 } 7379 7380 static void tracing_err_log_seq_stop(struct seq_file *m, void *v) 7381 { 7382 mutex_unlock(&tracing_err_log_lock); 7383 } 7384 7385 static void tracing_err_log_show_pos(struct seq_file *m, u8 pos) 7386 { 7387 u8 i; 7388 7389 for (i = 0; i < sizeof(CMD_PREFIX) - 1; i++) 7390 seq_putc(m, ' '); 7391 for (i = 0; i < pos; i++) 7392 seq_putc(m, ' '); 7393 seq_puts(m, "^\n"); 7394 } 7395 7396 static int tracing_err_log_seq_show(struct seq_file *m, void *v) 7397 { 7398 struct tracing_log_err *err = v; 7399 7400 if (err) { 7401 const char *err_text = err->info.errs[err->info.type]; 7402 u64 sec = err->info.ts; 7403 u32 nsec; 7404 7405 nsec = do_div(sec, NSEC_PER_SEC); 7406 seq_printf(m, "[%5llu.%06u] %s%s", sec, nsec / 1000, 7407 err->loc, err_text); 7408 seq_printf(m, "%s", err->cmd); 7409 tracing_err_log_show_pos(m, err->info.pos); 7410 } 7411 7412 return 0; 7413 } 7414 7415 static const struct seq_operations tracing_err_log_seq_ops = { 7416 .start = tracing_err_log_seq_start, 7417 .next = tracing_err_log_seq_next, 7418 .stop = tracing_err_log_seq_stop, 7419 .show = tracing_err_log_seq_show 7420 }; 7421 7422 static int tracing_err_log_open(struct inode *inode, struct file *file) 7423 { 7424 struct trace_array *tr = inode->i_private; 7425 int ret = 0; 7426 7427 ret = tracing_check_open_get_tr(tr); 7428 if (ret) 7429 return ret; 7430 7431 /* If this file was opened for write, then erase contents */ 7432 if ((file->f_mode & FMODE_WRITE) && (file->f_flags & O_TRUNC)) 7433 clear_tracing_err_log(tr); 7434 7435 if (file->f_mode & FMODE_READ) { 7436 ret = seq_open(file, &tracing_err_log_seq_ops); 7437 if (!ret) { 7438 struct seq_file *m = file->private_data; 7439 m->private = tr; 7440 } else { 7441 trace_array_put(tr); 7442 } 7443 } 7444 return ret; 7445 } 7446 7447 static ssize_t tracing_err_log_write(struct file *file, 7448 const char __user *buffer, 7449 size_t count, loff_t *ppos) 7450 { 7451 return count; 7452 } 7453 7454 static int tracing_err_log_release(struct inode *inode, struct file *file) 7455 { 7456 struct trace_array *tr = inode->i_private; 7457 7458 trace_array_put(tr); 7459 7460 if (file->f_mode & FMODE_READ) 7461 seq_release(inode, file); 7462 7463 return 0; 7464 } 7465 7466 static const struct file_operations tracing_err_log_fops = { 7467 .open = tracing_err_log_open, 7468 .write = tracing_err_log_write, 7469 .read = seq_read, 7470 .llseek = seq_lseek, 7471 .release = tracing_err_log_release, 7472 }; 7473 7474 static int tracing_buffers_open(struct inode *inode, struct file *filp) 7475 { 7476 struct trace_array *tr = inode->i_private; 7477 struct ftrace_buffer_info *info; 7478 int ret; 7479 7480 ret = tracing_check_open_get_tr(tr); 7481 if (ret) 7482 return ret; 7483 7484 info = kvzalloc(sizeof(*info), GFP_KERNEL); 7485 if (!info) { 7486 trace_array_put(tr); 7487 return -ENOMEM; 7488 } 7489 7490 mutex_lock(&trace_types_lock); 7491 7492 info->iter.tr = tr; 7493 info->iter.cpu_file = tracing_get_cpu(inode); 7494 info->iter.trace = tr->current_trace; 7495 info->iter.array_buffer = &tr->array_buffer; 7496 info->spare = NULL; 7497 /* Force reading ring buffer for first read */ 7498 info->read = (unsigned int)-1; 7499 7500 filp->private_data = info; 7501 7502 tr->trace_ref++; 7503 7504 mutex_unlock(&trace_types_lock); 7505 7506 ret = nonseekable_open(inode, filp); 7507 if (ret < 0) 7508 trace_array_put(tr); 7509 7510 return ret; 7511 } 7512 7513 static __poll_t 7514 tracing_buffers_poll(struct file *filp, poll_table *poll_table) 7515 { 7516 struct ftrace_buffer_info *info = filp->private_data; 7517 struct trace_iterator *iter = &info->iter; 7518 7519 return trace_poll(iter, filp, poll_table); 7520 } 7521 7522 static ssize_t 7523 tracing_buffers_read(struct file *filp, char __user *ubuf, 7524 size_t count, loff_t *ppos) 7525 { 7526 struct ftrace_buffer_info *info = filp->private_data; 7527 struct trace_iterator *iter = &info->iter; 7528 ssize_t ret = 0; 7529 ssize_t size; 7530 7531 if (!count) 7532 return 0; 7533 7534 #ifdef CONFIG_TRACER_MAX_TRACE 7535 if (iter->snapshot && iter->tr->current_trace->use_max_tr) 7536 return -EBUSY; 7537 #endif 7538 7539 if (!info->spare) { 7540 info->spare = ring_buffer_alloc_read_page(iter->array_buffer->buffer, 7541 iter->cpu_file); 7542 if (IS_ERR(info->spare)) { 7543 ret = PTR_ERR(info->spare); 7544 info->spare = NULL; 7545 } else { 7546 info->spare_cpu = iter->cpu_file; 7547 } 7548 } 7549 if (!info->spare) 7550 return ret; 7551 7552 /* Do we have previous read data to read? */ 7553 if (info->read < PAGE_SIZE) 7554 goto read; 7555 7556 again: 7557 trace_access_lock(iter->cpu_file); 7558 ret = ring_buffer_read_page(iter->array_buffer->buffer, 7559 &info->spare, 7560 count, 7561 iter->cpu_file, 0); 7562 trace_access_unlock(iter->cpu_file); 7563 7564 if (ret < 0) { 7565 if (trace_empty(iter)) { 7566 if ((filp->f_flags & O_NONBLOCK)) 7567 return -EAGAIN; 7568 7569 ret = wait_on_pipe(iter, 0); 7570 if (ret) 7571 return ret; 7572 7573 goto again; 7574 } 7575 return 0; 7576 } 7577 7578 info->read = 0; 7579 read: 7580 size = PAGE_SIZE - info->read; 7581 if (size > count) 7582 size = count; 7583 7584 ret = copy_to_user(ubuf, info->spare + info->read, size); 7585 if (ret == size) 7586 return -EFAULT; 7587 7588 size -= ret; 7589 7590 *ppos += size; 7591 info->read += size; 7592 7593 return size; 7594 } 7595 7596 static int tracing_buffers_release(struct inode *inode, struct file *file) 7597 { 7598 struct ftrace_buffer_info *info = file->private_data; 7599 struct trace_iterator *iter = &info->iter; 7600 7601 mutex_lock(&trace_types_lock); 7602 7603 iter->tr->trace_ref--; 7604 7605 __trace_array_put(iter->tr); 7606 7607 if (info->spare) 7608 ring_buffer_free_read_page(iter->array_buffer->buffer, 7609 info->spare_cpu, info->spare); 7610 kvfree(info); 7611 7612 mutex_unlock(&trace_types_lock); 7613 7614 return 0; 7615 } 7616 7617 struct buffer_ref { 7618 struct trace_buffer *buffer; 7619 void *page; 7620 int cpu; 7621 refcount_t refcount; 7622 }; 7623 7624 static void buffer_ref_release(struct buffer_ref *ref) 7625 { 7626 if (!refcount_dec_and_test(&ref->refcount)) 7627 return; 7628 ring_buffer_free_read_page(ref->buffer, ref->cpu, ref->page); 7629 kfree(ref); 7630 } 7631 7632 static void buffer_pipe_buf_release(struct pipe_inode_info *pipe, 7633 struct pipe_buffer *buf) 7634 { 7635 struct buffer_ref *ref = (struct buffer_ref *)buf->private; 7636 7637 buffer_ref_release(ref); 7638 buf->private = 0; 7639 } 7640 7641 static bool buffer_pipe_buf_get(struct pipe_inode_info *pipe, 7642 struct pipe_buffer *buf) 7643 { 7644 struct buffer_ref *ref = (struct buffer_ref *)buf->private; 7645 7646 if (refcount_read(&ref->refcount) > INT_MAX/2) 7647 return false; 7648 7649 refcount_inc(&ref->refcount); 7650 return true; 7651 } 7652 7653 /* Pipe buffer operations for a buffer. */ 7654 static const struct pipe_buf_operations buffer_pipe_buf_ops = { 7655 .release = buffer_pipe_buf_release, 7656 .get = buffer_pipe_buf_get, 7657 }; 7658 7659 /* 7660 * Callback from splice_to_pipe(), if we need to release some pages 7661 * at the end of the spd in case we error'ed out in filling the pipe. 7662 */ 7663 static void buffer_spd_release(struct splice_pipe_desc *spd, unsigned int i) 7664 { 7665 struct buffer_ref *ref = 7666 (struct buffer_ref *)spd->partial[i].private; 7667 7668 buffer_ref_release(ref); 7669 spd->partial[i].private = 0; 7670 } 7671 7672 static ssize_t 7673 tracing_buffers_splice_read(struct file *file, loff_t *ppos, 7674 struct pipe_inode_info *pipe, size_t len, 7675 unsigned int flags) 7676 { 7677 struct ftrace_buffer_info *info = file->private_data; 7678 struct trace_iterator *iter = &info->iter; 7679 struct partial_page partial_def[PIPE_DEF_BUFFERS]; 7680 struct page *pages_def[PIPE_DEF_BUFFERS]; 7681 struct splice_pipe_desc spd = { 7682 .pages = pages_def, 7683 .partial = partial_def, 7684 .nr_pages_max = PIPE_DEF_BUFFERS, 7685 .ops = &buffer_pipe_buf_ops, 7686 .spd_release = buffer_spd_release, 7687 }; 7688 struct buffer_ref *ref; 7689 int entries, i; 7690 ssize_t ret = 0; 7691 7692 #ifdef CONFIG_TRACER_MAX_TRACE 7693 if (iter->snapshot && iter->tr->current_trace->use_max_tr) 7694 return -EBUSY; 7695 #endif 7696 7697 if (*ppos & (PAGE_SIZE - 1)) 7698 return -EINVAL; 7699 7700 if (len & (PAGE_SIZE - 1)) { 7701 if (len < PAGE_SIZE) 7702 return -EINVAL; 7703 len &= PAGE_MASK; 7704 } 7705 7706 if (splice_grow_spd(pipe, &spd)) 7707 return -ENOMEM; 7708 7709 again: 7710 trace_access_lock(iter->cpu_file); 7711 entries = ring_buffer_entries_cpu(iter->array_buffer->buffer, iter->cpu_file); 7712 7713 for (i = 0; i < spd.nr_pages_max && len && entries; i++, len -= PAGE_SIZE) { 7714 struct page *page; 7715 int r; 7716 7717 ref = kzalloc(sizeof(*ref), GFP_KERNEL); 7718 if (!ref) { 7719 ret = -ENOMEM; 7720 break; 7721 } 7722 7723 refcount_set(&ref->refcount, 1); 7724 ref->buffer = iter->array_buffer->buffer; 7725 ref->page = ring_buffer_alloc_read_page(ref->buffer, iter->cpu_file); 7726 if (IS_ERR(ref->page)) { 7727 ret = PTR_ERR(ref->page); 7728 ref->page = NULL; 7729 kfree(ref); 7730 break; 7731 } 7732 ref->cpu = iter->cpu_file; 7733 7734 r = ring_buffer_read_page(ref->buffer, &ref->page, 7735 len, iter->cpu_file, 1); 7736 if (r < 0) { 7737 ring_buffer_free_read_page(ref->buffer, ref->cpu, 7738 ref->page); 7739 kfree(ref); 7740 break; 7741 } 7742 7743 page = virt_to_page(ref->page); 7744 7745 spd.pages[i] = page; 7746 spd.partial[i].len = PAGE_SIZE; 7747 spd.partial[i].offset = 0; 7748 spd.partial[i].private = (unsigned long)ref; 7749 spd.nr_pages++; 7750 *ppos += PAGE_SIZE; 7751 7752 entries = ring_buffer_entries_cpu(iter->array_buffer->buffer, iter->cpu_file); 7753 } 7754 7755 trace_access_unlock(iter->cpu_file); 7756 spd.nr_pages = i; 7757 7758 /* did we read anything? */ 7759 if (!spd.nr_pages) { 7760 if (ret) 7761 goto out; 7762 7763 ret = -EAGAIN; 7764 if ((file->f_flags & O_NONBLOCK) || (flags & SPLICE_F_NONBLOCK)) 7765 goto out; 7766 7767 ret = wait_on_pipe(iter, iter->tr->buffer_percent); 7768 if (ret) 7769 goto out; 7770 7771 goto again; 7772 } 7773 7774 ret = splice_to_pipe(pipe, &spd); 7775 out: 7776 splice_shrink_spd(&spd); 7777 7778 return ret; 7779 } 7780 7781 static const struct file_operations tracing_buffers_fops = { 7782 .open = tracing_buffers_open, 7783 .read = tracing_buffers_read, 7784 .poll = tracing_buffers_poll, 7785 .release = tracing_buffers_release, 7786 .splice_read = tracing_buffers_splice_read, 7787 .llseek = no_llseek, 7788 }; 7789 7790 static ssize_t 7791 tracing_stats_read(struct file *filp, char __user *ubuf, 7792 size_t count, loff_t *ppos) 7793 { 7794 struct inode *inode = file_inode(filp); 7795 struct trace_array *tr = inode->i_private; 7796 struct array_buffer *trace_buf = &tr->array_buffer; 7797 int cpu = tracing_get_cpu(inode); 7798 struct trace_seq *s; 7799 unsigned long cnt; 7800 unsigned long long t; 7801 unsigned long usec_rem; 7802 7803 s = kmalloc(sizeof(*s), GFP_KERNEL); 7804 if (!s) 7805 return -ENOMEM; 7806 7807 trace_seq_init(s); 7808 7809 cnt = ring_buffer_entries_cpu(trace_buf->buffer, cpu); 7810 trace_seq_printf(s, "entries: %ld\n", cnt); 7811 7812 cnt = ring_buffer_overrun_cpu(trace_buf->buffer, cpu); 7813 trace_seq_printf(s, "overrun: %ld\n", cnt); 7814 7815 cnt = ring_buffer_commit_overrun_cpu(trace_buf->buffer, cpu); 7816 trace_seq_printf(s, "commit overrun: %ld\n", cnt); 7817 7818 cnt = ring_buffer_bytes_cpu(trace_buf->buffer, cpu); 7819 trace_seq_printf(s, "bytes: %ld\n", cnt); 7820 7821 if (trace_clocks[tr->clock_id].in_ns) { 7822 /* local or global for trace_clock */ 7823 t = ns2usecs(ring_buffer_oldest_event_ts(trace_buf->buffer, cpu)); 7824 usec_rem = do_div(t, USEC_PER_SEC); 7825 trace_seq_printf(s, "oldest event ts: %5llu.%06lu\n", 7826 t, usec_rem); 7827 7828 t = ns2usecs(ring_buffer_time_stamp(trace_buf->buffer, cpu)); 7829 usec_rem = do_div(t, USEC_PER_SEC); 7830 trace_seq_printf(s, "now ts: %5llu.%06lu\n", t, usec_rem); 7831 } else { 7832 /* counter or tsc mode for trace_clock */ 7833 trace_seq_printf(s, "oldest event ts: %llu\n", 7834 ring_buffer_oldest_event_ts(trace_buf->buffer, cpu)); 7835 7836 trace_seq_printf(s, "now ts: %llu\n", 7837 ring_buffer_time_stamp(trace_buf->buffer, cpu)); 7838 } 7839 7840 cnt = ring_buffer_dropped_events_cpu(trace_buf->buffer, cpu); 7841 trace_seq_printf(s, "dropped events: %ld\n", cnt); 7842 7843 cnt = ring_buffer_read_events_cpu(trace_buf->buffer, cpu); 7844 trace_seq_printf(s, "read events: %ld\n", cnt); 7845 7846 count = simple_read_from_buffer(ubuf, count, ppos, 7847 s->buffer, trace_seq_used(s)); 7848 7849 kfree(s); 7850 7851 return count; 7852 } 7853 7854 static const struct file_operations tracing_stats_fops = { 7855 .open = tracing_open_generic_tr, 7856 .read = tracing_stats_read, 7857 .llseek = generic_file_llseek, 7858 .release = tracing_release_generic_tr, 7859 }; 7860 7861 #ifdef CONFIG_DYNAMIC_FTRACE 7862 7863 static ssize_t 7864 tracing_read_dyn_info(struct file *filp, char __user *ubuf, 7865 size_t cnt, loff_t *ppos) 7866 { 7867 ssize_t ret; 7868 char *buf; 7869 int r; 7870 7871 /* 256 should be plenty to hold the amount needed */ 7872 buf = kmalloc(256, GFP_KERNEL); 7873 if (!buf) 7874 return -ENOMEM; 7875 7876 r = scnprintf(buf, 256, "%ld pages:%ld groups: %ld\n", 7877 ftrace_update_tot_cnt, 7878 ftrace_number_of_pages, 7879 ftrace_number_of_groups); 7880 7881 ret = simple_read_from_buffer(ubuf, cnt, ppos, buf, r); 7882 kfree(buf); 7883 return ret; 7884 } 7885 7886 static const struct file_operations tracing_dyn_info_fops = { 7887 .open = tracing_open_generic, 7888 .read = tracing_read_dyn_info, 7889 .llseek = generic_file_llseek, 7890 }; 7891 #endif /* CONFIG_DYNAMIC_FTRACE */ 7892 7893 #if defined(CONFIG_TRACER_SNAPSHOT) && defined(CONFIG_DYNAMIC_FTRACE) 7894 static void 7895 ftrace_snapshot(unsigned long ip, unsigned long parent_ip, 7896 struct trace_array *tr, struct ftrace_probe_ops *ops, 7897 void *data) 7898 { 7899 tracing_snapshot_instance(tr); 7900 } 7901 7902 static void 7903 ftrace_count_snapshot(unsigned long ip, unsigned long parent_ip, 7904 struct trace_array *tr, struct ftrace_probe_ops *ops, 7905 void *data) 7906 { 7907 struct ftrace_func_mapper *mapper = data; 7908 long *count = NULL; 7909 7910 if (mapper) 7911 count = (long *)ftrace_func_mapper_find_ip(mapper, ip); 7912 7913 if (count) { 7914 7915 if (*count <= 0) 7916 return; 7917 7918 (*count)--; 7919 } 7920 7921 tracing_snapshot_instance(tr); 7922 } 7923 7924 static int 7925 ftrace_snapshot_print(struct seq_file *m, unsigned long ip, 7926 struct ftrace_probe_ops *ops, void *data) 7927 { 7928 struct ftrace_func_mapper *mapper = data; 7929 long *count = NULL; 7930 7931 seq_printf(m, "%ps:", (void *)ip); 7932 7933 seq_puts(m, "snapshot"); 7934 7935 if (mapper) 7936 count = (long *)ftrace_func_mapper_find_ip(mapper, ip); 7937 7938 if (count) 7939 seq_printf(m, ":count=%ld\n", *count); 7940 else 7941 seq_puts(m, ":unlimited\n"); 7942 7943 return 0; 7944 } 7945 7946 static int 7947 ftrace_snapshot_init(struct ftrace_probe_ops *ops, struct trace_array *tr, 7948 unsigned long ip, void *init_data, void **data) 7949 { 7950 struct ftrace_func_mapper *mapper = *data; 7951 7952 if (!mapper) { 7953 mapper = allocate_ftrace_func_mapper(); 7954 if (!mapper) 7955 return -ENOMEM; 7956 *data = mapper; 7957 } 7958 7959 return ftrace_func_mapper_add_ip(mapper, ip, init_data); 7960 } 7961 7962 static void 7963 ftrace_snapshot_free(struct ftrace_probe_ops *ops, struct trace_array *tr, 7964 unsigned long ip, void *data) 7965 { 7966 struct ftrace_func_mapper *mapper = data; 7967 7968 if (!ip) { 7969 if (!mapper) 7970 return; 7971 free_ftrace_func_mapper(mapper, NULL); 7972 return; 7973 } 7974 7975 ftrace_func_mapper_remove_ip(mapper, ip); 7976 } 7977 7978 static struct ftrace_probe_ops snapshot_probe_ops = { 7979 .func = ftrace_snapshot, 7980 .print = ftrace_snapshot_print, 7981 }; 7982 7983 static struct ftrace_probe_ops snapshot_count_probe_ops = { 7984 .func = ftrace_count_snapshot, 7985 .print = ftrace_snapshot_print, 7986 .init = ftrace_snapshot_init, 7987 .free = ftrace_snapshot_free, 7988 }; 7989 7990 static int 7991 ftrace_trace_snapshot_callback(struct trace_array *tr, struct ftrace_hash *hash, 7992 char *glob, char *cmd, char *param, int enable) 7993 { 7994 struct ftrace_probe_ops *ops; 7995 void *count = (void *)-1; 7996 char *number; 7997 int ret; 7998 7999 if (!tr) 8000 return -ENODEV; 8001 8002 /* hash funcs only work with set_ftrace_filter */ 8003 if (!enable) 8004 return -EINVAL; 8005 8006 ops = param ? &snapshot_count_probe_ops : &snapshot_probe_ops; 8007 8008 if (glob[0] == '!') 8009 return unregister_ftrace_function_probe_func(glob+1, tr, ops); 8010 8011 if (!param) 8012 goto out_reg; 8013 8014 number = strsep(¶m, ":"); 8015 8016 if (!strlen(number)) 8017 goto out_reg; 8018 8019 /* 8020 * We use the callback data field (which is a pointer) 8021 * as our counter. 8022 */ 8023 ret = kstrtoul(number, 0, (unsigned long *)&count); 8024 if (ret) 8025 return ret; 8026 8027 out_reg: 8028 ret = tracing_alloc_snapshot_instance(tr); 8029 if (ret < 0) 8030 goto out; 8031 8032 ret = register_ftrace_function_probe(glob, tr, ops, count); 8033 8034 out: 8035 return ret < 0 ? ret : 0; 8036 } 8037 8038 static struct ftrace_func_command ftrace_snapshot_cmd = { 8039 .name = "snapshot", 8040 .func = ftrace_trace_snapshot_callback, 8041 }; 8042 8043 static __init int register_snapshot_cmd(void) 8044 { 8045 return register_ftrace_command(&ftrace_snapshot_cmd); 8046 } 8047 #else 8048 static inline __init int register_snapshot_cmd(void) { return 0; } 8049 #endif /* defined(CONFIG_TRACER_SNAPSHOT) && defined(CONFIG_DYNAMIC_FTRACE) */ 8050 8051 static struct dentry *tracing_get_dentry(struct trace_array *tr) 8052 { 8053 if (WARN_ON(!tr->dir)) 8054 return ERR_PTR(-ENODEV); 8055 8056 /* Top directory uses NULL as the parent */ 8057 if (tr->flags & TRACE_ARRAY_FL_GLOBAL) 8058 return NULL; 8059 8060 /* All sub buffers have a descriptor */ 8061 return tr->dir; 8062 } 8063 8064 static struct dentry *tracing_dentry_percpu(struct trace_array *tr, int cpu) 8065 { 8066 struct dentry *d_tracer; 8067 8068 if (tr->percpu_dir) 8069 return tr->percpu_dir; 8070 8071 d_tracer = tracing_get_dentry(tr); 8072 if (IS_ERR(d_tracer)) 8073 return NULL; 8074 8075 tr->percpu_dir = tracefs_create_dir("per_cpu", d_tracer); 8076 8077 MEM_FAIL(!tr->percpu_dir, 8078 "Could not create tracefs directory 'per_cpu/%d'\n", cpu); 8079 8080 return tr->percpu_dir; 8081 } 8082 8083 static struct dentry * 8084 trace_create_cpu_file(const char *name, umode_t mode, struct dentry *parent, 8085 void *data, long cpu, const struct file_operations *fops) 8086 { 8087 struct dentry *ret = trace_create_file(name, mode, parent, data, fops); 8088 8089 if (ret) /* See tracing_get_cpu() */ 8090 d_inode(ret)->i_cdev = (void *)(cpu + 1); 8091 return ret; 8092 } 8093 8094 static void 8095 tracing_init_tracefs_percpu(struct trace_array *tr, long cpu) 8096 { 8097 struct dentry *d_percpu = tracing_dentry_percpu(tr, cpu); 8098 struct dentry *d_cpu; 8099 char cpu_dir[30]; /* 30 characters should be more than enough */ 8100 8101 if (!d_percpu) 8102 return; 8103 8104 snprintf(cpu_dir, 30, "cpu%ld", cpu); 8105 d_cpu = tracefs_create_dir(cpu_dir, d_percpu); 8106 if (!d_cpu) { 8107 pr_warn("Could not create tracefs '%s' entry\n", cpu_dir); 8108 return; 8109 } 8110 8111 /* per cpu trace_pipe */ 8112 trace_create_cpu_file("trace_pipe", 0444, d_cpu, 8113 tr, cpu, &tracing_pipe_fops); 8114 8115 /* per cpu trace */ 8116 trace_create_cpu_file("trace", 0644, d_cpu, 8117 tr, cpu, &tracing_fops); 8118 8119 trace_create_cpu_file("trace_pipe_raw", 0444, d_cpu, 8120 tr, cpu, &tracing_buffers_fops); 8121 8122 trace_create_cpu_file("stats", 0444, d_cpu, 8123 tr, cpu, &tracing_stats_fops); 8124 8125 trace_create_cpu_file("buffer_size_kb", 0444, d_cpu, 8126 tr, cpu, &tracing_entries_fops); 8127 8128 #ifdef CONFIG_TRACER_SNAPSHOT 8129 trace_create_cpu_file("snapshot", 0644, d_cpu, 8130 tr, cpu, &snapshot_fops); 8131 8132 trace_create_cpu_file("snapshot_raw", 0444, d_cpu, 8133 tr, cpu, &snapshot_raw_fops); 8134 #endif 8135 } 8136 8137 #ifdef CONFIG_FTRACE_SELFTEST 8138 /* Let selftest have access to static functions in this file */ 8139 #include "trace_selftest.c" 8140 #endif 8141 8142 static ssize_t 8143 trace_options_read(struct file *filp, char __user *ubuf, size_t cnt, 8144 loff_t *ppos) 8145 { 8146 struct trace_option_dentry *topt = filp->private_data; 8147 char *buf; 8148 8149 if (topt->flags->val & topt->opt->bit) 8150 buf = "1\n"; 8151 else 8152 buf = "0\n"; 8153 8154 return simple_read_from_buffer(ubuf, cnt, ppos, buf, 2); 8155 } 8156 8157 static ssize_t 8158 trace_options_write(struct file *filp, const char __user *ubuf, size_t cnt, 8159 loff_t *ppos) 8160 { 8161 struct trace_option_dentry *topt = filp->private_data; 8162 unsigned long val; 8163 int ret; 8164 8165 ret = kstrtoul_from_user(ubuf, cnt, 10, &val); 8166 if (ret) 8167 return ret; 8168 8169 if (val != 0 && val != 1) 8170 return -EINVAL; 8171 8172 if (!!(topt->flags->val & topt->opt->bit) != val) { 8173 mutex_lock(&trace_types_lock); 8174 ret = __set_tracer_option(topt->tr, topt->flags, 8175 topt->opt, !val); 8176 mutex_unlock(&trace_types_lock); 8177 if (ret) 8178 return ret; 8179 } 8180 8181 *ppos += cnt; 8182 8183 return cnt; 8184 } 8185 8186 8187 static const struct file_operations trace_options_fops = { 8188 .open = tracing_open_generic, 8189 .read = trace_options_read, 8190 .write = trace_options_write, 8191 .llseek = generic_file_llseek, 8192 }; 8193 8194 /* 8195 * In order to pass in both the trace_array descriptor as well as the index 8196 * to the flag that the trace option file represents, the trace_array 8197 * has a character array of trace_flags_index[], which holds the index 8198 * of the bit for the flag it represents. index[0] == 0, index[1] == 1, etc. 8199 * The address of this character array is passed to the flag option file 8200 * read/write callbacks. 8201 * 8202 * In order to extract both the index and the trace_array descriptor, 8203 * get_tr_index() uses the following algorithm. 8204 * 8205 * idx = *ptr; 8206 * 8207 * As the pointer itself contains the address of the index (remember 8208 * index[1] == 1). 8209 * 8210 * Then to get the trace_array descriptor, by subtracting that index 8211 * from the ptr, we get to the start of the index itself. 8212 * 8213 * ptr - idx == &index[0] 8214 * 8215 * Then a simple container_of() from that pointer gets us to the 8216 * trace_array descriptor. 8217 */ 8218 static void get_tr_index(void *data, struct trace_array **ptr, 8219 unsigned int *pindex) 8220 { 8221 *pindex = *(unsigned char *)data; 8222 8223 *ptr = container_of(data - *pindex, struct trace_array, 8224 trace_flags_index); 8225 } 8226 8227 static ssize_t 8228 trace_options_core_read(struct file *filp, char __user *ubuf, size_t cnt, 8229 loff_t *ppos) 8230 { 8231 void *tr_index = filp->private_data; 8232 struct trace_array *tr; 8233 unsigned int index; 8234 char *buf; 8235 8236 get_tr_index(tr_index, &tr, &index); 8237 8238 if (tr->trace_flags & (1 << index)) 8239 buf = "1\n"; 8240 else 8241 buf = "0\n"; 8242 8243 return simple_read_from_buffer(ubuf, cnt, ppos, buf, 2); 8244 } 8245 8246 static ssize_t 8247 trace_options_core_write(struct file *filp, const char __user *ubuf, size_t cnt, 8248 loff_t *ppos) 8249 { 8250 void *tr_index = filp->private_data; 8251 struct trace_array *tr; 8252 unsigned int index; 8253 unsigned long val; 8254 int ret; 8255 8256 get_tr_index(tr_index, &tr, &index); 8257 8258 ret = kstrtoul_from_user(ubuf, cnt, 10, &val); 8259 if (ret) 8260 return ret; 8261 8262 if (val != 0 && val != 1) 8263 return -EINVAL; 8264 8265 mutex_lock(&event_mutex); 8266 mutex_lock(&trace_types_lock); 8267 ret = set_tracer_flag(tr, 1 << index, val); 8268 mutex_unlock(&trace_types_lock); 8269 mutex_unlock(&event_mutex); 8270 8271 if (ret < 0) 8272 return ret; 8273 8274 *ppos += cnt; 8275 8276 return cnt; 8277 } 8278 8279 static const struct file_operations trace_options_core_fops = { 8280 .open = tracing_open_generic, 8281 .read = trace_options_core_read, 8282 .write = trace_options_core_write, 8283 .llseek = generic_file_llseek, 8284 }; 8285 8286 struct dentry *trace_create_file(const char *name, 8287 umode_t mode, 8288 struct dentry *parent, 8289 void *data, 8290 const struct file_operations *fops) 8291 { 8292 struct dentry *ret; 8293 8294 ret = tracefs_create_file(name, mode, parent, data, fops); 8295 if (!ret) 8296 pr_warn("Could not create tracefs '%s' entry\n", name); 8297 8298 return ret; 8299 } 8300 8301 8302 static struct dentry *trace_options_init_dentry(struct trace_array *tr) 8303 { 8304 struct dentry *d_tracer; 8305 8306 if (tr->options) 8307 return tr->options; 8308 8309 d_tracer = tracing_get_dentry(tr); 8310 if (IS_ERR(d_tracer)) 8311 return NULL; 8312 8313 tr->options = tracefs_create_dir("options", d_tracer); 8314 if (!tr->options) { 8315 pr_warn("Could not create tracefs directory 'options'\n"); 8316 return NULL; 8317 } 8318 8319 return tr->options; 8320 } 8321 8322 static void 8323 create_trace_option_file(struct trace_array *tr, 8324 struct trace_option_dentry *topt, 8325 struct tracer_flags *flags, 8326 struct tracer_opt *opt) 8327 { 8328 struct dentry *t_options; 8329 8330 t_options = trace_options_init_dentry(tr); 8331 if (!t_options) 8332 return; 8333 8334 topt->flags = flags; 8335 topt->opt = opt; 8336 topt->tr = tr; 8337 8338 topt->entry = trace_create_file(opt->name, 0644, t_options, topt, 8339 &trace_options_fops); 8340 8341 } 8342 8343 static void 8344 create_trace_option_files(struct trace_array *tr, struct tracer *tracer) 8345 { 8346 struct trace_option_dentry *topts; 8347 struct trace_options *tr_topts; 8348 struct tracer_flags *flags; 8349 struct tracer_opt *opts; 8350 int cnt; 8351 int i; 8352 8353 if (!tracer) 8354 return; 8355 8356 flags = tracer->flags; 8357 8358 if (!flags || !flags->opts) 8359 return; 8360 8361 /* 8362 * If this is an instance, only create flags for tracers 8363 * the instance may have. 8364 */ 8365 if (!trace_ok_for_array(tracer, tr)) 8366 return; 8367 8368 for (i = 0; i < tr->nr_topts; i++) { 8369 /* Make sure there's no duplicate flags. */ 8370 if (WARN_ON_ONCE(tr->topts[i].tracer->flags == tracer->flags)) 8371 return; 8372 } 8373 8374 opts = flags->opts; 8375 8376 for (cnt = 0; opts[cnt].name; cnt++) 8377 ; 8378 8379 topts = kcalloc(cnt + 1, sizeof(*topts), GFP_KERNEL); 8380 if (!topts) 8381 return; 8382 8383 tr_topts = krealloc(tr->topts, sizeof(*tr->topts) * (tr->nr_topts + 1), 8384 GFP_KERNEL); 8385 if (!tr_topts) { 8386 kfree(topts); 8387 return; 8388 } 8389 8390 tr->topts = tr_topts; 8391 tr->topts[tr->nr_topts].tracer = tracer; 8392 tr->topts[tr->nr_topts].topts = topts; 8393 tr->nr_topts++; 8394 8395 for (cnt = 0; opts[cnt].name; cnt++) { 8396 create_trace_option_file(tr, &topts[cnt], flags, 8397 &opts[cnt]); 8398 MEM_FAIL(topts[cnt].entry == NULL, 8399 "Failed to create trace option: %s", 8400 opts[cnt].name); 8401 } 8402 } 8403 8404 static struct dentry * 8405 create_trace_option_core_file(struct trace_array *tr, 8406 const char *option, long index) 8407 { 8408 struct dentry *t_options; 8409 8410 t_options = trace_options_init_dentry(tr); 8411 if (!t_options) 8412 return NULL; 8413 8414 return trace_create_file(option, 0644, t_options, 8415 (void *)&tr->trace_flags_index[index], 8416 &trace_options_core_fops); 8417 } 8418 8419 static void create_trace_options_dir(struct trace_array *tr) 8420 { 8421 struct dentry *t_options; 8422 bool top_level = tr == &global_trace; 8423 int i; 8424 8425 t_options = trace_options_init_dentry(tr); 8426 if (!t_options) 8427 return; 8428 8429 for (i = 0; trace_options[i]; i++) { 8430 if (top_level || 8431 !((1 << i) & TOP_LEVEL_TRACE_FLAGS)) 8432 create_trace_option_core_file(tr, trace_options[i], i); 8433 } 8434 } 8435 8436 static ssize_t 8437 rb_simple_read(struct file *filp, char __user *ubuf, 8438 size_t cnt, loff_t *ppos) 8439 { 8440 struct trace_array *tr = filp->private_data; 8441 char buf[64]; 8442 int r; 8443 8444 r = tracer_tracing_is_on(tr); 8445 r = sprintf(buf, "%d\n", r); 8446 8447 return simple_read_from_buffer(ubuf, cnt, ppos, buf, r); 8448 } 8449 8450 static ssize_t 8451 rb_simple_write(struct file *filp, const char __user *ubuf, 8452 size_t cnt, loff_t *ppos) 8453 { 8454 struct trace_array *tr = filp->private_data; 8455 struct trace_buffer *buffer = tr->array_buffer.buffer; 8456 unsigned long val; 8457 int ret; 8458 8459 ret = kstrtoul_from_user(ubuf, cnt, 10, &val); 8460 if (ret) 8461 return ret; 8462 8463 if (buffer) { 8464 mutex_lock(&trace_types_lock); 8465 if (!!val == tracer_tracing_is_on(tr)) { 8466 val = 0; /* do nothing */ 8467 } else if (val) { 8468 tracer_tracing_on(tr); 8469 if (tr->current_trace->start) 8470 tr->current_trace->start(tr); 8471 } else { 8472 tracer_tracing_off(tr); 8473 if (tr->current_trace->stop) 8474 tr->current_trace->stop(tr); 8475 } 8476 mutex_unlock(&trace_types_lock); 8477 } 8478 8479 (*ppos)++; 8480 8481 return cnt; 8482 } 8483 8484 static const struct file_operations rb_simple_fops = { 8485 .open = tracing_open_generic_tr, 8486 .read = rb_simple_read, 8487 .write = rb_simple_write, 8488 .release = tracing_release_generic_tr, 8489 .llseek = default_llseek, 8490 }; 8491 8492 static ssize_t 8493 buffer_percent_read(struct file *filp, char __user *ubuf, 8494 size_t cnt, loff_t *ppos) 8495 { 8496 struct trace_array *tr = filp->private_data; 8497 char buf[64]; 8498 int r; 8499 8500 r = tr->buffer_percent; 8501 r = sprintf(buf, "%d\n", r); 8502 8503 return simple_read_from_buffer(ubuf, cnt, ppos, buf, r); 8504 } 8505 8506 static ssize_t 8507 buffer_percent_write(struct file *filp, const char __user *ubuf, 8508 size_t cnt, loff_t *ppos) 8509 { 8510 struct trace_array *tr = filp->private_data; 8511 unsigned long val; 8512 int ret; 8513 8514 ret = kstrtoul_from_user(ubuf, cnt, 10, &val); 8515 if (ret) 8516 return ret; 8517 8518 if (val > 100) 8519 return -EINVAL; 8520 8521 if (!val) 8522 val = 1; 8523 8524 tr->buffer_percent = val; 8525 8526 (*ppos)++; 8527 8528 return cnt; 8529 } 8530 8531 static const struct file_operations buffer_percent_fops = { 8532 .open = tracing_open_generic_tr, 8533 .read = buffer_percent_read, 8534 .write = buffer_percent_write, 8535 .release = tracing_release_generic_tr, 8536 .llseek = default_llseek, 8537 }; 8538 8539 static struct dentry *trace_instance_dir; 8540 8541 static void 8542 init_tracer_tracefs(struct trace_array *tr, struct dentry *d_tracer); 8543 8544 static int 8545 allocate_trace_buffer(struct trace_array *tr, struct array_buffer *buf, int size) 8546 { 8547 enum ring_buffer_flags rb_flags; 8548 8549 rb_flags = tr->trace_flags & TRACE_ITER_OVERWRITE ? RB_FL_OVERWRITE : 0; 8550 8551 buf->tr = tr; 8552 8553 buf->buffer = ring_buffer_alloc(size, rb_flags); 8554 if (!buf->buffer) 8555 return -ENOMEM; 8556 8557 buf->data = alloc_percpu(struct trace_array_cpu); 8558 if (!buf->data) { 8559 ring_buffer_free(buf->buffer); 8560 buf->buffer = NULL; 8561 return -ENOMEM; 8562 } 8563 8564 /* Allocate the first page for all buffers */ 8565 set_buffer_entries(&tr->array_buffer, 8566 ring_buffer_size(tr->array_buffer.buffer, 0)); 8567 8568 return 0; 8569 } 8570 8571 static int allocate_trace_buffers(struct trace_array *tr, int size) 8572 { 8573 int ret; 8574 8575 ret = allocate_trace_buffer(tr, &tr->array_buffer, size); 8576 if (ret) 8577 return ret; 8578 8579 #ifdef CONFIG_TRACER_MAX_TRACE 8580 ret = allocate_trace_buffer(tr, &tr->max_buffer, 8581 allocate_snapshot ? size : 1); 8582 if (MEM_FAIL(ret, "Failed to allocate trace buffer\n")) { 8583 ring_buffer_free(tr->array_buffer.buffer); 8584 tr->array_buffer.buffer = NULL; 8585 free_percpu(tr->array_buffer.data); 8586 tr->array_buffer.data = NULL; 8587 return -ENOMEM; 8588 } 8589 tr->allocated_snapshot = allocate_snapshot; 8590 8591 /* 8592 * Only the top level trace array gets its snapshot allocated 8593 * from the kernel command line. 8594 */ 8595 allocate_snapshot = false; 8596 #endif 8597 8598 return 0; 8599 } 8600 8601 static void free_trace_buffer(struct array_buffer *buf) 8602 { 8603 if (buf->buffer) { 8604 ring_buffer_free(buf->buffer); 8605 buf->buffer = NULL; 8606 free_percpu(buf->data); 8607 buf->data = NULL; 8608 } 8609 } 8610 8611 static void free_trace_buffers(struct trace_array *tr) 8612 { 8613 if (!tr) 8614 return; 8615 8616 free_trace_buffer(&tr->array_buffer); 8617 8618 #ifdef CONFIG_TRACER_MAX_TRACE 8619 free_trace_buffer(&tr->max_buffer); 8620 #endif 8621 } 8622 8623 static void init_trace_flags_index(struct trace_array *tr) 8624 { 8625 int i; 8626 8627 /* Used by the trace options files */ 8628 for (i = 0; i < TRACE_FLAGS_MAX_SIZE; i++) 8629 tr->trace_flags_index[i] = i; 8630 } 8631 8632 static void __update_tracer_options(struct trace_array *tr) 8633 { 8634 struct tracer *t; 8635 8636 for (t = trace_types; t; t = t->next) 8637 add_tracer_options(tr, t); 8638 } 8639 8640 static void update_tracer_options(struct trace_array *tr) 8641 { 8642 mutex_lock(&trace_types_lock); 8643 __update_tracer_options(tr); 8644 mutex_unlock(&trace_types_lock); 8645 } 8646 8647 /* Must have trace_types_lock held */ 8648 struct trace_array *trace_array_find(const char *instance) 8649 { 8650 struct trace_array *tr, *found = NULL; 8651 8652 list_for_each_entry(tr, &ftrace_trace_arrays, list) { 8653 if (tr->name && strcmp(tr->name, instance) == 0) { 8654 found = tr; 8655 break; 8656 } 8657 } 8658 8659 return found; 8660 } 8661 8662 struct trace_array *trace_array_find_get(const char *instance) 8663 { 8664 struct trace_array *tr; 8665 8666 mutex_lock(&trace_types_lock); 8667 tr = trace_array_find(instance); 8668 if (tr) 8669 tr->ref++; 8670 mutex_unlock(&trace_types_lock); 8671 8672 return tr; 8673 } 8674 8675 static int trace_array_create_dir(struct trace_array *tr) 8676 { 8677 int ret; 8678 8679 tr->dir = tracefs_create_dir(tr->name, trace_instance_dir); 8680 if (!tr->dir) 8681 return -EINVAL; 8682 8683 ret = event_trace_add_tracer(tr->dir, tr); 8684 if (ret) 8685 tracefs_remove(tr->dir); 8686 8687 init_tracer_tracefs(tr, tr->dir); 8688 __update_tracer_options(tr); 8689 8690 return ret; 8691 } 8692 8693 static struct trace_array *trace_array_create(const char *name) 8694 { 8695 struct trace_array *tr; 8696 int ret; 8697 8698 ret = -ENOMEM; 8699 tr = kzalloc(sizeof(*tr), GFP_KERNEL); 8700 if (!tr) 8701 return ERR_PTR(ret); 8702 8703 tr->name = kstrdup(name, GFP_KERNEL); 8704 if (!tr->name) 8705 goto out_free_tr; 8706 8707 if (!alloc_cpumask_var(&tr->tracing_cpumask, GFP_KERNEL)) 8708 goto out_free_tr; 8709 8710 tr->trace_flags = global_trace.trace_flags & ~ZEROED_TRACE_FLAGS; 8711 8712 cpumask_copy(tr->tracing_cpumask, cpu_all_mask); 8713 8714 raw_spin_lock_init(&tr->start_lock); 8715 8716 tr->max_lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED; 8717 8718 tr->current_trace = &nop_trace; 8719 8720 INIT_LIST_HEAD(&tr->systems); 8721 INIT_LIST_HEAD(&tr->events); 8722 INIT_LIST_HEAD(&tr->hist_vars); 8723 INIT_LIST_HEAD(&tr->err_log); 8724 8725 if (allocate_trace_buffers(tr, trace_buf_size) < 0) 8726 goto out_free_tr; 8727 8728 if (ftrace_allocate_ftrace_ops(tr) < 0) 8729 goto out_free_tr; 8730 8731 ftrace_init_trace_array(tr); 8732 8733 init_trace_flags_index(tr); 8734 8735 if (trace_instance_dir) { 8736 ret = trace_array_create_dir(tr); 8737 if (ret) 8738 goto out_free_tr; 8739 } else 8740 __trace_early_add_events(tr); 8741 8742 list_add(&tr->list, &ftrace_trace_arrays); 8743 8744 tr->ref++; 8745 8746 return tr; 8747 8748 out_free_tr: 8749 ftrace_free_ftrace_ops(tr); 8750 free_trace_buffers(tr); 8751 free_cpumask_var(tr->tracing_cpumask); 8752 kfree(tr->name); 8753 kfree(tr); 8754 8755 return ERR_PTR(ret); 8756 } 8757 8758 static int instance_mkdir(const char *name) 8759 { 8760 struct trace_array *tr; 8761 int ret; 8762 8763 mutex_lock(&event_mutex); 8764 mutex_lock(&trace_types_lock); 8765 8766 ret = -EEXIST; 8767 if (trace_array_find(name)) 8768 goto out_unlock; 8769 8770 tr = trace_array_create(name); 8771 8772 ret = PTR_ERR_OR_ZERO(tr); 8773 8774 out_unlock: 8775 mutex_unlock(&trace_types_lock); 8776 mutex_unlock(&event_mutex); 8777 return ret; 8778 } 8779 8780 /** 8781 * trace_array_get_by_name - Create/Lookup a trace array, given its name. 8782 * @name: The name of the trace array to be looked up/created. 8783 * 8784 * Returns pointer to trace array with given name. 8785 * NULL, if it cannot be created. 8786 * 8787 * NOTE: This function increments the reference counter associated with the 8788 * trace array returned. This makes sure it cannot be freed while in use. 8789 * Use trace_array_put() once the trace array is no longer needed. 8790 * If the trace_array is to be freed, trace_array_destroy() needs to 8791 * be called after the trace_array_put(), or simply let user space delete 8792 * it from the tracefs instances directory. But until the 8793 * trace_array_put() is called, user space can not delete it. 8794 * 8795 */ 8796 struct trace_array *trace_array_get_by_name(const char *name) 8797 { 8798 struct trace_array *tr; 8799 8800 mutex_lock(&event_mutex); 8801 mutex_lock(&trace_types_lock); 8802 8803 list_for_each_entry(tr, &ftrace_trace_arrays, list) { 8804 if (tr->name && strcmp(tr->name, name) == 0) 8805 goto out_unlock; 8806 } 8807 8808 tr = trace_array_create(name); 8809 8810 if (IS_ERR(tr)) 8811 tr = NULL; 8812 out_unlock: 8813 if (tr) 8814 tr->ref++; 8815 8816 mutex_unlock(&trace_types_lock); 8817 mutex_unlock(&event_mutex); 8818 return tr; 8819 } 8820 EXPORT_SYMBOL_GPL(trace_array_get_by_name); 8821 8822 static int __remove_instance(struct trace_array *tr) 8823 { 8824 int i; 8825 8826 /* Reference counter for a newly created trace array = 1. */ 8827 if (tr->ref > 1 || (tr->current_trace && tr->trace_ref)) 8828 return -EBUSY; 8829 8830 list_del(&tr->list); 8831 8832 /* Disable all the flags that were enabled coming in */ 8833 for (i = 0; i < TRACE_FLAGS_MAX_SIZE; i++) { 8834 if ((1 << i) & ZEROED_TRACE_FLAGS) 8835 set_tracer_flag(tr, 1 << i, 0); 8836 } 8837 8838 tracing_set_nop(tr); 8839 clear_ftrace_function_probes(tr); 8840 event_trace_del_tracer(tr); 8841 ftrace_clear_pids(tr); 8842 ftrace_destroy_function_files(tr); 8843 tracefs_remove(tr->dir); 8844 free_trace_buffers(tr); 8845 8846 for (i = 0; i < tr->nr_topts; i++) { 8847 kfree(tr->topts[i].topts); 8848 } 8849 kfree(tr->topts); 8850 8851 free_cpumask_var(tr->tracing_cpumask); 8852 kfree(tr->name); 8853 kfree(tr); 8854 8855 return 0; 8856 } 8857 8858 int trace_array_destroy(struct trace_array *this_tr) 8859 { 8860 struct trace_array *tr; 8861 int ret; 8862 8863 if (!this_tr) 8864 return -EINVAL; 8865 8866 mutex_lock(&event_mutex); 8867 mutex_lock(&trace_types_lock); 8868 8869 ret = -ENODEV; 8870 8871 /* Making sure trace array exists before destroying it. */ 8872 list_for_each_entry(tr, &ftrace_trace_arrays, list) { 8873 if (tr == this_tr) { 8874 ret = __remove_instance(tr); 8875 break; 8876 } 8877 } 8878 8879 mutex_unlock(&trace_types_lock); 8880 mutex_unlock(&event_mutex); 8881 8882 return ret; 8883 } 8884 EXPORT_SYMBOL_GPL(trace_array_destroy); 8885 8886 static int instance_rmdir(const char *name) 8887 { 8888 struct trace_array *tr; 8889 int ret; 8890 8891 mutex_lock(&event_mutex); 8892 mutex_lock(&trace_types_lock); 8893 8894 ret = -ENODEV; 8895 tr = trace_array_find(name); 8896 if (tr) 8897 ret = __remove_instance(tr); 8898 8899 mutex_unlock(&trace_types_lock); 8900 mutex_unlock(&event_mutex); 8901 8902 return ret; 8903 } 8904 8905 static __init void create_trace_instances(struct dentry *d_tracer) 8906 { 8907 struct trace_array *tr; 8908 8909 trace_instance_dir = tracefs_create_instance_dir("instances", d_tracer, 8910 instance_mkdir, 8911 instance_rmdir); 8912 if (MEM_FAIL(!trace_instance_dir, "Failed to create instances directory\n")) 8913 return; 8914 8915 mutex_lock(&event_mutex); 8916 mutex_lock(&trace_types_lock); 8917 8918 list_for_each_entry(tr, &ftrace_trace_arrays, list) { 8919 if (!tr->name) 8920 continue; 8921 if (MEM_FAIL(trace_array_create_dir(tr) < 0, 8922 "Failed to create instance directory\n")) 8923 break; 8924 } 8925 8926 mutex_unlock(&trace_types_lock); 8927 mutex_unlock(&event_mutex); 8928 } 8929 8930 static void 8931 init_tracer_tracefs(struct trace_array *tr, struct dentry *d_tracer) 8932 { 8933 struct trace_event_file *file; 8934 int cpu; 8935 8936 trace_create_file("available_tracers", 0444, d_tracer, 8937 tr, &show_traces_fops); 8938 8939 trace_create_file("current_tracer", 0644, d_tracer, 8940 tr, &set_tracer_fops); 8941 8942 trace_create_file("tracing_cpumask", 0644, d_tracer, 8943 tr, &tracing_cpumask_fops); 8944 8945 trace_create_file("trace_options", 0644, d_tracer, 8946 tr, &tracing_iter_fops); 8947 8948 trace_create_file("trace", 0644, d_tracer, 8949 tr, &tracing_fops); 8950 8951 trace_create_file("trace_pipe", 0444, d_tracer, 8952 tr, &tracing_pipe_fops); 8953 8954 trace_create_file("buffer_size_kb", 0644, d_tracer, 8955 tr, &tracing_entries_fops); 8956 8957 trace_create_file("buffer_total_size_kb", 0444, d_tracer, 8958 tr, &tracing_total_entries_fops); 8959 8960 trace_create_file("free_buffer", 0200, d_tracer, 8961 tr, &tracing_free_buffer_fops); 8962 8963 trace_create_file("trace_marker", 0220, d_tracer, 8964 tr, &tracing_mark_fops); 8965 8966 file = __find_event_file(tr, "ftrace", "print"); 8967 if (file && file->dir) 8968 trace_create_file("trigger", 0644, file->dir, file, 8969 &event_trigger_fops); 8970 tr->trace_marker_file = file; 8971 8972 trace_create_file("trace_marker_raw", 0220, d_tracer, 8973 tr, &tracing_mark_raw_fops); 8974 8975 trace_create_file("trace_clock", 0644, d_tracer, tr, 8976 &trace_clock_fops); 8977 8978 trace_create_file("tracing_on", 0644, d_tracer, 8979 tr, &rb_simple_fops); 8980 8981 trace_create_file("timestamp_mode", 0444, d_tracer, tr, 8982 &trace_time_stamp_mode_fops); 8983 8984 tr->buffer_percent = 50; 8985 8986 trace_create_file("buffer_percent", 0444, d_tracer, 8987 tr, &buffer_percent_fops); 8988 8989 create_trace_options_dir(tr); 8990 8991 #if defined(CONFIG_TRACER_MAX_TRACE) || defined(CONFIG_HWLAT_TRACER) 8992 trace_create_maxlat_file(tr, d_tracer); 8993 #endif 8994 8995 if (ftrace_create_function_files(tr, d_tracer)) 8996 MEM_FAIL(1, "Could not allocate function filter files"); 8997 8998 #ifdef CONFIG_TRACER_SNAPSHOT 8999 trace_create_file("snapshot", 0644, d_tracer, 9000 tr, &snapshot_fops); 9001 #endif 9002 9003 trace_create_file("error_log", 0644, d_tracer, 9004 tr, &tracing_err_log_fops); 9005 9006 for_each_tracing_cpu(cpu) 9007 tracing_init_tracefs_percpu(tr, cpu); 9008 9009 ftrace_init_tracefs(tr, d_tracer); 9010 } 9011 9012 static struct vfsmount *trace_automount(struct dentry *mntpt, void *ingore) 9013 { 9014 struct vfsmount *mnt; 9015 struct file_system_type *type; 9016 9017 /* 9018 * To maintain backward compatibility for tools that mount 9019 * debugfs to get to the tracing facility, tracefs is automatically 9020 * mounted to the debugfs/tracing directory. 9021 */ 9022 type = get_fs_type("tracefs"); 9023 if (!type) 9024 return NULL; 9025 mnt = vfs_submount(mntpt, type, "tracefs", NULL); 9026 put_filesystem(type); 9027 if (IS_ERR(mnt)) 9028 return NULL; 9029 mntget(mnt); 9030 9031 return mnt; 9032 } 9033 9034 /** 9035 * tracing_init_dentry - initialize top level trace array 9036 * 9037 * This is called when creating files or directories in the tracing 9038 * directory. It is called via fs_initcall() by any of the boot up code 9039 * and expects to return the dentry of the top level tracing directory. 9040 */ 9041 int tracing_init_dentry(void) 9042 { 9043 struct trace_array *tr = &global_trace; 9044 9045 if (security_locked_down(LOCKDOWN_TRACEFS)) { 9046 pr_warn("Tracing disabled due to lockdown\n"); 9047 return -EPERM; 9048 } 9049 9050 /* The top level trace array uses NULL as parent */ 9051 if (tr->dir) 9052 return 0; 9053 9054 if (WARN_ON(!tracefs_initialized())) 9055 return -ENODEV; 9056 9057 /* 9058 * As there may still be users that expect the tracing 9059 * files to exist in debugfs/tracing, we must automount 9060 * the tracefs file system there, so older tools still 9061 * work with the newer kerenl. 9062 */ 9063 tr->dir = debugfs_create_automount("tracing", NULL, 9064 trace_automount, NULL); 9065 9066 return 0; 9067 } 9068 9069 extern struct trace_eval_map *__start_ftrace_eval_maps[]; 9070 extern struct trace_eval_map *__stop_ftrace_eval_maps[]; 9071 9072 static struct workqueue_struct *eval_map_wq __initdata; 9073 static struct work_struct eval_map_work __initdata; 9074 9075 static void __init eval_map_work_func(struct work_struct *work) 9076 { 9077 int len; 9078 9079 len = __stop_ftrace_eval_maps - __start_ftrace_eval_maps; 9080 trace_insert_eval_map(NULL, __start_ftrace_eval_maps, len); 9081 } 9082 9083 static int __init trace_eval_init(void) 9084 { 9085 INIT_WORK(&eval_map_work, eval_map_work_func); 9086 9087 eval_map_wq = alloc_workqueue("eval_map_wq", WQ_UNBOUND, 0); 9088 if (!eval_map_wq) { 9089 pr_err("Unable to allocate eval_map_wq\n"); 9090 /* Do work here */ 9091 eval_map_work_func(&eval_map_work); 9092 return -ENOMEM; 9093 } 9094 9095 queue_work(eval_map_wq, &eval_map_work); 9096 return 0; 9097 } 9098 9099 static int __init trace_eval_sync(void) 9100 { 9101 /* Make sure the eval map updates are finished */ 9102 if (eval_map_wq) 9103 destroy_workqueue(eval_map_wq); 9104 return 0; 9105 } 9106 9107 late_initcall_sync(trace_eval_sync); 9108 9109 9110 #ifdef CONFIG_MODULES 9111 static void trace_module_add_evals(struct module *mod) 9112 { 9113 if (!mod->num_trace_evals) 9114 return; 9115 9116 /* 9117 * Modules with bad taint do not have events created, do 9118 * not bother with enums either. 9119 */ 9120 if (trace_module_has_bad_taint(mod)) 9121 return; 9122 9123 trace_insert_eval_map(mod, mod->trace_evals, mod->num_trace_evals); 9124 } 9125 9126 #ifdef CONFIG_TRACE_EVAL_MAP_FILE 9127 static void trace_module_remove_evals(struct module *mod) 9128 { 9129 union trace_eval_map_item *map; 9130 union trace_eval_map_item **last = &trace_eval_maps; 9131 9132 if (!mod->num_trace_evals) 9133 return; 9134 9135 mutex_lock(&trace_eval_mutex); 9136 9137 map = trace_eval_maps; 9138 9139 while (map) { 9140 if (map->head.mod == mod) 9141 break; 9142 map = trace_eval_jmp_to_tail(map); 9143 last = &map->tail.next; 9144 map = map->tail.next; 9145 } 9146 if (!map) 9147 goto out; 9148 9149 *last = trace_eval_jmp_to_tail(map)->tail.next; 9150 kfree(map); 9151 out: 9152 mutex_unlock(&trace_eval_mutex); 9153 } 9154 #else 9155 static inline void trace_module_remove_evals(struct module *mod) { } 9156 #endif /* CONFIG_TRACE_EVAL_MAP_FILE */ 9157 9158 static int trace_module_notify(struct notifier_block *self, 9159 unsigned long val, void *data) 9160 { 9161 struct module *mod = data; 9162 9163 switch (val) { 9164 case MODULE_STATE_COMING: 9165 trace_module_add_evals(mod); 9166 break; 9167 case MODULE_STATE_GOING: 9168 trace_module_remove_evals(mod); 9169 break; 9170 } 9171 9172 return NOTIFY_OK; 9173 } 9174 9175 static struct notifier_block trace_module_nb = { 9176 .notifier_call = trace_module_notify, 9177 .priority = 0, 9178 }; 9179 #endif /* CONFIG_MODULES */ 9180 9181 static __init int tracer_init_tracefs(void) 9182 { 9183 int ret; 9184 9185 trace_access_lock_init(); 9186 9187 ret = tracing_init_dentry(); 9188 if (ret) 9189 return 0; 9190 9191 event_trace_init(); 9192 9193 init_tracer_tracefs(&global_trace, NULL); 9194 ftrace_init_tracefs_toplevel(&global_trace, NULL); 9195 9196 trace_create_file("tracing_thresh", 0644, NULL, 9197 &global_trace, &tracing_thresh_fops); 9198 9199 trace_create_file("README", 0444, NULL, 9200 NULL, &tracing_readme_fops); 9201 9202 trace_create_file("saved_cmdlines", 0444, NULL, 9203 NULL, &tracing_saved_cmdlines_fops); 9204 9205 trace_create_file("saved_cmdlines_size", 0644, NULL, 9206 NULL, &tracing_saved_cmdlines_size_fops); 9207 9208 trace_create_file("saved_tgids", 0444, NULL, 9209 NULL, &tracing_saved_tgids_fops); 9210 9211 trace_eval_init(); 9212 9213 trace_create_eval_file(NULL); 9214 9215 #ifdef CONFIG_MODULES 9216 register_module_notifier(&trace_module_nb); 9217 #endif 9218 9219 #ifdef CONFIG_DYNAMIC_FTRACE 9220 trace_create_file("dyn_ftrace_total_info", 0444, NULL, 9221 NULL, &tracing_dyn_info_fops); 9222 #endif 9223 9224 create_trace_instances(NULL); 9225 9226 update_tracer_options(&global_trace); 9227 9228 return 0; 9229 } 9230 9231 static int trace_panic_handler(struct notifier_block *this, 9232 unsigned long event, void *unused) 9233 { 9234 if (ftrace_dump_on_oops) 9235 ftrace_dump(ftrace_dump_on_oops); 9236 return NOTIFY_OK; 9237 } 9238 9239 static struct notifier_block trace_panic_notifier = { 9240 .notifier_call = trace_panic_handler, 9241 .next = NULL, 9242 .priority = 150 /* priority: INT_MAX >= x >= 0 */ 9243 }; 9244 9245 static int trace_die_handler(struct notifier_block *self, 9246 unsigned long val, 9247 void *data) 9248 { 9249 switch (val) { 9250 case DIE_OOPS: 9251 if (ftrace_dump_on_oops) 9252 ftrace_dump(ftrace_dump_on_oops); 9253 break; 9254 default: 9255 break; 9256 } 9257 return NOTIFY_OK; 9258 } 9259 9260 static struct notifier_block trace_die_notifier = { 9261 .notifier_call = trace_die_handler, 9262 .priority = 200 9263 }; 9264 9265 /* 9266 * printk is set to max of 1024, we really don't need it that big. 9267 * Nothing should be printing 1000 characters anyway. 9268 */ 9269 #define TRACE_MAX_PRINT 1000 9270 9271 /* 9272 * Define here KERN_TRACE so that we have one place to modify 9273 * it if we decide to change what log level the ftrace dump 9274 * should be at. 9275 */ 9276 #define KERN_TRACE KERN_EMERG 9277 9278 void 9279 trace_printk_seq(struct trace_seq *s) 9280 { 9281 /* Probably should print a warning here. */ 9282 if (s->seq.len >= TRACE_MAX_PRINT) 9283 s->seq.len = TRACE_MAX_PRINT; 9284 9285 /* 9286 * More paranoid code. Although the buffer size is set to 9287 * PAGE_SIZE, and TRACE_MAX_PRINT is 1000, this is just 9288 * an extra layer of protection. 9289 */ 9290 if (WARN_ON_ONCE(s->seq.len >= s->seq.size)) 9291 s->seq.len = s->seq.size - 1; 9292 9293 /* should be zero ended, but we are paranoid. */ 9294 s->buffer[s->seq.len] = 0; 9295 9296 printk(KERN_TRACE "%s", s->buffer); 9297 9298 trace_seq_init(s); 9299 } 9300 9301 void trace_init_global_iter(struct trace_iterator *iter) 9302 { 9303 iter->tr = &global_trace; 9304 iter->trace = iter->tr->current_trace; 9305 iter->cpu_file = RING_BUFFER_ALL_CPUS; 9306 iter->array_buffer = &global_trace.array_buffer; 9307 9308 if (iter->trace && iter->trace->open) 9309 iter->trace->open(iter); 9310 9311 /* Annotate start of buffers if we had overruns */ 9312 if (ring_buffer_overruns(iter->array_buffer->buffer)) 9313 iter->iter_flags |= TRACE_FILE_ANNOTATE; 9314 9315 /* Output in nanoseconds only if we are using a clock in nanoseconds. */ 9316 if (trace_clocks[iter->tr->clock_id].in_ns) 9317 iter->iter_flags |= TRACE_FILE_TIME_IN_NS; 9318 } 9319 9320 void ftrace_dump(enum ftrace_dump_mode oops_dump_mode) 9321 { 9322 /* use static because iter can be a bit big for the stack */ 9323 static struct trace_iterator iter; 9324 static atomic_t dump_running; 9325 struct trace_array *tr = &global_trace; 9326 unsigned int old_userobj; 9327 unsigned long flags; 9328 int cnt = 0, cpu; 9329 9330 /* Only allow one dump user at a time. */ 9331 if (atomic_inc_return(&dump_running) != 1) { 9332 atomic_dec(&dump_running); 9333 return; 9334 } 9335 9336 /* 9337 * Always turn off tracing when we dump. 9338 * We don't need to show trace output of what happens 9339 * between multiple crashes. 9340 * 9341 * If the user does a sysrq-z, then they can re-enable 9342 * tracing with echo 1 > tracing_on. 9343 */ 9344 tracing_off(); 9345 9346 local_irq_save(flags); 9347 printk_nmi_direct_enter(); 9348 9349 /* Simulate the iterator */ 9350 trace_init_global_iter(&iter); 9351 /* Can not use kmalloc for iter.temp */ 9352 iter.temp = static_temp_buf; 9353 iter.temp_size = STATIC_TEMP_BUF_SIZE; 9354 9355 for_each_tracing_cpu(cpu) { 9356 atomic_inc(&per_cpu_ptr(iter.array_buffer->data, cpu)->disabled); 9357 } 9358 9359 old_userobj = tr->trace_flags & TRACE_ITER_SYM_USEROBJ; 9360 9361 /* don't look at user memory in panic mode */ 9362 tr->trace_flags &= ~TRACE_ITER_SYM_USEROBJ; 9363 9364 switch (oops_dump_mode) { 9365 case DUMP_ALL: 9366 iter.cpu_file = RING_BUFFER_ALL_CPUS; 9367 break; 9368 case DUMP_ORIG: 9369 iter.cpu_file = raw_smp_processor_id(); 9370 break; 9371 case DUMP_NONE: 9372 goto out_enable; 9373 default: 9374 printk(KERN_TRACE "Bad dumping mode, switching to all CPUs dump\n"); 9375 iter.cpu_file = RING_BUFFER_ALL_CPUS; 9376 } 9377 9378 printk(KERN_TRACE "Dumping ftrace buffer:\n"); 9379 9380 /* Did function tracer already get disabled? */ 9381 if (ftrace_is_dead()) { 9382 printk("# WARNING: FUNCTION TRACING IS CORRUPTED\n"); 9383 printk("# MAY BE MISSING FUNCTION EVENTS\n"); 9384 } 9385 9386 /* 9387 * We need to stop all tracing on all CPUS to read 9388 * the next buffer. This is a bit expensive, but is 9389 * not done often. We fill all what we can read, 9390 * and then release the locks again. 9391 */ 9392 9393 while (!trace_empty(&iter)) { 9394 9395 if (!cnt) 9396 printk(KERN_TRACE "---------------------------------\n"); 9397 9398 cnt++; 9399 9400 trace_iterator_reset(&iter); 9401 iter.iter_flags |= TRACE_FILE_LAT_FMT; 9402 9403 if (trace_find_next_entry_inc(&iter) != NULL) { 9404 int ret; 9405 9406 ret = print_trace_line(&iter); 9407 if (ret != TRACE_TYPE_NO_CONSUME) 9408 trace_consume(&iter); 9409 } 9410 touch_nmi_watchdog(); 9411 9412 trace_printk_seq(&iter.seq); 9413 } 9414 9415 if (!cnt) 9416 printk(KERN_TRACE " (ftrace buffer empty)\n"); 9417 else 9418 printk(KERN_TRACE "---------------------------------\n"); 9419 9420 out_enable: 9421 tr->trace_flags |= old_userobj; 9422 9423 for_each_tracing_cpu(cpu) { 9424 atomic_dec(&per_cpu_ptr(iter.array_buffer->data, cpu)->disabled); 9425 } 9426 atomic_dec(&dump_running); 9427 printk_nmi_direct_exit(); 9428 local_irq_restore(flags); 9429 } 9430 EXPORT_SYMBOL_GPL(ftrace_dump); 9431 9432 int trace_run_command(const char *buf, int (*createfn)(int, char **)) 9433 { 9434 char **argv; 9435 int argc, ret; 9436 9437 argc = 0; 9438 ret = 0; 9439 argv = argv_split(GFP_KERNEL, buf, &argc); 9440 if (!argv) 9441 return -ENOMEM; 9442 9443 if (argc) 9444 ret = createfn(argc, argv); 9445 9446 argv_free(argv); 9447 9448 return ret; 9449 } 9450 9451 #define WRITE_BUFSIZE 4096 9452 9453 ssize_t trace_parse_run_command(struct file *file, const char __user *buffer, 9454 size_t count, loff_t *ppos, 9455 int (*createfn)(int, char **)) 9456 { 9457 char *kbuf, *buf, *tmp; 9458 int ret = 0; 9459 size_t done = 0; 9460 size_t size; 9461 9462 kbuf = kmalloc(WRITE_BUFSIZE, GFP_KERNEL); 9463 if (!kbuf) 9464 return -ENOMEM; 9465 9466 while (done < count) { 9467 size = count - done; 9468 9469 if (size >= WRITE_BUFSIZE) 9470 size = WRITE_BUFSIZE - 1; 9471 9472 if (copy_from_user(kbuf, buffer + done, size)) { 9473 ret = -EFAULT; 9474 goto out; 9475 } 9476 kbuf[size] = '\0'; 9477 buf = kbuf; 9478 do { 9479 tmp = strchr(buf, '\n'); 9480 if (tmp) { 9481 *tmp = '\0'; 9482 size = tmp - buf + 1; 9483 } else { 9484 size = strlen(buf); 9485 if (done + size < count) { 9486 if (buf != kbuf) 9487 break; 9488 /* This can accept WRITE_BUFSIZE - 2 ('\n' + '\0') */ 9489 pr_warn("Line length is too long: Should be less than %d\n", 9490 WRITE_BUFSIZE - 2); 9491 ret = -EINVAL; 9492 goto out; 9493 } 9494 } 9495 done += size; 9496 9497 /* Remove comments */ 9498 tmp = strchr(buf, '#'); 9499 9500 if (tmp) 9501 *tmp = '\0'; 9502 9503 ret = trace_run_command(buf, createfn); 9504 if (ret) 9505 goto out; 9506 buf += size; 9507 9508 } while (done < count); 9509 } 9510 ret = done; 9511 9512 out: 9513 kfree(kbuf); 9514 9515 return ret; 9516 } 9517 9518 __init static int tracer_alloc_buffers(void) 9519 { 9520 int ring_buf_size; 9521 int ret = -ENOMEM; 9522 9523 9524 if (security_locked_down(LOCKDOWN_TRACEFS)) { 9525 pr_warn("Tracing disabled due to lockdown\n"); 9526 return -EPERM; 9527 } 9528 9529 /* 9530 * Make sure we don't accidentally add more trace options 9531 * than we have bits for. 9532 */ 9533 BUILD_BUG_ON(TRACE_ITER_LAST_BIT > TRACE_FLAGS_MAX_SIZE); 9534 9535 if (!alloc_cpumask_var(&tracing_buffer_mask, GFP_KERNEL)) 9536 goto out; 9537 9538 if (!alloc_cpumask_var(&global_trace.tracing_cpumask, GFP_KERNEL)) 9539 goto out_free_buffer_mask; 9540 9541 /* Only allocate trace_printk buffers if a trace_printk exists */ 9542 if (&__stop___trace_bprintk_fmt != &__start___trace_bprintk_fmt) 9543 /* Must be called before global_trace.buffer is allocated */ 9544 trace_printk_init_buffers(); 9545 9546 /* To save memory, keep the ring buffer size to its minimum */ 9547 if (ring_buffer_expanded) 9548 ring_buf_size = trace_buf_size; 9549 else 9550 ring_buf_size = 1; 9551 9552 cpumask_copy(tracing_buffer_mask, cpu_possible_mask); 9553 cpumask_copy(global_trace.tracing_cpumask, cpu_all_mask); 9554 9555 raw_spin_lock_init(&global_trace.start_lock); 9556 9557 /* 9558 * The prepare callbacks allocates some memory for the ring buffer. We 9559 * don't free the buffer if the CPU goes down. If we were to free 9560 * the buffer, then the user would lose any trace that was in the 9561 * buffer. The memory will be removed once the "instance" is removed. 9562 */ 9563 ret = cpuhp_setup_state_multi(CPUHP_TRACE_RB_PREPARE, 9564 "trace/RB:preapre", trace_rb_cpu_prepare, 9565 NULL); 9566 if (ret < 0) 9567 goto out_free_cpumask; 9568 /* Used for event triggers */ 9569 ret = -ENOMEM; 9570 temp_buffer = ring_buffer_alloc(PAGE_SIZE, RB_FL_OVERWRITE); 9571 if (!temp_buffer) 9572 goto out_rm_hp_state; 9573 9574 if (trace_create_savedcmd() < 0) 9575 goto out_free_temp_buffer; 9576 9577 /* TODO: make the number of buffers hot pluggable with CPUS */ 9578 if (allocate_trace_buffers(&global_trace, ring_buf_size) < 0) { 9579 MEM_FAIL(1, "tracer: failed to allocate ring buffer!\n"); 9580 goto out_free_savedcmd; 9581 } 9582 9583 if (global_trace.buffer_disabled) 9584 tracing_off(); 9585 9586 if (trace_boot_clock) { 9587 ret = tracing_set_clock(&global_trace, trace_boot_clock); 9588 if (ret < 0) 9589 pr_warn("Trace clock %s not defined, going back to default\n", 9590 trace_boot_clock); 9591 } 9592 9593 /* 9594 * register_tracer() might reference current_trace, so it 9595 * needs to be set before we register anything. This is 9596 * just a bootstrap of current_trace anyway. 9597 */ 9598 global_trace.current_trace = &nop_trace; 9599 9600 global_trace.max_lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED; 9601 9602 ftrace_init_global_array_ops(&global_trace); 9603 9604 init_trace_flags_index(&global_trace); 9605 9606 register_tracer(&nop_trace); 9607 9608 /* Function tracing may start here (via kernel command line) */ 9609 init_function_trace(); 9610 9611 /* All seems OK, enable tracing */ 9612 tracing_disabled = 0; 9613 9614 atomic_notifier_chain_register(&panic_notifier_list, 9615 &trace_panic_notifier); 9616 9617 register_die_notifier(&trace_die_notifier); 9618 9619 global_trace.flags = TRACE_ARRAY_FL_GLOBAL; 9620 9621 INIT_LIST_HEAD(&global_trace.systems); 9622 INIT_LIST_HEAD(&global_trace.events); 9623 INIT_LIST_HEAD(&global_trace.hist_vars); 9624 INIT_LIST_HEAD(&global_trace.err_log); 9625 list_add(&global_trace.list, &ftrace_trace_arrays); 9626 9627 apply_trace_boot_options(); 9628 9629 register_snapshot_cmd(); 9630 9631 return 0; 9632 9633 out_free_savedcmd: 9634 free_saved_cmdlines_buffer(savedcmd); 9635 out_free_temp_buffer: 9636 ring_buffer_free(temp_buffer); 9637 out_rm_hp_state: 9638 cpuhp_remove_multi_state(CPUHP_TRACE_RB_PREPARE); 9639 out_free_cpumask: 9640 free_cpumask_var(global_trace.tracing_cpumask); 9641 out_free_buffer_mask: 9642 free_cpumask_var(tracing_buffer_mask); 9643 out: 9644 return ret; 9645 } 9646 9647 void __init early_trace_init(void) 9648 { 9649 if (tracepoint_printk) { 9650 tracepoint_print_iter = 9651 kmalloc(sizeof(*tracepoint_print_iter), GFP_KERNEL); 9652 if (MEM_FAIL(!tracepoint_print_iter, 9653 "Failed to allocate trace iterator\n")) 9654 tracepoint_printk = 0; 9655 else 9656 static_key_enable(&tracepoint_printk_key.key); 9657 } 9658 tracer_alloc_buffers(); 9659 } 9660 9661 void __init trace_init(void) 9662 { 9663 trace_event_init(); 9664 } 9665 9666 __init static int clear_boot_tracer(void) 9667 { 9668 /* 9669 * The default tracer at boot buffer is an init section. 9670 * This function is called in lateinit. If we did not 9671 * find the boot tracer, then clear it out, to prevent 9672 * later registration from accessing the buffer that is 9673 * about to be freed. 9674 */ 9675 if (!default_bootup_tracer) 9676 return 0; 9677 9678 printk(KERN_INFO "ftrace bootup tracer '%s' not registered.\n", 9679 default_bootup_tracer); 9680 default_bootup_tracer = NULL; 9681 9682 return 0; 9683 } 9684 9685 fs_initcall(tracer_init_tracefs); 9686 late_initcall_sync(clear_boot_tracer); 9687 9688 #ifdef CONFIG_HAVE_UNSTABLE_SCHED_CLOCK 9689 __init static int tracing_set_default_clock(void) 9690 { 9691 /* sched_clock_stable() is determined in late_initcall */ 9692 if (!trace_boot_clock && !sched_clock_stable()) { 9693 if (security_locked_down(LOCKDOWN_TRACEFS)) { 9694 pr_warn("Can not set tracing clock due to lockdown\n"); 9695 return -EPERM; 9696 } 9697 9698 printk(KERN_WARNING 9699 "Unstable clock detected, switching default tracing clock to \"global\"\n" 9700 "If you want to keep using the local clock, then add:\n" 9701 " \"trace_clock=local\"\n" 9702 "on the kernel command line\n"); 9703 tracing_set_clock(&global_trace, "global"); 9704 } 9705 9706 return 0; 9707 } 9708 late_initcall_sync(tracing_set_default_clock); 9709 #endif 9710