xref: /linux/tools/perf/builtin-lock.c (revision 95db3b255fde4e830e5f8cc011eb404023f669d4)
1 #include "builtin.h"
2 #include "perf.h"
3 
4 #include "util/evlist.h"
5 #include "util/evsel.h"
6 #include "util/util.h"
7 #include "util/cache.h"
8 #include "util/symbol.h"
9 #include "util/thread.h"
10 #include "util/header.h"
11 
12 #include <subcmd/parse-options.h>
13 #include "util/trace-event.h"
14 
15 #include "util/debug.h"
16 #include "util/session.h"
17 #include "util/tool.h"
18 #include "util/data.h"
19 
20 #include <sys/types.h>
21 #include <sys/prctl.h>
22 #include <semaphore.h>
23 #include <pthread.h>
24 #include <math.h>
25 #include <limits.h>
26 
27 #include <linux/list.h>
28 #include <linux/hash.h>
29 
30 static struct perf_session *session;
31 
32 /* based on kernel/lockdep.c */
33 #define LOCKHASH_BITS		12
34 #define LOCKHASH_SIZE		(1UL << LOCKHASH_BITS)
35 
36 static struct list_head lockhash_table[LOCKHASH_SIZE];
37 
38 #define __lockhashfn(key)	hash_long((unsigned long)key, LOCKHASH_BITS)
39 #define lockhashentry(key)	(lockhash_table + __lockhashfn((key)))
40 
41 struct lock_stat {
42 	struct list_head	hash_entry;
43 	struct rb_node		rb;		/* used for sorting */
44 
45 	/*
46 	 * FIXME: perf_evsel__intval() returns u64,
47 	 * so address of lockdep_map should be dealed as 64bit.
48 	 * Is there more better solution?
49 	 */
50 	void			*addr;		/* address of lockdep_map, used as ID */
51 	char			*name;		/* for strcpy(), we cannot use const */
52 
53 	unsigned int		nr_acquire;
54 	unsigned int		nr_acquired;
55 	unsigned int		nr_contended;
56 	unsigned int		nr_release;
57 
58 	unsigned int		nr_readlock;
59 	unsigned int		nr_trylock;
60 
61 	/* these times are in nano sec. */
62 	u64                     avg_wait_time;
63 	u64			wait_time_total;
64 	u64			wait_time_min;
65 	u64			wait_time_max;
66 
67 	int			discard; /* flag of blacklist */
68 };
69 
70 /*
71  * States of lock_seq_stat
72  *
73  * UNINITIALIZED is required for detecting first event of acquire.
74  * As the nature of lock events, there is no guarantee
75  * that the first event for the locks are acquire,
76  * it can be acquired, contended or release.
77  */
78 #define SEQ_STATE_UNINITIALIZED      0	       /* initial state */
79 #define SEQ_STATE_RELEASED	1
80 #define SEQ_STATE_ACQUIRING	2
81 #define SEQ_STATE_ACQUIRED	3
82 #define SEQ_STATE_READ_ACQUIRED	4
83 #define SEQ_STATE_CONTENDED	5
84 
85 /*
86  * MAX_LOCK_DEPTH
87  * Imported from include/linux/sched.h.
88  * Should this be synchronized?
89  */
90 #define MAX_LOCK_DEPTH 48
91 
92 /*
93  * struct lock_seq_stat:
94  * Place to put on state of one lock sequence
95  * 1) acquire -> acquired -> release
96  * 2) acquire -> contended -> acquired -> release
97  * 3) acquire (with read or try) -> release
98  * 4) Are there other patterns?
99  */
100 struct lock_seq_stat {
101 	struct list_head        list;
102 	int			state;
103 	u64			prev_event_time;
104 	void                    *addr;
105 
106 	int                     read_count;
107 };
108 
109 struct thread_stat {
110 	struct rb_node		rb;
111 
112 	u32                     tid;
113 	struct list_head        seq_list;
114 };
115 
116 static struct rb_root		thread_stats;
117 
118 static struct thread_stat *thread_stat_find(u32 tid)
119 {
120 	struct rb_node *node;
121 	struct thread_stat *st;
122 
123 	node = thread_stats.rb_node;
124 	while (node) {
125 		st = container_of(node, struct thread_stat, rb);
126 		if (st->tid == tid)
127 			return st;
128 		else if (tid < st->tid)
129 			node = node->rb_left;
130 		else
131 			node = node->rb_right;
132 	}
133 
134 	return NULL;
135 }
136 
137 static void thread_stat_insert(struct thread_stat *new)
138 {
139 	struct rb_node **rb = &thread_stats.rb_node;
140 	struct rb_node *parent = NULL;
141 	struct thread_stat *p;
142 
143 	while (*rb) {
144 		p = container_of(*rb, struct thread_stat, rb);
145 		parent = *rb;
146 
147 		if (new->tid < p->tid)
148 			rb = &(*rb)->rb_left;
149 		else if (new->tid > p->tid)
150 			rb = &(*rb)->rb_right;
151 		else
152 			BUG_ON("inserting invalid thread_stat\n");
153 	}
154 
155 	rb_link_node(&new->rb, parent, rb);
156 	rb_insert_color(&new->rb, &thread_stats);
157 }
158 
159 static struct thread_stat *thread_stat_findnew_after_first(u32 tid)
160 {
161 	struct thread_stat *st;
162 
163 	st = thread_stat_find(tid);
164 	if (st)
165 		return st;
166 
167 	st = zalloc(sizeof(struct thread_stat));
168 	if (!st) {
169 		pr_err("memory allocation failed\n");
170 		return NULL;
171 	}
172 
173 	st->tid = tid;
174 	INIT_LIST_HEAD(&st->seq_list);
175 
176 	thread_stat_insert(st);
177 
178 	return st;
179 }
180 
181 static struct thread_stat *thread_stat_findnew_first(u32 tid);
182 static struct thread_stat *(*thread_stat_findnew)(u32 tid) =
183 	thread_stat_findnew_first;
184 
185 static struct thread_stat *thread_stat_findnew_first(u32 tid)
186 {
187 	struct thread_stat *st;
188 
189 	st = zalloc(sizeof(struct thread_stat));
190 	if (!st) {
191 		pr_err("memory allocation failed\n");
192 		return NULL;
193 	}
194 	st->tid = tid;
195 	INIT_LIST_HEAD(&st->seq_list);
196 
197 	rb_link_node(&st->rb, NULL, &thread_stats.rb_node);
198 	rb_insert_color(&st->rb, &thread_stats);
199 
200 	thread_stat_findnew = thread_stat_findnew_after_first;
201 	return st;
202 }
203 
204 /* build simple key function one is bigger than two */
205 #define SINGLE_KEY(member)						\
206 	static int lock_stat_key_ ## member(struct lock_stat *one,	\
207 					 struct lock_stat *two)		\
208 	{								\
209 		return one->member > two->member;			\
210 	}
211 
212 SINGLE_KEY(nr_acquired)
213 SINGLE_KEY(nr_contended)
214 SINGLE_KEY(avg_wait_time)
215 SINGLE_KEY(wait_time_total)
216 SINGLE_KEY(wait_time_max)
217 
218 static int lock_stat_key_wait_time_min(struct lock_stat *one,
219 					struct lock_stat *two)
220 {
221 	u64 s1 = one->wait_time_min;
222 	u64 s2 = two->wait_time_min;
223 	if (s1 == ULLONG_MAX)
224 		s1 = 0;
225 	if (s2 == ULLONG_MAX)
226 		s2 = 0;
227 	return s1 > s2;
228 }
229 
230 struct lock_key {
231 	/*
232 	 * name: the value for specify by user
233 	 * this should be simpler than raw name of member
234 	 * e.g. nr_acquired -> acquired, wait_time_total -> wait_total
235 	 */
236 	const char		*name;
237 	int			(*key)(struct lock_stat*, struct lock_stat*);
238 };
239 
240 static const char		*sort_key = "acquired";
241 
242 static int			(*compare)(struct lock_stat *, struct lock_stat *);
243 
244 static struct rb_root		result;	/* place to store sorted data */
245 
246 #define DEF_KEY_LOCK(name, fn_suffix)	\
247 	{ #name, lock_stat_key_ ## fn_suffix }
248 struct lock_key keys[] = {
249 	DEF_KEY_LOCK(acquired, nr_acquired),
250 	DEF_KEY_LOCK(contended, nr_contended),
251 	DEF_KEY_LOCK(avg_wait, avg_wait_time),
252 	DEF_KEY_LOCK(wait_total, wait_time_total),
253 	DEF_KEY_LOCK(wait_min, wait_time_min),
254 	DEF_KEY_LOCK(wait_max, wait_time_max),
255 
256 	/* extra comparisons much complicated should be here */
257 
258 	{ NULL, NULL }
259 };
260 
261 static int select_key(void)
262 {
263 	int i;
264 
265 	for (i = 0; keys[i].name; i++) {
266 		if (!strcmp(keys[i].name, sort_key)) {
267 			compare = keys[i].key;
268 			return 0;
269 		}
270 	}
271 
272 	pr_err("Unknown compare key: %s\n", sort_key);
273 
274 	return -1;
275 }
276 
277 static void insert_to_result(struct lock_stat *st,
278 			     int (*bigger)(struct lock_stat *, struct lock_stat *))
279 {
280 	struct rb_node **rb = &result.rb_node;
281 	struct rb_node *parent = NULL;
282 	struct lock_stat *p;
283 
284 	while (*rb) {
285 		p = container_of(*rb, struct lock_stat, rb);
286 		parent = *rb;
287 
288 		if (bigger(st, p))
289 			rb = &(*rb)->rb_left;
290 		else
291 			rb = &(*rb)->rb_right;
292 	}
293 
294 	rb_link_node(&st->rb, parent, rb);
295 	rb_insert_color(&st->rb, &result);
296 }
297 
298 /* returns left most element of result, and erase it */
299 static struct lock_stat *pop_from_result(void)
300 {
301 	struct rb_node *node = result.rb_node;
302 
303 	if (!node)
304 		return NULL;
305 
306 	while (node->rb_left)
307 		node = node->rb_left;
308 
309 	rb_erase(node, &result);
310 	return container_of(node, struct lock_stat, rb);
311 }
312 
313 static struct lock_stat *lock_stat_findnew(void *addr, const char *name)
314 {
315 	struct list_head *entry = lockhashentry(addr);
316 	struct lock_stat *ret, *new;
317 
318 	list_for_each_entry(ret, entry, hash_entry) {
319 		if (ret->addr == addr)
320 			return ret;
321 	}
322 
323 	new = zalloc(sizeof(struct lock_stat));
324 	if (!new)
325 		goto alloc_failed;
326 
327 	new->addr = addr;
328 	new->name = zalloc(sizeof(char) * strlen(name) + 1);
329 	if (!new->name) {
330 		free(new);
331 		goto alloc_failed;
332 	}
333 
334 	strcpy(new->name, name);
335 	new->wait_time_min = ULLONG_MAX;
336 
337 	list_add(&new->hash_entry, entry);
338 	return new;
339 
340 alloc_failed:
341 	pr_err("memory allocation failed\n");
342 	return NULL;
343 }
344 
345 struct trace_lock_handler {
346 	int (*acquire_event)(struct perf_evsel *evsel,
347 			     struct perf_sample *sample);
348 
349 	int (*acquired_event)(struct perf_evsel *evsel,
350 			      struct perf_sample *sample);
351 
352 	int (*contended_event)(struct perf_evsel *evsel,
353 			       struct perf_sample *sample);
354 
355 	int (*release_event)(struct perf_evsel *evsel,
356 			     struct perf_sample *sample);
357 };
358 
359 static struct lock_seq_stat *get_seq(struct thread_stat *ts, void *addr)
360 {
361 	struct lock_seq_stat *seq;
362 
363 	list_for_each_entry(seq, &ts->seq_list, list) {
364 		if (seq->addr == addr)
365 			return seq;
366 	}
367 
368 	seq = zalloc(sizeof(struct lock_seq_stat));
369 	if (!seq) {
370 		pr_err("memory allocation failed\n");
371 		return NULL;
372 	}
373 	seq->state = SEQ_STATE_UNINITIALIZED;
374 	seq->addr = addr;
375 
376 	list_add(&seq->list, &ts->seq_list);
377 	return seq;
378 }
379 
380 enum broken_state {
381 	BROKEN_ACQUIRE,
382 	BROKEN_ACQUIRED,
383 	BROKEN_CONTENDED,
384 	BROKEN_RELEASE,
385 	BROKEN_MAX,
386 };
387 
388 static int bad_hist[BROKEN_MAX];
389 
390 enum acquire_flags {
391 	TRY_LOCK = 1,
392 	READ_LOCK = 2,
393 };
394 
395 static int report_lock_acquire_event(struct perf_evsel *evsel,
396 				     struct perf_sample *sample)
397 {
398 	void *addr;
399 	struct lock_stat *ls;
400 	struct thread_stat *ts;
401 	struct lock_seq_stat *seq;
402 	const char *name = perf_evsel__strval(evsel, sample, "name");
403 	u64 tmp = perf_evsel__intval(evsel, sample, "lockdep_addr");
404 	int flag = perf_evsel__intval(evsel, sample, "flag");
405 
406 	memcpy(&addr, &tmp, sizeof(void *));
407 
408 	ls = lock_stat_findnew(addr, name);
409 	if (!ls)
410 		return -ENOMEM;
411 	if (ls->discard)
412 		return 0;
413 
414 	ts = thread_stat_findnew(sample->tid);
415 	if (!ts)
416 		return -ENOMEM;
417 
418 	seq = get_seq(ts, addr);
419 	if (!seq)
420 		return -ENOMEM;
421 
422 	switch (seq->state) {
423 	case SEQ_STATE_UNINITIALIZED:
424 	case SEQ_STATE_RELEASED:
425 		if (!flag) {
426 			seq->state = SEQ_STATE_ACQUIRING;
427 		} else {
428 			if (flag & TRY_LOCK)
429 				ls->nr_trylock++;
430 			if (flag & READ_LOCK)
431 				ls->nr_readlock++;
432 			seq->state = SEQ_STATE_READ_ACQUIRED;
433 			seq->read_count = 1;
434 			ls->nr_acquired++;
435 		}
436 		break;
437 	case SEQ_STATE_READ_ACQUIRED:
438 		if (flag & READ_LOCK) {
439 			seq->read_count++;
440 			ls->nr_acquired++;
441 			goto end;
442 		} else {
443 			goto broken;
444 		}
445 		break;
446 	case SEQ_STATE_ACQUIRED:
447 	case SEQ_STATE_ACQUIRING:
448 	case SEQ_STATE_CONTENDED:
449 broken:
450 		/* broken lock sequence, discard it */
451 		ls->discard = 1;
452 		bad_hist[BROKEN_ACQUIRE]++;
453 		list_del(&seq->list);
454 		free(seq);
455 		goto end;
456 	default:
457 		BUG_ON("Unknown state of lock sequence found!\n");
458 		break;
459 	}
460 
461 	ls->nr_acquire++;
462 	seq->prev_event_time = sample->time;
463 end:
464 	return 0;
465 }
466 
467 static int report_lock_acquired_event(struct perf_evsel *evsel,
468 				      struct perf_sample *sample)
469 {
470 	void *addr;
471 	struct lock_stat *ls;
472 	struct thread_stat *ts;
473 	struct lock_seq_stat *seq;
474 	u64 contended_term;
475 	const char *name = perf_evsel__strval(evsel, sample, "name");
476 	u64 tmp = perf_evsel__intval(evsel, sample, "lockdep_addr");
477 
478 	memcpy(&addr, &tmp, sizeof(void *));
479 
480 	ls = lock_stat_findnew(addr, name);
481 	if (!ls)
482 		return -ENOMEM;
483 	if (ls->discard)
484 		return 0;
485 
486 	ts = thread_stat_findnew(sample->tid);
487 	if (!ts)
488 		return -ENOMEM;
489 
490 	seq = get_seq(ts, addr);
491 	if (!seq)
492 		return -ENOMEM;
493 
494 	switch (seq->state) {
495 	case SEQ_STATE_UNINITIALIZED:
496 		/* orphan event, do nothing */
497 		return 0;
498 	case SEQ_STATE_ACQUIRING:
499 		break;
500 	case SEQ_STATE_CONTENDED:
501 		contended_term = sample->time - seq->prev_event_time;
502 		ls->wait_time_total += contended_term;
503 		if (contended_term < ls->wait_time_min)
504 			ls->wait_time_min = contended_term;
505 		if (ls->wait_time_max < contended_term)
506 			ls->wait_time_max = contended_term;
507 		break;
508 	case SEQ_STATE_RELEASED:
509 	case SEQ_STATE_ACQUIRED:
510 	case SEQ_STATE_READ_ACQUIRED:
511 		/* broken lock sequence, discard it */
512 		ls->discard = 1;
513 		bad_hist[BROKEN_ACQUIRED]++;
514 		list_del(&seq->list);
515 		free(seq);
516 		goto end;
517 	default:
518 		BUG_ON("Unknown state of lock sequence found!\n");
519 		break;
520 	}
521 
522 	seq->state = SEQ_STATE_ACQUIRED;
523 	ls->nr_acquired++;
524 	ls->avg_wait_time = ls->nr_contended ? ls->wait_time_total/ls->nr_contended : 0;
525 	seq->prev_event_time = sample->time;
526 end:
527 	return 0;
528 }
529 
530 static int report_lock_contended_event(struct perf_evsel *evsel,
531 				       struct perf_sample *sample)
532 {
533 	void *addr;
534 	struct lock_stat *ls;
535 	struct thread_stat *ts;
536 	struct lock_seq_stat *seq;
537 	const char *name = perf_evsel__strval(evsel, sample, "name");
538 	u64 tmp = perf_evsel__intval(evsel, sample, "lockdep_addr");
539 
540 	memcpy(&addr, &tmp, sizeof(void *));
541 
542 	ls = lock_stat_findnew(addr, name);
543 	if (!ls)
544 		return -ENOMEM;
545 	if (ls->discard)
546 		return 0;
547 
548 	ts = thread_stat_findnew(sample->tid);
549 	if (!ts)
550 		return -ENOMEM;
551 
552 	seq = get_seq(ts, addr);
553 	if (!seq)
554 		return -ENOMEM;
555 
556 	switch (seq->state) {
557 	case SEQ_STATE_UNINITIALIZED:
558 		/* orphan event, do nothing */
559 		return 0;
560 	case SEQ_STATE_ACQUIRING:
561 		break;
562 	case SEQ_STATE_RELEASED:
563 	case SEQ_STATE_ACQUIRED:
564 	case SEQ_STATE_READ_ACQUIRED:
565 	case SEQ_STATE_CONTENDED:
566 		/* broken lock sequence, discard it */
567 		ls->discard = 1;
568 		bad_hist[BROKEN_CONTENDED]++;
569 		list_del(&seq->list);
570 		free(seq);
571 		goto end;
572 	default:
573 		BUG_ON("Unknown state of lock sequence found!\n");
574 		break;
575 	}
576 
577 	seq->state = SEQ_STATE_CONTENDED;
578 	ls->nr_contended++;
579 	ls->avg_wait_time = ls->wait_time_total/ls->nr_contended;
580 	seq->prev_event_time = sample->time;
581 end:
582 	return 0;
583 }
584 
585 static int report_lock_release_event(struct perf_evsel *evsel,
586 				     struct perf_sample *sample)
587 {
588 	void *addr;
589 	struct lock_stat *ls;
590 	struct thread_stat *ts;
591 	struct lock_seq_stat *seq;
592 	const char *name = perf_evsel__strval(evsel, sample, "name");
593 	u64 tmp = perf_evsel__intval(evsel, sample, "lockdep_addr");
594 
595 	memcpy(&addr, &tmp, sizeof(void *));
596 
597 	ls = lock_stat_findnew(addr, name);
598 	if (!ls)
599 		return -ENOMEM;
600 	if (ls->discard)
601 		return 0;
602 
603 	ts = thread_stat_findnew(sample->tid);
604 	if (!ts)
605 		return -ENOMEM;
606 
607 	seq = get_seq(ts, addr);
608 	if (!seq)
609 		return -ENOMEM;
610 
611 	switch (seq->state) {
612 	case SEQ_STATE_UNINITIALIZED:
613 		goto end;
614 	case SEQ_STATE_ACQUIRED:
615 		break;
616 	case SEQ_STATE_READ_ACQUIRED:
617 		seq->read_count--;
618 		BUG_ON(seq->read_count < 0);
619 		if (!seq->read_count) {
620 			ls->nr_release++;
621 			goto end;
622 		}
623 		break;
624 	case SEQ_STATE_ACQUIRING:
625 	case SEQ_STATE_CONTENDED:
626 	case SEQ_STATE_RELEASED:
627 		/* broken lock sequence, discard it */
628 		ls->discard = 1;
629 		bad_hist[BROKEN_RELEASE]++;
630 		goto free_seq;
631 	default:
632 		BUG_ON("Unknown state of lock sequence found!\n");
633 		break;
634 	}
635 
636 	ls->nr_release++;
637 free_seq:
638 	list_del(&seq->list);
639 	free(seq);
640 end:
641 	return 0;
642 }
643 
644 /* lock oriented handlers */
645 /* TODO: handlers for CPU oriented, thread oriented */
646 static struct trace_lock_handler report_lock_ops  = {
647 	.acquire_event		= report_lock_acquire_event,
648 	.acquired_event		= report_lock_acquired_event,
649 	.contended_event	= report_lock_contended_event,
650 	.release_event		= report_lock_release_event,
651 };
652 
653 static struct trace_lock_handler *trace_handler;
654 
655 static int perf_evsel__process_lock_acquire(struct perf_evsel *evsel,
656 					     struct perf_sample *sample)
657 {
658 	if (trace_handler->acquire_event)
659 		return trace_handler->acquire_event(evsel, sample);
660 	return 0;
661 }
662 
663 static int perf_evsel__process_lock_acquired(struct perf_evsel *evsel,
664 					      struct perf_sample *sample)
665 {
666 	if (trace_handler->acquired_event)
667 		return trace_handler->acquired_event(evsel, sample);
668 	return 0;
669 }
670 
671 static int perf_evsel__process_lock_contended(struct perf_evsel *evsel,
672 					      struct perf_sample *sample)
673 {
674 	if (trace_handler->contended_event)
675 		return trace_handler->contended_event(evsel, sample);
676 	return 0;
677 }
678 
679 static int perf_evsel__process_lock_release(struct perf_evsel *evsel,
680 					    struct perf_sample *sample)
681 {
682 	if (trace_handler->release_event)
683 		return trace_handler->release_event(evsel, sample);
684 	return 0;
685 }
686 
687 static void print_bad_events(int bad, int total)
688 {
689 	/* Output for debug, this have to be removed */
690 	int i;
691 	const char *name[4] =
692 		{ "acquire", "acquired", "contended", "release" };
693 
694 	pr_info("\n=== output for debug===\n\n");
695 	pr_info("bad: %d, total: %d\n", bad, total);
696 	pr_info("bad rate: %.2f %%\n", (double)bad / (double)total * 100);
697 	pr_info("histogram of events caused bad sequence\n");
698 	for (i = 0; i < BROKEN_MAX; i++)
699 		pr_info(" %10s: %d\n", name[i], bad_hist[i]);
700 }
701 
702 /* TODO: various way to print, coloring, nano or milli sec */
703 static void print_result(void)
704 {
705 	struct lock_stat *st;
706 	char cut_name[20];
707 	int bad, total;
708 
709 	pr_info("%20s ", "Name");
710 	pr_info("%10s ", "acquired");
711 	pr_info("%10s ", "contended");
712 
713 	pr_info("%15s ", "avg wait (ns)");
714 	pr_info("%15s ", "total wait (ns)");
715 	pr_info("%15s ", "max wait (ns)");
716 	pr_info("%15s ", "min wait (ns)");
717 
718 	pr_info("\n\n");
719 
720 	bad = total = 0;
721 	while ((st = pop_from_result())) {
722 		total++;
723 		if (st->discard) {
724 			bad++;
725 			continue;
726 		}
727 		bzero(cut_name, 20);
728 
729 		if (strlen(st->name) < 16) {
730 			/* output raw name */
731 			pr_info("%20s ", st->name);
732 		} else {
733 			strncpy(cut_name, st->name, 16);
734 			cut_name[16] = '.';
735 			cut_name[17] = '.';
736 			cut_name[18] = '.';
737 			cut_name[19] = '\0';
738 			/* cut off name for saving output style */
739 			pr_info("%20s ", cut_name);
740 		}
741 
742 		pr_info("%10u ", st->nr_acquired);
743 		pr_info("%10u ", st->nr_contended);
744 
745 		pr_info("%15" PRIu64 " ", st->avg_wait_time);
746 		pr_info("%15" PRIu64 " ", st->wait_time_total);
747 		pr_info("%15" PRIu64 " ", st->wait_time_max);
748 		pr_info("%15" PRIu64 " ", st->wait_time_min == ULLONG_MAX ?
749 		       0 : st->wait_time_min);
750 		pr_info("\n");
751 	}
752 
753 	print_bad_events(bad, total);
754 }
755 
756 static bool info_threads, info_map;
757 
758 static void dump_threads(void)
759 {
760 	struct thread_stat *st;
761 	struct rb_node *node;
762 	struct thread *t;
763 
764 	pr_info("%10s: comm\n", "Thread ID");
765 
766 	node = rb_first(&thread_stats);
767 	while (node) {
768 		st = container_of(node, struct thread_stat, rb);
769 		t = perf_session__findnew(session, st->tid);
770 		pr_info("%10d: %s\n", st->tid, thread__comm_str(t));
771 		node = rb_next(node);
772 		thread__put(t);
773 	};
774 }
775 
776 static void dump_map(void)
777 {
778 	unsigned int i;
779 	struct lock_stat *st;
780 
781 	pr_info("Address of instance: name of class\n");
782 	for (i = 0; i < LOCKHASH_SIZE; i++) {
783 		list_for_each_entry(st, &lockhash_table[i], hash_entry) {
784 			pr_info(" %p: %s\n", st->addr, st->name);
785 		}
786 	}
787 }
788 
789 static int dump_info(void)
790 {
791 	int rc = 0;
792 
793 	if (info_threads)
794 		dump_threads();
795 	else if (info_map)
796 		dump_map();
797 	else {
798 		rc = -1;
799 		pr_err("Unknown type of information\n");
800 	}
801 
802 	return rc;
803 }
804 
805 typedef int (*tracepoint_handler)(struct perf_evsel *evsel,
806 				  struct perf_sample *sample);
807 
808 static int process_sample_event(struct perf_tool *tool __maybe_unused,
809 				union perf_event *event,
810 				struct perf_sample *sample,
811 				struct perf_evsel *evsel,
812 				struct machine *machine)
813 {
814 	int err = 0;
815 	struct thread *thread = machine__findnew_thread(machine, sample->pid,
816 							sample->tid);
817 
818 	if (thread == NULL) {
819 		pr_debug("problem processing %d event, skipping it.\n",
820 			event->header.type);
821 		return -1;
822 	}
823 
824 	if (evsel->handler != NULL) {
825 		tracepoint_handler f = evsel->handler;
826 		err = f(evsel, sample);
827 	}
828 
829 	thread__put(thread);
830 
831 	return err;
832 }
833 
834 static void sort_result(void)
835 {
836 	unsigned int i;
837 	struct lock_stat *st;
838 
839 	for (i = 0; i < LOCKHASH_SIZE; i++) {
840 		list_for_each_entry(st, &lockhash_table[i], hash_entry) {
841 			insert_to_result(st, compare);
842 		}
843 	}
844 }
845 
846 static const struct perf_evsel_str_handler lock_tracepoints[] = {
847 	{ "lock:lock_acquire",	 perf_evsel__process_lock_acquire,   }, /* CONFIG_LOCKDEP */
848 	{ "lock:lock_acquired",	 perf_evsel__process_lock_acquired,  }, /* CONFIG_LOCKDEP, CONFIG_LOCK_STAT */
849 	{ "lock:lock_contended", perf_evsel__process_lock_contended, }, /* CONFIG_LOCKDEP, CONFIG_LOCK_STAT */
850 	{ "lock:lock_release",	 perf_evsel__process_lock_release,   }, /* CONFIG_LOCKDEP */
851 };
852 
853 static bool force;
854 
855 static int __cmd_report(bool display_info)
856 {
857 	int err = -EINVAL;
858 	struct perf_tool eops = {
859 		.sample		 = process_sample_event,
860 		.comm		 = perf_event__process_comm,
861 		.ordered_events	 = true,
862 	};
863 	struct perf_data_file file = {
864 		.path = input_name,
865 		.mode = PERF_DATA_MODE_READ,
866 		.force = force,
867 	};
868 
869 	session = perf_session__new(&file, false, &eops);
870 	if (!session) {
871 		pr_err("Initializing perf session failed\n");
872 		return -1;
873 	}
874 
875 	symbol__init(&session->header.env);
876 
877 	if (!perf_session__has_traces(session, "lock record"))
878 		goto out_delete;
879 
880 	if (perf_session__set_tracepoints_handlers(session, lock_tracepoints)) {
881 		pr_err("Initializing perf session tracepoint handlers failed\n");
882 		goto out_delete;
883 	}
884 
885 	if (select_key())
886 		goto out_delete;
887 
888 	err = perf_session__process_events(session);
889 	if (err)
890 		goto out_delete;
891 
892 	setup_pager();
893 	if (display_info) /* used for info subcommand */
894 		err = dump_info();
895 	else {
896 		sort_result();
897 		print_result();
898 	}
899 
900 out_delete:
901 	perf_session__delete(session);
902 	return err;
903 }
904 
905 static int __cmd_record(int argc, const char **argv)
906 {
907 	const char *record_args[] = {
908 		"record", "-R", "-m", "1024", "-c", "1",
909 	};
910 	unsigned int rec_argc, i, j, ret;
911 	const char **rec_argv;
912 
913 	for (i = 0; i < ARRAY_SIZE(lock_tracepoints); i++) {
914 		if (!is_valid_tracepoint(lock_tracepoints[i].name)) {
915 				pr_err("tracepoint %s is not enabled. "
916 				       "Are CONFIG_LOCKDEP and CONFIG_LOCK_STAT enabled?\n",
917 				       lock_tracepoints[i].name);
918 				return 1;
919 		}
920 	}
921 
922 	rec_argc = ARRAY_SIZE(record_args) + argc - 1;
923 	/* factor of 2 is for -e in front of each tracepoint */
924 	rec_argc += 2 * ARRAY_SIZE(lock_tracepoints);
925 
926 	rec_argv = calloc(rec_argc + 1, sizeof(char *));
927 	if (!rec_argv)
928 		return -ENOMEM;
929 
930 	for (i = 0; i < ARRAY_SIZE(record_args); i++)
931 		rec_argv[i] = strdup(record_args[i]);
932 
933 	for (j = 0; j < ARRAY_SIZE(lock_tracepoints); j++) {
934 		rec_argv[i++] = "-e";
935 		rec_argv[i++] = strdup(lock_tracepoints[j].name);
936 	}
937 
938 	for (j = 1; j < (unsigned int)argc; j++, i++)
939 		rec_argv[i] = argv[j];
940 
941 	BUG_ON(i != rec_argc);
942 
943 	ret = cmd_record(i, rec_argv, NULL);
944 	free(rec_argv);
945 	return ret;
946 }
947 
948 int cmd_lock(int argc, const char **argv, const char *prefix __maybe_unused)
949 {
950 	const struct option info_options[] = {
951 	OPT_BOOLEAN('t', "threads", &info_threads,
952 		    "dump thread list in perf.data"),
953 	OPT_BOOLEAN('m', "map", &info_map,
954 		    "map of lock instances (address:name table)"),
955 	OPT_BOOLEAN('f', "force", &force, "don't complain, do it"),
956 	OPT_END()
957 	};
958 	const struct option lock_options[] = {
959 	OPT_STRING('i', "input", &input_name, "file", "input file name"),
960 	OPT_INCR('v', "verbose", &verbose, "be more verbose (show symbol address, etc)"),
961 	OPT_BOOLEAN('D', "dump-raw-trace", &dump_trace, "dump raw trace in ASCII"),
962 	OPT_END()
963 	};
964 	const struct option report_options[] = {
965 	OPT_STRING('k', "key", &sort_key, "acquired",
966 		    "key for sorting (acquired / contended / avg_wait / wait_total / wait_max / wait_min)"),
967 	OPT_BOOLEAN('f', "force", &force, "don't complain, do it"),
968 	/* TODO: type */
969 	OPT_END()
970 	};
971 	const char * const info_usage[] = {
972 		"perf lock info [<options>]",
973 		NULL
974 	};
975 	const char *const lock_subcommands[] = { "record", "report", "script",
976 						 "info", NULL };
977 	const char *lock_usage[] = {
978 		NULL,
979 		NULL
980 	};
981 	const char * const report_usage[] = {
982 		"perf lock report [<options>]",
983 		NULL
984 	};
985 	unsigned int i;
986 	int rc = 0;
987 
988 	for (i = 0; i < LOCKHASH_SIZE; i++)
989 		INIT_LIST_HEAD(lockhash_table + i);
990 
991 	argc = parse_options_subcommand(argc, argv, lock_options, lock_subcommands,
992 					lock_usage, PARSE_OPT_STOP_AT_NON_OPTION);
993 	if (!argc)
994 		usage_with_options(lock_usage, lock_options);
995 
996 	if (!strncmp(argv[0], "rec", 3)) {
997 		return __cmd_record(argc, argv);
998 	} else if (!strncmp(argv[0], "report", 6)) {
999 		trace_handler = &report_lock_ops;
1000 		if (argc) {
1001 			argc = parse_options(argc, argv,
1002 					     report_options, report_usage, 0);
1003 			if (argc)
1004 				usage_with_options(report_usage, report_options);
1005 		}
1006 		rc = __cmd_report(false);
1007 	} else if (!strcmp(argv[0], "script")) {
1008 		/* Aliased to 'perf script' */
1009 		return cmd_script(argc, argv, prefix);
1010 	} else if (!strcmp(argv[0], "info")) {
1011 		if (argc) {
1012 			argc = parse_options(argc, argv,
1013 					     info_options, info_usage, 0);
1014 			if (argc)
1015 				usage_with_options(info_usage, info_options);
1016 		}
1017 		/* recycling report_lock_ops */
1018 		trace_handler = &report_lock_ops;
1019 		rc = __cmd_report(true);
1020 	} else {
1021 		usage_with_options(lock_usage, lock_options);
1022 	}
1023 
1024 	return rc;
1025 }
1026