xref: /linux/kernel/locking/qspinlock_stat.h (revision e65e175b07bef5974045cc42238de99057669ca7)
1 /* SPDX-License-Identifier: GPL-2.0-or-later */
2 /*
3  *
4  * Authors: Waiman Long <longman@redhat.com>
5  */
6 
7 #include "lock_events.h"
8 
9 #ifdef CONFIG_LOCK_EVENT_COUNTS
10 #ifdef CONFIG_PARAVIRT_SPINLOCKS
11 /*
12  * Collect pvqspinlock locking event counts
13  */
14 #include <linux/sched.h>
15 #include <linux/sched/clock.h>
16 #include <linux/fs.h>
17 
18 #define EVENT_COUNT(ev)	lockevents[LOCKEVENT_ ## ev]
19 
20 /*
21  * PV specific per-cpu counter
22  */
23 static DEFINE_PER_CPU(u64, pv_kick_time);
24 
25 /*
26  * Function to read and return the PV qspinlock counts.
27  *
28  * The following counters are handled specially:
29  * 1. pv_latency_kick
30  *    Average kick latency (ns) = pv_latency_kick/pv_kick_unlock
31  * 2. pv_latency_wake
32  *    Average wake latency (ns) = pv_latency_wake/pv_kick_wake
33  * 3. pv_hash_hops
34  *    Average hops/hash = pv_hash_hops/pv_kick_unlock
35  */
36 ssize_t lockevent_read(struct file *file, char __user *user_buf,
37 		       size_t count, loff_t *ppos)
38 {
39 	char buf[64];
40 	int cpu, id, len;
41 	u64 sum = 0, kicks = 0;
42 
43 	/*
44 	 * Get the counter ID stored in file->f_inode->i_private
45 	 */
46 	id = (long)file_inode(file)->i_private;
47 
48 	if (id >= lockevent_num)
49 		return -EBADF;
50 
51 	for_each_possible_cpu(cpu) {
52 		sum += per_cpu(lockevents[id], cpu);
53 		/*
54 		 * Need to sum additional counters for some of them
55 		 */
56 		switch (id) {
57 
58 		case LOCKEVENT_pv_latency_kick:
59 		case LOCKEVENT_pv_hash_hops:
60 			kicks += per_cpu(EVENT_COUNT(pv_kick_unlock), cpu);
61 			break;
62 
63 		case LOCKEVENT_pv_latency_wake:
64 			kicks += per_cpu(EVENT_COUNT(pv_kick_wake), cpu);
65 			break;
66 		}
67 	}
68 
69 	if (id == LOCKEVENT_pv_hash_hops) {
70 		u64 frac = 0;
71 
72 		if (kicks) {
73 			frac = 100ULL * do_div(sum, kicks);
74 			frac = DIV_ROUND_CLOSEST_ULL(frac, kicks);
75 		}
76 
77 		/*
78 		 * Return a X.XX decimal number
79 		 */
80 		len = snprintf(buf, sizeof(buf) - 1, "%llu.%02llu\n",
81 			       sum, frac);
82 	} else {
83 		/*
84 		 * Round to the nearest ns
85 		 */
86 		if ((id == LOCKEVENT_pv_latency_kick) ||
87 		    (id == LOCKEVENT_pv_latency_wake)) {
88 			if (kicks)
89 				sum = DIV_ROUND_CLOSEST_ULL(sum, kicks);
90 		}
91 		len = snprintf(buf, sizeof(buf) - 1, "%llu\n", sum);
92 	}
93 
94 	return simple_read_from_buffer(user_buf, count, ppos, buf, len);
95 }
96 
97 /*
98  * PV hash hop count
99  */
100 static inline void lockevent_pv_hop(int hopcnt)
101 {
102 	this_cpu_add(EVENT_COUNT(pv_hash_hops), hopcnt);
103 }
104 
105 /*
106  * Replacement function for pv_kick()
107  */
108 static inline void __pv_kick(int cpu)
109 {
110 	u64 start = sched_clock();
111 
112 	per_cpu(pv_kick_time, cpu) = start;
113 	pv_kick(cpu);
114 	this_cpu_add(EVENT_COUNT(pv_latency_kick), sched_clock() - start);
115 }
116 
117 /*
118  * Replacement function for pv_wait()
119  */
120 static inline void __pv_wait(u8 *ptr, u8 val)
121 {
122 	u64 *pkick_time = this_cpu_ptr(&pv_kick_time);
123 
124 	*pkick_time = 0;
125 	pv_wait(ptr, val);
126 	if (*pkick_time) {
127 		this_cpu_add(EVENT_COUNT(pv_latency_wake),
128 			     sched_clock() - *pkick_time);
129 		lockevent_inc(pv_kick_wake);
130 	}
131 }
132 
133 #define pv_kick(c)	__pv_kick(c)
134 #define pv_wait(p, v)	__pv_wait(p, v)
135 
136 #endif /* CONFIG_PARAVIRT_SPINLOCKS */
137 
138 #else /* CONFIG_LOCK_EVENT_COUNTS */
139 
140 static inline void lockevent_pv_hop(int hopcnt)	{ }
141 
142 #endif /* CONFIG_LOCK_EVENT_COUNTS */
143