xref: /linux/arch/powerpc/perf/hv-gpci.c (revision 1f2367a39f17bd553a75e179a747f9b257bc9478)
1 /*
2  * Hypervisor supplied "gpci" ("get performance counter info") performance
3  * counter support
4  *
5  * Author: Cody P Schafer <cody@linux.vnet.ibm.com>
6  * Copyright 2014 IBM Corporation.
7  *
8  * This program is free software; you can redistribute it and/or
9  * modify it under the terms of the GNU General Public License
10  * as published by the Free Software Foundation; either version
11  * 2 of the License, or (at your option) any later version.
12  */
13 
14 #define pr_fmt(fmt) "hv-gpci: " fmt
15 
16 #include <linux/init.h>
17 #include <linux/perf_event.h>
18 #include <asm/firmware.h>
19 #include <asm/hvcall.h>
20 #include <asm/io.h>
21 
22 #include "hv-gpci.h"
23 #include "hv-common.h"
24 
25 /*
26  * Example usage:
27  *  perf stat -e 'hv_gpci/counter_info_version=3,offset=0,length=8,
28  *		  secondary_index=0,starting_index=0xffffffff,request=0x10/' ...
29  */
30 
31 /* u32 */
32 EVENT_DEFINE_RANGE_FORMAT(request, config, 0, 31);
33 /* u32 */
34 /*
35  * Note that starting_index, phys_processor_idx, sibling_part_id,
36  * hw_chip_id, partition_id all refer to the same bit range. They
37  * are basically aliases for the starting_index. The specific alias
38  * used depends on the event. See REQUEST_IDX_KIND in hv-gpci-requests.h
39  */
40 EVENT_DEFINE_RANGE_FORMAT(starting_index, config, 32, 63);
41 EVENT_DEFINE_RANGE_FORMAT_LITE(phys_processor_idx, config, 32, 63);
42 EVENT_DEFINE_RANGE_FORMAT_LITE(sibling_part_id, config, 32, 63);
43 EVENT_DEFINE_RANGE_FORMAT_LITE(hw_chip_id, config, 32, 63);
44 EVENT_DEFINE_RANGE_FORMAT_LITE(partition_id, config, 32, 63);
45 
46 /* u16 */
47 EVENT_DEFINE_RANGE_FORMAT(secondary_index, config1, 0, 15);
48 /* u8 */
49 EVENT_DEFINE_RANGE_FORMAT(counter_info_version, config1, 16, 23);
50 /* u8, bytes of data (1-8) */
51 EVENT_DEFINE_RANGE_FORMAT(length, config1, 24, 31);
52 /* u32, byte offset */
53 EVENT_DEFINE_RANGE_FORMAT(offset, config1, 32, 63);
54 
55 static struct attribute *format_attrs[] = {
56 	&format_attr_request.attr,
57 	&format_attr_starting_index.attr,
58 	&format_attr_phys_processor_idx.attr,
59 	&format_attr_sibling_part_id.attr,
60 	&format_attr_hw_chip_id.attr,
61 	&format_attr_partition_id.attr,
62 	&format_attr_secondary_index.attr,
63 	&format_attr_counter_info_version.attr,
64 
65 	&format_attr_offset.attr,
66 	&format_attr_length.attr,
67 	NULL,
68 };
69 
70 static struct attribute_group format_group = {
71 	.name = "format",
72 	.attrs = format_attrs,
73 };
74 
75 static struct attribute_group event_group = {
76 	.name  = "events",
77 	.attrs = hv_gpci_event_attrs,
78 };
79 
80 #define HV_CAPS_ATTR(_name, _format)				\
81 static ssize_t _name##_show(struct device *dev,			\
82 			    struct device_attribute *attr,	\
83 			    char *page)				\
84 {								\
85 	struct hv_perf_caps caps;				\
86 	unsigned long hret = hv_perf_caps_get(&caps);		\
87 	if (hret)						\
88 		return -EIO;					\
89 								\
90 	return sprintf(page, _format, caps._name);		\
91 }								\
92 static struct device_attribute hv_caps_attr_##_name = __ATTR_RO(_name)
93 
94 static ssize_t kernel_version_show(struct device *dev,
95 				   struct device_attribute *attr,
96 				   char *page)
97 {
98 	return sprintf(page, "0x%x\n", COUNTER_INFO_VERSION_CURRENT);
99 }
100 
101 static DEVICE_ATTR_RO(kernel_version);
102 HV_CAPS_ATTR(version, "0x%x\n");
103 HV_CAPS_ATTR(ga, "%d\n");
104 HV_CAPS_ATTR(expanded, "%d\n");
105 HV_CAPS_ATTR(lab, "%d\n");
106 HV_CAPS_ATTR(collect_privileged, "%d\n");
107 
108 static struct attribute *interface_attrs[] = {
109 	&dev_attr_kernel_version.attr,
110 	&hv_caps_attr_version.attr,
111 	&hv_caps_attr_ga.attr,
112 	&hv_caps_attr_expanded.attr,
113 	&hv_caps_attr_lab.attr,
114 	&hv_caps_attr_collect_privileged.attr,
115 	NULL,
116 };
117 
118 static struct attribute_group interface_group = {
119 	.name = "interface",
120 	.attrs = interface_attrs,
121 };
122 
123 static const struct attribute_group *attr_groups[] = {
124 	&format_group,
125 	&event_group,
126 	&interface_group,
127 	NULL,
128 };
129 
130 #define HGPCI_REQ_BUFFER_SIZE	4096
131 #define HGPCI_MAX_DATA_BYTES \
132 	(HGPCI_REQ_BUFFER_SIZE - sizeof(struct hv_get_perf_counter_info_params))
133 
134 static DEFINE_PER_CPU(char, hv_gpci_reqb[HGPCI_REQ_BUFFER_SIZE]) __aligned(sizeof(uint64_t));
135 
136 struct hv_gpci_request_buffer {
137 	struct hv_get_perf_counter_info_params params;
138 	uint8_t bytes[HGPCI_MAX_DATA_BYTES];
139 } __packed;
140 
141 static unsigned long single_gpci_request(u32 req, u32 starting_index,
142 		u16 secondary_index, u8 version_in, u32 offset, u8 length,
143 		u64 *value)
144 {
145 	unsigned long ret;
146 	size_t i;
147 	u64 count;
148 	struct hv_gpci_request_buffer *arg;
149 
150 	arg = (void *)get_cpu_var(hv_gpci_reqb);
151 	memset(arg, 0, HGPCI_REQ_BUFFER_SIZE);
152 
153 	arg->params.counter_request = cpu_to_be32(req);
154 	arg->params.starting_index = cpu_to_be32(starting_index);
155 	arg->params.secondary_index = cpu_to_be16(secondary_index);
156 	arg->params.counter_info_version_in = version_in;
157 
158 	ret = plpar_hcall_norets(H_GET_PERF_COUNTER_INFO,
159 			virt_to_phys(arg), HGPCI_REQ_BUFFER_SIZE);
160 	if (ret) {
161 		pr_devel("hcall failed: 0x%lx\n", ret);
162 		goto out;
163 	}
164 
165 	/*
166 	 * we verify offset and length are within the zeroed buffer at event
167 	 * init.
168 	 */
169 	count = 0;
170 	for (i = offset; i < offset + length; i++)
171 		count |= arg->bytes[i] << (i - offset);
172 
173 	*value = count;
174 out:
175 	put_cpu_var(hv_gpci_reqb);
176 	return ret;
177 }
178 
179 static u64 h_gpci_get_value(struct perf_event *event)
180 {
181 	u64 count;
182 	unsigned long ret = single_gpci_request(event_get_request(event),
183 					event_get_starting_index(event),
184 					event_get_secondary_index(event),
185 					event_get_counter_info_version(event),
186 					event_get_offset(event),
187 					event_get_length(event),
188 					&count);
189 	if (ret)
190 		return 0;
191 	return count;
192 }
193 
194 static void h_gpci_event_update(struct perf_event *event)
195 {
196 	s64 prev;
197 	u64 now = h_gpci_get_value(event);
198 	prev = local64_xchg(&event->hw.prev_count, now);
199 	local64_add(now - prev, &event->count);
200 }
201 
202 static void h_gpci_event_start(struct perf_event *event, int flags)
203 {
204 	local64_set(&event->hw.prev_count, h_gpci_get_value(event));
205 }
206 
207 static void h_gpci_event_stop(struct perf_event *event, int flags)
208 {
209 	h_gpci_event_update(event);
210 }
211 
212 static int h_gpci_event_add(struct perf_event *event, int flags)
213 {
214 	if (flags & PERF_EF_START)
215 		h_gpci_event_start(event, flags);
216 
217 	return 0;
218 }
219 
220 static int h_gpci_event_init(struct perf_event *event)
221 {
222 	u64 count;
223 	u8 length;
224 
225 	/* Not our event */
226 	if (event->attr.type != event->pmu->type)
227 		return -ENOENT;
228 
229 	/* config2 is unused */
230 	if (event->attr.config2) {
231 		pr_devel("config2 set when reserved\n");
232 		return -EINVAL;
233 	}
234 
235 	/* no branch sampling */
236 	if (has_branch_stack(event))
237 		return -EOPNOTSUPP;
238 
239 	length = event_get_length(event);
240 	if (length < 1 || length > 8) {
241 		pr_devel("length invalid\n");
242 		return -EINVAL;
243 	}
244 
245 	/* last byte within the buffer? */
246 	if ((event_get_offset(event) + length) > HGPCI_MAX_DATA_BYTES) {
247 		pr_devel("request outside of buffer: %zu > %zu\n",
248 				(size_t)event_get_offset(event) + length,
249 				HGPCI_MAX_DATA_BYTES);
250 		return -EINVAL;
251 	}
252 
253 	/* check if the request works... */
254 	if (single_gpci_request(event_get_request(event),
255 				event_get_starting_index(event),
256 				event_get_secondary_index(event),
257 				event_get_counter_info_version(event),
258 				event_get_offset(event),
259 				length,
260 				&count)) {
261 		pr_devel("gpci hcall failed\n");
262 		return -EINVAL;
263 	}
264 
265 	return 0;
266 }
267 
268 static struct pmu h_gpci_pmu = {
269 	.task_ctx_nr = perf_invalid_context,
270 
271 	.name = "hv_gpci",
272 	.attr_groups = attr_groups,
273 	.event_init  = h_gpci_event_init,
274 	.add         = h_gpci_event_add,
275 	.del         = h_gpci_event_stop,
276 	.start       = h_gpci_event_start,
277 	.stop        = h_gpci_event_stop,
278 	.read        = h_gpci_event_update,
279 	.capabilities = PERF_PMU_CAP_NO_EXCLUDE,
280 };
281 
282 static int hv_gpci_init(void)
283 {
284 	int r;
285 	unsigned long hret;
286 	struct hv_perf_caps caps;
287 
288 	hv_gpci_assert_offsets_correct();
289 
290 	if (!firmware_has_feature(FW_FEATURE_LPAR)) {
291 		pr_debug("not a virtualized system, not enabling\n");
292 		return -ENODEV;
293 	}
294 
295 	hret = hv_perf_caps_get(&caps);
296 	if (hret) {
297 		pr_debug("could not obtain capabilities, not enabling, rc=%ld\n",
298 				hret);
299 		return -ENODEV;
300 	}
301 
302 	/* sampling not supported */
303 	h_gpci_pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT;
304 
305 	r = perf_pmu_register(&h_gpci_pmu, h_gpci_pmu.name, -1);
306 	if (r)
307 		return r;
308 
309 	return 0;
310 }
311 
312 device_initcall(hv_gpci_init);
313