xref: /linux/kernel/trace/trace_events_hist.c (revision 5b95a50001f1fbcae6b7fe211a02f37a04dd9a4b)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * trace_events_hist - trace event hist triggers
4  *
5  * Copyright (C) 2015 Tom Zanussi <tom.zanussi@linux.intel.com>
6  */
7 
8 #include <linux/module.h>
9 #include <linux/kallsyms.h>
10 #include <linux/security.h>
11 #include <linux/mutex.h>
12 #include <linux/slab.h>
13 #include <linux/stacktrace.h>
14 #include <linux/rculist.h>
15 #include <linux/tracefs.h>
16 
17 /* for gfp flag names */
18 #include <linux/trace_events.h>
19 #include <trace/events/mmflags.h>
20 
21 #include "tracing_map.h"
22 #include "trace_synth.h"
23 
24 #define ERRORS								\
25 	C(NONE,			"No error"),				\
26 	C(DUPLICATE_VAR,	"Variable already defined"),		\
27 	C(VAR_NOT_UNIQUE,	"Variable name not unique, need to use fully qualified name (subsys.event.var) for variable"), \
28 	C(TOO_MANY_VARS,	"Too many variables defined"),		\
29 	C(MALFORMED_ASSIGNMENT,	"Malformed assignment"),		\
30 	C(NAMED_MISMATCH,	"Named hist trigger doesn't match existing named trigger (includes variables)"), \
31 	C(TRIGGER_EEXIST,	"Hist trigger already exists"),		\
32 	C(TRIGGER_ENOENT_CLEAR,	"Can't clear or continue a nonexistent hist trigger"), \
33 	C(SET_CLOCK_FAIL,	"Couldn't set trace_clock"),		\
34 	C(BAD_FIELD_MODIFIER,	"Invalid field modifier"),		\
35 	C(TOO_MANY_SUBEXPR,	"Too many subexpressions (3 max)"),	\
36 	C(TIMESTAMP_MISMATCH,	"Timestamp units in expression don't match"), \
37 	C(TOO_MANY_FIELD_VARS,	"Too many field variables defined"),	\
38 	C(EVENT_FILE_NOT_FOUND,	"Event file not found"),		\
39 	C(HIST_NOT_FOUND,	"Matching event histogram not found"),	\
40 	C(HIST_CREATE_FAIL,	"Couldn't create histogram for field"),	\
41 	C(SYNTH_VAR_NOT_FOUND,	"Couldn't find synthetic variable"),	\
42 	C(SYNTH_EVENT_NOT_FOUND,"Couldn't find synthetic event"),	\
43 	C(SYNTH_TYPE_MISMATCH,	"Param type doesn't match synthetic event field type"), \
44 	C(SYNTH_COUNT_MISMATCH,	"Param count doesn't match synthetic event field count"), \
45 	C(FIELD_VAR_PARSE_FAIL,	"Couldn't parse field variable"),	\
46 	C(VAR_CREATE_FIND_FAIL,	"Couldn't create or find variable"),	\
47 	C(ONX_NOT_VAR,		"For onmax(x) or onchange(x), x must be a variable"), \
48 	C(ONX_VAR_NOT_FOUND,	"Couldn't find onmax or onchange variable"), \
49 	C(ONX_VAR_CREATE_FAIL,	"Couldn't create onmax or onchange variable"), \
50 	C(FIELD_VAR_CREATE_FAIL,"Couldn't create field variable"),	\
51 	C(TOO_MANY_PARAMS,	"Too many action params"),		\
52 	C(PARAM_NOT_FOUND,	"Couldn't find param"),			\
53 	C(INVALID_PARAM,	"Invalid action param"),		\
54 	C(ACTION_NOT_FOUND,	"No action found"),			\
55 	C(NO_SAVE_PARAMS,	"No params found for save()"),		\
56 	C(TOO_MANY_SAVE_ACTIONS,"Can't have more than one save() action per hist"), \
57 	C(ACTION_MISMATCH,	"Handler doesn't support action"),	\
58 	C(NO_CLOSING_PAREN,	"No closing paren found"),		\
59 	C(SUBSYS_NOT_FOUND,	"Missing subsystem"),			\
60 	C(INVALID_SUBSYS_EVENT,	"Invalid subsystem or event name"),	\
61 	C(INVALID_REF_KEY,	"Using variable references in keys not supported"), \
62 	C(VAR_NOT_FOUND,	"Couldn't find variable"),		\
63 	C(FIELD_NOT_FOUND,	"Couldn't find field"),			\
64 	C(EMPTY_ASSIGNMENT,	"Empty assignment"),			\
65 	C(INVALID_SORT_MODIFIER,"Invalid sort modifier"),		\
66 	C(EMPTY_SORT_FIELD,	"Empty sort field"),			\
67 	C(TOO_MANY_SORT_FIELDS,	"Too many sort fields (Max = 2)"),	\
68 	C(INVALID_SORT_FIELD,	"Sort field must be a key or a val"),	\
69 	C(INVALID_STR_OPERAND,	"String type can not be an operand in expression"), \
70 	C(EXPECT_NUMBER,	"Expecting numeric literal"),		\
71 	C(UNARY_MINUS_SUBEXPR,	"Unary minus not supported in sub-expressions"), \
72 	C(DIVISION_BY_ZERO,	"Division by zero"),			\
73 	C(NEED_NOHC_VAL,	"Non-hitcount value is required for 'nohitcount'"),
74 
75 #undef C
76 #define C(a, b)		HIST_ERR_##a
77 
78 enum { ERRORS };
79 
80 #undef C
81 #define C(a, b)		b
82 
83 static const char *err_text[] = { ERRORS };
84 
85 struct hist_field;
86 
87 typedef u64 (*hist_field_fn_t) (struct hist_field *field,
88 				struct tracing_map_elt *elt,
89 				struct trace_buffer *buffer,
90 				struct ring_buffer_event *rbe,
91 				void *event);
92 
93 #define HIST_FIELD_OPERANDS_MAX	2
94 #define HIST_FIELDS_MAX		(TRACING_MAP_FIELDS_MAX + TRACING_MAP_VARS_MAX)
95 #define HIST_ACTIONS_MAX	8
96 #define HIST_CONST_DIGITS_MAX	21
97 #define HIST_DIV_SHIFT		20  /* For optimizing division by constants */
98 
99 enum field_op_id {
100 	FIELD_OP_NONE,
101 	FIELD_OP_PLUS,
102 	FIELD_OP_MINUS,
103 	FIELD_OP_UNARY_MINUS,
104 	FIELD_OP_DIV,
105 	FIELD_OP_MULT,
106 };
107 
108 enum hist_field_fn {
109 	HIST_FIELD_FN_NOP,
110 	HIST_FIELD_FN_VAR_REF,
111 	HIST_FIELD_FN_COUNTER,
112 	HIST_FIELD_FN_CONST,
113 	HIST_FIELD_FN_LOG2,
114 	HIST_FIELD_FN_BUCKET,
115 	HIST_FIELD_FN_TIMESTAMP,
116 	HIST_FIELD_FN_CPU,
117 	HIST_FIELD_FN_COMM,
118 	HIST_FIELD_FN_STRING,
119 	HIST_FIELD_FN_DYNSTRING,
120 	HIST_FIELD_FN_RELDYNSTRING,
121 	HIST_FIELD_FN_PSTRING,
122 	HIST_FIELD_FN_S64,
123 	HIST_FIELD_FN_U64,
124 	HIST_FIELD_FN_S32,
125 	HIST_FIELD_FN_U32,
126 	HIST_FIELD_FN_S16,
127 	HIST_FIELD_FN_U16,
128 	HIST_FIELD_FN_S8,
129 	HIST_FIELD_FN_U8,
130 	HIST_FIELD_FN_UMINUS,
131 	HIST_FIELD_FN_MINUS,
132 	HIST_FIELD_FN_PLUS,
133 	HIST_FIELD_FN_DIV,
134 	HIST_FIELD_FN_MULT,
135 	HIST_FIELD_FN_DIV_POWER2,
136 	HIST_FIELD_FN_DIV_NOT_POWER2,
137 	HIST_FIELD_FN_DIV_MULT_SHIFT,
138 	HIST_FIELD_FN_EXECNAME,
139 	HIST_FIELD_FN_STACK,
140 };
141 
142 /*
143  * A hist_var (histogram variable) contains variable information for
144  * hist_fields having the HIST_FIELD_FL_VAR or HIST_FIELD_FL_VAR_REF
145  * flag set.  A hist_var has a variable name e.g. ts0, and is
146  * associated with a given histogram trigger, as specified by
147  * hist_data.  The hist_var idx is the unique index assigned to the
148  * variable by the hist trigger's tracing_map.  The idx is what is
149  * used to set a variable's value and, by a variable reference, to
150  * retrieve it.
151  */
152 struct hist_var {
153 	char				*name;
154 	struct hist_trigger_data	*hist_data;
155 	unsigned int			idx;
156 };
157 
158 struct hist_field {
159 	struct ftrace_event_field	*field;
160 	unsigned long			flags;
161 	unsigned long			buckets;
162 	const char			*type;
163 	struct hist_field		*operands[HIST_FIELD_OPERANDS_MAX];
164 	struct hist_trigger_data	*hist_data;
165 	enum hist_field_fn		fn_num;
166 	unsigned int			ref;
167 	unsigned int			size;
168 	unsigned int			offset;
169 	unsigned int                    is_signed;
170 
171 	/*
172 	 * Variable fields contain variable-specific info in var.
173 	 */
174 	struct hist_var			var;
175 	enum field_op_id		operator;
176 	char				*system;
177 	char				*event_name;
178 
179 	/*
180 	 * The name field is used for EXPR and VAR_REF fields.  VAR
181 	 * fields contain the variable name in var.name.
182 	 */
183 	char				*name;
184 
185 	/*
186 	 * When a histogram trigger is hit, if it has any references
187 	 * to variables, the values of those variables are collected
188 	 * into a var_ref_vals array by resolve_var_refs().  The
189 	 * current value of each variable is read from the tracing_map
190 	 * using the hist field's hist_var.idx and entered into the
191 	 * var_ref_idx entry i.e. var_ref_vals[var_ref_idx].
192 	 */
193 	unsigned int			var_ref_idx;
194 	bool                            read_once;
195 
196 	unsigned int			var_str_idx;
197 
198 	/* Numeric literals are represented as u64 */
199 	u64				constant;
200 	/* Used to optimize division by constants */
201 	u64				div_multiplier;
202 };
203 
204 static u64 hist_fn_call(struct hist_field *hist_field,
205 			struct tracing_map_elt *elt,
206 			struct trace_buffer *buffer,
207 			struct ring_buffer_event *rbe,
208 			void *event);
209 
hist_field_const(struct hist_field * field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)210 static u64 hist_field_const(struct hist_field *field,
211 			   struct tracing_map_elt *elt,
212 			   struct trace_buffer *buffer,
213 			   struct ring_buffer_event *rbe,
214 			   void *event)
215 {
216 	return field->constant;
217 }
218 
hist_field_counter(struct hist_field * field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)219 static u64 hist_field_counter(struct hist_field *field,
220 			      struct tracing_map_elt *elt,
221 			      struct trace_buffer *buffer,
222 			      struct ring_buffer_event *rbe,
223 			      void *event)
224 {
225 	return 1;
226 }
227 
hist_field_string(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)228 static u64 hist_field_string(struct hist_field *hist_field,
229 			     struct tracing_map_elt *elt,
230 			     struct trace_buffer *buffer,
231 			     struct ring_buffer_event *rbe,
232 			     void *event)
233 {
234 	char *addr = (char *)(event + hist_field->field->offset);
235 
236 	return (u64)(unsigned long)addr;
237 }
238 
hist_field_dynstring(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)239 static u64 hist_field_dynstring(struct hist_field *hist_field,
240 				struct tracing_map_elt *elt,
241 				struct trace_buffer *buffer,
242 				struct ring_buffer_event *rbe,
243 				void *event)
244 {
245 	u32 str_item = *(u32 *)(event + hist_field->field->offset);
246 	int str_loc = str_item & 0xffff;
247 	char *addr = (char *)(event + str_loc);
248 
249 	return (u64)(unsigned long)addr;
250 }
251 
hist_field_reldynstring(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)252 static u64 hist_field_reldynstring(struct hist_field *hist_field,
253 				   struct tracing_map_elt *elt,
254 				   struct trace_buffer *buffer,
255 				   struct ring_buffer_event *rbe,
256 				   void *event)
257 {
258 	u32 *item = event + hist_field->field->offset;
259 	u32 str_item = *item;
260 	int str_loc = str_item & 0xffff;
261 	char *addr = (char *)&item[1] + str_loc;
262 
263 	return (u64)(unsigned long)addr;
264 }
265 
hist_field_pstring(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)266 static u64 hist_field_pstring(struct hist_field *hist_field,
267 			      struct tracing_map_elt *elt,
268 			      struct trace_buffer *buffer,
269 			      struct ring_buffer_event *rbe,
270 			      void *event)
271 {
272 	char **addr = (char **)(event + hist_field->field->offset);
273 
274 	return (u64)(unsigned long)*addr;
275 }
276 
hist_field_log2(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)277 static u64 hist_field_log2(struct hist_field *hist_field,
278 			   struct tracing_map_elt *elt,
279 			   struct trace_buffer *buffer,
280 			   struct ring_buffer_event *rbe,
281 			   void *event)
282 {
283 	struct hist_field *operand = hist_field->operands[0];
284 
285 	u64 val = hist_fn_call(operand, elt, buffer, rbe, event);
286 
287 	return (u64) ilog2(roundup_pow_of_two(val));
288 }
289 
hist_field_bucket(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)290 static u64 hist_field_bucket(struct hist_field *hist_field,
291 			     struct tracing_map_elt *elt,
292 			     struct trace_buffer *buffer,
293 			     struct ring_buffer_event *rbe,
294 			     void *event)
295 {
296 	struct hist_field *operand = hist_field->operands[0];
297 	unsigned long buckets = hist_field->buckets;
298 
299 	u64 val = hist_fn_call(operand, elt, buffer, rbe, event);
300 
301 	if (WARN_ON_ONCE(!buckets))
302 		return val;
303 
304 	if (val >= LONG_MAX)
305 		val = div64_ul(val, buckets);
306 	else
307 		val = (u64)((unsigned long)val / buckets);
308 	return val * buckets;
309 }
310 
hist_field_plus(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)311 static u64 hist_field_plus(struct hist_field *hist_field,
312 			   struct tracing_map_elt *elt,
313 			   struct trace_buffer *buffer,
314 			   struct ring_buffer_event *rbe,
315 			   void *event)
316 {
317 	struct hist_field *operand1 = hist_field->operands[0];
318 	struct hist_field *operand2 = hist_field->operands[1];
319 
320 	u64 val1 = hist_fn_call(operand1, elt, buffer, rbe, event);
321 	u64 val2 = hist_fn_call(operand2, elt, buffer, rbe, event);
322 
323 	return val1 + val2;
324 }
325 
hist_field_minus(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)326 static u64 hist_field_minus(struct hist_field *hist_field,
327 			    struct tracing_map_elt *elt,
328 			    struct trace_buffer *buffer,
329 			    struct ring_buffer_event *rbe,
330 			    void *event)
331 {
332 	struct hist_field *operand1 = hist_field->operands[0];
333 	struct hist_field *operand2 = hist_field->operands[1];
334 
335 	u64 val1 = hist_fn_call(operand1, elt, buffer, rbe, event);
336 	u64 val2 = hist_fn_call(operand2, elt, buffer, rbe, event);
337 
338 	return val1 - val2;
339 }
340 
hist_field_div(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)341 static u64 hist_field_div(struct hist_field *hist_field,
342 			   struct tracing_map_elt *elt,
343 			   struct trace_buffer *buffer,
344 			   struct ring_buffer_event *rbe,
345 			   void *event)
346 {
347 	struct hist_field *operand1 = hist_field->operands[0];
348 	struct hist_field *operand2 = hist_field->operands[1];
349 
350 	u64 val1 = hist_fn_call(operand1, elt, buffer, rbe, event);
351 	u64 val2 = hist_fn_call(operand2, elt, buffer, rbe, event);
352 
353 	/* Return -1 for the undefined case */
354 	if (!val2)
355 		return -1;
356 
357 	/* Use shift if the divisor is a power of 2 */
358 	if (!(val2 & (val2 - 1)))
359 		return val1 >> __ffs64(val2);
360 
361 	return div64_u64(val1, val2);
362 }
363 
div_by_power_of_two(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)364 static u64 div_by_power_of_two(struct hist_field *hist_field,
365 				struct tracing_map_elt *elt,
366 				struct trace_buffer *buffer,
367 				struct ring_buffer_event *rbe,
368 				void *event)
369 {
370 	struct hist_field *operand1 = hist_field->operands[0];
371 	struct hist_field *operand2 = hist_field->operands[1];
372 
373 	u64 val1 = hist_fn_call(operand1, elt, buffer, rbe, event);
374 
375 	return val1 >> __ffs64(operand2->constant);
376 }
377 
div_by_not_power_of_two(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)378 static u64 div_by_not_power_of_two(struct hist_field *hist_field,
379 				struct tracing_map_elt *elt,
380 				struct trace_buffer *buffer,
381 				struct ring_buffer_event *rbe,
382 				void *event)
383 {
384 	struct hist_field *operand1 = hist_field->operands[0];
385 	struct hist_field *operand2 = hist_field->operands[1];
386 
387 	u64 val1 = hist_fn_call(operand1, elt, buffer, rbe, event);
388 
389 	return div64_u64(val1, operand2->constant);
390 }
391 
div_by_mult_and_shift(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)392 static u64 div_by_mult_and_shift(struct hist_field *hist_field,
393 				struct tracing_map_elt *elt,
394 				struct trace_buffer *buffer,
395 				struct ring_buffer_event *rbe,
396 				void *event)
397 {
398 	struct hist_field *operand1 = hist_field->operands[0];
399 	struct hist_field *operand2 = hist_field->operands[1];
400 
401 	u64 val1 = hist_fn_call(operand1, elt, buffer, rbe, event);
402 
403 	/*
404 	 * If the divisor is a constant, do a multiplication and shift instead.
405 	 *
406 	 * Choose Z = some power of 2. If Y <= Z, then:
407 	 *     X / Y = (X * (Z / Y)) / Z
408 	 *
409 	 * (Z / Y) is a constant (mult) which is calculated at parse time, so:
410 	 *     X / Y = (X * mult) / Z
411 	 *
412 	 * The division by Z can be replaced by a shift since Z is a power of 2:
413 	 *     X / Y = (X * mult) >> HIST_DIV_SHIFT
414 	 *
415 	 * As long, as X < Z the results will not be off by more than 1.
416 	 */
417 	if (val1 < (1 << HIST_DIV_SHIFT)) {
418 		u64 mult = operand2->div_multiplier;
419 
420 		return (val1 * mult + ((1 << HIST_DIV_SHIFT) - 1)) >> HIST_DIV_SHIFT;
421 	}
422 
423 	return div64_u64(val1, operand2->constant);
424 }
425 
hist_field_mult(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)426 static u64 hist_field_mult(struct hist_field *hist_field,
427 			   struct tracing_map_elt *elt,
428 			   struct trace_buffer *buffer,
429 			   struct ring_buffer_event *rbe,
430 			   void *event)
431 {
432 	struct hist_field *operand1 = hist_field->operands[0];
433 	struct hist_field *operand2 = hist_field->operands[1];
434 
435 	u64 val1 = hist_fn_call(operand1, elt, buffer, rbe, event);
436 	u64 val2 = hist_fn_call(operand2, elt, buffer, rbe, event);
437 
438 	return val1 * val2;
439 }
440 
hist_field_unary_minus(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)441 static u64 hist_field_unary_minus(struct hist_field *hist_field,
442 				  struct tracing_map_elt *elt,
443 				  struct trace_buffer *buffer,
444 				  struct ring_buffer_event *rbe,
445 				  void *event)
446 {
447 	struct hist_field *operand = hist_field->operands[0];
448 
449 	s64 sval = (s64)hist_fn_call(operand, elt, buffer, rbe, event);
450 	u64 val = (u64)-sval;
451 
452 	return val;
453 }
454 
455 #define DEFINE_HIST_FIELD_FN(type)					\
456 	static u64 hist_field_##type(struct hist_field *hist_field,	\
457 				     struct tracing_map_elt *elt,	\
458 				     struct trace_buffer *buffer,	\
459 				     struct ring_buffer_event *rbe,	\
460 				     void *event)			\
461 {									\
462 	type *addr = (type *)(event + hist_field->field->offset);	\
463 									\
464 	return (u64)(unsigned long)*addr;				\
465 }
466 
467 DEFINE_HIST_FIELD_FN(s64);
468 DEFINE_HIST_FIELD_FN(u64);
469 DEFINE_HIST_FIELD_FN(s32);
470 DEFINE_HIST_FIELD_FN(u32);
471 DEFINE_HIST_FIELD_FN(s16);
472 DEFINE_HIST_FIELD_FN(u16);
473 DEFINE_HIST_FIELD_FN(s8);
474 DEFINE_HIST_FIELD_FN(u8);
475 
476 #define for_each_hist_field(i, hist_data)	\
477 	for ((i) = 0; (i) < (hist_data)->n_fields; (i)++)
478 
479 #define for_each_hist_val_field(i, hist_data)	\
480 	for ((i) = 0; (i) < (hist_data)->n_vals; (i)++)
481 
482 #define for_each_hist_key_field(i, hist_data)	\
483 	for ((i) = (hist_data)->n_vals; (i) < (hist_data)->n_fields; (i)++)
484 
485 #define HITCOUNT_IDX		0
486 #define HIST_KEY_SIZE_MAX	(MAX_FILTER_STR_VAL + HIST_STACKTRACE_SIZE)
487 
488 enum hist_field_flags {
489 	HIST_FIELD_FL_HITCOUNT		= 1 << 0,
490 	HIST_FIELD_FL_KEY		= 1 << 1,
491 	HIST_FIELD_FL_STRING		= 1 << 2,
492 	HIST_FIELD_FL_HEX		= 1 << 3,
493 	HIST_FIELD_FL_SYM		= 1 << 4,
494 	HIST_FIELD_FL_SYM_OFFSET	= 1 << 5,
495 	HIST_FIELD_FL_EXECNAME		= 1 << 6,
496 	HIST_FIELD_FL_SYSCALL		= 1 << 7,
497 	HIST_FIELD_FL_STACKTRACE	= 1 << 8,
498 	HIST_FIELD_FL_LOG2		= 1 << 9,
499 	HIST_FIELD_FL_TIMESTAMP		= 1 << 10,
500 	HIST_FIELD_FL_TIMESTAMP_USECS	= 1 << 11,
501 	HIST_FIELD_FL_VAR		= 1 << 12,
502 	HIST_FIELD_FL_EXPR		= 1 << 13,
503 	HIST_FIELD_FL_VAR_REF		= 1 << 14,
504 	HIST_FIELD_FL_CPU		= 1 << 15,
505 	HIST_FIELD_FL_ALIAS		= 1 << 16,
506 	HIST_FIELD_FL_BUCKET		= 1 << 17,
507 	HIST_FIELD_FL_CONST		= 1 << 18,
508 	HIST_FIELD_FL_PERCENT		= 1 << 19,
509 	HIST_FIELD_FL_GRAPH		= 1 << 20,
510 	HIST_FIELD_FL_COMM		= 1 << 21,
511 };
512 
513 struct var_defs {
514 	unsigned int	n_vars;
515 	char		*name[TRACING_MAP_VARS_MAX];
516 	char		*expr[TRACING_MAP_VARS_MAX];
517 };
518 
519 struct hist_trigger_attrs {
520 	char		*keys_str;
521 	char		*vals_str;
522 	char		*sort_key_str;
523 	char		*name;
524 	char		*clock;
525 	bool		pause;
526 	bool		cont;
527 	bool		clear;
528 	bool		ts_in_usecs;
529 	bool		no_hitcount;
530 	unsigned int	map_bits;
531 
532 	char		*assignment_str[TRACING_MAP_VARS_MAX];
533 	unsigned int	n_assignments;
534 
535 	char		*action_str[HIST_ACTIONS_MAX];
536 	unsigned int	n_actions;
537 
538 	struct var_defs	var_defs;
539 };
540 
541 struct field_var {
542 	struct hist_field	*var;
543 	struct hist_field	*val;
544 };
545 
546 struct field_var_hist {
547 	struct hist_trigger_data	*hist_data;
548 	char				*cmd;
549 };
550 
551 struct hist_trigger_data {
552 	struct hist_field               *fields[HIST_FIELDS_MAX];
553 	unsigned int			n_vals;
554 	unsigned int			n_keys;
555 	unsigned int			n_fields;
556 	unsigned int			n_vars;
557 	unsigned int			n_var_str;
558 	unsigned int			key_size;
559 	struct tracing_map_sort_key	sort_keys[TRACING_MAP_SORT_KEYS_MAX];
560 	unsigned int			n_sort_keys;
561 	struct trace_event_file		*event_file;
562 	struct hist_trigger_attrs	*attrs;
563 	struct tracing_map		*map;
564 	bool				enable_timestamps;
565 	bool				remove;
566 	struct hist_field               *var_refs[TRACING_MAP_VARS_MAX];
567 	unsigned int			n_var_refs;
568 
569 	struct action_data		*actions[HIST_ACTIONS_MAX];
570 	unsigned int			n_actions;
571 
572 	struct field_var		*field_vars[SYNTH_FIELDS_MAX];
573 	unsigned int			n_field_vars;
574 	unsigned int			n_field_var_str;
575 	struct field_var_hist		*field_var_hists[SYNTH_FIELDS_MAX];
576 	unsigned int			n_field_var_hists;
577 
578 	struct field_var		*save_vars[SYNTH_FIELDS_MAX];
579 	unsigned int			n_save_vars;
580 	unsigned int			n_save_var_str;
581 };
582 
583 struct action_data;
584 
585 typedef void (*action_fn_t) (struct hist_trigger_data *hist_data,
586 			     struct tracing_map_elt *elt,
587 			     struct trace_buffer *buffer, void *rec,
588 			     struct ring_buffer_event *rbe, void *key,
589 			     struct action_data *data, u64 *var_ref_vals);
590 
591 typedef bool (*check_track_val_fn_t) (u64 track_val, u64 var_val);
592 
593 enum handler_id {
594 	HANDLER_ONMATCH = 1,
595 	HANDLER_ONMAX,
596 	HANDLER_ONCHANGE,
597 };
598 
599 enum action_id {
600 	ACTION_SAVE = 1,
601 	ACTION_TRACE,
602 	ACTION_SNAPSHOT,
603 };
604 
605 struct action_data {
606 	enum handler_id		handler;
607 	enum action_id		action;
608 	char			*action_name;
609 	action_fn_t		fn;
610 
611 	unsigned int		n_params;
612 	char			*params[SYNTH_FIELDS_MAX];
613 
614 	/*
615 	 * When a histogram trigger is hit, the values of any
616 	 * references to variables, including variables being passed
617 	 * as parameters to synthetic events, are collected into a
618 	 * var_ref_vals array.  This var_ref_idx array is an array of
619 	 * indices into the var_ref_vals array, one for each synthetic
620 	 * event param, and is passed to the synthetic event
621 	 * invocation.
622 	 */
623 	unsigned int		var_ref_idx[SYNTH_FIELDS_MAX];
624 	struct synth_event	*synth_event;
625 	bool			use_trace_keyword;
626 	char			*synth_event_name;
627 
628 	union {
629 		struct {
630 			char			*event;
631 			char			*event_system;
632 		} match_data;
633 
634 		struct {
635 			/*
636 			 * var_str contains the $-unstripped variable
637 			 * name referenced by var_ref, and used when
638 			 * printing the action.  Because var_ref
639 			 * creation is deferred to create_actions(),
640 			 * we need a per-action way to save it until
641 			 * then, thus var_str.
642 			 */
643 			char			*var_str;
644 
645 			/*
646 			 * var_ref refers to the variable being
647 			 * tracked e.g onmax($var).
648 			 */
649 			struct hist_field	*var_ref;
650 
651 			/*
652 			 * track_var contains the 'invisible' tracking
653 			 * variable created to keep the current
654 			 * e.g. max value.
655 			 */
656 			struct hist_field	*track_var;
657 
658 			check_track_val_fn_t	check_val;
659 			action_fn_t		save_data;
660 		} track_data;
661 	};
662 };
663 
664 struct track_data {
665 	u64				track_val;
666 	bool				updated;
667 
668 	unsigned int			key_len;
669 	void				*key;
670 	struct tracing_map_elt		elt;
671 
672 	struct action_data		*action_data;
673 	struct hist_trigger_data	*hist_data;
674 };
675 
676 struct hist_elt_data {
677 	char *comm;
678 	u64 *var_ref_vals;
679 	char **field_var_str;
680 	int n_field_var_str;
681 };
682 
683 struct snapshot_context {
684 	struct tracing_map_elt	*elt;
685 	void			*key;
686 };
687 
688 /*
689  * Returns the specific division function to use if the divisor
690  * is constant. This avoids extra branches when the trigger is hit.
691  */
hist_field_get_div_fn(struct hist_field * divisor)692 static enum hist_field_fn hist_field_get_div_fn(struct hist_field *divisor)
693 {
694 	u64 div = divisor->constant;
695 
696 	if (!(div & (div - 1)))
697 		return HIST_FIELD_FN_DIV_POWER2;
698 
699 	/* If the divisor is too large, do a regular division */
700 	if (div > (1 << HIST_DIV_SHIFT))
701 		return HIST_FIELD_FN_DIV_NOT_POWER2;
702 
703 	divisor->div_multiplier = div64_u64((u64)(1 << HIST_DIV_SHIFT), div);
704 	return HIST_FIELD_FN_DIV_MULT_SHIFT;
705 }
706 
track_data_free(struct track_data * track_data)707 static void track_data_free(struct track_data *track_data)
708 {
709 	struct hist_elt_data *elt_data;
710 
711 	if (!track_data)
712 		return;
713 
714 	kfree(track_data->key);
715 
716 	elt_data = track_data->elt.private_data;
717 	if (elt_data) {
718 		kfree(elt_data->comm);
719 		kfree(elt_data);
720 	}
721 
722 	kfree(track_data);
723 }
724 
track_data_alloc(unsigned int key_len,struct action_data * action_data,struct hist_trigger_data * hist_data)725 static struct track_data *track_data_alloc(unsigned int key_len,
726 					   struct action_data *action_data,
727 					   struct hist_trigger_data *hist_data)
728 {
729 	struct track_data *data = kzalloc(sizeof(*data), GFP_KERNEL);
730 	struct hist_elt_data *elt_data;
731 
732 	if (!data)
733 		return ERR_PTR(-ENOMEM);
734 
735 	data->key = kzalloc(key_len, GFP_KERNEL);
736 	if (!data->key) {
737 		track_data_free(data);
738 		return ERR_PTR(-ENOMEM);
739 	}
740 
741 	data->key_len = key_len;
742 	data->action_data = action_data;
743 	data->hist_data = hist_data;
744 
745 	elt_data = kzalloc(sizeof(*elt_data), GFP_KERNEL);
746 	if (!elt_data) {
747 		track_data_free(data);
748 		return ERR_PTR(-ENOMEM);
749 	}
750 
751 	data->elt.private_data = elt_data;
752 
753 	elt_data->comm = kzalloc(TASK_COMM_LEN, GFP_KERNEL);
754 	if (!elt_data->comm) {
755 		track_data_free(data);
756 		return ERR_PTR(-ENOMEM);
757 	}
758 
759 	return data;
760 }
761 
762 #define HIST_PREFIX "hist:"
763 
764 static char *last_cmd;
765 static char last_cmd_loc[MAX_FILTER_STR_VAL];
766 
errpos(char * str)767 static int errpos(char *str)
768 {
769 	if (!str || !last_cmd)
770 		return 0;
771 
772 	return err_pos(last_cmd, str);
773 }
774 
last_cmd_set(struct trace_event_file * file,char * str)775 static void last_cmd_set(struct trace_event_file *file, char *str)
776 {
777 	const char *system = NULL, *name = NULL;
778 	struct trace_event_call *call;
779 
780 	if (!str)
781 		return;
782 
783 	kfree(last_cmd);
784 
785 	last_cmd = kasprintf(GFP_KERNEL, HIST_PREFIX "%s", str);
786 	if (!last_cmd)
787 		return;
788 
789 	if (file) {
790 		call = file->event_call;
791 		system = call->class->system;
792 		if (system) {
793 			name = trace_event_name(call);
794 			if (!name)
795 				system = NULL;
796 		}
797 	}
798 
799 	if (system)
800 		snprintf(last_cmd_loc, MAX_FILTER_STR_VAL, HIST_PREFIX "%s:%s", system, name);
801 }
802 
hist_err(struct trace_array * tr,u8 err_type,u16 err_pos)803 static void hist_err(struct trace_array *tr, u8 err_type, u16 err_pos)
804 {
805 	if (!last_cmd)
806 		return;
807 
808 	tracing_log_err(tr, last_cmd_loc, last_cmd, err_text,
809 			err_type, err_pos);
810 }
811 
hist_err_clear(void)812 static void hist_err_clear(void)
813 {
814 	if (last_cmd)
815 		last_cmd[0] = '\0';
816 	last_cmd_loc[0] = '\0';
817 }
818 
819 typedef void (*synth_probe_func_t) (void *__data, u64 *var_ref_vals,
820 				    unsigned int *var_ref_idx);
821 
trace_synth(struct synth_event * event,u64 * var_ref_vals,unsigned int * var_ref_idx)822 static inline void trace_synth(struct synth_event *event, u64 *var_ref_vals,
823 			       unsigned int *var_ref_idx)
824 {
825 	struct tracepoint *tp = event->tp;
826 
827 	if (unlikely(static_key_enabled(&tp->key))) {
828 		struct tracepoint_func *probe_func_ptr;
829 		synth_probe_func_t probe_func;
830 		void *__data;
831 
832 		if (!(cpu_online(raw_smp_processor_id())))
833 			return;
834 
835 		probe_func_ptr = rcu_dereference_sched((tp)->funcs);
836 		if (probe_func_ptr) {
837 			do {
838 				probe_func = probe_func_ptr->func;
839 				__data = probe_func_ptr->data;
840 				probe_func(__data, var_ref_vals, var_ref_idx);
841 			} while ((++probe_func_ptr)->func);
842 		}
843 	}
844 }
845 
action_trace(struct hist_trigger_data * hist_data,struct tracing_map_elt * elt,struct trace_buffer * buffer,void * rec,struct ring_buffer_event * rbe,void * key,struct action_data * data,u64 * var_ref_vals)846 static void action_trace(struct hist_trigger_data *hist_data,
847 			 struct tracing_map_elt *elt,
848 			 struct trace_buffer *buffer, void *rec,
849 			 struct ring_buffer_event *rbe, void *key,
850 			 struct action_data *data, u64 *var_ref_vals)
851 {
852 	struct synth_event *event = data->synth_event;
853 
854 	trace_synth(event, var_ref_vals, data->var_ref_idx);
855 }
856 
857 struct hist_var_data {
858 	struct list_head list;
859 	struct hist_trigger_data *hist_data;
860 };
861 
hist_field_timestamp(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)862 static u64 hist_field_timestamp(struct hist_field *hist_field,
863 				struct tracing_map_elt *elt,
864 				struct trace_buffer *buffer,
865 				struct ring_buffer_event *rbe,
866 				void *event)
867 {
868 	struct hist_trigger_data *hist_data = hist_field->hist_data;
869 	struct trace_array *tr = hist_data->event_file->tr;
870 
871 	u64 ts = ring_buffer_event_time_stamp(buffer, rbe);
872 
873 	if (hist_data->attrs->ts_in_usecs && trace_clock_in_ns(tr))
874 		ts = ns2usecs(ts);
875 
876 	return ts;
877 }
878 
hist_field_cpu(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)879 static u64 hist_field_cpu(struct hist_field *hist_field,
880 			  struct tracing_map_elt *elt,
881 			  struct trace_buffer *buffer,
882 			  struct ring_buffer_event *rbe,
883 			  void *event)
884 {
885 	int cpu = smp_processor_id();
886 
887 	return cpu;
888 }
889 
hist_field_comm(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)890 static u64 hist_field_comm(struct hist_field *hist_field,
891 			  struct tracing_map_elt *elt,
892 			  struct trace_buffer *buffer,
893 			  struct ring_buffer_event *rbe,
894 			  void *event)
895 {
896 	return (u64)(unsigned long)current->comm;
897 }
898 
899 /**
900  * check_field_for_var_ref - Check if a VAR_REF field references a variable
901  * @hist_field: The VAR_REF field to check
902  * @var_data: The hist trigger that owns the variable
903  * @var_idx: The trigger variable identifier
904  *
905  * Check the given VAR_REF field to see whether or not it references
906  * the given variable associated with the given trigger.
907  *
908  * Return: The VAR_REF field if it does reference the variable, NULL if not
909  */
910 static struct hist_field *
check_field_for_var_ref(struct hist_field * hist_field,struct hist_trigger_data * var_data,unsigned int var_idx)911 check_field_for_var_ref(struct hist_field *hist_field,
912 			struct hist_trigger_data *var_data,
913 			unsigned int var_idx)
914 {
915 	WARN_ON(!(hist_field && hist_field->flags & HIST_FIELD_FL_VAR_REF));
916 
917 	if (hist_field && hist_field->var.idx == var_idx &&
918 	    hist_field->var.hist_data == var_data)
919 		return hist_field;
920 
921 	return NULL;
922 }
923 
924 /**
925  * find_var_ref - Check if a trigger has a reference to a trigger variable
926  * @hist_data: The hist trigger that might have a reference to the variable
927  * @var_data: The hist trigger that owns the variable
928  * @var_idx: The trigger variable identifier
929  *
930  * Check the list of var_refs[] on the first hist trigger to see
931  * whether any of them are references to the variable on the second
932  * trigger.
933  *
934  * Return: The VAR_REF field referencing the variable if so, NULL if not
935  */
find_var_ref(struct hist_trigger_data * hist_data,struct hist_trigger_data * var_data,unsigned int var_idx)936 static struct hist_field *find_var_ref(struct hist_trigger_data *hist_data,
937 				       struct hist_trigger_data *var_data,
938 				       unsigned int var_idx)
939 {
940 	struct hist_field *hist_field;
941 	unsigned int i;
942 
943 	for (i = 0; i < hist_data->n_var_refs; i++) {
944 		hist_field = hist_data->var_refs[i];
945 		if (check_field_for_var_ref(hist_field, var_data, var_idx))
946 			return hist_field;
947 	}
948 
949 	return NULL;
950 }
951 
952 /**
953  * find_any_var_ref - Check if there is a reference to a given trigger variable
954  * @hist_data: The hist trigger
955  * @var_idx: The trigger variable identifier
956  *
957  * Check to see whether the given variable is currently referenced by
958  * any other trigger.
959  *
960  * The trigger the variable is defined on is explicitly excluded - the
961  * assumption being that a self-reference doesn't prevent a trigger
962  * from being removed.
963  *
964  * Return: The VAR_REF field referencing the variable if so, NULL if not
965  */
find_any_var_ref(struct hist_trigger_data * hist_data,unsigned int var_idx)966 static struct hist_field *find_any_var_ref(struct hist_trigger_data *hist_data,
967 					   unsigned int var_idx)
968 {
969 	struct trace_array *tr = hist_data->event_file->tr;
970 	struct hist_field *found = NULL;
971 	struct hist_var_data *var_data;
972 
973 	list_for_each_entry(var_data, &tr->hist_vars, list) {
974 		if (var_data->hist_data == hist_data)
975 			continue;
976 		found = find_var_ref(var_data->hist_data, hist_data, var_idx);
977 		if (found)
978 			break;
979 	}
980 
981 	return found;
982 }
983 
984 /**
985  * check_var_refs - Check if there is a reference to any of trigger's variables
986  * @hist_data: The hist trigger
987  *
988  * A trigger can define one or more variables.  If any one of them is
989  * currently referenced by any other trigger, this function will
990  * determine that.
991  *
992  * Typically used to determine whether or not a trigger can be removed
993  * - if there are any references to a trigger's variables, it cannot.
994  *
995  * Return: True if there is a reference to any of trigger's variables
996  */
check_var_refs(struct hist_trigger_data * hist_data)997 static bool check_var_refs(struct hist_trigger_data *hist_data)
998 {
999 	struct hist_field *field;
1000 	bool found = false;
1001 	int i;
1002 
1003 	for_each_hist_field(i, hist_data) {
1004 		field = hist_data->fields[i];
1005 		if (field && field->flags & HIST_FIELD_FL_VAR) {
1006 			if (find_any_var_ref(hist_data, field->var.idx)) {
1007 				found = true;
1008 				break;
1009 			}
1010 		}
1011 	}
1012 
1013 	return found;
1014 }
1015 
find_hist_vars(struct hist_trigger_data * hist_data)1016 static struct hist_var_data *find_hist_vars(struct hist_trigger_data *hist_data)
1017 {
1018 	struct trace_array *tr = hist_data->event_file->tr;
1019 	struct hist_var_data *var_data, *found = NULL;
1020 
1021 	list_for_each_entry(var_data, &tr->hist_vars, list) {
1022 		if (var_data->hist_data == hist_data) {
1023 			found = var_data;
1024 			break;
1025 		}
1026 	}
1027 
1028 	return found;
1029 }
1030 
field_has_hist_vars(struct hist_field * hist_field,unsigned int level)1031 static bool field_has_hist_vars(struct hist_field *hist_field,
1032 				unsigned int level)
1033 {
1034 	int i;
1035 
1036 	if (level > 3)
1037 		return false;
1038 
1039 	if (!hist_field)
1040 		return false;
1041 
1042 	if (hist_field->flags & HIST_FIELD_FL_VAR ||
1043 	    hist_field->flags & HIST_FIELD_FL_VAR_REF)
1044 		return true;
1045 
1046 	for (i = 0; i < HIST_FIELD_OPERANDS_MAX; i++) {
1047 		struct hist_field *operand;
1048 
1049 		operand = hist_field->operands[i];
1050 		if (field_has_hist_vars(operand, level + 1))
1051 			return true;
1052 	}
1053 
1054 	return false;
1055 }
1056 
has_hist_vars(struct hist_trigger_data * hist_data)1057 static bool has_hist_vars(struct hist_trigger_data *hist_data)
1058 {
1059 	struct hist_field *hist_field;
1060 	int i;
1061 
1062 	for_each_hist_field(i, hist_data) {
1063 		hist_field = hist_data->fields[i];
1064 		if (field_has_hist_vars(hist_field, 0))
1065 			return true;
1066 	}
1067 
1068 	return false;
1069 }
1070 
save_hist_vars(struct hist_trigger_data * hist_data)1071 static int save_hist_vars(struct hist_trigger_data *hist_data)
1072 {
1073 	struct trace_array *tr = hist_data->event_file->tr;
1074 	struct hist_var_data *var_data;
1075 
1076 	var_data = find_hist_vars(hist_data);
1077 	if (var_data)
1078 		return 0;
1079 
1080 	if (tracing_check_open_get_tr(tr))
1081 		return -ENODEV;
1082 
1083 	var_data = kzalloc(sizeof(*var_data), GFP_KERNEL);
1084 	if (!var_data) {
1085 		trace_array_put(tr);
1086 		return -ENOMEM;
1087 	}
1088 
1089 	var_data->hist_data = hist_data;
1090 	list_add(&var_data->list, &tr->hist_vars);
1091 
1092 	return 0;
1093 }
1094 
remove_hist_vars(struct hist_trigger_data * hist_data)1095 static void remove_hist_vars(struct hist_trigger_data *hist_data)
1096 {
1097 	struct trace_array *tr = hist_data->event_file->tr;
1098 	struct hist_var_data *var_data;
1099 
1100 	var_data = find_hist_vars(hist_data);
1101 	if (!var_data)
1102 		return;
1103 
1104 	if (WARN_ON(check_var_refs(hist_data)))
1105 		return;
1106 
1107 	list_del(&var_data->list);
1108 
1109 	kfree(var_data);
1110 
1111 	trace_array_put(tr);
1112 }
1113 
find_var_field(struct hist_trigger_data * hist_data,const char * var_name)1114 static struct hist_field *find_var_field(struct hist_trigger_data *hist_data,
1115 					 const char *var_name)
1116 {
1117 	struct hist_field *hist_field, *found = NULL;
1118 	int i;
1119 
1120 	for_each_hist_field(i, hist_data) {
1121 		hist_field = hist_data->fields[i];
1122 		if (hist_field && hist_field->flags & HIST_FIELD_FL_VAR &&
1123 		    strcmp(hist_field->var.name, var_name) == 0) {
1124 			found = hist_field;
1125 			break;
1126 		}
1127 	}
1128 
1129 	return found;
1130 }
1131 
find_var(struct hist_trigger_data * hist_data,struct trace_event_file * file,const char * var_name)1132 static struct hist_field *find_var(struct hist_trigger_data *hist_data,
1133 				   struct trace_event_file *file,
1134 				   const char *var_name)
1135 {
1136 	struct hist_trigger_data *test_data;
1137 	struct event_trigger_data *test;
1138 	struct hist_field *hist_field;
1139 
1140 	lockdep_assert_held(&event_mutex);
1141 
1142 	hist_field = find_var_field(hist_data, var_name);
1143 	if (hist_field)
1144 		return hist_field;
1145 
1146 	list_for_each_entry(test, &file->triggers, list) {
1147 		if (test->cmd_ops->trigger_type == ETT_EVENT_HIST) {
1148 			test_data = test->private_data;
1149 			hist_field = find_var_field(test_data, var_name);
1150 			if (hist_field)
1151 				return hist_field;
1152 		}
1153 	}
1154 
1155 	return NULL;
1156 }
1157 
find_var_file(struct trace_array * tr,char * system,char * event_name,char * var_name)1158 static struct trace_event_file *find_var_file(struct trace_array *tr,
1159 					      char *system,
1160 					      char *event_name,
1161 					      char *var_name)
1162 {
1163 	struct hist_trigger_data *var_hist_data;
1164 	struct hist_var_data *var_data;
1165 	struct trace_event_file *file, *found = NULL;
1166 
1167 	if (system)
1168 		return find_event_file(tr, system, event_name);
1169 
1170 	list_for_each_entry(var_data, &tr->hist_vars, list) {
1171 		var_hist_data = var_data->hist_data;
1172 		file = var_hist_data->event_file;
1173 		if (file == found)
1174 			continue;
1175 
1176 		if (find_var_field(var_hist_data, var_name)) {
1177 			if (found) {
1178 				hist_err(tr, HIST_ERR_VAR_NOT_UNIQUE, errpos(var_name));
1179 				return NULL;
1180 			}
1181 
1182 			found = file;
1183 		}
1184 	}
1185 
1186 	return found;
1187 }
1188 
find_file_var(struct trace_event_file * file,const char * var_name)1189 static struct hist_field *find_file_var(struct trace_event_file *file,
1190 					const char *var_name)
1191 {
1192 	struct hist_trigger_data *test_data;
1193 	struct event_trigger_data *test;
1194 	struct hist_field *hist_field;
1195 
1196 	lockdep_assert_held(&event_mutex);
1197 
1198 	list_for_each_entry(test, &file->triggers, list) {
1199 		if (test->cmd_ops->trigger_type == ETT_EVENT_HIST) {
1200 			test_data = test->private_data;
1201 			hist_field = find_var_field(test_data, var_name);
1202 			if (hist_field)
1203 				return hist_field;
1204 		}
1205 	}
1206 
1207 	return NULL;
1208 }
1209 
1210 static struct hist_field *
find_match_var(struct hist_trigger_data * hist_data,char * var_name)1211 find_match_var(struct hist_trigger_data *hist_data, char *var_name)
1212 {
1213 	struct trace_array *tr = hist_data->event_file->tr;
1214 	struct hist_field *hist_field, *found = NULL;
1215 	struct trace_event_file *file;
1216 	unsigned int i;
1217 
1218 	for (i = 0; i < hist_data->n_actions; i++) {
1219 		struct action_data *data = hist_data->actions[i];
1220 
1221 		if (data->handler == HANDLER_ONMATCH) {
1222 			char *system = data->match_data.event_system;
1223 			char *event_name = data->match_data.event;
1224 
1225 			file = find_var_file(tr, system, event_name, var_name);
1226 			if (!file)
1227 				continue;
1228 			hist_field = find_file_var(file, var_name);
1229 			if (hist_field) {
1230 				if (found) {
1231 					hist_err(tr, HIST_ERR_VAR_NOT_UNIQUE,
1232 						 errpos(var_name));
1233 					return ERR_PTR(-EINVAL);
1234 				}
1235 
1236 				found = hist_field;
1237 			}
1238 		}
1239 	}
1240 	return found;
1241 }
1242 
find_event_var(struct hist_trigger_data * hist_data,char * system,char * event_name,char * var_name)1243 static struct hist_field *find_event_var(struct hist_trigger_data *hist_data,
1244 					 char *system,
1245 					 char *event_name,
1246 					 char *var_name)
1247 {
1248 	struct trace_array *tr = hist_data->event_file->tr;
1249 	struct hist_field *hist_field = NULL;
1250 	struct trace_event_file *file;
1251 
1252 	if (!system || !event_name) {
1253 		hist_field = find_match_var(hist_data, var_name);
1254 		if (IS_ERR(hist_field))
1255 			return NULL;
1256 		if (hist_field)
1257 			return hist_field;
1258 	}
1259 
1260 	file = find_var_file(tr, system, event_name, var_name);
1261 	if (!file)
1262 		return NULL;
1263 
1264 	hist_field = find_file_var(file, var_name);
1265 
1266 	return hist_field;
1267 }
1268 
hist_field_var_ref(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)1269 static u64 hist_field_var_ref(struct hist_field *hist_field,
1270 			      struct tracing_map_elt *elt,
1271 			      struct trace_buffer *buffer,
1272 			      struct ring_buffer_event *rbe,
1273 			      void *event)
1274 {
1275 	struct hist_elt_data *elt_data;
1276 	u64 var_val = 0;
1277 
1278 	if (WARN_ON_ONCE(!elt))
1279 		return var_val;
1280 
1281 	elt_data = elt->private_data;
1282 	var_val = elt_data->var_ref_vals[hist_field->var_ref_idx];
1283 
1284 	return var_val;
1285 }
1286 
resolve_var_refs(struct hist_trigger_data * hist_data,void * key,u64 * var_ref_vals,bool self)1287 static bool resolve_var_refs(struct hist_trigger_data *hist_data, void *key,
1288 			     u64 *var_ref_vals, bool self)
1289 {
1290 	struct hist_trigger_data *var_data;
1291 	struct tracing_map_elt *var_elt;
1292 	struct hist_field *hist_field;
1293 	unsigned int i, var_idx;
1294 	bool resolved = true;
1295 	u64 var_val = 0;
1296 
1297 	for (i = 0; i < hist_data->n_var_refs; i++) {
1298 		hist_field = hist_data->var_refs[i];
1299 		var_idx = hist_field->var.idx;
1300 		var_data = hist_field->var.hist_data;
1301 
1302 		if (var_data == NULL) {
1303 			resolved = false;
1304 			break;
1305 		}
1306 
1307 		if ((self && var_data != hist_data) ||
1308 		    (!self && var_data == hist_data))
1309 			continue;
1310 
1311 		var_elt = tracing_map_lookup(var_data->map, key);
1312 		if (!var_elt) {
1313 			resolved = false;
1314 			break;
1315 		}
1316 
1317 		if (!tracing_map_var_set(var_elt, var_idx)) {
1318 			resolved = false;
1319 			break;
1320 		}
1321 
1322 		if (self || !hist_field->read_once)
1323 			var_val = tracing_map_read_var(var_elt, var_idx);
1324 		else
1325 			var_val = tracing_map_read_var_once(var_elt, var_idx);
1326 
1327 		var_ref_vals[i] = var_val;
1328 	}
1329 
1330 	return resolved;
1331 }
1332 
hist_field_name(struct hist_field * field,unsigned int level)1333 static const char *hist_field_name(struct hist_field *field,
1334 				   unsigned int level)
1335 {
1336 	const char *field_name = "";
1337 
1338 	if (WARN_ON_ONCE(!field))
1339 		return field_name;
1340 
1341 	if (level > 1)
1342 		return field_name;
1343 
1344 	if (field->field)
1345 		field_name = field->field->name;
1346 	else if (field->flags & HIST_FIELD_FL_LOG2 ||
1347 		 field->flags & HIST_FIELD_FL_ALIAS ||
1348 		 field->flags & HIST_FIELD_FL_BUCKET)
1349 		field_name = hist_field_name(field->operands[0], ++level);
1350 	else if (field->flags & HIST_FIELD_FL_CPU)
1351 		field_name = "common_cpu";
1352 	else if (field->flags & HIST_FIELD_FL_COMM)
1353 		field_name = "common_comm";
1354 	else if (field->flags & HIST_FIELD_FL_EXPR ||
1355 		 field->flags & HIST_FIELD_FL_VAR_REF) {
1356 		if (field->system) {
1357 			static char full_name[MAX_FILTER_STR_VAL];
1358 
1359 			strcat(full_name, field->system);
1360 			strcat(full_name, ".");
1361 			strcat(full_name, field->event_name);
1362 			strcat(full_name, ".");
1363 			strcat(full_name, field->name);
1364 			field_name = full_name;
1365 		} else
1366 			field_name = field->name;
1367 	} else if (field->flags & HIST_FIELD_FL_TIMESTAMP)
1368 		field_name = "common_timestamp";
1369 	else if (field->flags & HIST_FIELD_FL_STACKTRACE) {
1370 		field_name = "common_stacktrace";
1371 	} else if (field->flags & HIST_FIELD_FL_HITCOUNT)
1372 		field_name = "hitcount";
1373 
1374 	if (field_name == NULL)
1375 		field_name = "";
1376 
1377 	return field_name;
1378 }
1379 
select_value_fn(int field_size,int field_is_signed)1380 static enum hist_field_fn select_value_fn(int field_size, int field_is_signed)
1381 {
1382 	switch (field_size) {
1383 	case 8:
1384 		if (field_is_signed)
1385 			return HIST_FIELD_FN_S64;
1386 		else
1387 			return HIST_FIELD_FN_U64;
1388 	case 4:
1389 		if (field_is_signed)
1390 			return HIST_FIELD_FN_S32;
1391 		else
1392 			return HIST_FIELD_FN_U32;
1393 	case 2:
1394 		if (field_is_signed)
1395 			return HIST_FIELD_FN_S16;
1396 		else
1397 			return HIST_FIELD_FN_U16;
1398 	case 1:
1399 		if (field_is_signed)
1400 			return HIST_FIELD_FN_S8;
1401 		else
1402 			return HIST_FIELD_FN_U8;
1403 	}
1404 
1405 	return HIST_FIELD_FN_NOP;
1406 }
1407 
parse_map_size(char * str)1408 static int parse_map_size(char *str)
1409 {
1410 	unsigned long size, map_bits;
1411 	int ret;
1412 
1413 	ret = kstrtoul(str, 0, &size);
1414 	if (ret)
1415 		goto out;
1416 
1417 	map_bits = ilog2(roundup_pow_of_two(size));
1418 	if (map_bits < TRACING_MAP_BITS_MIN ||
1419 	    map_bits > TRACING_MAP_BITS_MAX)
1420 		ret = -EINVAL;
1421 	else
1422 		ret = map_bits;
1423  out:
1424 	return ret;
1425 }
1426 
destroy_hist_trigger_attrs(struct hist_trigger_attrs * attrs)1427 static void destroy_hist_trigger_attrs(struct hist_trigger_attrs *attrs)
1428 {
1429 	unsigned int i;
1430 
1431 	if (!attrs)
1432 		return;
1433 
1434 	for (i = 0; i < attrs->n_assignments; i++)
1435 		kfree(attrs->assignment_str[i]);
1436 
1437 	for (i = 0; i < attrs->n_actions; i++)
1438 		kfree(attrs->action_str[i]);
1439 
1440 	kfree(attrs->name);
1441 	kfree(attrs->sort_key_str);
1442 	kfree(attrs->keys_str);
1443 	kfree(attrs->vals_str);
1444 	kfree(attrs->clock);
1445 	kfree(attrs);
1446 }
1447 
parse_action(char * str,struct hist_trigger_attrs * attrs)1448 static int parse_action(char *str, struct hist_trigger_attrs *attrs)
1449 {
1450 	int ret = -EINVAL;
1451 
1452 	if (attrs->n_actions >= HIST_ACTIONS_MAX)
1453 		return ret;
1454 
1455 	if ((str_has_prefix(str, "onmatch(")) ||
1456 	    (str_has_prefix(str, "onmax(")) ||
1457 	    (str_has_prefix(str, "onchange("))) {
1458 		attrs->action_str[attrs->n_actions] = kstrdup(str, GFP_KERNEL);
1459 		if (!attrs->action_str[attrs->n_actions]) {
1460 			ret = -ENOMEM;
1461 			return ret;
1462 		}
1463 		attrs->n_actions++;
1464 		ret = 0;
1465 	}
1466 	return ret;
1467 }
1468 
parse_assignment(struct trace_array * tr,char * str,struct hist_trigger_attrs * attrs)1469 static int parse_assignment(struct trace_array *tr,
1470 			    char *str, struct hist_trigger_attrs *attrs)
1471 {
1472 	int len, ret = 0;
1473 
1474 	if ((len = str_has_prefix(str, "key=")) ||
1475 	    (len = str_has_prefix(str, "keys="))) {
1476 		attrs->keys_str = kstrdup(str + len, GFP_KERNEL);
1477 		if (!attrs->keys_str) {
1478 			ret = -ENOMEM;
1479 			goto out;
1480 		}
1481 	} else if ((len = str_has_prefix(str, "val=")) ||
1482 		   (len = str_has_prefix(str, "vals=")) ||
1483 		   (len = str_has_prefix(str, "values="))) {
1484 		attrs->vals_str = kstrdup(str + len, GFP_KERNEL);
1485 		if (!attrs->vals_str) {
1486 			ret = -ENOMEM;
1487 			goto out;
1488 		}
1489 	} else if ((len = str_has_prefix(str, "sort="))) {
1490 		attrs->sort_key_str = kstrdup(str + len, GFP_KERNEL);
1491 		if (!attrs->sort_key_str) {
1492 			ret = -ENOMEM;
1493 			goto out;
1494 		}
1495 	} else if (str_has_prefix(str, "name=")) {
1496 		attrs->name = kstrdup(str, GFP_KERNEL);
1497 		if (!attrs->name) {
1498 			ret = -ENOMEM;
1499 			goto out;
1500 		}
1501 	} else if ((len = str_has_prefix(str, "clock="))) {
1502 		str += len;
1503 
1504 		str = strstrip(str);
1505 		attrs->clock = kstrdup(str, GFP_KERNEL);
1506 		if (!attrs->clock) {
1507 			ret = -ENOMEM;
1508 			goto out;
1509 		}
1510 	} else if ((len = str_has_prefix(str, "size="))) {
1511 		int map_bits = parse_map_size(str + len);
1512 
1513 		if (map_bits < 0) {
1514 			ret = map_bits;
1515 			goto out;
1516 		}
1517 		attrs->map_bits = map_bits;
1518 	} else {
1519 		char *assignment;
1520 
1521 		if (attrs->n_assignments == TRACING_MAP_VARS_MAX) {
1522 			hist_err(tr, HIST_ERR_TOO_MANY_VARS, errpos(str));
1523 			ret = -EINVAL;
1524 			goto out;
1525 		}
1526 
1527 		assignment = kstrdup(str, GFP_KERNEL);
1528 		if (!assignment) {
1529 			ret = -ENOMEM;
1530 			goto out;
1531 		}
1532 
1533 		attrs->assignment_str[attrs->n_assignments++] = assignment;
1534 	}
1535  out:
1536 	return ret;
1537 }
1538 
1539 static struct hist_trigger_attrs *
parse_hist_trigger_attrs(struct trace_array * tr,char * trigger_str)1540 parse_hist_trigger_attrs(struct trace_array *tr, char *trigger_str)
1541 {
1542 	struct hist_trigger_attrs *attrs;
1543 	int ret = 0;
1544 
1545 	attrs = kzalloc(sizeof(*attrs), GFP_KERNEL);
1546 	if (!attrs)
1547 		return ERR_PTR(-ENOMEM);
1548 
1549 	while (trigger_str) {
1550 		char *str = strsep(&trigger_str, ":");
1551 		char *rhs;
1552 
1553 		rhs = strchr(str, '=');
1554 		if (rhs) {
1555 			if (!strlen(++rhs)) {
1556 				ret = -EINVAL;
1557 				hist_err(tr, HIST_ERR_EMPTY_ASSIGNMENT, errpos(str));
1558 				goto free;
1559 			}
1560 			ret = parse_assignment(tr, str, attrs);
1561 			if (ret)
1562 				goto free;
1563 		} else if (strcmp(str, "nohitcount") == 0 ||
1564 			   strcmp(str, "NOHC") == 0)
1565 			attrs->no_hitcount = true;
1566 		else if (strcmp(str, "pause") == 0)
1567 			attrs->pause = true;
1568 		else if ((strcmp(str, "cont") == 0) ||
1569 			 (strcmp(str, "continue") == 0))
1570 			attrs->cont = true;
1571 		else if (strcmp(str, "clear") == 0)
1572 			attrs->clear = true;
1573 		else {
1574 			ret = parse_action(str, attrs);
1575 			if (ret)
1576 				goto free;
1577 		}
1578 	}
1579 
1580 	if (!attrs->keys_str) {
1581 		ret = -EINVAL;
1582 		goto free;
1583 	}
1584 
1585 	if (!attrs->clock) {
1586 		attrs->clock = kstrdup("global", GFP_KERNEL);
1587 		if (!attrs->clock) {
1588 			ret = -ENOMEM;
1589 			goto free;
1590 		}
1591 	}
1592 
1593 	return attrs;
1594  free:
1595 	destroy_hist_trigger_attrs(attrs);
1596 
1597 	return ERR_PTR(ret);
1598 }
1599 
save_comm(char * comm,struct task_struct * task)1600 static inline void save_comm(char *comm, struct task_struct *task)
1601 {
1602 	if (!task->pid) {
1603 		strcpy(comm, "<idle>");
1604 		return;
1605 	}
1606 
1607 	if (WARN_ON_ONCE(task->pid < 0)) {
1608 		strcpy(comm, "<XXX>");
1609 		return;
1610 	}
1611 
1612 	strscpy(comm, task->comm, TASK_COMM_LEN);
1613 }
1614 
hist_elt_data_free(struct hist_elt_data * elt_data)1615 static void hist_elt_data_free(struct hist_elt_data *elt_data)
1616 {
1617 	unsigned int i;
1618 
1619 	for (i = 0; i < elt_data->n_field_var_str; i++)
1620 		kfree(elt_data->field_var_str[i]);
1621 
1622 	kfree(elt_data->field_var_str);
1623 
1624 	kfree(elt_data->comm);
1625 	kfree(elt_data);
1626 }
1627 
hist_trigger_elt_data_free(struct tracing_map_elt * elt)1628 static void hist_trigger_elt_data_free(struct tracing_map_elt *elt)
1629 {
1630 	struct hist_elt_data *elt_data = elt->private_data;
1631 
1632 	hist_elt_data_free(elt_data);
1633 }
1634 
hist_trigger_elt_data_alloc(struct tracing_map_elt * elt)1635 static int hist_trigger_elt_data_alloc(struct tracing_map_elt *elt)
1636 {
1637 	struct hist_trigger_data *hist_data = elt->map->private_data;
1638 	unsigned int size = TASK_COMM_LEN;
1639 	struct hist_elt_data *elt_data;
1640 	struct hist_field *hist_field;
1641 	unsigned int i, n_str;
1642 
1643 	elt_data = kzalloc(sizeof(*elt_data), GFP_KERNEL);
1644 	if (!elt_data)
1645 		return -ENOMEM;
1646 
1647 	for_each_hist_field(i, hist_data) {
1648 		hist_field = hist_data->fields[i];
1649 
1650 		if (hist_field->flags & HIST_FIELD_FL_EXECNAME) {
1651 			elt_data->comm = kzalloc(size, GFP_KERNEL);
1652 			if (!elt_data->comm) {
1653 				kfree(elt_data);
1654 				return -ENOMEM;
1655 			}
1656 			break;
1657 		}
1658 	}
1659 
1660 	n_str = hist_data->n_field_var_str + hist_data->n_save_var_str +
1661 		hist_data->n_var_str;
1662 	if (n_str > SYNTH_FIELDS_MAX) {
1663 		hist_elt_data_free(elt_data);
1664 		return -EINVAL;
1665 	}
1666 
1667 	BUILD_BUG_ON(STR_VAR_LEN_MAX & (sizeof(u64) - 1));
1668 
1669 	size = STR_VAR_LEN_MAX;
1670 
1671 	elt_data->field_var_str = kcalloc(n_str, sizeof(char *), GFP_KERNEL);
1672 	if (!elt_data->field_var_str) {
1673 		hist_elt_data_free(elt_data);
1674 		return -EINVAL;
1675 	}
1676 	elt_data->n_field_var_str = n_str;
1677 
1678 	for (i = 0; i < n_str; i++) {
1679 		elt_data->field_var_str[i] = kzalloc(size, GFP_KERNEL);
1680 		if (!elt_data->field_var_str[i]) {
1681 			hist_elt_data_free(elt_data);
1682 			return -ENOMEM;
1683 		}
1684 	}
1685 
1686 	elt->private_data = elt_data;
1687 
1688 	return 0;
1689 }
1690 
hist_trigger_elt_data_init(struct tracing_map_elt * elt)1691 static void hist_trigger_elt_data_init(struct tracing_map_elt *elt)
1692 {
1693 	struct hist_elt_data *elt_data = elt->private_data;
1694 
1695 	if (elt_data->comm)
1696 		save_comm(elt_data->comm, current);
1697 }
1698 
1699 static const struct tracing_map_ops hist_trigger_elt_data_ops = {
1700 	.elt_alloc	= hist_trigger_elt_data_alloc,
1701 	.elt_free	= hist_trigger_elt_data_free,
1702 	.elt_init	= hist_trigger_elt_data_init,
1703 };
1704 
get_hist_field_flags(struct hist_field * hist_field)1705 static const char *get_hist_field_flags(struct hist_field *hist_field)
1706 {
1707 	const char *flags_str = NULL;
1708 
1709 	if (hist_field->flags & HIST_FIELD_FL_HEX)
1710 		flags_str = "hex";
1711 	else if (hist_field->flags & HIST_FIELD_FL_SYM)
1712 		flags_str = "sym";
1713 	else if (hist_field->flags & HIST_FIELD_FL_SYM_OFFSET)
1714 		flags_str = "sym-offset";
1715 	else if (hist_field->flags & HIST_FIELD_FL_EXECNAME)
1716 		flags_str = "execname";
1717 	else if (hist_field->flags & HIST_FIELD_FL_SYSCALL)
1718 		flags_str = "syscall";
1719 	else if (hist_field->flags & HIST_FIELD_FL_LOG2)
1720 		flags_str = "log2";
1721 	else if (hist_field->flags & HIST_FIELD_FL_BUCKET)
1722 		flags_str = "buckets";
1723 	else if (hist_field->flags & HIST_FIELD_FL_TIMESTAMP_USECS)
1724 		flags_str = "usecs";
1725 	else if (hist_field->flags & HIST_FIELD_FL_PERCENT)
1726 		flags_str = "percent";
1727 	else if (hist_field->flags & HIST_FIELD_FL_GRAPH)
1728 		flags_str = "graph";
1729 	else if (hist_field->flags & HIST_FIELD_FL_STACKTRACE)
1730 		flags_str = "stacktrace";
1731 
1732 	return flags_str;
1733 }
1734 
expr_field_str(struct hist_field * field,char * expr)1735 static void expr_field_str(struct hist_field *field, char *expr)
1736 {
1737 	if (field->flags & HIST_FIELD_FL_VAR_REF)
1738 		strcat(expr, "$");
1739 	else if (field->flags & HIST_FIELD_FL_CONST) {
1740 		char str[HIST_CONST_DIGITS_MAX];
1741 
1742 		snprintf(str, HIST_CONST_DIGITS_MAX, "%llu", field->constant);
1743 		strcat(expr, str);
1744 	}
1745 
1746 	strcat(expr, hist_field_name(field, 0));
1747 
1748 	if (field->flags && !(field->flags & HIST_FIELD_FL_VAR_REF)) {
1749 		const char *flags_str = get_hist_field_flags(field);
1750 
1751 		if (flags_str) {
1752 			strcat(expr, ".");
1753 			strcat(expr, flags_str);
1754 		}
1755 	}
1756 }
1757 
expr_str(struct hist_field * field,unsigned int level)1758 static char *expr_str(struct hist_field *field, unsigned int level)
1759 {
1760 	char *expr;
1761 
1762 	if (level > 1)
1763 		return NULL;
1764 
1765 	expr = kzalloc(MAX_FILTER_STR_VAL, GFP_KERNEL);
1766 	if (!expr)
1767 		return NULL;
1768 
1769 	if (!field->operands[0]) {
1770 		expr_field_str(field, expr);
1771 		return expr;
1772 	}
1773 
1774 	if (field->operator == FIELD_OP_UNARY_MINUS) {
1775 		char *subexpr;
1776 
1777 		strcat(expr, "-(");
1778 		subexpr = expr_str(field->operands[0], ++level);
1779 		if (!subexpr) {
1780 			kfree(expr);
1781 			return NULL;
1782 		}
1783 		strcat(expr, subexpr);
1784 		strcat(expr, ")");
1785 
1786 		kfree(subexpr);
1787 
1788 		return expr;
1789 	}
1790 
1791 	expr_field_str(field->operands[0], expr);
1792 
1793 	switch (field->operator) {
1794 	case FIELD_OP_MINUS:
1795 		strcat(expr, "-");
1796 		break;
1797 	case FIELD_OP_PLUS:
1798 		strcat(expr, "+");
1799 		break;
1800 	case FIELD_OP_DIV:
1801 		strcat(expr, "/");
1802 		break;
1803 	case FIELD_OP_MULT:
1804 		strcat(expr, "*");
1805 		break;
1806 	default:
1807 		kfree(expr);
1808 		return NULL;
1809 	}
1810 
1811 	expr_field_str(field->operands[1], expr);
1812 
1813 	return expr;
1814 }
1815 
1816 /*
1817  * If field_op != FIELD_OP_NONE, *sep points to the root operator
1818  * of the expression tree to be evaluated.
1819  */
contains_operator(char * str,char ** sep)1820 static int contains_operator(char *str, char **sep)
1821 {
1822 	enum field_op_id field_op = FIELD_OP_NONE;
1823 	char *minus_op, *plus_op, *div_op, *mult_op;
1824 
1825 
1826 	/*
1827 	 * Report the last occurrence of the operators first, so that the
1828 	 * expression is evaluated left to right. This is important since
1829 	 * subtraction and division are not associative.
1830 	 *
1831 	 *	e.g
1832 	 *		64/8/4/2 is 1, i.e 64/8/4/2 = ((64/8)/4)/2
1833 	 *		14-7-5-2 is 0, i.e 14-7-5-2 = ((14-7)-5)-2
1834 	 */
1835 
1836 	/*
1837 	 * First, find lower precedence addition and subtraction
1838 	 * since the expression will be evaluated recursively.
1839 	 */
1840 	minus_op = strrchr(str, '-');
1841 	if (minus_op) {
1842 		/*
1843 		 * Unary minus is not supported in sub-expressions. If
1844 		 * present, it is always the next root operator.
1845 		 */
1846 		if (minus_op == str) {
1847 			field_op = FIELD_OP_UNARY_MINUS;
1848 			goto out;
1849 		}
1850 
1851 		field_op = FIELD_OP_MINUS;
1852 	}
1853 
1854 	plus_op = strrchr(str, '+');
1855 	if (plus_op || minus_op) {
1856 		/*
1857 		 * For operators of the same precedence use to rightmost as the
1858 		 * root, so that the expression is evaluated left to right.
1859 		 */
1860 		if (plus_op > minus_op)
1861 			field_op = FIELD_OP_PLUS;
1862 		goto out;
1863 	}
1864 
1865 	/*
1866 	 * Multiplication and division have higher precedence than addition and
1867 	 * subtraction.
1868 	 */
1869 	div_op = strrchr(str, '/');
1870 	if (div_op)
1871 		field_op = FIELD_OP_DIV;
1872 
1873 	mult_op = strrchr(str, '*');
1874 	/*
1875 	 * For operators of the same precedence use to rightmost as the
1876 	 * root, so that the expression is evaluated left to right.
1877 	 */
1878 	if (mult_op > div_op)
1879 		field_op = FIELD_OP_MULT;
1880 
1881 out:
1882 	if (sep) {
1883 		switch (field_op) {
1884 		case FIELD_OP_UNARY_MINUS:
1885 		case FIELD_OP_MINUS:
1886 			*sep = minus_op;
1887 			break;
1888 		case FIELD_OP_PLUS:
1889 			*sep = plus_op;
1890 			break;
1891 		case FIELD_OP_DIV:
1892 			*sep = div_op;
1893 			break;
1894 		case FIELD_OP_MULT:
1895 			*sep = mult_op;
1896 			break;
1897 		case FIELD_OP_NONE:
1898 		default:
1899 			*sep = NULL;
1900 			break;
1901 		}
1902 	}
1903 
1904 	return field_op;
1905 }
1906 
get_hist_field(struct hist_field * hist_field)1907 static void get_hist_field(struct hist_field *hist_field)
1908 {
1909 	hist_field->ref++;
1910 }
1911 
__destroy_hist_field(struct hist_field * hist_field)1912 static void __destroy_hist_field(struct hist_field *hist_field)
1913 {
1914 	if (--hist_field->ref > 1)
1915 		return;
1916 
1917 	kfree(hist_field->var.name);
1918 	kfree(hist_field->name);
1919 
1920 	/* Can likely be a const */
1921 	kfree_const(hist_field->type);
1922 
1923 	kfree(hist_field->system);
1924 	kfree(hist_field->event_name);
1925 
1926 	kfree(hist_field);
1927 }
1928 
destroy_hist_field(struct hist_field * hist_field,unsigned int level)1929 static void destroy_hist_field(struct hist_field *hist_field,
1930 			       unsigned int level)
1931 {
1932 	unsigned int i;
1933 
1934 	if (level > 3)
1935 		return;
1936 
1937 	if (!hist_field)
1938 		return;
1939 
1940 	if (hist_field->flags & HIST_FIELD_FL_VAR_REF)
1941 		return; /* var refs will be destroyed separately */
1942 
1943 	for (i = 0; i < HIST_FIELD_OPERANDS_MAX; i++)
1944 		destroy_hist_field(hist_field->operands[i], level + 1);
1945 
1946 	__destroy_hist_field(hist_field);
1947 }
1948 
create_hist_field(struct hist_trigger_data * hist_data,struct ftrace_event_field * field,unsigned long flags,char * var_name)1949 static struct hist_field *create_hist_field(struct hist_trigger_data *hist_data,
1950 					    struct ftrace_event_field *field,
1951 					    unsigned long flags,
1952 					    char *var_name)
1953 {
1954 	struct hist_field *hist_field;
1955 
1956 	if (field && is_function_field(field))
1957 		return NULL;
1958 
1959 	hist_field = kzalloc(sizeof(struct hist_field), GFP_KERNEL);
1960 	if (!hist_field)
1961 		return NULL;
1962 
1963 	hist_field->ref = 1;
1964 
1965 	hist_field->hist_data = hist_data;
1966 
1967 	if (flags & HIST_FIELD_FL_EXPR || flags & HIST_FIELD_FL_ALIAS)
1968 		goto out; /* caller will populate */
1969 
1970 	if (flags & HIST_FIELD_FL_VAR_REF) {
1971 		hist_field->fn_num = HIST_FIELD_FN_VAR_REF;
1972 		goto out;
1973 	}
1974 
1975 	if (flags & HIST_FIELD_FL_HITCOUNT) {
1976 		hist_field->fn_num = HIST_FIELD_FN_COUNTER;
1977 		hist_field->size = sizeof(u64);
1978 		hist_field->type = "u64";
1979 		goto out;
1980 	}
1981 
1982 	if (flags & HIST_FIELD_FL_CONST) {
1983 		hist_field->fn_num = HIST_FIELD_FN_CONST;
1984 		hist_field->size = sizeof(u64);
1985 		hist_field->type = kstrdup("u64", GFP_KERNEL);
1986 		if (!hist_field->type)
1987 			goto free;
1988 		goto out;
1989 	}
1990 
1991 	if (flags & HIST_FIELD_FL_STACKTRACE) {
1992 		if (field)
1993 			hist_field->fn_num = HIST_FIELD_FN_STACK;
1994 		else
1995 			hist_field->fn_num = HIST_FIELD_FN_NOP;
1996 		hist_field->size = HIST_STACKTRACE_SIZE;
1997 		hist_field->type = kstrdup_const("unsigned long[]", GFP_KERNEL);
1998 		if (!hist_field->type)
1999 			goto free;
2000 		goto out;
2001 	}
2002 
2003 	if (flags & (HIST_FIELD_FL_LOG2 | HIST_FIELD_FL_BUCKET)) {
2004 		unsigned long fl = flags & ~(HIST_FIELD_FL_LOG2 | HIST_FIELD_FL_BUCKET);
2005 		hist_field->fn_num = flags & HIST_FIELD_FL_LOG2 ? HIST_FIELD_FN_LOG2 :
2006 			HIST_FIELD_FN_BUCKET;
2007 		hist_field->operands[0] = create_hist_field(hist_data, field, fl, NULL);
2008 		if (!hist_field->operands[0])
2009 			goto free;
2010 		hist_field->size = hist_field->operands[0]->size;
2011 		hist_field->type = kstrdup_const(hist_field->operands[0]->type, GFP_KERNEL);
2012 		if (!hist_field->type)
2013 			goto free;
2014 		goto out;
2015 	}
2016 
2017 	if (flags & HIST_FIELD_FL_TIMESTAMP) {
2018 		hist_field->fn_num = HIST_FIELD_FN_TIMESTAMP;
2019 		hist_field->size = sizeof(u64);
2020 		hist_field->type = "u64";
2021 		goto out;
2022 	}
2023 
2024 	if (flags & HIST_FIELD_FL_CPU) {
2025 		hist_field->fn_num = HIST_FIELD_FN_CPU;
2026 		hist_field->size = sizeof(int);
2027 		hist_field->type = "unsigned int";
2028 		goto out;
2029 	}
2030 
2031 	if (flags & HIST_FIELD_FL_COMM) {
2032 		hist_field->fn_num = HIST_FIELD_FN_COMM;
2033 		hist_field->size = MAX_FILTER_STR_VAL;
2034 		hist_field->type = "char[]";
2035 		goto out;
2036 	}
2037 
2038 	if (WARN_ON_ONCE(!field))
2039 		goto out;
2040 
2041 	/* Pointers to strings are just pointers and dangerous to dereference */
2042 	if (is_string_field(field) &&
2043 	    (field->filter_type != FILTER_PTR_STRING)) {
2044 		flags |= HIST_FIELD_FL_STRING;
2045 
2046 		hist_field->size = MAX_FILTER_STR_VAL;
2047 		hist_field->type = kstrdup_const(field->type, GFP_KERNEL);
2048 		if (!hist_field->type)
2049 			goto free;
2050 
2051 		if (field->filter_type == FILTER_STATIC_STRING) {
2052 			hist_field->fn_num = HIST_FIELD_FN_STRING;
2053 			hist_field->size = field->size;
2054 		} else if (field->filter_type == FILTER_DYN_STRING) {
2055 			hist_field->fn_num = HIST_FIELD_FN_DYNSTRING;
2056 		} else if (field->filter_type == FILTER_RDYN_STRING)
2057 			hist_field->fn_num = HIST_FIELD_FN_RELDYNSTRING;
2058 		else
2059 			hist_field->fn_num = HIST_FIELD_FN_PSTRING;
2060 	} else {
2061 		hist_field->size = field->size;
2062 		hist_field->is_signed = field->is_signed;
2063 		hist_field->type = kstrdup_const(field->type, GFP_KERNEL);
2064 		if (!hist_field->type)
2065 			goto free;
2066 
2067 		hist_field->fn_num = select_value_fn(field->size,
2068 						     field->is_signed);
2069 		if (hist_field->fn_num == HIST_FIELD_FN_NOP) {
2070 			destroy_hist_field(hist_field, 0);
2071 			return NULL;
2072 		}
2073 	}
2074  out:
2075 	hist_field->field = field;
2076 	hist_field->flags = flags;
2077 
2078 	if (var_name) {
2079 		hist_field->var.name = kstrdup(var_name, GFP_KERNEL);
2080 		if (!hist_field->var.name)
2081 			goto free;
2082 	}
2083 
2084 	return hist_field;
2085  free:
2086 	destroy_hist_field(hist_field, 0);
2087 	return NULL;
2088 }
2089 
destroy_hist_fields(struct hist_trigger_data * hist_data)2090 static void destroy_hist_fields(struct hist_trigger_data *hist_data)
2091 {
2092 	unsigned int i;
2093 
2094 	for (i = 0; i < HIST_FIELDS_MAX; i++) {
2095 		if (hist_data->fields[i]) {
2096 			destroy_hist_field(hist_data->fields[i], 0);
2097 			hist_data->fields[i] = NULL;
2098 		}
2099 	}
2100 
2101 	for (i = 0; i < hist_data->n_var_refs; i++) {
2102 		WARN_ON(!(hist_data->var_refs[i]->flags & HIST_FIELD_FL_VAR_REF));
2103 		__destroy_hist_field(hist_data->var_refs[i]);
2104 		hist_data->var_refs[i] = NULL;
2105 	}
2106 }
2107 
init_var_ref(struct hist_field * ref_field,struct hist_field * var_field,char * system,char * event_name)2108 static int init_var_ref(struct hist_field *ref_field,
2109 			struct hist_field *var_field,
2110 			char *system, char *event_name)
2111 {
2112 	int err = 0;
2113 
2114 	ref_field->var.idx = var_field->var.idx;
2115 	ref_field->var.hist_data = var_field->hist_data;
2116 	ref_field->size = var_field->size;
2117 	ref_field->is_signed = var_field->is_signed;
2118 	ref_field->flags |= var_field->flags &
2119 		(HIST_FIELD_FL_TIMESTAMP | HIST_FIELD_FL_TIMESTAMP_USECS);
2120 
2121 	if (system) {
2122 		ref_field->system = kstrdup(system, GFP_KERNEL);
2123 		if (!ref_field->system)
2124 			return -ENOMEM;
2125 	}
2126 
2127 	if (event_name) {
2128 		ref_field->event_name = kstrdup(event_name, GFP_KERNEL);
2129 		if (!ref_field->event_name) {
2130 			err = -ENOMEM;
2131 			goto free;
2132 		}
2133 	}
2134 
2135 	if (var_field->var.name) {
2136 		ref_field->name = kstrdup(var_field->var.name, GFP_KERNEL);
2137 		if (!ref_field->name) {
2138 			err = -ENOMEM;
2139 			goto free;
2140 		}
2141 	} else if (var_field->name) {
2142 		ref_field->name = kstrdup(var_field->name, GFP_KERNEL);
2143 		if (!ref_field->name) {
2144 			err = -ENOMEM;
2145 			goto free;
2146 		}
2147 	}
2148 
2149 	ref_field->type = kstrdup_const(var_field->type, GFP_KERNEL);
2150 	if (!ref_field->type) {
2151 		err = -ENOMEM;
2152 		goto free;
2153 	}
2154  out:
2155 	return err;
2156  free:
2157 	kfree(ref_field->system);
2158 	ref_field->system = NULL;
2159 	kfree(ref_field->event_name);
2160 	ref_field->event_name = NULL;
2161 	kfree(ref_field->name);
2162 	ref_field->name = NULL;
2163 
2164 	goto out;
2165 }
2166 
find_var_ref_idx(struct hist_trigger_data * hist_data,struct hist_field * var_field)2167 static int find_var_ref_idx(struct hist_trigger_data *hist_data,
2168 			    struct hist_field *var_field)
2169 {
2170 	struct hist_field *ref_field;
2171 	int i;
2172 
2173 	for (i = 0; i < hist_data->n_var_refs; i++) {
2174 		ref_field = hist_data->var_refs[i];
2175 		if (ref_field->var.idx == var_field->var.idx &&
2176 		    ref_field->var.hist_data == var_field->hist_data)
2177 			return i;
2178 	}
2179 
2180 	return -ENOENT;
2181 }
2182 
2183 /**
2184  * create_var_ref - Create a variable reference and attach it to trigger
2185  * @hist_data: The trigger that will be referencing the variable
2186  * @var_field: The VAR field to create a reference to
2187  * @system: The optional system string
2188  * @event_name: The optional event_name string
2189  *
2190  * Given a variable hist_field, create a VAR_REF hist_field that
2191  * represents a reference to it.
2192  *
2193  * This function also adds the reference to the trigger that
2194  * now references the variable.
2195  *
2196  * Return: The VAR_REF field if successful, NULL if not
2197  */
create_var_ref(struct hist_trigger_data * hist_data,struct hist_field * var_field,char * system,char * event_name)2198 static struct hist_field *create_var_ref(struct hist_trigger_data *hist_data,
2199 					 struct hist_field *var_field,
2200 					 char *system, char *event_name)
2201 {
2202 	unsigned long flags = HIST_FIELD_FL_VAR_REF;
2203 	struct hist_field *ref_field;
2204 	int i;
2205 
2206 	/* Check if the variable already exists */
2207 	for (i = 0; i < hist_data->n_var_refs; i++) {
2208 		ref_field = hist_data->var_refs[i];
2209 		if (ref_field->var.idx == var_field->var.idx &&
2210 		    ref_field->var.hist_data == var_field->hist_data) {
2211 			get_hist_field(ref_field);
2212 			return ref_field;
2213 		}
2214 	}
2215 	/* Sanity check to avoid out-of-bound write on 'hist_data->var_refs' */
2216 	if (hist_data->n_var_refs >= TRACING_MAP_VARS_MAX)
2217 		return NULL;
2218 	ref_field = create_hist_field(var_field->hist_data, NULL, flags, NULL);
2219 	if (ref_field) {
2220 		if (init_var_ref(ref_field, var_field, system, event_name)) {
2221 			destroy_hist_field(ref_field, 0);
2222 			return NULL;
2223 		}
2224 
2225 		hist_data->var_refs[hist_data->n_var_refs] = ref_field;
2226 		ref_field->var_ref_idx = hist_data->n_var_refs++;
2227 	}
2228 
2229 	return ref_field;
2230 }
2231 
is_var_ref(char * var_name)2232 static bool is_var_ref(char *var_name)
2233 {
2234 	if (!var_name || strlen(var_name) < 2 || var_name[0] != '$')
2235 		return false;
2236 
2237 	return true;
2238 }
2239 
field_name_from_var(struct hist_trigger_data * hist_data,char * var_name)2240 static char *field_name_from_var(struct hist_trigger_data *hist_data,
2241 				 char *var_name)
2242 {
2243 	char *name, *field;
2244 	unsigned int i;
2245 
2246 	for (i = 0; i < hist_data->attrs->var_defs.n_vars; i++) {
2247 		name = hist_data->attrs->var_defs.name[i];
2248 
2249 		if (strcmp(var_name, name) == 0) {
2250 			field = hist_data->attrs->var_defs.expr[i];
2251 			if (contains_operator(field, NULL) || is_var_ref(field))
2252 				continue;
2253 			return field;
2254 		}
2255 	}
2256 
2257 	return NULL;
2258 }
2259 
local_field_var_ref(struct hist_trigger_data * hist_data,char * system,char * event_name,char * var_name)2260 static char *local_field_var_ref(struct hist_trigger_data *hist_data,
2261 				 char *system, char *event_name,
2262 				 char *var_name)
2263 {
2264 	struct trace_event_call *call;
2265 
2266 	if (system && event_name) {
2267 		call = hist_data->event_file->event_call;
2268 
2269 		if (strcmp(system, call->class->system) != 0)
2270 			return NULL;
2271 
2272 		if (strcmp(event_name, trace_event_name(call)) != 0)
2273 			return NULL;
2274 	}
2275 
2276 	if (!!system != !!event_name)
2277 		return NULL;
2278 
2279 	if (!is_var_ref(var_name))
2280 		return NULL;
2281 
2282 	var_name++;
2283 
2284 	return field_name_from_var(hist_data, var_name);
2285 }
2286 
parse_var_ref(struct hist_trigger_data * hist_data,char * system,char * event_name,char * var_name)2287 static struct hist_field *parse_var_ref(struct hist_trigger_data *hist_data,
2288 					char *system, char *event_name,
2289 					char *var_name)
2290 {
2291 	struct hist_field *var_field = NULL, *ref_field = NULL;
2292 	struct trace_array *tr = hist_data->event_file->tr;
2293 
2294 	if (!is_var_ref(var_name))
2295 		return NULL;
2296 
2297 	var_name++;
2298 
2299 	var_field = find_event_var(hist_data, system, event_name, var_name);
2300 	if (var_field)
2301 		ref_field = create_var_ref(hist_data, var_field,
2302 					   system, event_name);
2303 
2304 	if (!ref_field)
2305 		hist_err(tr, HIST_ERR_VAR_NOT_FOUND, errpos(var_name));
2306 
2307 	return ref_field;
2308 }
2309 
2310 static struct ftrace_event_field *
parse_field(struct hist_trigger_data * hist_data,struct trace_event_file * file,char * field_str,unsigned long * flags,unsigned long * buckets)2311 parse_field(struct hist_trigger_data *hist_data, struct trace_event_file *file,
2312 	    char *field_str, unsigned long *flags, unsigned long *buckets)
2313 {
2314 	struct ftrace_event_field *field = NULL;
2315 	char *field_name, *modifier, *str;
2316 	struct trace_array *tr = file->tr;
2317 
2318 	modifier = str = kstrdup(field_str, GFP_KERNEL);
2319 	if (!modifier)
2320 		return ERR_PTR(-ENOMEM);
2321 
2322 	field_name = strsep(&modifier, ".");
2323 	if (modifier) {
2324 		if (strcmp(modifier, "hex") == 0)
2325 			*flags |= HIST_FIELD_FL_HEX;
2326 		else if (strcmp(modifier, "sym") == 0)
2327 			*flags |= HIST_FIELD_FL_SYM;
2328 		/*
2329 		 * 'sym-offset' occurrences in the trigger string are modified
2330 		 * to 'symXoffset' to simplify arithmetic expression parsing.
2331 		 */
2332 		else if (strcmp(modifier, "symXoffset") == 0)
2333 			*flags |= HIST_FIELD_FL_SYM_OFFSET;
2334 		else if ((strcmp(modifier, "execname") == 0) &&
2335 			 (strcmp(field_name, "common_pid") == 0))
2336 			*flags |= HIST_FIELD_FL_EXECNAME;
2337 		else if (strcmp(modifier, "syscall") == 0)
2338 			*flags |= HIST_FIELD_FL_SYSCALL;
2339 		else if (strcmp(modifier, "stacktrace") == 0)
2340 			*flags |= HIST_FIELD_FL_STACKTRACE;
2341 		else if (strcmp(modifier, "log2") == 0)
2342 			*flags |= HIST_FIELD_FL_LOG2;
2343 		else if (strcmp(modifier, "usecs") == 0)
2344 			*flags |= HIST_FIELD_FL_TIMESTAMP_USECS;
2345 		else if (strncmp(modifier, "bucket", 6) == 0) {
2346 			int ret;
2347 
2348 			modifier += 6;
2349 
2350 			if (*modifier == 's')
2351 				modifier++;
2352 			if (*modifier != '=')
2353 				goto error;
2354 			modifier++;
2355 			ret = kstrtoul(modifier, 0, buckets);
2356 			if (ret || !(*buckets))
2357 				goto error;
2358 			*flags |= HIST_FIELD_FL_BUCKET;
2359 		} else if (strncmp(modifier, "percent", 7) == 0) {
2360 			if (*flags & (HIST_FIELD_FL_VAR | HIST_FIELD_FL_KEY))
2361 				goto error;
2362 			*flags |= HIST_FIELD_FL_PERCENT;
2363 		} else if (strncmp(modifier, "graph", 5) == 0) {
2364 			if (*flags & (HIST_FIELD_FL_VAR | HIST_FIELD_FL_KEY))
2365 				goto error;
2366 			*flags |= HIST_FIELD_FL_GRAPH;
2367 		} else {
2368  error:
2369 			hist_err(tr, HIST_ERR_BAD_FIELD_MODIFIER, errpos(modifier));
2370 			field = ERR_PTR(-EINVAL);
2371 			goto out;
2372 		}
2373 	}
2374 
2375 	if (strcmp(field_name, "common_timestamp") == 0) {
2376 		*flags |= HIST_FIELD_FL_TIMESTAMP;
2377 		hist_data->enable_timestamps = true;
2378 		if (*flags & HIST_FIELD_FL_TIMESTAMP_USECS)
2379 			hist_data->attrs->ts_in_usecs = true;
2380 	} else if (strcmp(field_name, "common_stacktrace") == 0) {
2381 		*flags |= HIST_FIELD_FL_STACKTRACE;
2382 	} else if (strcmp(field_name, "common_cpu") == 0) {
2383 		*flags |= HIST_FIELD_FL_CPU;
2384 	} else if (strcmp(field_name, "common_comm") == 0) {
2385 		*flags |= HIST_FIELD_FL_COMM | HIST_FIELD_FL_STRING;
2386 	} else if (strcmp(field_name, "hitcount") == 0)
2387 		*flags |= HIST_FIELD_FL_HITCOUNT;
2388 	else {
2389 		field = trace_find_event_field(file->event_call, field_name);
2390 		if (!field || !field->size) {
2391 			/*
2392 			 * For backward compatibility, if field_name
2393 			 * was "cpu" or "stacktrace", then we treat this
2394 			 * the same as common_cpu and common_stacktrace
2395 			 * respectively. This also works for "CPU", and
2396 			 * "STACKTRACE".
2397 			 */
2398 			if (field && field->filter_type == FILTER_CPU) {
2399 				*flags |= HIST_FIELD_FL_CPU;
2400 			} else if (field && field->filter_type == FILTER_STACKTRACE) {
2401 				*flags |= HIST_FIELD_FL_STACKTRACE;
2402 			} else if (field && field->filter_type == FILTER_COMM) {
2403 				*flags |= HIST_FIELD_FL_COMM | HIST_FIELD_FL_STRING;
2404 			} else {
2405 				hist_err(tr, HIST_ERR_FIELD_NOT_FOUND,
2406 					 errpos(field_name));
2407 				field = ERR_PTR(-EINVAL);
2408 				goto out;
2409 			}
2410 		}
2411 	}
2412  out:
2413 	kfree(str);
2414 
2415 	return field;
2416 }
2417 
create_alias(struct hist_trigger_data * hist_data,struct hist_field * var_ref,char * var_name)2418 static struct hist_field *create_alias(struct hist_trigger_data *hist_data,
2419 				       struct hist_field *var_ref,
2420 				       char *var_name)
2421 {
2422 	struct hist_field *alias = NULL;
2423 	unsigned long flags = HIST_FIELD_FL_ALIAS | HIST_FIELD_FL_VAR;
2424 
2425 	alias = create_hist_field(hist_data, NULL, flags, var_name);
2426 	if (!alias)
2427 		return NULL;
2428 
2429 	alias->fn_num = var_ref->fn_num;
2430 	alias->operands[0] = var_ref;
2431 
2432 	if (init_var_ref(alias, var_ref, var_ref->system, var_ref->event_name)) {
2433 		destroy_hist_field(alias, 0);
2434 		return NULL;
2435 	}
2436 
2437 	alias->var_ref_idx = var_ref->var_ref_idx;
2438 
2439 	return alias;
2440 }
2441 
parse_const(struct hist_trigger_data * hist_data,char * str,char * var_name,unsigned long * flags)2442 static struct hist_field *parse_const(struct hist_trigger_data *hist_data,
2443 				      char *str, char *var_name,
2444 				      unsigned long *flags)
2445 {
2446 	struct trace_array *tr = hist_data->event_file->tr;
2447 	struct hist_field *field = NULL;
2448 	u64 constant;
2449 
2450 	if (kstrtoull(str, 0, &constant)) {
2451 		hist_err(tr, HIST_ERR_EXPECT_NUMBER, errpos(str));
2452 		return NULL;
2453 	}
2454 
2455 	*flags |= HIST_FIELD_FL_CONST;
2456 	field = create_hist_field(hist_data, NULL, *flags, var_name);
2457 	if (!field)
2458 		return NULL;
2459 
2460 	field->constant = constant;
2461 
2462 	return field;
2463 }
2464 
parse_atom(struct hist_trigger_data * hist_data,struct trace_event_file * file,char * str,unsigned long * flags,char * var_name)2465 static struct hist_field *parse_atom(struct hist_trigger_data *hist_data,
2466 				     struct trace_event_file *file, char *str,
2467 				     unsigned long *flags, char *var_name)
2468 {
2469 	char *s, *ref_system = NULL, *ref_event = NULL, *ref_var = str;
2470 	struct ftrace_event_field *field = NULL;
2471 	struct hist_field *hist_field = NULL;
2472 	unsigned long buckets = 0;
2473 	int ret = 0;
2474 
2475 	if (isdigit(str[0])) {
2476 		hist_field = parse_const(hist_data, str, var_name, flags);
2477 		if (!hist_field) {
2478 			ret = -EINVAL;
2479 			goto out;
2480 		}
2481 		return hist_field;
2482 	}
2483 
2484 	s = strchr(str, '.');
2485 	if (s) {
2486 		s = strchr(++s, '.');
2487 		if (s) {
2488 			ref_system = strsep(&str, ".");
2489 			if (!str) {
2490 				ret = -EINVAL;
2491 				goto out;
2492 			}
2493 			ref_event = strsep(&str, ".");
2494 			if (!str) {
2495 				ret = -EINVAL;
2496 				goto out;
2497 			}
2498 			ref_var = str;
2499 		}
2500 	}
2501 
2502 	s = local_field_var_ref(hist_data, ref_system, ref_event, ref_var);
2503 	if (!s) {
2504 		hist_field = parse_var_ref(hist_data, ref_system,
2505 					   ref_event, ref_var);
2506 		if (hist_field) {
2507 			if (var_name) {
2508 				hist_field = create_alias(hist_data, hist_field, var_name);
2509 				if (!hist_field) {
2510 					ret = -ENOMEM;
2511 					goto out;
2512 				}
2513 			}
2514 			return hist_field;
2515 		}
2516 	} else
2517 		str = s;
2518 
2519 	field = parse_field(hist_data, file, str, flags, &buckets);
2520 	if (IS_ERR(field)) {
2521 		ret = PTR_ERR(field);
2522 		goto out;
2523 	}
2524 
2525 	hist_field = create_hist_field(hist_data, field, *flags, var_name);
2526 	if (!hist_field) {
2527 		ret = -ENOMEM;
2528 		goto out;
2529 	}
2530 	hist_field->buckets = buckets;
2531 
2532 	return hist_field;
2533  out:
2534 	return ERR_PTR(ret);
2535 }
2536 
2537 static struct hist_field *parse_expr(struct hist_trigger_data *hist_data,
2538 				     struct trace_event_file *file,
2539 				     char *str, unsigned long flags,
2540 				     char *var_name, unsigned int *n_subexprs);
2541 
parse_unary(struct hist_trigger_data * hist_data,struct trace_event_file * file,char * str,unsigned long flags,char * var_name,unsigned int * n_subexprs)2542 static struct hist_field *parse_unary(struct hist_trigger_data *hist_data,
2543 				      struct trace_event_file *file,
2544 				      char *str, unsigned long flags,
2545 				      char *var_name, unsigned int *n_subexprs)
2546 {
2547 	struct hist_field *operand1, *expr = NULL;
2548 	unsigned long operand_flags;
2549 	int ret = 0;
2550 	char *s;
2551 
2552 	/* Unary minus operator, increment n_subexprs */
2553 	++*n_subexprs;
2554 
2555 	/* we support only -(xxx) i.e. explicit parens required */
2556 
2557 	if (*n_subexprs > 3) {
2558 		hist_err(file->tr, HIST_ERR_TOO_MANY_SUBEXPR, errpos(str));
2559 		ret = -EINVAL;
2560 		goto free;
2561 	}
2562 
2563 	str++; /* skip leading '-' */
2564 
2565 	s = strchr(str, '(');
2566 	if (s)
2567 		str++;
2568 	else {
2569 		ret = -EINVAL;
2570 		goto free;
2571 	}
2572 
2573 	s = strrchr(str, ')');
2574 	if (s) {
2575 		 /* unary minus not supported in sub-expressions */
2576 		if (*(s+1) != '\0') {
2577 			hist_err(file->tr, HIST_ERR_UNARY_MINUS_SUBEXPR,
2578 				 errpos(str));
2579 			ret = -EINVAL;
2580 			goto free;
2581 		}
2582 		*s = '\0';
2583 	}
2584 	else {
2585 		ret = -EINVAL; /* no closing ')' */
2586 		goto free;
2587 	}
2588 
2589 	flags |= HIST_FIELD_FL_EXPR;
2590 	expr = create_hist_field(hist_data, NULL, flags, var_name);
2591 	if (!expr) {
2592 		ret = -ENOMEM;
2593 		goto free;
2594 	}
2595 
2596 	operand_flags = 0;
2597 	operand1 = parse_expr(hist_data, file, str, operand_flags, NULL, n_subexprs);
2598 	if (IS_ERR(operand1)) {
2599 		ret = PTR_ERR(operand1);
2600 		goto free;
2601 	}
2602 	if (operand1->flags & HIST_FIELD_FL_STRING) {
2603 		/* String type can not be the operand of unary operator. */
2604 		hist_err(file->tr, HIST_ERR_INVALID_STR_OPERAND, errpos(str));
2605 		destroy_hist_field(operand1, 0);
2606 		ret = -EINVAL;
2607 		goto free;
2608 	}
2609 
2610 	expr->flags |= operand1->flags &
2611 		(HIST_FIELD_FL_TIMESTAMP | HIST_FIELD_FL_TIMESTAMP_USECS);
2612 	expr->fn_num = HIST_FIELD_FN_UMINUS;
2613 	expr->operands[0] = operand1;
2614 	expr->size = operand1->size;
2615 	expr->is_signed = operand1->is_signed;
2616 	expr->operator = FIELD_OP_UNARY_MINUS;
2617 	expr->name = expr_str(expr, 0);
2618 	expr->type = kstrdup_const(operand1->type, GFP_KERNEL);
2619 	if (!expr->type) {
2620 		ret = -ENOMEM;
2621 		goto free;
2622 	}
2623 
2624 	return expr;
2625  free:
2626 	destroy_hist_field(expr, 0);
2627 	return ERR_PTR(ret);
2628 }
2629 
2630 /*
2631  * If the operands are var refs, return pointers the
2632  * variable(s) referenced in var1 and var2, else NULL.
2633  */
check_expr_operands(struct trace_array * tr,struct hist_field * operand1,struct hist_field * operand2,struct hist_field ** var1,struct hist_field ** var2)2634 static int check_expr_operands(struct trace_array *tr,
2635 			       struct hist_field *operand1,
2636 			       struct hist_field *operand2,
2637 			       struct hist_field **var1,
2638 			       struct hist_field **var2)
2639 {
2640 	unsigned long operand1_flags = operand1->flags;
2641 	unsigned long operand2_flags = operand2->flags;
2642 
2643 	if ((operand1_flags & HIST_FIELD_FL_VAR_REF) ||
2644 	    (operand1_flags & HIST_FIELD_FL_ALIAS)) {
2645 		struct hist_field *var;
2646 
2647 		var = find_var_field(operand1->var.hist_data, operand1->name);
2648 		if (!var)
2649 			return -EINVAL;
2650 		operand1_flags = var->flags;
2651 		*var1 = var;
2652 	}
2653 
2654 	if ((operand2_flags & HIST_FIELD_FL_VAR_REF) ||
2655 	    (operand2_flags & HIST_FIELD_FL_ALIAS)) {
2656 		struct hist_field *var;
2657 
2658 		var = find_var_field(operand2->var.hist_data, operand2->name);
2659 		if (!var)
2660 			return -EINVAL;
2661 		operand2_flags = var->flags;
2662 		*var2 = var;
2663 	}
2664 
2665 	if ((operand1_flags & HIST_FIELD_FL_TIMESTAMP_USECS) !=
2666 	    (operand2_flags & HIST_FIELD_FL_TIMESTAMP_USECS)) {
2667 		hist_err(tr, HIST_ERR_TIMESTAMP_MISMATCH, 0);
2668 		return -EINVAL;
2669 	}
2670 
2671 	return 0;
2672 }
2673 
parse_expr(struct hist_trigger_data * hist_data,struct trace_event_file * file,char * str,unsigned long flags,char * var_name,unsigned int * n_subexprs)2674 static struct hist_field *parse_expr(struct hist_trigger_data *hist_data,
2675 				     struct trace_event_file *file,
2676 				     char *str, unsigned long flags,
2677 				     char *var_name, unsigned int *n_subexprs)
2678 {
2679 	struct hist_field *operand1 = NULL, *operand2 = NULL, *expr = NULL;
2680 	struct hist_field *var1 = NULL, *var2 = NULL;
2681 	unsigned long operand_flags, operand2_flags;
2682 	int field_op, ret = -EINVAL;
2683 	char *sep, *operand1_str;
2684 	enum hist_field_fn op_fn;
2685 	bool combine_consts;
2686 
2687 	if (*n_subexprs > 3) {
2688 		hist_err(file->tr, HIST_ERR_TOO_MANY_SUBEXPR, errpos(str));
2689 		return ERR_PTR(-EINVAL);
2690 	}
2691 
2692 	field_op = contains_operator(str, &sep);
2693 
2694 	if (field_op == FIELD_OP_NONE)
2695 		return parse_atom(hist_data, file, str, &flags, var_name);
2696 
2697 	if (field_op == FIELD_OP_UNARY_MINUS)
2698 		return parse_unary(hist_data, file, str, flags, var_name, n_subexprs);
2699 
2700 	/* Binary operator found, increment n_subexprs */
2701 	++*n_subexprs;
2702 
2703 	/* Split the expression string at the root operator */
2704 	if (!sep)
2705 		return ERR_PTR(-EINVAL);
2706 
2707 	*sep = '\0';
2708 	operand1_str = str;
2709 	str = sep+1;
2710 
2711 	/* Binary operator requires both operands */
2712 	if (*operand1_str == '\0' || *str == '\0')
2713 		return ERR_PTR(-EINVAL);
2714 
2715 	operand_flags = 0;
2716 
2717 	/* LHS of string is an expression e.g. a+b in a+b+c */
2718 	operand1 = parse_expr(hist_data, file, operand1_str, operand_flags, NULL, n_subexprs);
2719 	if (IS_ERR(operand1))
2720 		return ERR_CAST(operand1);
2721 
2722 	if (operand1->flags & HIST_FIELD_FL_STRING) {
2723 		hist_err(file->tr, HIST_ERR_INVALID_STR_OPERAND, errpos(operand1_str));
2724 		ret = -EINVAL;
2725 		goto free_op1;
2726 	}
2727 
2728 	/* RHS of string is another expression e.g. c in a+b+c */
2729 	operand_flags = 0;
2730 	operand2 = parse_expr(hist_data, file, str, operand_flags, NULL, n_subexprs);
2731 	if (IS_ERR(operand2)) {
2732 		ret = PTR_ERR(operand2);
2733 		goto free_op1;
2734 	}
2735 	if (operand2->flags & HIST_FIELD_FL_STRING) {
2736 		hist_err(file->tr, HIST_ERR_INVALID_STR_OPERAND, errpos(str));
2737 		ret = -EINVAL;
2738 		goto free_operands;
2739 	}
2740 
2741 	switch (field_op) {
2742 	case FIELD_OP_MINUS:
2743 		op_fn = HIST_FIELD_FN_MINUS;
2744 		break;
2745 	case FIELD_OP_PLUS:
2746 		op_fn = HIST_FIELD_FN_PLUS;
2747 		break;
2748 	case FIELD_OP_DIV:
2749 		op_fn = HIST_FIELD_FN_DIV;
2750 		break;
2751 	case FIELD_OP_MULT:
2752 		op_fn = HIST_FIELD_FN_MULT;
2753 		break;
2754 	default:
2755 		ret = -EINVAL;
2756 		goto free_operands;
2757 	}
2758 
2759 	ret = check_expr_operands(file->tr, operand1, operand2, &var1, &var2);
2760 	if (ret)
2761 		goto free_operands;
2762 
2763 	operand_flags = var1 ? var1->flags : operand1->flags;
2764 	operand2_flags = var2 ? var2->flags : operand2->flags;
2765 
2766 	/*
2767 	 * If both operands are constant, the expression can be
2768 	 * collapsed to a single constant.
2769 	 */
2770 	combine_consts = operand_flags & operand2_flags & HIST_FIELD_FL_CONST;
2771 
2772 	flags |= combine_consts ? HIST_FIELD_FL_CONST : HIST_FIELD_FL_EXPR;
2773 
2774 	flags |= operand1->flags &
2775 		(HIST_FIELD_FL_TIMESTAMP | HIST_FIELD_FL_TIMESTAMP_USECS);
2776 
2777 	expr = create_hist_field(hist_data, NULL, flags, var_name);
2778 	if (!expr) {
2779 		ret = -ENOMEM;
2780 		goto free_operands;
2781 	}
2782 
2783 	operand1->read_once = true;
2784 	operand2->read_once = true;
2785 
2786 	/* The operands are now owned and free'd by 'expr' */
2787 	expr->operands[0] = operand1;
2788 	expr->operands[1] = operand2;
2789 
2790 	if (field_op == FIELD_OP_DIV &&
2791 			operand2_flags & HIST_FIELD_FL_CONST) {
2792 		u64 divisor = var2 ? var2->constant : operand2->constant;
2793 
2794 		if (!divisor) {
2795 			hist_err(file->tr, HIST_ERR_DIVISION_BY_ZERO, errpos(str));
2796 			ret = -EDOM;
2797 			goto free_expr;
2798 		}
2799 
2800 		/*
2801 		 * Copy the divisor here so we don't have to look it up
2802 		 * later if this is a var ref
2803 		 */
2804 		operand2->constant = divisor;
2805 		op_fn = hist_field_get_div_fn(operand2);
2806 	}
2807 
2808 	expr->fn_num = op_fn;
2809 
2810 	if (combine_consts) {
2811 		if (var1)
2812 			expr->operands[0] = var1;
2813 		if (var2)
2814 			expr->operands[1] = var2;
2815 
2816 		expr->constant = hist_fn_call(expr, NULL, NULL, NULL, NULL);
2817 		expr->fn_num = HIST_FIELD_FN_CONST;
2818 
2819 		expr->operands[0] = NULL;
2820 		expr->operands[1] = NULL;
2821 
2822 		/*
2823 		 * var refs won't be destroyed immediately
2824 		 * See: destroy_hist_field()
2825 		 */
2826 		destroy_hist_field(operand2, 0);
2827 		destroy_hist_field(operand1, 0);
2828 
2829 		expr->name = expr_str(expr, 0);
2830 	} else {
2831 		/* The operand sizes should be the same, so just pick one */
2832 		expr->size = operand1->size;
2833 		expr->is_signed = operand1->is_signed;
2834 
2835 		expr->operator = field_op;
2836 		expr->type = kstrdup_const(operand1->type, GFP_KERNEL);
2837 		if (!expr->type) {
2838 			ret = -ENOMEM;
2839 			goto free_expr;
2840 		}
2841 
2842 		expr->name = expr_str(expr, 0);
2843 	}
2844 
2845 	return expr;
2846 
2847 free_operands:
2848 	destroy_hist_field(operand2, 0);
2849 free_op1:
2850 	destroy_hist_field(operand1, 0);
2851 	return ERR_PTR(ret);
2852 
2853 free_expr:
2854 	destroy_hist_field(expr, 0);
2855 	return ERR_PTR(ret);
2856 }
2857 
find_trigger_filter(struct hist_trigger_data * hist_data,struct trace_event_file * file)2858 static char *find_trigger_filter(struct hist_trigger_data *hist_data,
2859 				 struct trace_event_file *file)
2860 {
2861 	struct event_trigger_data *test;
2862 
2863 	lockdep_assert_held(&event_mutex);
2864 
2865 	list_for_each_entry(test, &file->triggers, list) {
2866 		if (test->cmd_ops->trigger_type == ETT_EVENT_HIST) {
2867 			if (test->private_data == hist_data)
2868 				return test->filter_str;
2869 		}
2870 	}
2871 
2872 	return NULL;
2873 }
2874 
2875 static struct event_command trigger_hist_cmd;
2876 static int event_hist_trigger_parse(struct event_command *cmd_ops,
2877 				    struct trace_event_file *file,
2878 				    char *glob, char *cmd,
2879 				    char *param_and_filter);
2880 
compatible_keys(struct hist_trigger_data * target_hist_data,struct hist_trigger_data * hist_data,unsigned int n_keys)2881 static bool compatible_keys(struct hist_trigger_data *target_hist_data,
2882 			    struct hist_trigger_data *hist_data,
2883 			    unsigned int n_keys)
2884 {
2885 	struct hist_field *target_hist_field, *hist_field;
2886 	unsigned int n, i, j;
2887 
2888 	if (hist_data->n_fields - hist_data->n_vals != n_keys)
2889 		return false;
2890 
2891 	i = hist_data->n_vals;
2892 	j = target_hist_data->n_vals;
2893 
2894 	for (n = 0; n < n_keys; n++) {
2895 		hist_field = hist_data->fields[i + n];
2896 		target_hist_field = target_hist_data->fields[j + n];
2897 
2898 		if (strcmp(hist_field->type, target_hist_field->type) != 0)
2899 			return false;
2900 		if (hist_field->size != target_hist_field->size)
2901 			return false;
2902 		if (hist_field->is_signed != target_hist_field->is_signed)
2903 			return false;
2904 	}
2905 
2906 	return true;
2907 }
2908 
2909 static struct hist_trigger_data *
find_compatible_hist(struct hist_trigger_data * target_hist_data,struct trace_event_file * file)2910 find_compatible_hist(struct hist_trigger_data *target_hist_data,
2911 		     struct trace_event_file *file)
2912 {
2913 	struct hist_trigger_data *hist_data;
2914 	struct event_trigger_data *test;
2915 	unsigned int n_keys;
2916 
2917 	lockdep_assert_held(&event_mutex);
2918 
2919 	n_keys = target_hist_data->n_fields - target_hist_data->n_vals;
2920 
2921 	list_for_each_entry(test, &file->triggers, list) {
2922 		if (test->cmd_ops->trigger_type == ETT_EVENT_HIST) {
2923 			hist_data = test->private_data;
2924 
2925 			if (compatible_keys(target_hist_data, hist_data, n_keys))
2926 				return hist_data;
2927 		}
2928 	}
2929 
2930 	return NULL;
2931 }
2932 
event_file(struct trace_array * tr,char * system,char * event_name)2933 static struct trace_event_file *event_file(struct trace_array *tr,
2934 					   char *system, char *event_name)
2935 {
2936 	struct trace_event_file *file;
2937 
2938 	file = __find_event_file(tr, system, event_name);
2939 	if (!file)
2940 		return ERR_PTR(-EINVAL);
2941 
2942 	return file;
2943 }
2944 
2945 static struct hist_field *
find_synthetic_field_var(struct hist_trigger_data * target_hist_data,char * system,char * event_name,char * field_name)2946 find_synthetic_field_var(struct hist_trigger_data *target_hist_data,
2947 			 char *system, char *event_name, char *field_name)
2948 {
2949 	struct hist_field *event_var;
2950 	char *synthetic_name;
2951 
2952 	synthetic_name = kzalloc(MAX_FILTER_STR_VAL, GFP_KERNEL);
2953 	if (!synthetic_name)
2954 		return ERR_PTR(-ENOMEM);
2955 
2956 	strcpy(synthetic_name, "synthetic_");
2957 	strcat(synthetic_name, field_name);
2958 
2959 	event_var = find_event_var(target_hist_data, system, event_name, synthetic_name);
2960 
2961 	kfree(synthetic_name);
2962 
2963 	return event_var;
2964 }
2965 
2966 /**
2967  * create_field_var_hist - Automatically create a histogram and var for a field
2968  * @target_hist_data: The target hist trigger
2969  * @subsys_name: Optional subsystem name
2970  * @event_name: Optional event name
2971  * @field_name: The name of the field (and the resulting variable)
2972  *
2973  * Hist trigger actions fetch data from variables, not directly from
2974  * events.  However, for convenience, users are allowed to directly
2975  * specify an event field in an action, which will be automatically
2976  * converted into a variable on their behalf.
2977  *
2978  * If a user specifies a field on an event that isn't the event the
2979  * histogram currently being defined (the target event histogram), the
2980  * only way that can be accomplished is if a new hist trigger is
2981  * created and the field variable defined on that.
2982  *
2983  * This function creates a new histogram compatible with the target
2984  * event (meaning a histogram with the same key as the target
2985  * histogram), and creates a variable for the specified field, but
2986  * with 'synthetic_' prepended to the variable name in order to avoid
2987  * collision with normal field variables.
2988  *
2989  * Return: The variable created for the field.
2990  */
2991 static struct hist_field *
create_field_var_hist(struct hist_trigger_data * target_hist_data,char * subsys_name,char * event_name,char * field_name)2992 create_field_var_hist(struct hist_trigger_data *target_hist_data,
2993 		      char *subsys_name, char *event_name, char *field_name)
2994 {
2995 	struct trace_array *tr = target_hist_data->event_file->tr;
2996 	struct hist_trigger_data *hist_data;
2997 	unsigned int i, n, first = true;
2998 	struct field_var_hist *var_hist;
2999 	struct trace_event_file *file;
3000 	struct hist_field *key_field;
3001 	struct hist_field *event_var;
3002 	char *saved_filter;
3003 	char *cmd;
3004 	int ret;
3005 
3006 	if (target_hist_data->n_field_var_hists >= SYNTH_FIELDS_MAX) {
3007 		hist_err(tr, HIST_ERR_TOO_MANY_FIELD_VARS, errpos(field_name));
3008 		return ERR_PTR(-EINVAL);
3009 	}
3010 
3011 	file = event_file(tr, subsys_name, event_name);
3012 
3013 	if (IS_ERR(file)) {
3014 		hist_err(tr, HIST_ERR_EVENT_FILE_NOT_FOUND, errpos(field_name));
3015 		ret = PTR_ERR(file);
3016 		return ERR_PTR(ret);
3017 	}
3018 
3019 	/*
3020 	 * Look for a histogram compatible with target.  We'll use the
3021 	 * found histogram specification to create a new matching
3022 	 * histogram with our variable on it.  target_hist_data is not
3023 	 * yet a registered histogram so we can't use that.
3024 	 */
3025 	hist_data = find_compatible_hist(target_hist_data, file);
3026 	if (!hist_data) {
3027 		hist_err(tr, HIST_ERR_HIST_NOT_FOUND, errpos(field_name));
3028 		return ERR_PTR(-EINVAL);
3029 	}
3030 
3031 	/* See if a synthetic field variable has already been created */
3032 	event_var = find_synthetic_field_var(target_hist_data, subsys_name,
3033 					     event_name, field_name);
3034 	if (!IS_ERR_OR_NULL(event_var))
3035 		return event_var;
3036 
3037 	var_hist = kzalloc(sizeof(*var_hist), GFP_KERNEL);
3038 	if (!var_hist)
3039 		return ERR_PTR(-ENOMEM);
3040 
3041 	cmd = kzalloc(MAX_FILTER_STR_VAL, GFP_KERNEL);
3042 	if (!cmd) {
3043 		kfree(var_hist);
3044 		return ERR_PTR(-ENOMEM);
3045 	}
3046 
3047 	/* Use the same keys as the compatible histogram */
3048 	strcat(cmd, "keys=");
3049 
3050 	for_each_hist_key_field(i, hist_data) {
3051 		key_field = hist_data->fields[i];
3052 		if (!first)
3053 			strcat(cmd, ",");
3054 		strcat(cmd, key_field->field->name);
3055 		first = false;
3056 	}
3057 
3058 	/* Create the synthetic field variable specification */
3059 	strcat(cmd, ":synthetic_");
3060 	strcat(cmd, field_name);
3061 	strcat(cmd, "=");
3062 	strcat(cmd, field_name);
3063 
3064 	/* Use the same filter as the compatible histogram */
3065 	saved_filter = find_trigger_filter(hist_data, file);
3066 	if (saved_filter) {
3067 		strcat(cmd, " if ");
3068 		strcat(cmd, saved_filter);
3069 	}
3070 
3071 	var_hist->cmd = kstrdup(cmd, GFP_KERNEL);
3072 	if (!var_hist->cmd) {
3073 		kfree(cmd);
3074 		kfree(var_hist);
3075 		return ERR_PTR(-ENOMEM);
3076 	}
3077 
3078 	/* Save the compatible histogram information */
3079 	var_hist->hist_data = hist_data;
3080 
3081 	/* Create the new histogram with our variable */
3082 	ret = event_hist_trigger_parse(&trigger_hist_cmd, file,
3083 				       "", "hist", cmd);
3084 	if (ret) {
3085 		kfree(cmd);
3086 		kfree(var_hist->cmd);
3087 		kfree(var_hist);
3088 		hist_err(tr, HIST_ERR_HIST_CREATE_FAIL, errpos(field_name));
3089 		return ERR_PTR(ret);
3090 	}
3091 
3092 	kfree(cmd);
3093 
3094 	/* If we can't find the variable, something went wrong */
3095 	event_var = find_synthetic_field_var(target_hist_data, subsys_name,
3096 					     event_name, field_name);
3097 	if (IS_ERR_OR_NULL(event_var)) {
3098 		kfree(var_hist->cmd);
3099 		kfree(var_hist);
3100 		hist_err(tr, HIST_ERR_SYNTH_VAR_NOT_FOUND, errpos(field_name));
3101 		return ERR_PTR(-EINVAL);
3102 	}
3103 
3104 	n = target_hist_data->n_field_var_hists;
3105 	target_hist_data->field_var_hists[n] = var_hist;
3106 	target_hist_data->n_field_var_hists++;
3107 
3108 	return event_var;
3109 }
3110 
3111 static struct hist_field *
find_target_event_var(struct hist_trigger_data * hist_data,char * subsys_name,char * event_name,char * var_name)3112 find_target_event_var(struct hist_trigger_data *hist_data,
3113 		      char *subsys_name, char *event_name, char *var_name)
3114 {
3115 	struct trace_event_file *file = hist_data->event_file;
3116 	struct hist_field *hist_field = NULL;
3117 
3118 	if (subsys_name) {
3119 		struct trace_event_call *call;
3120 
3121 		if (!event_name)
3122 			return NULL;
3123 
3124 		call = file->event_call;
3125 
3126 		if (strcmp(subsys_name, call->class->system) != 0)
3127 			return NULL;
3128 
3129 		if (strcmp(event_name, trace_event_name(call)) != 0)
3130 			return NULL;
3131 	}
3132 
3133 	hist_field = find_var_field(hist_data, var_name);
3134 
3135 	return hist_field;
3136 }
3137 
__update_field_vars(struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * rec,struct field_var ** field_vars,unsigned int n_field_vars,unsigned int field_var_str_start)3138 static inline void __update_field_vars(struct tracing_map_elt *elt,
3139 				       struct trace_buffer *buffer,
3140 				       struct ring_buffer_event *rbe,
3141 				       void *rec,
3142 				       struct field_var **field_vars,
3143 				       unsigned int n_field_vars,
3144 				       unsigned int field_var_str_start)
3145 {
3146 	struct hist_elt_data *elt_data = elt->private_data;
3147 	unsigned int i, j, var_idx;
3148 	u64 var_val;
3149 
3150 	/* Make sure stacktrace can fit in the string variable length */
3151 	BUILD_BUG_ON((HIST_STACKTRACE_DEPTH + 1) * sizeof(long) >= STR_VAR_LEN_MAX);
3152 
3153 	for (i = 0, j = field_var_str_start; i < n_field_vars; i++) {
3154 		struct field_var *field_var = field_vars[i];
3155 		struct hist_field *var = field_var->var;
3156 		struct hist_field *val = field_var->val;
3157 
3158 		var_val = hist_fn_call(val, elt, buffer, rbe, rec);
3159 		var_idx = var->var.idx;
3160 
3161 		if (val->flags & (HIST_FIELD_FL_STRING |
3162 				  HIST_FIELD_FL_STACKTRACE)) {
3163 			char *str = elt_data->field_var_str[j++];
3164 			char *val_str = (char *)(uintptr_t)var_val;
3165 			unsigned int size;
3166 
3167 			if (val->flags & HIST_FIELD_FL_STRING) {
3168 				size = min(val->size, STR_VAR_LEN_MAX);
3169 				strscpy(str, val_str, size);
3170 			} else {
3171 				char *stack_start = str + sizeof(unsigned long);
3172 				int e;
3173 
3174 				e = stack_trace_save((void *)stack_start,
3175 						     HIST_STACKTRACE_DEPTH,
3176 						     HIST_STACKTRACE_SKIP);
3177 				if (e < HIST_STACKTRACE_DEPTH - 1)
3178 					((unsigned long *)stack_start)[e] = 0;
3179 				*((unsigned long *)str) = e;
3180 			}
3181 			var_val = (u64)(uintptr_t)str;
3182 		}
3183 		tracing_map_set_var(elt, var_idx, var_val);
3184 	}
3185 }
3186 
update_field_vars(struct hist_trigger_data * hist_data,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * rec)3187 static void update_field_vars(struct hist_trigger_data *hist_data,
3188 			      struct tracing_map_elt *elt,
3189 			      struct trace_buffer *buffer,
3190 			      struct ring_buffer_event *rbe,
3191 			      void *rec)
3192 {
3193 	__update_field_vars(elt, buffer, rbe, rec, hist_data->field_vars,
3194 			    hist_data->n_field_vars, 0);
3195 }
3196 
save_track_data_vars(struct hist_trigger_data * hist_data,struct tracing_map_elt * elt,struct trace_buffer * buffer,void * rec,struct ring_buffer_event * rbe,void * key,struct action_data * data,u64 * var_ref_vals)3197 static void save_track_data_vars(struct hist_trigger_data *hist_data,
3198 				 struct tracing_map_elt *elt,
3199 				 struct trace_buffer *buffer,  void *rec,
3200 				 struct ring_buffer_event *rbe, void *key,
3201 				 struct action_data *data, u64 *var_ref_vals)
3202 {
3203 	__update_field_vars(elt, buffer, rbe, rec, hist_data->save_vars,
3204 			    hist_data->n_save_vars, hist_data->n_field_var_str);
3205 }
3206 
create_var(struct hist_trigger_data * hist_data,struct trace_event_file * file,char * name,int size,const char * type)3207 static struct hist_field *create_var(struct hist_trigger_data *hist_data,
3208 				     struct trace_event_file *file,
3209 				     char *name, int size, const char *type)
3210 {
3211 	struct hist_field *var;
3212 	int idx;
3213 
3214 	if (find_var(hist_data, file, name) && !hist_data->remove) {
3215 		var = ERR_PTR(-EINVAL);
3216 		goto out;
3217 	}
3218 
3219 	var = kzalloc(sizeof(struct hist_field), GFP_KERNEL);
3220 	if (!var) {
3221 		var = ERR_PTR(-ENOMEM);
3222 		goto out;
3223 	}
3224 
3225 	idx = tracing_map_add_var(hist_data->map);
3226 	if (idx < 0) {
3227 		kfree(var);
3228 		var = ERR_PTR(-EINVAL);
3229 		goto out;
3230 	}
3231 
3232 	var->ref = 1;
3233 	var->flags = HIST_FIELD_FL_VAR;
3234 	var->var.idx = idx;
3235 	var->var.hist_data = var->hist_data = hist_data;
3236 	var->size = size;
3237 	var->var.name = kstrdup(name, GFP_KERNEL);
3238 	var->type = kstrdup_const(type, GFP_KERNEL);
3239 	if (!var->var.name || !var->type) {
3240 		kfree_const(var->type);
3241 		kfree(var->var.name);
3242 		kfree(var);
3243 		var = ERR_PTR(-ENOMEM);
3244 	}
3245  out:
3246 	return var;
3247 }
3248 
create_field_var(struct hist_trigger_data * hist_data,struct trace_event_file * file,char * field_name)3249 static struct field_var *create_field_var(struct hist_trigger_data *hist_data,
3250 					  struct trace_event_file *file,
3251 					  char *field_name)
3252 {
3253 	struct hist_field *val = NULL, *var = NULL;
3254 	unsigned long flags = HIST_FIELD_FL_VAR;
3255 	struct trace_array *tr = file->tr;
3256 	struct field_var *field_var;
3257 	int ret = 0;
3258 
3259 	if (hist_data->n_field_vars >= SYNTH_FIELDS_MAX) {
3260 		hist_err(tr, HIST_ERR_TOO_MANY_FIELD_VARS, errpos(field_name));
3261 		ret = -EINVAL;
3262 		goto err;
3263 	}
3264 
3265 	val = parse_atom(hist_data, file, field_name, &flags, NULL);
3266 	if (IS_ERR(val)) {
3267 		hist_err(tr, HIST_ERR_FIELD_VAR_PARSE_FAIL, errpos(field_name));
3268 		ret = PTR_ERR(val);
3269 		goto err;
3270 	}
3271 
3272 	var = create_var(hist_data, file, field_name, val->size, val->type);
3273 	if (IS_ERR(var)) {
3274 		hist_err(tr, HIST_ERR_VAR_CREATE_FIND_FAIL, errpos(field_name));
3275 		destroy_hist_field(val, 0);
3276 		ret = PTR_ERR(var);
3277 		goto err;
3278 	}
3279 
3280 	field_var = kzalloc(sizeof(struct field_var), GFP_KERNEL);
3281 	if (!field_var) {
3282 		destroy_hist_field(val, 0);
3283 		kfree_const(var->type);
3284 		kfree(var->var.name);
3285 		kfree(var);
3286 		ret =  -ENOMEM;
3287 		goto err;
3288 	}
3289 
3290 	field_var->var = var;
3291 	field_var->val = val;
3292  out:
3293 	return field_var;
3294  err:
3295 	field_var = ERR_PTR(ret);
3296 	goto out;
3297 }
3298 
3299 /**
3300  * create_target_field_var - Automatically create a variable for a field
3301  * @target_hist_data: The target hist trigger
3302  * @subsys_name: Optional subsystem name
3303  * @event_name: Optional event name
3304  * @var_name: The name of the field (and the resulting variable)
3305  *
3306  * Hist trigger actions fetch data from variables, not directly from
3307  * events.  However, for convenience, users are allowed to directly
3308  * specify an event field in an action, which will be automatically
3309  * converted into a variable on their behalf.
3310  *
3311  * This function creates a field variable with the name var_name on
3312  * the hist trigger currently being defined on the target event.  If
3313  * subsys_name and event_name are specified, this function simply
3314  * verifies that they do in fact match the target event subsystem and
3315  * event name.
3316  *
3317  * Return: The variable created for the field.
3318  */
3319 static struct field_var *
create_target_field_var(struct hist_trigger_data * target_hist_data,char * subsys_name,char * event_name,char * var_name)3320 create_target_field_var(struct hist_trigger_data *target_hist_data,
3321 			char *subsys_name, char *event_name, char *var_name)
3322 {
3323 	struct trace_event_file *file = target_hist_data->event_file;
3324 
3325 	if (subsys_name) {
3326 		struct trace_event_call *call;
3327 
3328 		if (!event_name)
3329 			return NULL;
3330 
3331 		call = file->event_call;
3332 
3333 		if (strcmp(subsys_name, call->class->system) != 0)
3334 			return NULL;
3335 
3336 		if (strcmp(event_name, trace_event_name(call)) != 0)
3337 			return NULL;
3338 	}
3339 
3340 	return create_field_var(target_hist_data, file, var_name);
3341 }
3342 
check_track_val_max(u64 track_val,u64 var_val)3343 static bool check_track_val_max(u64 track_val, u64 var_val)
3344 {
3345 	if (var_val <= track_val)
3346 		return false;
3347 
3348 	return true;
3349 }
3350 
check_track_val_changed(u64 track_val,u64 var_val)3351 static bool check_track_val_changed(u64 track_val, u64 var_val)
3352 {
3353 	if (var_val == track_val)
3354 		return false;
3355 
3356 	return true;
3357 }
3358 
get_track_val(struct hist_trigger_data * hist_data,struct tracing_map_elt * elt,struct action_data * data)3359 static u64 get_track_val(struct hist_trigger_data *hist_data,
3360 			 struct tracing_map_elt *elt,
3361 			 struct action_data *data)
3362 {
3363 	unsigned int track_var_idx = data->track_data.track_var->var.idx;
3364 	u64 track_val;
3365 
3366 	track_val = tracing_map_read_var(elt, track_var_idx);
3367 
3368 	return track_val;
3369 }
3370 
save_track_val(struct hist_trigger_data * hist_data,struct tracing_map_elt * elt,struct action_data * data,u64 var_val)3371 static void save_track_val(struct hist_trigger_data *hist_data,
3372 			   struct tracing_map_elt *elt,
3373 			   struct action_data *data, u64 var_val)
3374 {
3375 	unsigned int track_var_idx = data->track_data.track_var->var.idx;
3376 
3377 	tracing_map_set_var(elt, track_var_idx, var_val);
3378 }
3379 
save_track_data(struct hist_trigger_data * hist_data,struct tracing_map_elt * elt,struct trace_buffer * buffer,void * rec,struct ring_buffer_event * rbe,void * key,struct action_data * data,u64 * var_ref_vals)3380 static void save_track_data(struct hist_trigger_data *hist_data,
3381 			    struct tracing_map_elt *elt,
3382 			    struct trace_buffer *buffer, void *rec,
3383 			    struct ring_buffer_event *rbe, void *key,
3384 			    struct action_data *data, u64 *var_ref_vals)
3385 {
3386 	if (data->track_data.save_data)
3387 		data->track_data.save_data(hist_data, elt, buffer, rec, rbe,
3388 					   key, data, var_ref_vals);
3389 }
3390 
check_track_val(struct tracing_map_elt * elt,struct action_data * data,u64 var_val)3391 static bool check_track_val(struct tracing_map_elt *elt,
3392 			    struct action_data *data,
3393 			    u64 var_val)
3394 {
3395 	struct hist_trigger_data *hist_data;
3396 	u64 track_val;
3397 
3398 	hist_data = data->track_data.track_var->hist_data;
3399 	track_val = get_track_val(hist_data, elt, data);
3400 
3401 	return data->track_data.check_val(track_val, var_val);
3402 }
3403 
3404 #ifdef CONFIG_TRACER_SNAPSHOT
cond_snapshot_update(struct trace_array * tr,void * cond_data)3405 static bool cond_snapshot_update(struct trace_array *tr, void *cond_data)
3406 {
3407 	/* called with tr->max_lock held */
3408 	struct track_data *track_data = tr->cond_snapshot->cond_data;
3409 	struct hist_elt_data *elt_data, *track_elt_data;
3410 	struct snapshot_context *context = cond_data;
3411 	struct action_data *action;
3412 	u64 track_val;
3413 
3414 	if (!track_data)
3415 		return false;
3416 
3417 	action = track_data->action_data;
3418 
3419 	track_val = get_track_val(track_data->hist_data, context->elt,
3420 				  track_data->action_data);
3421 
3422 	if (!action->track_data.check_val(track_data->track_val, track_val))
3423 		return false;
3424 
3425 	track_data->track_val = track_val;
3426 	memcpy(track_data->key, context->key, track_data->key_len);
3427 
3428 	elt_data = context->elt->private_data;
3429 	track_elt_data = track_data->elt.private_data;
3430 	if (elt_data->comm)
3431 		strscpy(track_elt_data->comm, elt_data->comm, TASK_COMM_LEN);
3432 
3433 	track_data->updated = true;
3434 
3435 	return true;
3436 }
3437 
save_track_data_snapshot(struct hist_trigger_data * hist_data,struct tracing_map_elt * elt,struct trace_buffer * buffer,void * rec,struct ring_buffer_event * rbe,void * key,struct action_data * data,u64 * var_ref_vals)3438 static void save_track_data_snapshot(struct hist_trigger_data *hist_data,
3439 				     struct tracing_map_elt *elt,
3440 				     struct trace_buffer *buffer, void *rec,
3441 				     struct ring_buffer_event *rbe, void *key,
3442 				     struct action_data *data,
3443 				     u64 *var_ref_vals)
3444 {
3445 	struct trace_event_file *file = hist_data->event_file;
3446 	struct snapshot_context context;
3447 
3448 	context.elt = elt;
3449 	context.key = key;
3450 
3451 	tracing_snapshot_cond(file->tr, &context);
3452 }
3453 
3454 static void hist_trigger_print_key(struct seq_file *m,
3455 				   struct hist_trigger_data *hist_data,
3456 				   void *key,
3457 				   struct tracing_map_elt *elt);
3458 
snapshot_action(struct hist_trigger_data * hist_data)3459 static struct action_data *snapshot_action(struct hist_trigger_data *hist_data)
3460 {
3461 	unsigned int i;
3462 
3463 	if (!hist_data->n_actions)
3464 		return NULL;
3465 
3466 	for (i = 0; i < hist_data->n_actions; i++) {
3467 		struct action_data *data = hist_data->actions[i];
3468 
3469 		if (data->action == ACTION_SNAPSHOT)
3470 			return data;
3471 	}
3472 
3473 	return NULL;
3474 }
3475 
track_data_snapshot_print(struct seq_file * m,struct hist_trigger_data * hist_data)3476 static void track_data_snapshot_print(struct seq_file *m,
3477 				      struct hist_trigger_data *hist_data)
3478 {
3479 	struct trace_event_file *file = hist_data->event_file;
3480 	struct track_data *track_data;
3481 	struct action_data *action;
3482 
3483 	track_data = tracing_cond_snapshot_data(file->tr);
3484 	if (!track_data)
3485 		return;
3486 
3487 	if (!track_data->updated)
3488 		return;
3489 
3490 	action = snapshot_action(hist_data);
3491 	if (!action)
3492 		return;
3493 
3494 	seq_puts(m, "\nSnapshot taken (see tracing/snapshot).  Details:\n");
3495 	seq_printf(m, "\ttriggering value { %s(%s) }: %10llu",
3496 		   action->handler == HANDLER_ONMAX ? "onmax" : "onchange",
3497 		   action->track_data.var_str, track_data->track_val);
3498 
3499 	seq_puts(m, "\ttriggered by event with key: ");
3500 	hist_trigger_print_key(m, hist_data, track_data->key, &track_data->elt);
3501 	seq_putc(m, '\n');
3502 }
3503 #else
cond_snapshot_update(struct trace_array * tr,void * cond_data)3504 static bool cond_snapshot_update(struct trace_array *tr, void *cond_data)
3505 {
3506 	return false;
3507 }
save_track_data_snapshot(struct hist_trigger_data * hist_data,struct tracing_map_elt * elt,struct trace_buffer * buffer,void * rec,struct ring_buffer_event * rbe,void * key,struct action_data * data,u64 * var_ref_vals)3508 static void save_track_data_snapshot(struct hist_trigger_data *hist_data,
3509 				     struct tracing_map_elt *elt,
3510 				     struct trace_buffer *buffer, void *rec,
3511 				     struct ring_buffer_event *rbe, void *key,
3512 				     struct action_data *data,
3513 				     u64 *var_ref_vals) {}
track_data_snapshot_print(struct seq_file * m,struct hist_trigger_data * hist_data)3514 static void track_data_snapshot_print(struct seq_file *m,
3515 				      struct hist_trigger_data *hist_data) {}
3516 #endif /* CONFIG_TRACER_SNAPSHOT */
3517 
track_data_print(struct seq_file * m,struct hist_trigger_data * hist_data,struct tracing_map_elt * elt,struct action_data * data)3518 static void track_data_print(struct seq_file *m,
3519 			     struct hist_trigger_data *hist_data,
3520 			     struct tracing_map_elt *elt,
3521 			     struct action_data *data)
3522 {
3523 	u64 track_val = get_track_val(hist_data, elt, data);
3524 	unsigned int i, save_var_idx;
3525 
3526 	if (data->handler == HANDLER_ONMAX)
3527 		seq_printf(m, "\n\tmax: %10llu", track_val);
3528 	else if (data->handler == HANDLER_ONCHANGE)
3529 		seq_printf(m, "\n\tchanged: %10llu", track_val);
3530 
3531 	if (data->action == ACTION_SNAPSHOT)
3532 		return;
3533 
3534 	for (i = 0; i < hist_data->n_save_vars; i++) {
3535 		struct hist_field *save_val = hist_data->save_vars[i]->val;
3536 		struct hist_field *save_var = hist_data->save_vars[i]->var;
3537 		u64 val;
3538 
3539 		save_var_idx = save_var->var.idx;
3540 
3541 		val = tracing_map_read_var(elt, save_var_idx);
3542 
3543 		if (save_val->flags & HIST_FIELD_FL_STRING) {
3544 			seq_printf(m, "  %s: %-32s", save_var->var.name,
3545 				   (char *)(uintptr_t)(val));
3546 		} else
3547 			seq_printf(m, "  %s: %10llu", save_var->var.name, val);
3548 	}
3549 }
3550 
ontrack_action(struct hist_trigger_data * hist_data,struct tracing_map_elt * elt,struct trace_buffer * buffer,void * rec,struct ring_buffer_event * rbe,void * key,struct action_data * data,u64 * var_ref_vals)3551 static void ontrack_action(struct hist_trigger_data *hist_data,
3552 			   struct tracing_map_elt *elt,
3553 			   struct trace_buffer *buffer, void *rec,
3554 			   struct ring_buffer_event *rbe, void *key,
3555 			   struct action_data *data, u64 *var_ref_vals)
3556 {
3557 	u64 var_val = var_ref_vals[data->track_data.var_ref->var_ref_idx];
3558 
3559 	if (check_track_val(elt, data, var_val)) {
3560 		save_track_val(hist_data, elt, data, var_val);
3561 		save_track_data(hist_data, elt, buffer, rec, rbe,
3562 				key, data, var_ref_vals);
3563 	}
3564 }
3565 
action_data_destroy(struct action_data * data)3566 static void action_data_destroy(struct action_data *data)
3567 {
3568 	unsigned int i;
3569 
3570 	lockdep_assert_held(&event_mutex);
3571 
3572 	kfree(data->action_name);
3573 
3574 	for (i = 0; i < data->n_params; i++)
3575 		kfree(data->params[i]);
3576 
3577 	if (data->synth_event)
3578 		data->synth_event->ref--;
3579 
3580 	kfree(data->synth_event_name);
3581 
3582 	kfree(data);
3583 }
3584 
track_data_destroy(struct hist_trigger_data * hist_data,struct action_data * data)3585 static void track_data_destroy(struct hist_trigger_data *hist_data,
3586 			       struct action_data *data)
3587 {
3588 	struct trace_event_file *file = hist_data->event_file;
3589 
3590 	destroy_hist_field(data->track_data.track_var, 0);
3591 
3592 	if (data->action == ACTION_SNAPSHOT) {
3593 		struct track_data *track_data;
3594 
3595 		track_data = tracing_cond_snapshot_data(file->tr);
3596 		if (track_data && track_data->hist_data == hist_data) {
3597 			tracing_snapshot_cond_disable(file->tr);
3598 			track_data_free(track_data);
3599 		}
3600 	}
3601 
3602 	kfree(data->track_data.var_str);
3603 
3604 	action_data_destroy(data);
3605 }
3606 
3607 static int action_create(struct hist_trigger_data *hist_data,
3608 			 struct action_data *data);
3609 
track_data_create(struct hist_trigger_data * hist_data,struct action_data * data)3610 static int track_data_create(struct hist_trigger_data *hist_data,
3611 			     struct action_data *data)
3612 {
3613 	struct hist_field *var_field, *ref_field, *track_var = NULL;
3614 	struct trace_event_file *file = hist_data->event_file;
3615 	struct trace_array *tr = file->tr;
3616 	char *track_data_var_str;
3617 	int ret = 0;
3618 
3619 	track_data_var_str = data->track_data.var_str;
3620 	if (track_data_var_str[0] != '$') {
3621 		hist_err(tr, HIST_ERR_ONX_NOT_VAR, errpos(track_data_var_str));
3622 		return -EINVAL;
3623 	}
3624 	track_data_var_str++;
3625 
3626 	var_field = find_target_event_var(hist_data, NULL, NULL, track_data_var_str);
3627 	if (!var_field) {
3628 		hist_err(tr, HIST_ERR_ONX_VAR_NOT_FOUND, errpos(track_data_var_str));
3629 		return -EINVAL;
3630 	}
3631 
3632 	ref_field = create_var_ref(hist_data, var_field, NULL, NULL);
3633 	if (!ref_field)
3634 		return -ENOMEM;
3635 
3636 	data->track_data.var_ref = ref_field;
3637 
3638 	if (data->handler == HANDLER_ONMAX)
3639 		track_var = create_var(hist_data, file, "__max", sizeof(u64), "u64");
3640 	if (IS_ERR(track_var)) {
3641 		hist_err(tr, HIST_ERR_ONX_VAR_CREATE_FAIL, 0);
3642 		ret = PTR_ERR(track_var);
3643 		goto out;
3644 	}
3645 
3646 	if (data->handler == HANDLER_ONCHANGE)
3647 		track_var = create_var(hist_data, file, "__change", sizeof(u64), "u64");
3648 	if (IS_ERR(track_var)) {
3649 		hist_err(tr, HIST_ERR_ONX_VAR_CREATE_FAIL, 0);
3650 		ret = PTR_ERR(track_var);
3651 		goto out;
3652 	}
3653 	data->track_data.track_var = track_var;
3654 
3655 	ret = action_create(hist_data, data);
3656  out:
3657 	return ret;
3658 }
3659 
parse_action_params(struct trace_array * tr,char * params,struct action_data * data)3660 static int parse_action_params(struct trace_array *tr, char *params,
3661 			       struct action_data *data)
3662 {
3663 	char *param, *saved_param;
3664 	bool first_param = true;
3665 	int ret = 0;
3666 
3667 	while (params) {
3668 		if (data->n_params >= SYNTH_FIELDS_MAX) {
3669 			hist_err(tr, HIST_ERR_TOO_MANY_PARAMS, 0);
3670 			ret = -EINVAL;
3671 			goto out;
3672 		}
3673 
3674 		param = strsep(&params, ",");
3675 		if (!param) {
3676 			hist_err(tr, HIST_ERR_PARAM_NOT_FOUND, 0);
3677 			ret = -EINVAL;
3678 			goto out;
3679 		}
3680 
3681 		param = strstrip(param);
3682 		if (strlen(param) < 2) {
3683 			hist_err(tr, HIST_ERR_INVALID_PARAM, errpos(param));
3684 			ret = -EINVAL;
3685 			goto out;
3686 		}
3687 
3688 		saved_param = kstrdup(param, GFP_KERNEL);
3689 		if (!saved_param) {
3690 			ret = -ENOMEM;
3691 			goto out;
3692 		}
3693 
3694 		if (first_param && data->use_trace_keyword) {
3695 			data->synth_event_name = saved_param;
3696 			first_param = false;
3697 			continue;
3698 		}
3699 		first_param = false;
3700 
3701 		data->params[data->n_params++] = saved_param;
3702 	}
3703  out:
3704 	return ret;
3705 }
3706 
action_parse(struct trace_array * tr,char * str,struct action_data * data,enum handler_id handler)3707 static int action_parse(struct trace_array *tr, char *str, struct action_data *data,
3708 			enum handler_id handler)
3709 {
3710 	char *action_name;
3711 	int ret = 0;
3712 
3713 	strsep(&str, ".");
3714 	if (!str) {
3715 		hist_err(tr, HIST_ERR_ACTION_NOT_FOUND, 0);
3716 		ret = -EINVAL;
3717 		goto out;
3718 	}
3719 
3720 	action_name = strsep(&str, "(");
3721 	if (!action_name || !str) {
3722 		hist_err(tr, HIST_ERR_ACTION_NOT_FOUND, 0);
3723 		ret = -EINVAL;
3724 		goto out;
3725 	}
3726 
3727 	if (str_has_prefix(action_name, "save")) {
3728 		char *params = strsep(&str, ")");
3729 
3730 		if (!params) {
3731 			hist_err(tr, HIST_ERR_NO_SAVE_PARAMS, 0);
3732 			ret = -EINVAL;
3733 			goto out;
3734 		}
3735 
3736 		ret = parse_action_params(tr, params, data);
3737 		if (ret)
3738 			goto out;
3739 
3740 		if (handler == HANDLER_ONMAX)
3741 			data->track_data.check_val = check_track_val_max;
3742 		else if (handler == HANDLER_ONCHANGE)
3743 			data->track_data.check_val = check_track_val_changed;
3744 		else {
3745 			hist_err(tr, HIST_ERR_ACTION_MISMATCH, errpos(action_name));
3746 			ret = -EINVAL;
3747 			goto out;
3748 		}
3749 
3750 		data->track_data.save_data = save_track_data_vars;
3751 		data->fn = ontrack_action;
3752 		data->action = ACTION_SAVE;
3753 	} else if (str_has_prefix(action_name, "snapshot")) {
3754 		char *params = strsep(&str, ")");
3755 
3756 		if (!str) {
3757 			hist_err(tr, HIST_ERR_NO_CLOSING_PAREN, errpos(params));
3758 			ret = -EINVAL;
3759 			goto out;
3760 		}
3761 
3762 		if (handler == HANDLER_ONMAX)
3763 			data->track_data.check_val = check_track_val_max;
3764 		else if (handler == HANDLER_ONCHANGE)
3765 			data->track_data.check_val = check_track_val_changed;
3766 		else {
3767 			hist_err(tr, HIST_ERR_ACTION_MISMATCH, errpos(action_name));
3768 			ret = -EINVAL;
3769 			goto out;
3770 		}
3771 
3772 		data->track_data.save_data = save_track_data_snapshot;
3773 		data->fn = ontrack_action;
3774 		data->action = ACTION_SNAPSHOT;
3775 	} else {
3776 		char *params = strsep(&str, ")");
3777 
3778 		if (str_has_prefix(action_name, "trace"))
3779 			data->use_trace_keyword = true;
3780 
3781 		if (params) {
3782 			ret = parse_action_params(tr, params, data);
3783 			if (ret)
3784 				goto out;
3785 		}
3786 
3787 		if (handler == HANDLER_ONMAX)
3788 			data->track_data.check_val = check_track_val_max;
3789 		else if (handler == HANDLER_ONCHANGE)
3790 			data->track_data.check_val = check_track_val_changed;
3791 
3792 		if (handler != HANDLER_ONMATCH) {
3793 			data->track_data.save_data = action_trace;
3794 			data->fn = ontrack_action;
3795 		} else
3796 			data->fn = action_trace;
3797 
3798 		data->action = ACTION_TRACE;
3799 	}
3800 
3801 	data->action_name = kstrdup(action_name, GFP_KERNEL);
3802 	if (!data->action_name) {
3803 		ret = -ENOMEM;
3804 		goto out;
3805 	}
3806 
3807 	data->handler = handler;
3808  out:
3809 	return ret;
3810 }
3811 
track_data_parse(struct hist_trigger_data * hist_data,char * str,enum handler_id handler)3812 static struct action_data *track_data_parse(struct hist_trigger_data *hist_data,
3813 					    char *str, enum handler_id handler)
3814 {
3815 	struct action_data *data;
3816 	int ret = -EINVAL;
3817 	char *var_str;
3818 
3819 	data = kzalloc(sizeof(*data), GFP_KERNEL);
3820 	if (!data)
3821 		return ERR_PTR(-ENOMEM);
3822 
3823 	var_str = strsep(&str, ")");
3824 	if (!var_str || !str) {
3825 		ret = -EINVAL;
3826 		goto free;
3827 	}
3828 
3829 	data->track_data.var_str = kstrdup(var_str, GFP_KERNEL);
3830 	if (!data->track_data.var_str) {
3831 		ret = -ENOMEM;
3832 		goto free;
3833 	}
3834 
3835 	ret = action_parse(hist_data->event_file->tr, str, data, handler);
3836 	if (ret)
3837 		goto free;
3838  out:
3839 	return data;
3840  free:
3841 	track_data_destroy(hist_data, data);
3842 	data = ERR_PTR(ret);
3843 	goto out;
3844 }
3845 
onmatch_destroy(struct action_data * data)3846 static void onmatch_destroy(struct action_data *data)
3847 {
3848 	kfree(data->match_data.event);
3849 	kfree(data->match_data.event_system);
3850 
3851 	action_data_destroy(data);
3852 }
3853 
destroy_field_var(struct field_var * field_var)3854 static void destroy_field_var(struct field_var *field_var)
3855 {
3856 	if (!field_var)
3857 		return;
3858 
3859 	destroy_hist_field(field_var->var, 0);
3860 	destroy_hist_field(field_var->val, 0);
3861 
3862 	kfree(field_var);
3863 }
3864 
destroy_field_vars(struct hist_trigger_data * hist_data)3865 static void destroy_field_vars(struct hist_trigger_data *hist_data)
3866 {
3867 	unsigned int i;
3868 
3869 	for (i = 0; i < hist_data->n_field_vars; i++)
3870 		destroy_field_var(hist_data->field_vars[i]);
3871 
3872 	for (i = 0; i < hist_data->n_save_vars; i++)
3873 		destroy_field_var(hist_data->save_vars[i]);
3874 }
3875 
save_field_var(struct hist_trigger_data * hist_data,struct field_var * field_var)3876 static void save_field_var(struct hist_trigger_data *hist_data,
3877 			   struct field_var *field_var)
3878 {
3879 	hist_data->field_vars[hist_data->n_field_vars++] = field_var;
3880 
3881 	/* Stack traces are saved in the string storage too */
3882 	if (field_var->val->flags & (HIST_FIELD_FL_STRING | HIST_FIELD_FL_STACKTRACE))
3883 		hist_data->n_field_var_str++;
3884 }
3885 
3886 
check_synth_field(struct synth_event * event,struct hist_field * hist_field,unsigned int field_pos)3887 static int check_synth_field(struct synth_event *event,
3888 			     struct hist_field *hist_field,
3889 			     unsigned int field_pos)
3890 {
3891 	struct synth_field *field;
3892 
3893 	if (field_pos >= event->n_fields)
3894 		return -EINVAL;
3895 
3896 	field = event->fields[field_pos];
3897 
3898 	/*
3899 	 * A dynamic string synth field can accept static or
3900 	 * dynamic. A static string synth field can only accept a
3901 	 * same-sized static string, which is checked for later.
3902 	 */
3903 	if (strstr(hist_field->type, "char[") && field->is_string
3904 	    && field->is_dynamic)
3905 		return 0;
3906 
3907 	if (strstr(hist_field->type, "long[") && field->is_stack)
3908 		return 0;
3909 
3910 	if (strcmp(field->type, hist_field->type) != 0) {
3911 		if (field->size != hist_field->size ||
3912 		    (!field->is_string && field->is_signed != hist_field->is_signed))
3913 			return -EINVAL;
3914 	}
3915 
3916 	return 0;
3917 }
3918 
3919 static struct hist_field *
trace_action_find_var(struct hist_trigger_data * hist_data,struct action_data * data,char * system,char * event,char * var)3920 trace_action_find_var(struct hist_trigger_data *hist_data,
3921 		      struct action_data *data,
3922 		      char *system, char *event, char *var)
3923 {
3924 	struct trace_array *tr = hist_data->event_file->tr;
3925 	struct hist_field *hist_field;
3926 
3927 	var++; /* skip '$' */
3928 
3929 	hist_field = find_target_event_var(hist_data, system, event, var);
3930 	if (!hist_field) {
3931 		if (!system && data->handler == HANDLER_ONMATCH) {
3932 			system = data->match_data.event_system;
3933 			event = data->match_data.event;
3934 		}
3935 
3936 		hist_field = find_event_var(hist_data, system, event, var);
3937 	}
3938 
3939 	if (!hist_field)
3940 		hist_err(tr, HIST_ERR_PARAM_NOT_FOUND, errpos(var));
3941 
3942 	return hist_field;
3943 }
3944 
3945 static struct hist_field *
trace_action_create_field_var(struct hist_trigger_data * hist_data,struct action_data * data,char * system,char * event,char * var)3946 trace_action_create_field_var(struct hist_trigger_data *hist_data,
3947 			      struct action_data *data, char *system,
3948 			      char *event, char *var)
3949 {
3950 	struct hist_field *hist_field = NULL;
3951 	struct field_var *field_var;
3952 
3953 	/*
3954 	 * First try to create a field var on the target event (the
3955 	 * currently being defined).  This will create a variable for
3956 	 * unqualified fields on the target event, or if qualified,
3957 	 * target fields that have qualified names matching the target.
3958 	 */
3959 	field_var = create_target_field_var(hist_data, system, event, var);
3960 
3961 	if (field_var && !IS_ERR(field_var)) {
3962 		save_field_var(hist_data, field_var);
3963 		hist_field = field_var->var;
3964 	} else {
3965 		field_var = NULL;
3966 		/*
3967 		 * If no explicit system.event is specified, default to
3968 		 * looking for fields on the onmatch(system.event.xxx)
3969 		 * event.
3970 		 */
3971 		if (!system && data->handler == HANDLER_ONMATCH) {
3972 			system = data->match_data.event_system;
3973 			event = data->match_data.event;
3974 		}
3975 
3976 		if (!event)
3977 			goto free;
3978 		/*
3979 		 * At this point, we're looking at a field on another
3980 		 * event.  Because we can't modify a hist trigger on
3981 		 * another event to add a variable for a field, we need
3982 		 * to create a new trigger on that event and create the
3983 		 * variable at the same time.
3984 		 */
3985 		hist_field = create_field_var_hist(hist_data, system, event, var);
3986 		if (IS_ERR(hist_field))
3987 			goto free;
3988 	}
3989  out:
3990 	return hist_field;
3991  free:
3992 	destroy_field_var(field_var);
3993 	hist_field = NULL;
3994 	goto out;
3995 }
3996 
trace_action_create(struct hist_trigger_data * hist_data,struct action_data * data)3997 static int trace_action_create(struct hist_trigger_data *hist_data,
3998 			       struct action_data *data)
3999 {
4000 	struct trace_array *tr = hist_data->event_file->tr;
4001 	char *event_name, *param, *system = NULL;
4002 	struct hist_field *hist_field, *var_ref;
4003 	unsigned int i;
4004 	unsigned int field_pos = 0;
4005 	struct synth_event *event;
4006 	char *synth_event_name;
4007 	int var_ref_idx, ret = 0;
4008 
4009 	lockdep_assert_held(&event_mutex);
4010 
4011 	/* Sanity check to avoid out-of-bound write on 'data->var_ref_idx' */
4012 	if (data->n_params > SYNTH_FIELDS_MAX)
4013 		return -EINVAL;
4014 
4015 	if (data->use_trace_keyword)
4016 		synth_event_name = data->synth_event_name;
4017 	else
4018 		synth_event_name = data->action_name;
4019 
4020 	event = find_synth_event(synth_event_name);
4021 	if (!event) {
4022 		hist_err(tr, HIST_ERR_SYNTH_EVENT_NOT_FOUND, errpos(synth_event_name));
4023 		return -EINVAL;
4024 	}
4025 
4026 	event->ref++;
4027 
4028 	for (i = 0; i < data->n_params; i++) {
4029 		char *p;
4030 
4031 		p = param = kstrdup(data->params[i], GFP_KERNEL);
4032 		if (!param) {
4033 			ret = -ENOMEM;
4034 			goto err;
4035 		}
4036 
4037 		system = strsep(&param, ".");
4038 		if (!param) {
4039 			param = (char *)system;
4040 			system = event_name = NULL;
4041 		} else {
4042 			event_name = strsep(&param, ".");
4043 			if (!param) {
4044 				kfree(p);
4045 				ret = -EINVAL;
4046 				goto err;
4047 			}
4048 		}
4049 
4050 		if (param[0] == '$')
4051 			hist_field = trace_action_find_var(hist_data, data,
4052 							   system, event_name,
4053 							   param);
4054 		else
4055 			hist_field = trace_action_create_field_var(hist_data,
4056 								   data,
4057 								   system,
4058 								   event_name,
4059 								   param);
4060 
4061 		if (!hist_field) {
4062 			kfree(p);
4063 			ret = -EINVAL;
4064 			goto err;
4065 		}
4066 
4067 		if (check_synth_field(event, hist_field, field_pos) == 0) {
4068 			var_ref = create_var_ref(hist_data, hist_field,
4069 						 system, event_name);
4070 			if (!var_ref) {
4071 				kfree(p);
4072 				ret = -ENOMEM;
4073 				goto err;
4074 			}
4075 
4076 			var_ref_idx = find_var_ref_idx(hist_data, var_ref);
4077 			if (WARN_ON(var_ref_idx < 0)) {
4078 				kfree(p);
4079 				ret = var_ref_idx;
4080 				goto err;
4081 			}
4082 
4083 			data->var_ref_idx[i] = var_ref_idx;
4084 
4085 			field_pos++;
4086 			kfree(p);
4087 			continue;
4088 		}
4089 
4090 		hist_err(tr, HIST_ERR_SYNTH_TYPE_MISMATCH, errpos(param));
4091 		kfree(p);
4092 		ret = -EINVAL;
4093 		goto err;
4094 	}
4095 
4096 	if (field_pos != event->n_fields) {
4097 		hist_err(tr, HIST_ERR_SYNTH_COUNT_MISMATCH, errpos(event->name));
4098 		ret = -EINVAL;
4099 		goto err;
4100 	}
4101 
4102 	data->synth_event = event;
4103  out:
4104 	return ret;
4105  err:
4106 	event->ref--;
4107 
4108 	goto out;
4109 }
4110 
action_create(struct hist_trigger_data * hist_data,struct action_data * data)4111 static int action_create(struct hist_trigger_data *hist_data,
4112 			 struct action_data *data)
4113 {
4114 	struct trace_event_file *file = hist_data->event_file;
4115 	struct trace_array *tr = file->tr;
4116 	struct track_data *track_data;
4117 	struct field_var *field_var;
4118 	unsigned int i;
4119 	char *param;
4120 	int ret = 0;
4121 
4122 	if (data->action == ACTION_TRACE)
4123 		return trace_action_create(hist_data, data);
4124 
4125 	if (data->action == ACTION_SNAPSHOT) {
4126 		track_data = track_data_alloc(hist_data->key_size, data, hist_data);
4127 		if (IS_ERR(track_data)) {
4128 			ret = PTR_ERR(track_data);
4129 			goto out;
4130 		}
4131 
4132 		ret = tracing_snapshot_cond_enable(file->tr, track_data,
4133 						   cond_snapshot_update);
4134 		if (ret)
4135 			track_data_free(track_data);
4136 
4137 		goto out;
4138 	}
4139 
4140 	if (data->action == ACTION_SAVE) {
4141 		if (hist_data->n_save_vars) {
4142 			ret = -EEXIST;
4143 			hist_err(tr, HIST_ERR_TOO_MANY_SAVE_ACTIONS, 0);
4144 			goto out;
4145 		}
4146 
4147 		for (i = 0; i < data->n_params; i++) {
4148 			param = kstrdup(data->params[i], GFP_KERNEL);
4149 			if (!param) {
4150 				ret = -ENOMEM;
4151 				goto out;
4152 			}
4153 
4154 			field_var = create_target_field_var(hist_data, NULL, NULL, param);
4155 			if (IS_ERR(field_var)) {
4156 				hist_err(tr, HIST_ERR_FIELD_VAR_CREATE_FAIL,
4157 					 errpos(param));
4158 				ret = PTR_ERR(field_var);
4159 				kfree(param);
4160 				goto out;
4161 			}
4162 
4163 			hist_data->save_vars[hist_data->n_save_vars++] = field_var;
4164 			if (field_var->val->flags &
4165 			    (HIST_FIELD_FL_STRING | HIST_FIELD_FL_STACKTRACE))
4166 				hist_data->n_save_var_str++;
4167 			kfree(param);
4168 		}
4169 	}
4170  out:
4171 	return ret;
4172 }
4173 
onmatch_create(struct hist_trigger_data * hist_data,struct action_data * data)4174 static int onmatch_create(struct hist_trigger_data *hist_data,
4175 			  struct action_data *data)
4176 {
4177 	return action_create(hist_data, data);
4178 }
4179 
onmatch_parse(struct trace_array * tr,char * str)4180 static struct action_data *onmatch_parse(struct trace_array *tr, char *str)
4181 {
4182 	char *match_event, *match_event_system;
4183 	struct action_data *data;
4184 	int ret = -EINVAL;
4185 
4186 	data = kzalloc(sizeof(*data), GFP_KERNEL);
4187 	if (!data)
4188 		return ERR_PTR(-ENOMEM);
4189 
4190 	match_event = strsep(&str, ")");
4191 	if (!match_event || !str) {
4192 		hist_err(tr, HIST_ERR_NO_CLOSING_PAREN, errpos(match_event));
4193 		goto free;
4194 	}
4195 
4196 	match_event_system = strsep(&match_event, ".");
4197 	if (!match_event) {
4198 		hist_err(tr, HIST_ERR_SUBSYS_NOT_FOUND, errpos(match_event_system));
4199 		goto free;
4200 	}
4201 
4202 	if (IS_ERR(event_file(tr, match_event_system, match_event))) {
4203 		hist_err(tr, HIST_ERR_INVALID_SUBSYS_EVENT, errpos(match_event));
4204 		goto free;
4205 	}
4206 
4207 	data->match_data.event = kstrdup(match_event, GFP_KERNEL);
4208 	if (!data->match_data.event) {
4209 		ret = -ENOMEM;
4210 		goto free;
4211 	}
4212 
4213 	data->match_data.event_system = kstrdup(match_event_system, GFP_KERNEL);
4214 	if (!data->match_data.event_system) {
4215 		ret = -ENOMEM;
4216 		goto free;
4217 	}
4218 
4219 	ret = action_parse(tr, str, data, HANDLER_ONMATCH);
4220 	if (ret)
4221 		goto free;
4222  out:
4223 	return data;
4224  free:
4225 	onmatch_destroy(data);
4226 	data = ERR_PTR(ret);
4227 	goto out;
4228 }
4229 
create_hitcount_val(struct hist_trigger_data * hist_data)4230 static int create_hitcount_val(struct hist_trigger_data *hist_data)
4231 {
4232 	hist_data->fields[HITCOUNT_IDX] =
4233 		create_hist_field(hist_data, NULL, HIST_FIELD_FL_HITCOUNT, NULL);
4234 	if (!hist_data->fields[HITCOUNT_IDX])
4235 		return -ENOMEM;
4236 
4237 	hist_data->n_vals++;
4238 	hist_data->n_fields++;
4239 
4240 	if (WARN_ON(hist_data->n_vals > TRACING_MAP_VALS_MAX))
4241 		return -EINVAL;
4242 
4243 	return 0;
4244 }
4245 
__create_val_field(struct hist_trigger_data * hist_data,unsigned int val_idx,struct trace_event_file * file,char * var_name,char * field_str,unsigned long flags)4246 static int __create_val_field(struct hist_trigger_data *hist_data,
4247 			      unsigned int val_idx,
4248 			      struct trace_event_file *file,
4249 			      char *var_name, char *field_str,
4250 			      unsigned long flags)
4251 {
4252 	struct hist_field *hist_field;
4253 	int ret = 0, n_subexprs = 0;
4254 
4255 	hist_field = parse_expr(hist_data, file, field_str, flags, var_name, &n_subexprs);
4256 	if (IS_ERR(hist_field)) {
4257 		ret = PTR_ERR(hist_field);
4258 		goto out;
4259 	}
4260 
4261 	/* values and variables should not have some modifiers */
4262 	if (hist_field->flags & HIST_FIELD_FL_VAR) {
4263 		/* Variable */
4264 		if (hist_field->flags & (HIST_FIELD_FL_GRAPH | HIST_FIELD_FL_PERCENT |
4265 					 HIST_FIELD_FL_BUCKET | HIST_FIELD_FL_LOG2))
4266 			goto err;
4267 	} else {
4268 		/* Value */
4269 		if (hist_field->flags & (HIST_FIELD_FL_GRAPH | HIST_FIELD_FL_PERCENT |
4270 					 HIST_FIELD_FL_BUCKET | HIST_FIELD_FL_LOG2 |
4271 					 HIST_FIELD_FL_SYM | HIST_FIELD_FL_SYM_OFFSET |
4272 					 HIST_FIELD_FL_SYSCALL | HIST_FIELD_FL_STACKTRACE))
4273 			goto err;
4274 	}
4275 
4276 	hist_data->fields[val_idx] = hist_field;
4277 
4278 	++hist_data->n_vals;
4279 	++hist_data->n_fields;
4280 
4281 	if (WARN_ON(hist_data->n_vals > TRACING_MAP_VALS_MAX + TRACING_MAP_VARS_MAX))
4282 		ret = -EINVAL;
4283  out:
4284 	return ret;
4285  err:
4286 	hist_err(file->tr, HIST_ERR_BAD_FIELD_MODIFIER, errpos(field_str));
4287 	return -EINVAL;
4288 }
4289 
create_val_field(struct hist_trigger_data * hist_data,unsigned int val_idx,struct trace_event_file * file,char * field_str)4290 static int create_val_field(struct hist_trigger_data *hist_data,
4291 			    unsigned int val_idx,
4292 			    struct trace_event_file *file,
4293 			    char *field_str)
4294 {
4295 	if (WARN_ON(val_idx >= TRACING_MAP_VALS_MAX))
4296 		return -EINVAL;
4297 
4298 	return __create_val_field(hist_data, val_idx, file, NULL, field_str, 0);
4299 }
4300 
4301 static const char no_comm[] = "(no comm)";
4302 
hist_field_execname(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)4303 static u64 hist_field_execname(struct hist_field *hist_field,
4304 			       struct tracing_map_elt *elt,
4305 			       struct trace_buffer *buffer,
4306 			       struct ring_buffer_event *rbe,
4307 			       void *event)
4308 {
4309 	struct hist_elt_data *elt_data;
4310 
4311 	if (WARN_ON_ONCE(!elt))
4312 		return (u64)(unsigned long)no_comm;
4313 
4314 	elt_data = elt->private_data;
4315 
4316 	if (WARN_ON_ONCE(!elt_data->comm))
4317 		return (u64)(unsigned long)no_comm;
4318 
4319 	return (u64)(unsigned long)(elt_data->comm);
4320 }
4321 
hist_field_stack(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)4322 static u64 hist_field_stack(struct hist_field *hist_field,
4323 			    struct tracing_map_elt *elt,
4324 			    struct trace_buffer *buffer,
4325 			    struct ring_buffer_event *rbe,
4326 			    void *event)
4327 {
4328 	u32 str_item = *(u32 *)(event + hist_field->field->offset);
4329 	int str_loc = str_item & 0xffff;
4330 	char *addr = (char *)(event + str_loc);
4331 
4332 	return (u64)(unsigned long)addr;
4333 }
4334 
hist_fn_call(struct hist_field * hist_field,struct tracing_map_elt * elt,struct trace_buffer * buffer,struct ring_buffer_event * rbe,void * event)4335 static u64 hist_fn_call(struct hist_field *hist_field,
4336 			struct tracing_map_elt *elt,
4337 			struct trace_buffer *buffer,
4338 			struct ring_buffer_event *rbe,
4339 			void *event)
4340 {
4341 	switch (hist_field->fn_num) {
4342 	case HIST_FIELD_FN_VAR_REF:
4343 		return hist_field_var_ref(hist_field, elt, buffer, rbe, event);
4344 	case HIST_FIELD_FN_COUNTER:
4345 		return hist_field_counter(hist_field, elt, buffer, rbe, event);
4346 	case HIST_FIELD_FN_CONST:
4347 		return hist_field_const(hist_field, elt, buffer, rbe, event);
4348 	case HIST_FIELD_FN_LOG2:
4349 		return hist_field_log2(hist_field, elt, buffer, rbe, event);
4350 	case HIST_FIELD_FN_BUCKET:
4351 		return hist_field_bucket(hist_field, elt, buffer, rbe, event);
4352 	case HIST_FIELD_FN_TIMESTAMP:
4353 		return hist_field_timestamp(hist_field, elt, buffer, rbe, event);
4354 	case HIST_FIELD_FN_CPU:
4355 		return hist_field_cpu(hist_field, elt, buffer, rbe, event);
4356 	case HIST_FIELD_FN_COMM:
4357 		return hist_field_comm(hist_field, elt, buffer, rbe, event);
4358 	case HIST_FIELD_FN_STRING:
4359 		return hist_field_string(hist_field, elt, buffer, rbe, event);
4360 	case HIST_FIELD_FN_DYNSTRING:
4361 		return hist_field_dynstring(hist_field, elt, buffer, rbe, event);
4362 	case HIST_FIELD_FN_RELDYNSTRING:
4363 		return hist_field_reldynstring(hist_field, elt, buffer, rbe, event);
4364 	case HIST_FIELD_FN_PSTRING:
4365 		return hist_field_pstring(hist_field, elt, buffer, rbe, event);
4366 	case HIST_FIELD_FN_S64:
4367 		return hist_field_s64(hist_field, elt, buffer, rbe, event);
4368 	case HIST_FIELD_FN_U64:
4369 		return hist_field_u64(hist_field, elt, buffer, rbe, event);
4370 	case HIST_FIELD_FN_S32:
4371 		return hist_field_s32(hist_field, elt, buffer, rbe, event);
4372 	case HIST_FIELD_FN_U32:
4373 		return hist_field_u32(hist_field, elt, buffer, rbe, event);
4374 	case HIST_FIELD_FN_S16:
4375 		return hist_field_s16(hist_field, elt, buffer, rbe, event);
4376 	case HIST_FIELD_FN_U16:
4377 		return hist_field_u16(hist_field, elt, buffer, rbe, event);
4378 	case HIST_FIELD_FN_S8:
4379 		return hist_field_s8(hist_field, elt, buffer, rbe, event);
4380 	case HIST_FIELD_FN_U8:
4381 		return hist_field_u8(hist_field, elt, buffer, rbe, event);
4382 	case HIST_FIELD_FN_UMINUS:
4383 		return hist_field_unary_minus(hist_field, elt, buffer, rbe, event);
4384 	case HIST_FIELD_FN_MINUS:
4385 		return hist_field_minus(hist_field, elt, buffer, rbe, event);
4386 	case HIST_FIELD_FN_PLUS:
4387 		return hist_field_plus(hist_field, elt, buffer, rbe, event);
4388 	case HIST_FIELD_FN_DIV:
4389 		return hist_field_div(hist_field, elt, buffer, rbe, event);
4390 	case HIST_FIELD_FN_MULT:
4391 		return hist_field_mult(hist_field, elt, buffer, rbe, event);
4392 	case HIST_FIELD_FN_DIV_POWER2:
4393 		return div_by_power_of_two(hist_field, elt, buffer, rbe, event);
4394 	case HIST_FIELD_FN_DIV_NOT_POWER2:
4395 		return div_by_not_power_of_two(hist_field, elt, buffer, rbe, event);
4396 	case HIST_FIELD_FN_DIV_MULT_SHIFT:
4397 		return div_by_mult_and_shift(hist_field, elt, buffer, rbe, event);
4398 	case HIST_FIELD_FN_EXECNAME:
4399 		return hist_field_execname(hist_field, elt, buffer, rbe, event);
4400 	case HIST_FIELD_FN_STACK:
4401 		return hist_field_stack(hist_field, elt, buffer, rbe, event);
4402 	default:
4403 		return 0;
4404 	}
4405 }
4406 
4407 /* Convert a var that points to common_pid.execname to a string */
update_var_execname(struct hist_field * hist_field)4408 static void update_var_execname(struct hist_field *hist_field)
4409 {
4410 	hist_field->flags = HIST_FIELD_FL_STRING | HIST_FIELD_FL_VAR |
4411 		HIST_FIELD_FL_EXECNAME;
4412 	hist_field->size = MAX_FILTER_STR_VAL;
4413 	hist_field->is_signed = 0;
4414 
4415 	kfree_const(hist_field->type);
4416 	hist_field->type = "char[]";
4417 
4418 	hist_field->fn_num = HIST_FIELD_FN_EXECNAME;
4419 }
4420 
create_var_field(struct hist_trigger_data * hist_data,unsigned int val_idx,struct trace_event_file * file,char * var_name,char * expr_str)4421 static int create_var_field(struct hist_trigger_data *hist_data,
4422 			    unsigned int val_idx,
4423 			    struct trace_event_file *file,
4424 			    char *var_name, char *expr_str)
4425 {
4426 	struct trace_array *tr = hist_data->event_file->tr;
4427 	unsigned long flags = 0;
4428 	int ret;
4429 
4430 	if (WARN_ON(val_idx >= TRACING_MAP_VALS_MAX + TRACING_MAP_VARS_MAX))
4431 		return -EINVAL;
4432 
4433 	if (find_var(hist_data, file, var_name) && !hist_data->remove) {
4434 		hist_err(tr, HIST_ERR_DUPLICATE_VAR, errpos(var_name));
4435 		return -EINVAL;
4436 	}
4437 
4438 	flags |= HIST_FIELD_FL_VAR;
4439 	hist_data->n_vars++;
4440 	if (WARN_ON(hist_data->n_vars > TRACING_MAP_VARS_MAX))
4441 		return -EINVAL;
4442 
4443 	ret = __create_val_field(hist_data, val_idx, file, var_name, expr_str, flags);
4444 
4445 	if (!ret && hist_data->fields[val_idx]->flags & HIST_FIELD_FL_EXECNAME)
4446 		update_var_execname(hist_data->fields[val_idx]);
4447 
4448 	if (!ret && hist_data->fields[val_idx]->flags &
4449 	    (HIST_FIELD_FL_STRING | HIST_FIELD_FL_STACKTRACE))
4450 		hist_data->fields[val_idx]->var_str_idx = hist_data->n_var_str++;
4451 
4452 	return ret;
4453 }
4454 
create_val_fields(struct hist_trigger_data * hist_data,struct trace_event_file * file)4455 static int create_val_fields(struct hist_trigger_data *hist_data,
4456 			     struct trace_event_file *file)
4457 {
4458 	unsigned int i, j = 1, n_hitcount = 0;
4459 	char *fields_str, *field_str;
4460 	int ret;
4461 
4462 	ret = create_hitcount_val(hist_data);
4463 	if (ret)
4464 		goto out;
4465 
4466 	fields_str = hist_data->attrs->vals_str;
4467 	if (!fields_str)
4468 		goto out;
4469 
4470 	for (i = 0, j = 1; i < TRACING_MAP_VALS_MAX &&
4471 		     j < TRACING_MAP_VALS_MAX; i++) {
4472 		field_str = strsep(&fields_str, ",");
4473 		if (!field_str)
4474 			break;
4475 
4476 		if (strcmp(field_str, "hitcount") == 0) {
4477 			if (!n_hitcount++)
4478 				continue;
4479 		}
4480 
4481 		ret = create_val_field(hist_data, j++, file, field_str);
4482 		if (ret)
4483 			goto out;
4484 	}
4485 
4486 	if (fields_str && (strcmp(fields_str, "hitcount") != 0))
4487 		ret = -EINVAL;
4488  out:
4489 	/* There is only raw hitcount but nohitcount suppresses it. */
4490 	if (j == 1 && hist_data->attrs->no_hitcount) {
4491 		hist_err(hist_data->event_file->tr, HIST_ERR_NEED_NOHC_VAL, 0);
4492 		ret = -ENOENT;
4493 	}
4494 
4495 	return ret;
4496 }
4497 
create_key_field(struct hist_trigger_data * hist_data,unsigned int key_idx,unsigned int key_offset,struct trace_event_file * file,char * field_str)4498 static int create_key_field(struct hist_trigger_data *hist_data,
4499 			    unsigned int key_idx,
4500 			    unsigned int key_offset,
4501 			    struct trace_event_file *file,
4502 			    char *field_str)
4503 {
4504 	struct trace_array *tr = hist_data->event_file->tr;
4505 	struct hist_field *hist_field = NULL;
4506 	unsigned long flags = 0;
4507 	unsigned int key_size;
4508 	int ret = 0, n_subexprs = 0;
4509 
4510 	if (WARN_ON(key_idx >= HIST_FIELDS_MAX))
4511 		return -EINVAL;
4512 
4513 	flags |= HIST_FIELD_FL_KEY;
4514 
4515 	if (strcmp(field_str, "stacktrace") == 0) {
4516 		flags |= HIST_FIELD_FL_STACKTRACE;
4517 		key_size = sizeof(unsigned long) * HIST_STACKTRACE_DEPTH;
4518 		hist_field = create_hist_field(hist_data, NULL, flags, NULL);
4519 	} else {
4520 		hist_field = parse_expr(hist_data, file, field_str, flags,
4521 					NULL, &n_subexprs);
4522 		if (IS_ERR(hist_field)) {
4523 			ret = PTR_ERR(hist_field);
4524 			goto out;
4525 		}
4526 
4527 		if (field_has_hist_vars(hist_field, 0))	{
4528 			hist_err(tr, HIST_ERR_INVALID_REF_KEY, errpos(field_str));
4529 			destroy_hist_field(hist_field, 0);
4530 			ret = -EINVAL;
4531 			goto out;
4532 		}
4533 
4534 		key_size = hist_field->size;
4535 	}
4536 
4537 	hist_data->fields[key_idx] = hist_field;
4538 
4539 	key_size = ALIGN(key_size, sizeof(u64));
4540 	hist_data->fields[key_idx]->size = key_size;
4541 	hist_data->fields[key_idx]->offset = key_offset;
4542 
4543 	hist_data->key_size += key_size;
4544 
4545 	if (hist_data->key_size > HIST_KEY_SIZE_MAX) {
4546 		ret = -EINVAL;
4547 		goto out;
4548 	}
4549 
4550 	hist_data->n_keys++;
4551 	hist_data->n_fields++;
4552 
4553 	if (WARN_ON(hist_data->n_keys > TRACING_MAP_KEYS_MAX))
4554 		return -EINVAL;
4555 
4556 	ret = key_size;
4557  out:
4558 	return ret;
4559 }
4560 
create_key_fields(struct hist_trigger_data * hist_data,struct trace_event_file * file)4561 static int create_key_fields(struct hist_trigger_data *hist_data,
4562 			     struct trace_event_file *file)
4563 {
4564 	unsigned int i, key_offset = 0, n_vals = hist_data->n_vals;
4565 	char *fields_str, *field_str;
4566 	int ret = -EINVAL;
4567 
4568 	fields_str = hist_data->attrs->keys_str;
4569 	if (!fields_str)
4570 		goto out;
4571 
4572 	for (i = n_vals; i < n_vals + TRACING_MAP_KEYS_MAX; i++) {
4573 		field_str = strsep(&fields_str, ",");
4574 		if (!field_str)
4575 			break;
4576 		ret = create_key_field(hist_data, i, key_offset,
4577 				       file, field_str);
4578 		if (ret < 0)
4579 			goto out;
4580 		key_offset += ret;
4581 	}
4582 	if (fields_str) {
4583 		ret = -EINVAL;
4584 		goto out;
4585 	}
4586 	ret = 0;
4587  out:
4588 	return ret;
4589 }
4590 
create_var_fields(struct hist_trigger_data * hist_data,struct trace_event_file * file)4591 static int create_var_fields(struct hist_trigger_data *hist_data,
4592 			     struct trace_event_file *file)
4593 {
4594 	unsigned int i, j = hist_data->n_vals;
4595 	int ret = 0;
4596 
4597 	unsigned int n_vars = hist_data->attrs->var_defs.n_vars;
4598 
4599 	for (i = 0; i < n_vars; i++) {
4600 		char *var_name = hist_data->attrs->var_defs.name[i];
4601 		char *expr = hist_data->attrs->var_defs.expr[i];
4602 
4603 		ret = create_var_field(hist_data, j++, file, var_name, expr);
4604 		if (ret)
4605 			goto out;
4606 	}
4607  out:
4608 	return ret;
4609 }
4610 
free_var_defs(struct hist_trigger_data * hist_data)4611 static void free_var_defs(struct hist_trigger_data *hist_data)
4612 {
4613 	unsigned int i;
4614 
4615 	for (i = 0; i < hist_data->attrs->var_defs.n_vars; i++) {
4616 		kfree(hist_data->attrs->var_defs.name[i]);
4617 		kfree(hist_data->attrs->var_defs.expr[i]);
4618 	}
4619 
4620 	hist_data->attrs->var_defs.n_vars = 0;
4621 }
4622 
parse_var_defs(struct hist_trigger_data * hist_data)4623 static int parse_var_defs(struct hist_trigger_data *hist_data)
4624 {
4625 	struct trace_array *tr = hist_data->event_file->tr;
4626 	char *s, *str, *var_name, *field_str;
4627 	unsigned int i, j, n_vars = 0;
4628 	int ret = 0;
4629 
4630 	for (i = 0; i < hist_data->attrs->n_assignments; i++) {
4631 		str = hist_data->attrs->assignment_str[i];
4632 		for (j = 0; j < TRACING_MAP_VARS_MAX; j++) {
4633 			field_str = strsep(&str, ",");
4634 			if (!field_str)
4635 				break;
4636 
4637 			var_name = strsep(&field_str, "=");
4638 			if (!var_name || !field_str) {
4639 				hist_err(tr, HIST_ERR_MALFORMED_ASSIGNMENT,
4640 					 errpos(var_name));
4641 				ret = -EINVAL;
4642 				goto free;
4643 			}
4644 
4645 			if (n_vars == TRACING_MAP_VARS_MAX) {
4646 				hist_err(tr, HIST_ERR_TOO_MANY_VARS, errpos(var_name));
4647 				ret = -EINVAL;
4648 				goto free;
4649 			}
4650 
4651 			s = kstrdup(var_name, GFP_KERNEL);
4652 			if (!s) {
4653 				ret = -ENOMEM;
4654 				goto free;
4655 			}
4656 			hist_data->attrs->var_defs.name[n_vars] = s;
4657 
4658 			s = kstrdup(field_str, GFP_KERNEL);
4659 			if (!s) {
4660 				kfree(hist_data->attrs->var_defs.name[n_vars]);
4661 				hist_data->attrs->var_defs.name[n_vars] = NULL;
4662 				ret = -ENOMEM;
4663 				goto free;
4664 			}
4665 			hist_data->attrs->var_defs.expr[n_vars++] = s;
4666 
4667 			hist_data->attrs->var_defs.n_vars = n_vars;
4668 		}
4669 	}
4670 
4671 	return ret;
4672  free:
4673 	free_var_defs(hist_data);
4674 
4675 	return ret;
4676 }
4677 
create_hist_fields(struct hist_trigger_data * hist_data,struct trace_event_file * file)4678 static int create_hist_fields(struct hist_trigger_data *hist_data,
4679 			      struct trace_event_file *file)
4680 {
4681 	int ret;
4682 
4683 	ret = parse_var_defs(hist_data);
4684 	if (ret)
4685 		return ret;
4686 
4687 	ret = create_val_fields(hist_data, file);
4688 	if (ret)
4689 		goto out;
4690 
4691 	ret = create_var_fields(hist_data, file);
4692 	if (ret)
4693 		goto out;
4694 
4695 	ret = create_key_fields(hist_data, file);
4696 
4697  out:
4698 	free_var_defs(hist_data);
4699 
4700 	return ret;
4701 }
4702 
is_descending(struct trace_array * tr,const char * str)4703 static int is_descending(struct trace_array *tr, const char *str)
4704 {
4705 	if (!str)
4706 		return 0;
4707 
4708 	if (strcmp(str, "descending") == 0)
4709 		return 1;
4710 
4711 	if (strcmp(str, "ascending") == 0)
4712 		return 0;
4713 
4714 	hist_err(tr, HIST_ERR_INVALID_SORT_MODIFIER, errpos((char *)str));
4715 
4716 	return -EINVAL;
4717 }
4718 
create_sort_keys(struct hist_trigger_data * hist_data)4719 static int create_sort_keys(struct hist_trigger_data *hist_data)
4720 {
4721 	struct trace_array *tr = hist_data->event_file->tr;
4722 	char *fields_str = hist_data->attrs->sort_key_str;
4723 	struct tracing_map_sort_key *sort_key;
4724 	int descending, ret = 0;
4725 	unsigned int i, j, k;
4726 
4727 	hist_data->n_sort_keys = 1; /* we always have at least one, hitcount */
4728 
4729 	if (!fields_str)
4730 		goto out;
4731 
4732 	for (i = 0; i < TRACING_MAP_SORT_KEYS_MAX; i++) {
4733 		struct hist_field *hist_field;
4734 		char *field_str, *field_name;
4735 		const char *test_name;
4736 
4737 		sort_key = &hist_data->sort_keys[i];
4738 
4739 		field_str = strsep(&fields_str, ",");
4740 		if (!field_str)
4741 			break;
4742 
4743 		if (!*field_str) {
4744 			ret = -EINVAL;
4745 			hist_err(tr, HIST_ERR_EMPTY_SORT_FIELD, errpos("sort="));
4746 			break;
4747 		}
4748 
4749 		if ((i == TRACING_MAP_SORT_KEYS_MAX - 1) && fields_str) {
4750 			hist_err(tr, HIST_ERR_TOO_MANY_SORT_FIELDS, errpos("sort="));
4751 			ret = -EINVAL;
4752 			break;
4753 		}
4754 
4755 		field_name = strsep(&field_str, ".");
4756 		if (!field_name || !*field_name) {
4757 			ret = -EINVAL;
4758 			hist_err(tr, HIST_ERR_EMPTY_SORT_FIELD, errpos("sort="));
4759 			break;
4760 		}
4761 
4762 		if (strcmp(field_name, "hitcount") == 0) {
4763 			descending = is_descending(tr, field_str);
4764 			if (descending < 0) {
4765 				ret = descending;
4766 				break;
4767 			}
4768 			sort_key->descending = descending;
4769 			continue;
4770 		}
4771 
4772 		for (j = 1, k = 1; j < hist_data->n_fields; j++) {
4773 			unsigned int idx;
4774 
4775 			hist_field = hist_data->fields[j];
4776 			if (hist_field->flags & HIST_FIELD_FL_VAR)
4777 				continue;
4778 
4779 			idx = k++;
4780 
4781 			test_name = hist_field_name(hist_field, 0);
4782 
4783 			if (strcmp(field_name, test_name) == 0) {
4784 				sort_key->field_idx = idx;
4785 				descending = is_descending(tr, field_str);
4786 				if (descending < 0) {
4787 					ret = descending;
4788 					goto out;
4789 				}
4790 				sort_key->descending = descending;
4791 				break;
4792 			}
4793 		}
4794 		if (j == hist_data->n_fields) {
4795 			ret = -EINVAL;
4796 			hist_err(tr, HIST_ERR_INVALID_SORT_FIELD, errpos(field_name));
4797 			break;
4798 		}
4799 	}
4800 
4801 	hist_data->n_sort_keys = i;
4802  out:
4803 	return ret;
4804 }
4805 
destroy_actions(struct hist_trigger_data * hist_data)4806 static void destroy_actions(struct hist_trigger_data *hist_data)
4807 {
4808 	unsigned int i;
4809 
4810 	for (i = 0; i < hist_data->n_actions; i++) {
4811 		struct action_data *data = hist_data->actions[i];
4812 
4813 		if (data->handler == HANDLER_ONMATCH)
4814 			onmatch_destroy(data);
4815 		else if (data->handler == HANDLER_ONMAX ||
4816 			 data->handler == HANDLER_ONCHANGE)
4817 			track_data_destroy(hist_data, data);
4818 		else
4819 			kfree(data);
4820 	}
4821 }
4822 
parse_actions(struct hist_trigger_data * hist_data)4823 static int parse_actions(struct hist_trigger_data *hist_data)
4824 {
4825 	struct trace_array *tr = hist_data->event_file->tr;
4826 	struct action_data *data;
4827 	unsigned int i;
4828 	int ret = 0;
4829 	char *str;
4830 	int len;
4831 
4832 	for (i = 0; i < hist_data->attrs->n_actions; i++) {
4833 		enum handler_id hid = 0;
4834 		char *action_str;
4835 
4836 		str = hist_data->attrs->action_str[i];
4837 
4838 		if ((len = str_has_prefix(str, "onmatch(")))
4839 			hid = HANDLER_ONMATCH;
4840 		else if ((len = str_has_prefix(str, "onmax(")))
4841 			hid = HANDLER_ONMAX;
4842 		else if ((len = str_has_prefix(str, "onchange(")))
4843 			hid = HANDLER_ONCHANGE;
4844 
4845 		action_str = str + len;
4846 
4847 		switch (hid) {
4848 		case HANDLER_ONMATCH:
4849 			data = onmatch_parse(tr, action_str);
4850 			break;
4851 		case HANDLER_ONMAX:
4852 		case HANDLER_ONCHANGE:
4853 			data = track_data_parse(hist_data, action_str, hid);
4854 			break;
4855 		default:
4856 			data = ERR_PTR(-EINVAL);
4857 			break;
4858 		}
4859 
4860 		if (IS_ERR(data)) {
4861 			ret = PTR_ERR(data);
4862 			break;
4863 		}
4864 
4865 		hist_data->actions[hist_data->n_actions++] = data;
4866 	}
4867 
4868 	return ret;
4869 }
4870 
create_actions(struct hist_trigger_data * hist_data)4871 static int create_actions(struct hist_trigger_data *hist_data)
4872 {
4873 	struct action_data *data;
4874 	unsigned int i;
4875 	int ret = 0;
4876 
4877 	for (i = 0; i < hist_data->attrs->n_actions; i++) {
4878 		data = hist_data->actions[i];
4879 
4880 		if (data->handler == HANDLER_ONMATCH) {
4881 			ret = onmatch_create(hist_data, data);
4882 			if (ret)
4883 				break;
4884 		} else if (data->handler == HANDLER_ONMAX ||
4885 			   data->handler == HANDLER_ONCHANGE) {
4886 			ret = track_data_create(hist_data, data);
4887 			if (ret)
4888 				break;
4889 		} else {
4890 			ret = -EINVAL;
4891 			break;
4892 		}
4893 	}
4894 
4895 	return ret;
4896 }
4897 
print_actions(struct seq_file * m,struct hist_trigger_data * hist_data,struct tracing_map_elt * elt)4898 static void print_actions(struct seq_file *m,
4899 			  struct hist_trigger_data *hist_data,
4900 			  struct tracing_map_elt *elt)
4901 {
4902 	unsigned int i;
4903 
4904 	for (i = 0; i < hist_data->n_actions; i++) {
4905 		struct action_data *data = hist_data->actions[i];
4906 
4907 		if (data->action == ACTION_SNAPSHOT)
4908 			continue;
4909 
4910 		if (data->handler == HANDLER_ONMAX ||
4911 		    data->handler == HANDLER_ONCHANGE)
4912 			track_data_print(m, hist_data, elt, data);
4913 	}
4914 }
4915 
print_action_spec(struct seq_file * m,struct hist_trigger_data * hist_data,struct action_data * data)4916 static void print_action_spec(struct seq_file *m,
4917 			      struct hist_trigger_data *hist_data,
4918 			      struct action_data *data)
4919 {
4920 	unsigned int i;
4921 
4922 	if (data->action == ACTION_SAVE) {
4923 		for (i = 0; i < hist_data->n_save_vars; i++) {
4924 			seq_printf(m, "%s", hist_data->save_vars[i]->var->var.name);
4925 			if (i < hist_data->n_save_vars - 1)
4926 				seq_puts(m, ",");
4927 		}
4928 	} else if (data->action == ACTION_TRACE) {
4929 		if (data->use_trace_keyword)
4930 			seq_printf(m, "%s", data->synth_event_name);
4931 		for (i = 0; i < data->n_params; i++) {
4932 			if (i || data->use_trace_keyword)
4933 				seq_puts(m, ",");
4934 			seq_printf(m, "%s", data->params[i]);
4935 		}
4936 	}
4937 }
4938 
print_track_data_spec(struct seq_file * m,struct hist_trigger_data * hist_data,struct action_data * data)4939 static void print_track_data_spec(struct seq_file *m,
4940 				  struct hist_trigger_data *hist_data,
4941 				  struct action_data *data)
4942 {
4943 	if (data->handler == HANDLER_ONMAX)
4944 		seq_puts(m, ":onmax(");
4945 	else if (data->handler == HANDLER_ONCHANGE)
4946 		seq_puts(m, ":onchange(");
4947 	seq_printf(m, "%s", data->track_data.var_str);
4948 	seq_printf(m, ").%s(", data->action_name);
4949 
4950 	print_action_spec(m, hist_data, data);
4951 
4952 	seq_puts(m, ")");
4953 }
4954 
print_onmatch_spec(struct seq_file * m,struct hist_trigger_data * hist_data,struct action_data * data)4955 static void print_onmatch_spec(struct seq_file *m,
4956 			       struct hist_trigger_data *hist_data,
4957 			       struct action_data *data)
4958 {
4959 	seq_printf(m, ":onmatch(%s.%s).", data->match_data.event_system,
4960 		   data->match_data.event);
4961 
4962 	seq_printf(m, "%s(", data->action_name);
4963 
4964 	print_action_spec(m, hist_data, data);
4965 
4966 	seq_puts(m, ")");
4967 }
4968 
actions_match(struct hist_trigger_data * hist_data,struct hist_trigger_data * hist_data_test)4969 static bool actions_match(struct hist_trigger_data *hist_data,
4970 			  struct hist_trigger_data *hist_data_test)
4971 {
4972 	unsigned int i, j;
4973 
4974 	if (hist_data->n_actions != hist_data_test->n_actions)
4975 		return false;
4976 
4977 	for (i = 0; i < hist_data->n_actions; i++) {
4978 		struct action_data *data = hist_data->actions[i];
4979 		struct action_data *data_test = hist_data_test->actions[i];
4980 		char *action_name, *action_name_test;
4981 
4982 		if (data->handler != data_test->handler)
4983 			return false;
4984 		if (data->action != data_test->action)
4985 			return false;
4986 
4987 		if (data->n_params != data_test->n_params)
4988 			return false;
4989 
4990 		for (j = 0; j < data->n_params; j++) {
4991 			if (strcmp(data->params[j], data_test->params[j]) != 0)
4992 				return false;
4993 		}
4994 
4995 		if (data->use_trace_keyword)
4996 			action_name = data->synth_event_name;
4997 		else
4998 			action_name = data->action_name;
4999 
5000 		if (data_test->use_trace_keyword)
5001 			action_name_test = data_test->synth_event_name;
5002 		else
5003 			action_name_test = data_test->action_name;
5004 
5005 		if (strcmp(action_name, action_name_test) != 0)
5006 			return false;
5007 
5008 		if (data->handler == HANDLER_ONMATCH) {
5009 			if (strcmp(data->match_data.event_system,
5010 				   data_test->match_data.event_system) != 0)
5011 				return false;
5012 			if (strcmp(data->match_data.event,
5013 				   data_test->match_data.event) != 0)
5014 				return false;
5015 		} else if (data->handler == HANDLER_ONMAX ||
5016 			   data->handler == HANDLER_ONCHANGE) {
5017 			if (strcmp(data->track_data.var_str,
5018 				   data_test->track_data.var_str) != 0)
5019 				return false;
5020 		}
5021 	}
5022 
5023 	return true;
5024 }
5025 
5026 
print_actions_spec(struct seq_file * m,struct hist_trigger_data * hist_data)5027 static void print_actions_spec(struct seq_file *m,
5028 			       struct hist_trigger_data *hist_data)
5029 {
5030 	unsigned int i;
5031 
5032 	for (i = 0; i < hist_data->n_actions; i++) {
5033 		struct action_data *data = hist_data->actions[i];
5034 
5035 		if (data->handler == HANDLER_ONMATCH)
5036 			print_onmatch_spec(m, hist_data, data);
5037 		else if (data->handler == HANDLER_ONMAX ||
5038 			 data->handler == HANDLER_ONCHANGE)
5039 			print_track_data_spec(m, hist_data, data);
5040 	}
5041 }
5042 
destroy_field_var_hists(struct hist_trigger_data * hist_data)5043 static void destroy_field_var_hists(struct hist_trigger_data *hist_data)
5044 {
5045 	unsigned int i;
5046 
5047 	for (i = 0; i < hist_data->n_field_var_hists; i++) {
5048 		kfree(hist_data->field_var_hists[i]->cmd);
5049 		kfree(hist_data->field_var_hists[i]);
5050 	}
5051 }
5052 
destroy_hist_data(struct hist_trigger_data * hist_data)5053 static void destroy_hist_data(struct hist_trigger_data *hist_data)
5054 {
5055 	if (!hist_data)
5056 		return;
5057 
5058 	destroy_hist_trigger_attrs(hist_data->attrs);
5059 	destroy_hist_fields(hist_data);
5060 	tracing_map_destroy(hist_data->map);
5061 
5062 	destroy_actions(hist_data);
5063 	destroy_field_vars(hist_data);
5064 	destroy_field_var_hists(hist_data);
5065 
5066 	kfree(hist_data);
5067 }
5068 
create_tracing_map_fields(struct hist_trigger_data * hist_data)5069 static int create_tracing_map_fields(struct hist_trigger_data *hist_data)
5070 {
5071 	struct tracing_map *map = hist_data->map;
5072 	struct ftrace_event_field *field;
5073 	struct hist_field *hist_field;
5074 	int i, idx = 0;
5075 
5076 	for_each_hist_field(i, hist_data) {
5077 		hist_field = hist_data->fields[i];
5078 		if (hist_field->flags & HIST_FIELD_FL_KEY) {
5079 			tracing_map_cmp_fn_t cmp_fn;
5080 
5081 			field = hist_field->field;
5082 
5083 			if (hist_field->flags & HIST_FIELD_FL_STACKTRACE)
5084 				cmp_fn = tracing_map_cmp_none;
5085 			else if (!field || hist_field->flags & HIST_FIELD_FL_CPU)
5086 				cmp_fn = tracing_map_cmp_num(hist_field->size,
5087 							     hist_field->is_signed);
5088 			else if (is_string_field(field))
5089 				cmp_fn = tracing_map_cmp_string;
5090 			else
5091 				cmp_fn = tracing_map_cmp_num(field->size,
5092 							     field->is_signed);
5093 			idx = tracing_map_add_key_field(map,
5094 							hist_field->offset,
5095 							cmp_fn);
5096 		} else if (!(hist_field->flags & HIST_FIELD_FL_VAR))
5097 			idx = tracing_map_add_sum_field(map);
5098 
5099 		if (idx < 0)
5100 			return idx;
5101 
5102 		if (hist_field->flags & HIST_FIELD_FL_VAR) {
5103 			idx = tracing_map_add_var(map);
5104 			if (idx < 0)
5105 				return idx;
5106 			hist_field->var.idx = idx;
5107 			hist_field->var.hist_data = hist_data;
5108 		}
5109 	}
5110 
5111 	return 0;
5112 }
5113 
5114 static struct hist_trigger_data *
create_hist_data(unsigned int map_bits,struct hist_trigger_attrs * attrs,struct trace_event_file * file,bool remove)5115 create_hist_data(unsigned int map_bits,
5116 		 struct hist_trigger_attrs *attrs,
5117 		 struct trace_event_file *file,
5118 		 bool remove)
5119 {
5120 	const struct tracing_map_ops *map_ops = NULL;
5121 	struct hist_trigger_data *hist_data;
5122 	int ret = 0;
5123 
5124 	hist_data = kzalloc(sizeof(*hist_data), GFP_KERNEL);
5125 	if (!hist_data)
5126 		return ERR_PTR(-ENOMEM);
5127 
5128 	hist_data->attrs = attrs;
5129 	hist_data->remove = remove;
5130 	hist_data->event_file = file;
5131 
5132 	ret = parse_actions(hist_data);
5133 	if (ret)
5134 		goto free;
5135 
5136 	ret = create_hist_fields(hist_data, file);
5137 	if (ret)
5138 		goto free;
5139 
5140 	ret = create_sort_keys(hist_data);
5141 	if (ret)
5142 		goto free;
5143 
5144 	map_ops = &hist_trigger_elt_data_ops;
5145 
5146 	hist_data->map = tracing_map_create(map_bits, hist_data->key_size,
5147 					    map_ops, hist_data);
5148 	if (IS_ERR(hist_data->map)) {
5149 		ret = PTR_ERR(hist_data->map);
5150 		hist_data->map = NULL;
5151 		goto free;
5152 	}
5153 
5154 	ret = create_tracing_map_fields(hist_data);
5155 	if (ret)
5156 		goto free;
5157  out:
5158 	return hist_data;
5159  free:
5160 	hist_data->attrs = NULL;
5161 
5162 	destroy_hist_data(hist_data);
5163 
5164 	hist_data = ERR_PTR(ret);
5165 
5166 	goto out;
5167 }
5168 
hist_trigger_elt_update(struct hist_trigger_data * hist_data,struct tracing_map_elt * elt,struct trace_buffer * buffer,void * rec,struct ring_buffer_event * rbe,u64 * var_ref_vals)5169 static void hist_trigger_elt_update(struct hist_trigger_data *hist_data,
5170 				    struct tracing_map_elt *elt,
5171 				    struct trace_buffer *buffer, void *rec,
5172 				    struct ring_buffer_event *rbe,
5173 				    u64 *var_ref_vals)
5174 {
5175 	struct hist_elt_data *elt_data;
5176 	struct hist_field *hist_field;
5177 	unsigned int i, var_idx;
5178 	u64 hist_val;
5179 
5180 	elt_data = elt->private_data;
5181 	elt_data->var_ref_vals = var_ref_vals;
5182 
5183 	for_each_hist_val_field(i, hist_data) {
5184 		hist_field = hist_data->fields[i];
5185 		hist_val = hist_fn_call(hist_field, elt, buffer, rbe, rec);
5186 		if (hist_field->flags & HIST_FIELD_FL_VAR) {
5187 			var_idx = hist_field->var.idx;
5188 
5189 			if (hist_field->flags &
5190 			    (HIST_FIELD_FL_STRING | HIST_FIELD_FL_STACKTRACE)) {
5191 				unsigned int str_start, var_str_idx, idx;
5192 				char *str, *val_str;
5193 				unsigned int size;
5194 
5195 				str_start = hist_data->n_field_var_str +
5196 					hist_data->n_save_var_str;
5197 				var_str_idx = hist_field->var_str_idx;
5198 				idx = str_start + var_str_idx;
5199 
5200 				str = elt_data->field_var_str[idx];
5201 				val_str = (char *)(uintptr_t)hist_val;
5202 
5203 				if (hist_field->flags & HIST_FIELD_FL_STRING) {
5204 					size = min(hist_field->size, STR_VAR_LEN_MAX);
5205 					strscpy(str, val_str, size);
5206 				} else {
5207 					char *stack_start = str + sizeof(unsigned long);
5208 					int e;
5209 
5210 					e = stack_trace_save((void *)stack_start,
5211 							     HIST_STACKTRACE_DEPTH,
5212 							     HIST_STACKTRACE_SKIP);
5213 					if (e < HIST_STACKTRACE_DEPTH - 1)
5214 						((unsigned long *)stack_start)[e] = 0;
5215 					*((unsigned long *)str) = e;
5216 				}
5217 				hist_val = (u64)(uintptr_t)str;
5218 			}
5219 			tracing_map_set_var(elt, var_idx, hist_val);
5220 			continue;
5221 		}
5222 		tracing_map_update_sum(elt, i, hist_val);
5223 	}
5224 
5225 	for_each_hist_key_field(i, hist_data) {
5226 		hist_field = hist_data->fields[i];
5227 		if (hist_field->flags & HIST_FIELD_FL_VAR) {
5228 			hist_val = hist_fn_call(hist_field, elt, buffer, rbe, rec);
5229 			var_idx = hist_field->var.idx;
5230 			tracing_map_set_var(elt, var_idx, hist_val);
5231 		}
5232 	}
5233 
5234 	update_field_vars(hist_data, elt, buffer, rbe, rec);
5235 }
5236 
add_to_key(char * compound_key,void * key,struct hist_field * key_field,void * rec)5237 static inline void add_to_key(char *compound_key, void *key,
5238 			      struct hist_field *key_field, void *rec)
5239 {
5240 	size_t size = key_field->size;
5241 
5242 	if (key_field->flags & HIST_FIELD_FL_STRING) {
5243 
5244 		if (key_field->flags & HIST_FIELD_FL_COMM) {
5245 			size = strlen((char *)key);
5246 		} else {
5247 			struct ftrace_event_field *field;
5248 
5249 			field = key_field->field;
5250 			if (field->filter_type == FILTER_DYN_STRING ||
5251 			    field->filter_type == FILTER_RDYN_STRING)
5252 				size = *(u32 *)(rec + field->offset) >> 16;
5253 			else if (field->filter_type == FILTER_STATIC_STRING)
5254 				size = field->size;
5255 		}
5256 
5257 		/* ensure NULL-termination */
5258 		if (size > key_field->size - 1)
5259 			size = key_field->size - 1;
5260 	}
5261 	memcpy(compound_key + key_field->offset, key, size);
5262 }
5263 
5264 static void
hist_trigger_actions(struct hist_trigger_data * hist_data,struct tracing_map_elt * elt,struct trace_buffer * buffer,void * rec,struct ring_buffer_event * rbe,void * key,u64 * var_ref_vals)5265 hist_trigger_actions(struct hist_trigger_data *hist_data,
5266 		     struct tracing_map_elt *elt,
5267 		     struct trace_buffer *buffer, void *rec,
5268 		     struct ring_buffer_event *rbe, void *key,
5269 		     u64 *var_ref_vals)
5270 {
5271 	struct action_data *data;
5272 	unsigned int i;
5273 
5274 	for (i = 0; i < hist_data->n_actions; i++) {
5275 		data = hist_data->actions[i];
5276 		data->fn(hist_data, elt, buffer, rec, rbe, key, data, var_ref_vals);
5277 	}
5278 }
5279 
5280 /*
5281  * The hist_pad structure is used to save information to create
5282  * a histogram from the histogram trigger. It's too big to store
5283  * on the stack, so when the histogram trigger is initialized
5284  * a percpu array of 4 hist_pad structures is allocated.
5285  * This will cover every context from normal, softirq, irq and NMI
5286  * in the very unlikely event that a tigger happens at each of
5287  * these contexts and interrupts a currently active trigger.
5288  */
5289 struct hist_pad {
5290 	unsigned long		entries[HIST_STACKTRACE_DEPTH];
5291 	u64			var_ref_vals[TRACING_MAP_VARS_MAX];
5292 	char			compound_key[HIST_KEY_SIZE_MAX];
5293 };
5294 
5295 static struct hist_pad __percpu *hist_pads;
5296 static DEFINE_PER_CPU(int, hist_pad_cnt);
5297 static refcount_t hist_pad_ref;
5298 
5299 /* One hist_pad for every context (normal, softirq, irq, NMI) */
5300 #define MAX_HIST_CNT 4
5301 
alloc_hist_pad(void)5302 static int alloc_hist_pad(void)
5303 {
5304 	lockdep_assert_held(&event_mutex);
5305 
5306 	if (refcount_read(&hist_pad_ref)) {
5307 		refcount_inc(&hist_pad_ref);
5308 		return 0;
5309 	}
5310 
5311 	hist_pads = __alloc_percpu(sizeof(struct hist_pad) * MAX_HIST_CNT,
5312 				   __alignof__(struct hist_pad));
5313 	if (!hist_pads)
5314 		return -ENOMEM;
5315 
5316 	refcount_set(&hist_pad_ref, 1);
5317 	return 0;
5318 }
5319 
free_hist_pad(void)5320 static void free_hist_pad(void)
5321 {
5322 	lockdep_assert_held(&event_mutex);
5323 
5324 	if (!refcount_dec_and_test(&hist_pad_ref))
5325 		return;
5326 
5327 	free_percpu(hist_pads);
5328 	hist_pads = NULL;
5329 }
5330 
get_hist_pad(void)5331 static struct hist_pad *get_hist_pad(void)
5332 {
5333 	struct hist_pad *hist_pad;
5334 	int cnt;
5335 
5336 	if (WARN_ON_ONCE(!hist_pads))
5337 		return NULL;
5338 
5339 	preempt_disable();
5340 
5341 	hist_pad = per_cpu_ptr(hist_pads, smp_processor_id());
5342 
5343 	if (this_cpu_read(hist_pad_cnt) == MAX_HIST_CNT) {
5344 		preempt_enable();
5345 		return NULL;
5346 	}
5347 
5348 	cnt = this_cpu_inc_return(hist_pad_cnt) - 1;
5349 
5350 	return &hist_pad[cnt];
5351 }
5352 
put_hist_pad(void)5353 static void put_hist_pad(void)
5354 {
5355 	this_cpu_dec(hist_pad_cnt);
5356 	preempt_enable();
5357 }
5358 
event_hist_trigger(struct event_trigger_data * data,struct trace_buffer * buffer,void * rec,struct ring_buffer_event * rbe)5359 static void event_hist_trigger(struct event_trigger_data *data,
5360 			       struct trace_buffer *buffer, void *rec,
5361 			       struct ring_buffer_event *rbe)
5362 {
5363 	struct hist_trigger_data *hist_data = data->private_data;
5364 	bool use_compound_key = (hist_data->n_keys > 1);
5365 	struct tracing_map_elt *elt = NULL;
5366 	struct hist_field *key_field;
5367 	struct hist_pad *hist_pad;
5368 	u64 field_contents;
5369 	void *key = NULL;
5370 	unsigned int i;
5371 
5372 	if (unlikely(!rbe))
5373 		return;
5374 
5375 	hist_pad = get_hist_pad();
5376 	if (!hist_pad)
5377 		return;
5378 
5379 	memset(hist_pad->compound_key, 0, hist_data->key_size);
5380 
5381 	for_each_hist_key_field(i, hist_data) {
5382 		key_field = hist_data->fields[i];
5383 
5384 		if (key_field->flags & HIST_FIELD_FL_STACKTRACE) {
5385 			unsigned long *entries = hist_pad->entries;
5386 
5387 			memset(entries, 0, HIST_STACKTRACE_SIZE);
5388 			if (key_field->field) {
5389 				unsigned long *stack, n_entries;
5390 
5391 				field_contents = hist_fn_call(key_field, elt, buffer, rbe, rec);
5392 				stack = (unsigned long *)(long)field_contents;
5393 				n_entries = *stack;
5394 				memcpy(entries, ++stack, n_entries * sizeof(unsigned long));
5395 			} else {
5396 				stack_trace_save(entries, HIST_STACKTRACE_DEPTH,
5397 						 HIST_STACKTRACE_SKIP);
5398 			}
5399 			key = entries;
5400 		} else {
5401 			field_contents = hist_fn_call(key_field, elt, buffer, rbe, rec);
5402 			if (key_field->flags & HIST_FIELD_FL_STRING) {
5403 				key = (void *)(unsigned long)field_contents;
5404 				use_compound_key = true;
5405 			} else
5406 				key = (void *)&field_contents;
5407 		}
5408 
5409 		if (use_compound_key)
5410 			add_to_key(hist_pad->compound_key, key, key_field, rec);
5411 	}
5412 
5413 	if (use_compound_key)
5414 		key = hist_pad->compound_key;
5415 
5416 	if (hist_data->n_var_refs &&
5417 	    !resolve_var_refs(hist_data, key, hist_pad->var_ref_vals, false))
5418 		goto out;
5419 
5420 	elt = tracing_map_insert(hist_data->map, key);
5421 	if (!elt)
5422 		goto out;
5423 
5424 	hist_trigger_elt_update(hist_data, elt, buffer, rec, rbe, hist_pad->var_ref_vals);
5425 
5426 	if (resolve_var_refs(hist_data, key, hist_pad->var_ref_vals, true)) {
5427 		hist_trigger_actions(hist_data, elt, buffer, rec, rbe,
5428 				     key, hist_pad->var_ref_vals);
5429 	}
5430 
5431 	hist_poll_wakeup();
5432 
5433  out:
5434 	put_hist_pad();
5435 }
5436 
hist_trigger_stacktrace_print(struct seq_file * m,unsigned long * stacktrace_entries,unsigned int max_entries)5437 static void hist_trigger_stacktrace_print(struct seq_file *m,
5438 					  unsigned long *stacktrace_entries,
5439 					  unsigned int max_entries)
5440 {
5441 	unsigned int spaces = 8;
5442 	unsigned int i;
5443 
5444 	for (i = 0; i < max_entries; i++) {
5445 		if (!stacktrace_entries[i])
5446 			return;
5447 
5448 		seq_printf(m, "%*c", 1 + spaces, ' ');
5449 		seq_printf(m, "%pS\n", (void*)stacktrace_entries[i]);
5450 	}
5451 }
5452 
hist_trigger_print_key(struct seq_file * m,struct hist_trigger_data * hist_data,void * key,struct tracing_map_elt * elt)5453 static void hist_trigger_print_key(struct seq_file *m,
5454 				   struct hist_trigger_data *hist_data,
5455 				   void *key,
5456 				   struct tracing_map_elt *elt)
5457 {
5458 	struct hist_field *key_field;
5459 	bool multiline = false;
5460 	const char *field_name;
5461 	unsigned int i;
5462 	u64 uval;
5463 
5464 	seq_puts(m, "{ ");
5465 
5466 	for_each_hist_key_field(i, hist_data) {
5467 		key_field = hist_data->fields[i];
5468 
5469 		if (i > hist_data->n_vals)
5470 			seq_puts(m, ", ");
5471 
5472 		field_name = hist_field_name(key_field, 0);
5473 
5474 		if (key_field->flags & HIST_FIELD_FL_HEX) {
5475 			uval = *(u64 *)(key + key_field->offset);
5476 			seq_printf(m, "%s: %llx", field_name, uval);
5477 		} else if (key_field->flags & HIST_FIELD_FL_SYM) {
5478 			uval = *(u64 *)(key + key_field->offset);
5479 			seq_printf(m, "%s: [%llx] %-45ps", field_name,
5480 				   uval, (void *)(uintptr_t)uval);
5481 		} else if (key_field->flags & HIST_FIELD_FL_SYM_OFFSET) {
5482 			uval = *(u64 *)(key + key_field->offset);
5483 			seq_printf(m, "%s: [%llx] %-55pS", field_name,
5484 				   uval, (void *)(uintptr_t)uval);
5485 		} else if (key_field->flags & HIST_FIELD_FL_EXECNAME) {
5486 			struct hist_elt_data *elt_data = elt->private_data;
5487 			char *comm;
5488 
5489 			if (WARN_ON_ONCE(!elt_data))
5490 				return;
5491 
5492 			comm = elt_data->comm;
5493 
5494 			uval = *(u64 *)(key + key_field->offset);
5495 			seq_printf(m, "%s: %-16s[%10llu]", field_name,
5496 				   comm, uval);
5497 		} else if (key_field->flags & HIST_FIELD_FL_SYSCALL) {
5498 			const char *syscall_name;
5499 
5500 			uval = *(u64 *)(key + key_field->offset);
5501 			syscall_name = get_syscall_name(uval);
5502 			if (!syscall_name)
5503 				syscall_name = "unknown_syscall";
5504 
5505 			seq_printf(m, "%s: %-30s[%3llu]", field_name,
5506 				   syscall_name, uval);
5507 		} else if (key_field->flags & HIST_FIELD_FL_STACKTRACE) {
5508 			if (key_field->field)
5509 				seq_printf(m, "%s.stacktrace", key_field->field->name);
5510 			else
5511 				seq_puts(m, "common_stacktrace:\n");
5512 			hist_trigger_stacktrace_print(m,
5513 						      key + key_field->offset,
5514 						      HIST_STACKTRACE_DEPTH);
5515 			multiline = true;
5516 		} else if (key_field->flags & HIST_FIELD_FL_LOG2) {
5517 			seq_printf(m, "%s: ~ 2^%-2llu", field_name,
5518 				   *(u64 *)(key + key_field->offset));
5519 		} else if (key_field->flags & HIST_FIELD_FL_BUCKET) {
5520 			unsigned long buckets = key_field->buckets;
5521 			uval = *(u64 *)(key + key_field->offset);
5522 			seq_printf(m, "%s: ~ %llu-%llu", field_name,
5523 				   uval, uval + buckets -1);
5524 		} else if (key_field->flags & HIST_FIELD_FL_STRING) {
5525 			seq_printf(m, "%s: %-50s", field_name,
5526 				   (char *)(key + key_field->offset));
5527 		} else {
5528 			uval = *(u64 *)(key + key_field->offset);
5529 			seq_printf(m, "%s: %10llu", field_name, uval);
5530 		}
5531 	}
5532 
5533 	if (!multiline)
5534 		seq_puts(m, " ");
5535 
5536 	seq_puts(m, "}");
5537 }
5538 
5539 /* Get the 100 times of the percentage of @val in @total */
__get_percentage(u64 val,u64 total)5540 static inline unsigned int __get_percentage(u64 val, u64 total)
5541 {
5542 	if (!total)
5543 		goto div0;
5544 
5545 	if (val < (U64_MAX / 10000))
5546 		return (unsigned int)div64_ul(val * 10000, total);
5547 
5548 	total = div64_u64(total, 10000);
5549 	if (!total)
5550 		goto div0;
5551 
5552 	return (unsigned int)div64_ul(val, total);
5553 div0:
5554 	return val ? UINT_MAX : 0;
5555 }
5556 
5557 #define BAR_CHAR '#'
5558 
__fill_bar_str(char * buf,int size,u64 val,u64 max)5559 static inline const char *__fill_bar_str(char *buf, int size, u64 val, u64 max)
5560 {
5561 	unsigned int len = __get_percentage(val, max);
5562 	int i;
5563 
5564 	if (len == UINT_MAX) {
5565 		snprintf(buf, size, "[ERROR]");
5566 		return buf;
5567 	}
5568 
5569 	len = len * size / 10000;
5570 	for (i = 0; i < len && i < size; i++)
5571 		buf[i] = BAR_CHAR;
5572 	while (i < size)
5573 		buf[i++] = ' ';
5574 	buf[size] = '\0';
5575 
5576 	return buf;
5577 }
5578 
5579 struct hist_val_stat {
5580 	u64 max;
5581 	u64 total;
5582 };
5583 
hist_trigger_print_val(struct seq_file * m,unsigned int idx,const char * field_name,unsigned long flags,struct hist_val_stat * stats,struct tracing_map_elt * elt)5584 static void hist_trigger_print_val(struct seq_file *m, unsigned int idx,
5585 				   const char *field_name, unsigned long flags,
5586 				   struct hist_val_stat *stats,
5587 				   struct tracing_map_elt *elt)
5588 {
5589 	u64 val = tracing_map_read_sum(elt, idx);
5590 	unsigned int pc;
5591 	char bar[21];
5592 
5593 	if (flags & HIST_FIELD_FL_PERCENT) {
5594 		pc = __get_percentage(val, stats[idx].total);
5595 		if (pc == UINT_MAX)
5596 			seq_printf(m, " %s (%%):[ERROR]", field_name);
5597 		else
5598 			seq_printf(m, " %s (%%): %3u.%02u", field_name,
5599 					pc / 100, pc % 100);
5600 	} else if (flags & HIST_FIELD_FL_GRAPH) {
5601 		seq_printf(m, " %s: %20s", field_name,
5602 			   __fill_bar_str(bar, 20, val, stats[idx].max));
5603 	} else if (flags & HIST_FIELD_FL_HEX) {
5604 		seq_printf(m, " %s: %10llx", field_name, val);
5605 	} else {
5606 		seq_printf(m, " %s: %10llu", field_name, val);
5607 	}
5608 }
5609 
hist_trigger_entry_print(struct seq_file * m,struct hist_trigger_data * hist_data,struct hist_val_stat * stats,void * key,struct tracing_map_elt * elt)5610 static void hist_trigger_entry_print(struct seq_file *m,
5611 				     struct hist_trigger_data *hist_data,
5612 				     struct hist_val_stat *stats,
5613 				     void *key,
5614 				     struct tracing_map_elt *elt)
5615 {
5616 	const char *field_name;
5617 	unsigned int i = HITCOUNT_IDX;
5618 	unsigned long flags;
5619 
5620 	hist_trigger_print_key(m, hist_data, key, elt);
5621 
5622 	/* At first, show the raw hitcount if !nohitcount */
5623 	if (!hist_data->attrs->no_hitcount)
5624 		hist_trigger_print_val(m, i, "hitcount", 0, stats, elt);
5625 
5626 	for (i = 1; i < hist_data->n_vals; i++) {
5627 		field_name = hist_field_name(hist_data->fields[i], 0);
5628 		flags = hist_data->fields[i]->flags;
5629 		if (flags & HIST_FIELD_FL_VAR || flags & HIST_FIELD_FL_EXPR)
5630 			continue;
5631 
5632 		seq_puts(m, " ");
5633 		hist_trigger_print_val(m, i, field_name, flags, stats, elt);
5634 	}
5635 
5636 	print_actions(m, hist_data, elt);
5637 
5638 	seq_puts(m, "\n");
5639 }
5640 
print_entries(struct seq_file * m,struct hist_trigger_data * hist_data)5641 static int print_entries(struct seq_file *m,
5642 			 struct hist_trigger_data *hist_data)
5643 {
5644 	struct tracing_map_sort_entry **sort_entries = NULL;
5645 	struct tracing_map *map = hist_data->map;
5646 	int i, j, n_entries;
5647 	struct hist_val_stat *stats = NULL;
5648 	u64 val;
5649 
5650 	n_entries = tracing_map_sort_entries(map, hist_data->sort_keys,
5651 					     hist_data->n_sort_keys,
5652 					     &sort_entries);
5653 	if (n_entries < 0)
5654 		return n_entries;
5655 
5656 	/* Calculate the max and the total for each field if needed. */
5657 	for (j = 0; j < hist_data->n_vals; j++) {
5658 		if (!(hist_data->fields[j]->flags &
5659 			(HIST_FIELD_FL_PERCENT | HIST_FIELD_FL_GRAPH)))
5660 			continue;
5661 		if (!stats) {
5662 			stats = kcalloc(hist_data->n_vals, sizeof(*stats),
5663 				       GFP_KERNEL);
5664 			if (!stats) {
5665 				n_entries = -ENOMEM;
5666 				goto out;
5667 			}
5668 		}
5669 		for (i = 0; i < n_entries; i++) {
5670 			val = tracing_map_read_sum(sort_entries[i]->elt, j);
5671 			stats[j].total += val;
5672 			if (stats[j].max < val)
5673 				stats[j].max = val;
5674 		}
5675 	}
5676 
5677 	for (i = 0; i < n_entries; i++)
5678 		hist_trigger_entry_print(m, hist_data, stats,
5679 					 sort_entries[i]->key,
5680 					 sort_entries[i]->elt);
5681 
5682 	kfree(stats);
5683 out:
5684 	tracing_map_destroy_sort_entries(sort_entries, n_entries);
5685 
5686 	return n_entries;
5687 }
5688 
hist_trigger_show(struct seq_file * m,struct event_trigger_data * data,int n)5689 static void hist_trigger_show(struct seq_file *m,
5690 			      struct event_trigger_data *data, int n)
5691 {
5692 	struct hist_trigger_data *hist_data;
5693 	int n_entries;
5694 
5695 	if (n > 0)
5696 		seq_puts(m, "\n\n");
5697 
5698 	seq_puts(m, "# event histogram\n#\n# trigger info: ");
5699 	data->ops->print(m, data);
5700 	seq_puts(m, "#\n\n");
5701 
5702 	hist_data = data->private_data;
5703 	n_entries = print_entries(m, hist_data);
5704 	if (n_entries < 0)
5705 		n_entries = 0;
5706 
5707 	track_data_snapshot_print(m, hist_data);
5708 
5709 	seq_printf(m, "\nTotals:\n    Hits: %llu\n    Entries: %u\n    Dropped: %llu\n",
5710 		   (u64)atomic64_read(&hist_data->map->hits),
5711 		   n_entries, (u64)atomic64_read(&hist_data->map->drops));
5712 }
5713 
5714 struct hist_file_data {
5715 	struct file *file;
5716 	u64 last_read;
5717 	u64 last_act;
5718 };
5719 
get_hist_hit_count(struct trace_event_file * event_file)5720 static u64 get_hist_hit_count(struct trace_event_file *event_file)
5721 {
5722 	struct hist_trigger_data *hist_data;
5723 	struct event_trigger_data *data;
5724 	u64 ret = 0;
5725 
5726 	list_for_each_entry(data, &event_file->triggers, list) {
5727 		if (data->cmd_ops->trigger_type == ETT_EVENT_HIST) {
5728 			hist_data = data->private_data;
5729 			ret += atomic64_read(&hist_data->map->hits);
5730 		}
5731 	}
5732 	return ret;
5733 }
5734 
hist_show(struct seq_file * m,void * v)5735 static int hist_show(struct seq_file *m, void *v)
5736 {
5737 	struct hist_file_data *hist_file = m->private;
5738 	struct event_trigger_data *data;
5739 	struct trace_event_file *event_file;
5740 	int n = 0;
5741 
5742 	guard(mutex)(&event_mutex);
5743 
5744 	event_file = event_file_file(hist_file->file);
5745 	if (unlikely(!event_file))
5746 		return -ENODEV;
5747 
5748 	list_for_each_entry(data, &event_file->triggers, list) {
5749 		if (data->cmd_ops->trigger_type == ETT_EVENT_HIST)
5750 			hist_trigger_show(m, data, n++);
5751 	}
5752 	hist_file->last_read = get_hist_hit_count(event_file);
5753 	/*
5754 	 * Update last_act too so that poll()/POLLPRI can wait for the next
5755 	 * event after any syscall on hist file.
5756 	 */
5757 	hist_file->last_act = hist_file->last_read;
5758 
5759 	return 0;
5760 }
5761 
event_hist_poll(struct file * file,struct poll_table_struct * wait)5762 static __poll_t event_hist_poll(struct file *file, struct poll_table_struct *wait)
5763 {
5764 	struct trace_event_file *event_file;
5765 	struct seq_file *m = file->private_data;
5766 	struct hist_file_data *hist_file = m->private;
5767 	__poll_t ret = 0;
5768 	u64 cnt;
5769 
5770 	guard(mutex)(&event_mutex);
5771 
5772 	event_file = event_file_data(file);
5773 	if (!event_file)
5774 		return EPOLLERR;
5775 
5776 	hist_poll_wait(file, wait);
5777 
5778 	cnt = get_hist_hit_count(event_file);
5779 	if (hist_file->last_read != cnt)
5780 		ret |= EPOLLIN | EPOLLRDNORM;
5781 	if (hist_file->last_act != cnt) {
5782 		hist_file->last_act = cnt;
5783 		ret |= EPOLLPRI;
5784 	}
5785 
5786 	return ret;
5787 }
5788 
event_hist_release(struct inode * inode,struct file * file)5789 static int event_hist_release(struct inode *inode, struct file *file)
5790 {
5791 	struct seq_file *m = file->private_data;
5792 	struct hist_file_data *hist_file = m->private;
5793 
5794 	kfree(hist_file);
5795 	return tracing_single_release_file_tr(inode, file);
5796 }
5797 
event_hist_open(struct inode * inode,struct file * file)5798 static int event_hist_open(struct inode *inode, struct file *file)
5799 {
5800 	struct trace_event_file *event_file;
5801 	struct hist_file_data *hist_file;
5802 	int ret;
5803 
5804 	ret = tracing_open_file_tr(inode, file);
5805 	if (ret)
5806 		return ret;
5807 
5808 	guard(mutex)(&event_mutex);
5809 
5810 	event_file = event_file_data(file);
5811 	if (!event_file) {
5812 		ret = -ENODEV;
5813 		goto err;
5814 	}
5815 
5816 	hist_file = kzalloc(sizeof(*hist_file), GFP_KERNEL);
5817 	if (!hist_file) {
5818 		ret = -ENOMEM;
5819 		goto err;
5820 	}
5821 
5822 	hist_file->file = file;
5823 	hist_file->last_act = get_hist_hit_count(event_file);
5824 
5825 	/* Clear private_data to avoid warning in single_open() */
5826 	file->private_data = NULL;
5827 	ret = single_open(file, hist_show, hist_file);
5828 	if (ret) {
5829 		kfree(hist_file);
5830 		goto err;
5831 	}
5832 
5833 	return 0;
5834 err:
5835 	tracing_release_file_tr(inode, file);
5836 	return ret;
5837 }
5838 
5839 const struct file_operations event_hist_fops = {
5840 	.open = event_hist_open,
5841 	.read = seq_read,
5842 	.llseek = seq_lseek,
5843 	.release = event_hist_release,
5844 	.poll = event_hist_poll,
5845 };
5846 
5847 #ifdef CONFIG_HIST_TRIGGERS_DEBUG
hist_field_debug_show_flags(struct seq_file * m,unsigned long flags)5848 static void hist_field_debug_show_flags(struct seq_file *m,
5849 					unsigned long flags)
5850 {
5851 	seq_puts(m, "      flags:\n");
5852 
5853 	if (flags & HIST_FIELD_FL_KEY)
5854 		seq_puts(m, "        HIST_FIELD_FL_KEY\n");
5855 	else if (flags & HIST_FIELD_FL_HITCOUNT)
5856 		seq_puts(m, "        VAL: HIST_FIELD_FL_HITCOUNT\n");
5857 	else if (flags & HIST_FIELD_FL_VAR)
5858 		seq_puts(m, "        HIST_FIELD_FL_VAR\n");
5859 	else if (flags & HIST_FIELD_FL_VAR_REF)
5860 		seq_puts(m, "        HIST_FIELD_FL_VAR_REF\n");
5861 	else
5862 		seq_puts(m, "        VAL: normal u64 value\n");
5863 
5864 	if (flags & HIST_FIELD_FL_ALIAS)
5865 		seq_puts(m, "        HIST_FIELD_FL_ALIAS\n");
5866 	else if (flags & HIST_FIELD_FL_CONST)
5867 		seq_puts(m, "        HIST_FIELD_FL_CONST\n");
5868 }
5869 
hist_field_debug_show(struct seq_file * m,struct hist_field * field,unsigned long flags)5870 static int hist_field_debug_show(struct seq_file *m,
5871 				 struct hist_field *field, unsigned long flags)
5872 {
5873 	if ((field->flags & flags) != flags) {
5874 		seq_printf(m, "ERROR: bad flags - %lx\n", flags);
5875 		return -EINVAL;
5876 	}
5877 
5878 	hist_field_debug_show_flags(m, field->flags);
5879 	if (field->field)
5880 		seq_printf(m, "      ftrace_event_field name: %s\n",
5881 			   field->field->name);
5882 
5883 	if (field->flags & HIST_FIELD_FL_VAR) {
5884 		seq_printf(m, "      var.name: %s\n", field->var.name);
5885 		seq_printf(m, "      var.idx (into tracing_map_elt.vars[]): %u\n",
5886 			   field->var.idx);
5887 	}
5888 
5889 	if (field->flags & HIST_FIELD_FL_CONST)
5890 		seq_printf(m, "      constant: %llu\n", field->constant);
5891 
5892 	if (field->flags & HIST_FIELD_FL_ALIAS)
5893 		seq_printf(m, "      var_ref_idx (into hist_data->var_refs[]): %u\n",
5894 			   field->var_ref_idx);
5895 
5896 	if (field->flags & HIST_FIELD_FL_VAR_REF) {
5897 		seq_printf(m, "      name: %s\n", field->name);
5898 		seq_printf(m, "      var.idx (into tracing_map_elt.vars[]): %u\n",
5899 			   field->var.idx);
5900 		seq_printf(m, "      var.hist_data: %p\n", field->var.hist_data);
5901 		seq_printf(m, "      var_ref_idx (into hist_data->var_refs[]): %u\n",
5902 			   field->var_ref_idx);
5903 		if (field->system)
5904 			seq_printf(m, "      system: %s\n", field->system);
5905 		if (field->event_name)
5906 			seq_printf(m, "      event_name: %s\n", field->event_name);
5907 	}
5908 
5909 	seq_printf(m, "      type: %s\n", field->type);
5910 	seq_printf(m, "      size: %u\n", field->size);
5911 	seq_printf(m, "      is_signed: %u\n", field->is_signed);
5912 
5913 	return 0;
5914 }
5915 
field_var_debug_show(struct seq_file * m,struct field_var * field_var,unsigned int i,bool save_vars)5916 static int field_var_debug_show(struct seq_file *m,
5917 				struct field_var *field_var, unsigned int i,
5918 				bool save_vars)
5919 {
5920 	const char *vars_name = save_vars ? "save_vars" : "field_vars";
5921 	struct hist_field *field;
5922 	int ret = 0;
5923 
5924 	seq_printf(m, "\n    hist_data->%s[%d]:\n", vars_name, i);
5925 
5926 	field = field_var->var;
5927 
5928 	seq_printf(m, "\n      %s[%d].var:\n", vars_name, i);
5929 
5930 	hist_field_debug_show_flags(m, field->flags);
5931 	seq_printf(m, "      var.name: %s\n", field->var.name);
5932 	seq_printf(m, "      var.idx (into tracing_map_elt.vars[]): %u\n",
5933 		   field->var.idx);
5934 
5935 	field = field_var->val;
5936 
5937 	seq_printf(m, "\n      %s[%d].val:\n", vars_name, i);
5938 	if (field->field)
5939 		seq_printf(m, "      ftrace_event_field name: %s\n",
5940 			   field->field->name);
5941 	else {
5942 		ret = -EINVAL;
5943 		goto out;
5944 	}
5945 
5946 	seq_printf(m, "      type: %s\n", field->type);
5947 	seq_printf(m, "      size: %u\n", field->size);
5948 	seq_printf(m, "      is_signed: %u\n", field->is_signed);
5949 out:
5950 	return ret;
5951 }
5952 
hist_action_debug_show(struct seq_file * m,struct action_data * data,int i)5953 static int hist_action_debug_show(struct seq_file *m,
5954 				  struct action_data *data, int i)
5955 {
5956 	int ret = 0;
5957 
5958 	if (data->handler == HANDLER_ONMAX ||
5959 	    data->handler == HANDLER_ONCHANGE) {
5960 		seq_printf(m, "\n    hist_data->actions[%d].track_data.var_ref:\n", i);
5961 		ret = hist_field_debug_show(m, data->track_data.var_ref,
5962 					    HIST_FIELD_FL_VAR_REF);
5963 		if (ret)
5964 			goto out;
5965 
5966 		seq_printf(m, "\n    hist_data->actions[%d].track_data.track_var:\n", i);
5967 		ret = hist_field_debug_show(m, data->track_data.track_var,
5968 					    HIST_FIELD_FL_VAR);
5969 		if (ret)
5970 			goto out;
5971 	}
5972 
5973 	if (data->handler == HANDLER_ONMATCH) {
5974 		seq_printf(m, "\n    hist_data->actions[%d].match_data.event_system: %s\n",
5975 			   i, data->match_data.event_system);
5976 		seq_printf(m, "    hist_data->actions[%d].match_data.event: %s\n",
5977 			   i, data->match_data.event);
5978 	}
5979 out:
5980 	return ret;
5981 }
5982 
hist_actions_debug_show(struct seq_file * m,struct hist_trigger_data * hist_data)5983 static int hist_actions_debug_show(struct seq_file *m,
5984 				   struct hist_trigger_data *hist_data)
5985 {
5986 	int i, ret = 0;
5987 
5988 	if (hist_data->n_actions)
5989 		seq_puts(m, "\n  action tracking variables (for onmax()/onchange()/onmatch()):\n");
5990 
5991 	for (i = 0; i < hist_data->n_actions; i++) {
5992 		struct action_data *action = hist_data->actions[i];
5993 
5994 		ret = hist_action_debug_show(m, action, i);
5995 		if (ret)
5996 			goto out;
5997 	}
5998 
5999 	if (hist_data->n_save_vars)
6000 		seq_puts(m, "\n  save action variables (save() params):\n");
6001 
6002 	for (i = 0; i < hist_data->n_save_vars; i++) {
6003 		ret = field_var_debug_show(m, hist_data->save_vars[i], i, true);
6004 		if (ret)
6005 			goto out;
6006 	}
6007 out:
6008 	return ret;
6009 }
6010 
hist_trigger_debug_show(struct seq_file * m,struct event_trigger_data * data,int n)6011 static void hist_trigger_debug_show(struct seq_file *m,
6012 				    struct event_trigger_data *data, int n)
6013 {
6014 	struct hist_trigger_data *hist_data;
6015 	int i, ret;
6016 
6017 	if (n > 0)
6018 		seq_puts(m, "\n\n");
6019 
6020 	seq_puts(m, "# event histogram\n#\n# trigger info: ");
6021 	data->ops->print(m, data);
6022 	seq_puts(m, "#\n\n");
6023 
6024 	hist_data = data->private_data;
6025 
6026 	seq_printf(m, "hist_data: %p\n\n", hist_data);
6027 	seq_printf(m, "  n_vals: %u\n", hist_data->n_vals);
6028 	seq_printf(m, "  n_keys: %u\n", hist_data->n_keys);
6029 	seq_printf(m, "  n_fields: %u\n", hist_data->n_fields);
6030 
6031 	seq_puts(m, "\n  val fields:\n\n");
6032 
6033 	seq_puts(m, "    hist_data->fields[0]:\n");
6034 	ret = hist_field_debug_show(m, hist_data->fields[0],
6035 				    HIST_FIELD_FL_HITCOUNT);
6036 	if (ret)
6037 		return;
6038 
6039 	for (i = 1; i < hist_data->n_vals; i++) {
6040 		seq_printf(m, "\n    hist_data->fields[%d]:\n", i);
6041 		ret = hist_field_debug_show(m, hist_data->fields[i], 0);
6042 		if (ret)
6043 			return;
6044 	}
6045 
6046 	seq_puts(m, "\n  key fields:\n");
6047 
6048 	for (i = hist_data->n_vals; i < hist_data->n_fields; i++) {
6049 		seq_printf(m, "\n    hist_data->fields[%d]:\n", i);
6050 		ret = hist_field_debug_show(m, hist_data->fields[i],
6051 					    HIST_FIELD_FL_KEY);
6052 		if (ret)
6053 			return;
6054 	}
6055 
6056 	if (hist_data->n_var_refs)
6057 		seq_puts(m, "\n  variable reference fields:\n");
6058 
6059 	for (i = 0; i < hist_data->n_var_refs; i++) {
6060 		seq_printf(m, "\n    hist_data->var_refs[%d]:\n", i);
6061 		ret = hist_field_debug_show(m, hist_data->var_refs[i],
6062 					    HIST_FIELD_FL_VAR_REF);
6063 		if (ret)
6064 			return;
6065 	}
6066 
6067 	if (hist_data->n_field_vars)
6068 		seq_puts(m, "\n  field variables:\n");
6069 
6070 	for (i = 0; i < hist_data->n_field_vars; i++) {
6071 		ret = field_var_debug_show(m, hist_data->field_vars[i], i, false);
6072 		if (ret)
6073 			return;
6074 	}
6075 
6076 	ret = hist_actions_debug_show(m, hist_data);
6077 	if (ret)
6078 		return;
6079 }
6080 
hist_debug_show(struct seq_file * m,void * v)6081 static int hist_debug_show(struct seq_file *m, void *v)
6082 {
6083 	struct event_trigger_data *data;
6084 	struct trace_event_file *event_file;
6085 	int n = 0;
6086 
6087 	guard(mutex)(&event_mutex);
6088 
6089 	event_file = event_file_file(m->private);
6090 	if (unlikely(!event_file))
6091 		return -ENODEV;
6092 
6093 	list_for_each_entry(data, &event_file->triggers, list) {
6094 		if (data->cmd_ops->trigger_type == ETT_EVENT_HIST)
6095 			hist_trigger_debug_show(m, data, n++);
6096 	}
6097 	return 0;
6098 }
6099 
event_hist_debug_open(struct inode * inode,struct file * file)6100 static int event_hist_debug_open(struct inode *inode, struct file *file)
6101 {
6102 	int ret;
6103 
6104 	ret = tracing_open_file_tr(inode, file);
6105 	if (ret)
6106 		return ret;
6107 
6108 	/* Clear private_data to avoid warning in single_open() */
6109 	file->private_data = NULL;
6110 	ret = single_open(file, hist_debug_show, file);
6111 	if (ret)
6112 		tracing_release_file_tr(inode, file);
6113 	return ret;
6114 }
6115 
6116 const struct file_operations event_hist_debug_fops = {
6117 	.open = event_hist_debug_open,
6118 	.read = seq_read,
6119 	.llseek = seq_lseek,
6120 	.release = tracing_single_release_file_tr,
6121 };
6122 #endif
6123 
hist_field_print(struct seq_file * m,struct hist_field * hist_field)6124 static void hist_field_print(struct seq_file *m, struct hist_field *hist_field)
6125 {
6126 	const char *field_name = hist_field_name(hist_field, 0);
6127 
6128 	if (hist_field->var.name)
6129 		seq_printf(m, "%s=", hist_field->var.name);
6130 
6131 	if (hist_field->flags & HIST_FIELD_FL_CPU)
6132 		seq_puts(m, "common_cpu");
6133 	if (hist_field->flags & HIST_FIELD_FL_COMM)
6134 		seq_puts(m, "common_comm");
6135 	else if (hist_field->flags & HIST_FIELD_FL_CONST)
6136 		seq_printf(m, "%llu", hist_field->constant);
6137 	else if (field_name) {
6138 		if (hist_field->flags & HIST_FIELD_FL_VAR_REF ||
6139 		    hist_field->flags & HIST_FIELD_FL_ALIAS)
6140 			seq_putc(m, '$');
6141 		seq_printf(m, "%s", field_name);
6142 	} else if (hist_field->flags & HIST_FIELD_FL_TIMESTAMP)
6143 		seq_puts(m, "common_timestamp");
6144 
6145 	if (hist_field->flags) {
6146 		if (!(hist_field->flags & HIST_FIELD_FL_VAR_REF) &&
6147 		    !(hist_field->flags & HIST_FIELD_FL_EXPR) &&
6148 		    !(hist_field->flags & HIST_FIELD_FL_STACKTRACE)) {
6149 			const char *flags = get_hist_field_flags(hist_field);
6150 
6151 			if (flags)
6152 				seq_printf(m, ".%s", flags);
6153 		}
6154 	}
6155 	if (hist_field->buckets)
6156 		seq_printf(m, "=%ld", hist_field->buckets);
6157 }
6158 
event_hist_trigger_print(struct seq_file * m,struct event_trigger_data * data)6159 static int event_hist_trigger_print(struct seq_file *m,
6160 				    struct event_trigger_data *data)
6161 {
6162 	struct hist_trigger_data *hist_data = data->private_data;
6163 	struct hist_field *field;
6164 	bool have_var = false;
6165 	bool show_val = false;
6166 	unsigned int i;
6167 
6168 	seq_puts(m, HIST_PREFIX);
6169 
6170 	if (data->name)
6171 		seq_printf(m, "%s:", data->name);
6172 
6173 	seq_puts(m, "keys=");
6174 
6175 	for_each_hist_key_field(i, hist_data) {
6176 		field = hist_data->fields[i];
6177 
6178 		if (i > hist_data->n_vals)
6179 			seq_puts(m, ",");
6180 
6181 		if (field->flags & HIST_FIELD_FL_STACKTRACE) {
6182 			if (field->field)
6183 				seq_printf(m, "%s.stacktrace", field->field->name);
6184 			else
6185 				seq_puts(m, "common_stacktrace");
6186 		} else
6187 			hist_field_print(m, field);
6188 	}
6189 
6190 	seq_puts(m, ":vals=");
6191 
6192 	for_each_hist_val_field(i, hist_data) {
6193 		field = hist_data->fields[i];
6194 		if (field->flags & HIST_FIELD_FL_VAR) {
6195 			have_var = true;
6196 			continue;
6197 		}
6198 
6199 		if (i == HITCOUNT_IDX) {
6200 			if (hist_data->attrs->no_hitcount)
6201 				continue;
6202 			seq_puts(m, "hitcount");
6203 		} else {
6204 			if (show_val)
6205 				seq_puts(m, ",");
6206 			hist_field_print(m, field);
6207 		}
6208 		show_val = true;
6209 	}
6210 
6211 	if (have_var) {
6212 		unsigned int n = 0;
6213 
6214 		seq_puts(m, ":");
6215 
6216 		for_each_hist_val_field(i, hist_data) {
6217 			field = hist_data->fields[i];
6218 
6219 			if (field->flags & HIST_FIELD_FL_VAR) {
6220 				if (n++)
6221 					seq_puts(m, ",");
6222 				hist_field_print(m, field);
6223 			}
6224 		}
6225 	}
6226 
6227 	seq_puts(m, ":sort=");
6228 
6229 	for (i = 0; i < hist_data->n_sort_keys; i++) {
6230 		struct tracing_map_sort_key *sort_key;
6231 		unsigned int idx, first_key_idx;
6232 
6233 		/* skip VAR vals */
6234 		first_key_idx = hist_data->n_vals - hist_data->n_vars;
6235 
6236 		sort_key = &hist_data->sort_keys[i];
6237 		idx = sort_key->field_idx;
6238 
6239 		if (WARN_ON(idx >= HIST_FIELDS_MAX))
6240 			return -EINVAL;
6241 
6242 		if (i > 0)
6243 			seq_puts(m, ",");
6244 
6245 		if (idx == HITCOUNT_IDX)
6246 			seq_puts(m, "hitcount");
6247 		else {
6248 			if (idx >= first_key_idx)
6249 				idx += hist_data->n_vars;
6250 			hist_field_print(m, hist_data->fields[idx]);
6251 		}
6252 
6253 		if (sort_key->descending)
6254 			seq_puts(m, ".descending");
6255 	}
6256 	seq_printf(m, ":size=%u", (1 << hist_data->map->map_bits));
6257 	if (hist_data->enable_timestamps)
6258 		seq_printf(m, ":clock=%s", hist_data->attrs->clock);
6259 	if (hist_data->attrs->no_hitcount)
6260 		seq_puts(m, ":nohitcount");
6261 
6262 	print_actions_spec(m, hist_data);
6263 
6264 	if (data->filter_str)
6265 		seq_printf(m, " if %s", data->filter_str);
6266 
6267 	if (data->paused)
6268 		seq_puts(m, " [paused]");
6269 	else
6270 		seq_puts(m, " [active]");
6271 
6272 	seq_putc(m, '\n');
6273 
6274 	return 0;
6275 }
6276 
event_hist_trigger_init(struct event_trigger_data * data)6277 static int event_hist_trigger_init(struct event_trigger_data *data)
6278 {
6279 	struct hist_trigger_data *hist_data = data->private_data;
6280 
6281 	if (alloc_hist_pad() < 0)
6282 		return -ENOMEM;
6283 
6284 	if (!data->ref && hist_data->attrs->name)
6285 		save_named_trigger(hist_data->attrs->name, data);
6286 
6287 	data->ref++;
6288 
6289 	return 0;
6290 }
6291 
unregister_field_var_hists(struct hist_trigger_data * hist_data)6292 static void unregister_field_var_hists(struct hist_trigger_data *hist_data)
6293 {
6294 	struct trace_event_file *file;
6295 	unsigned int i;
6296 	char *cmd;
6297 	int ret;
6298 
6299 	for (i = 0; i < hist_data->n_field_var_hists; i++) {
6300 		file = hist_data->field_var_hists[i]->hist_data->event_file;
6301 		cmd = hist_data->field_var_hists[i]->cmd;
6302 		ret = event_hist_trigger_parse(&trigger_hist_cmd, file,
6303 					       "!hist", "hist", cmd);
6304 		WARN_ON_ONCE(ret < 0);
6305 	}
6306 }
6307 
event_hist_trigger_free(struct event_trigger_data * data)6308 static void event_hist_trigger_free(struct event_trigger_data *data)
6309 {
6310 	struct hist_trigger_data *hist_data = data->private_data;
6311 
6312 	if (WARN_ON_ONCE(data->ref <= 0))
6313 		return;
6314 
6315 	data->ref--;
6316 	if (!data->ref) {
6317 		if (data->name)
6318 			del_named_trigger(data);
6319 
6320 		trigger_data_free(data);
6321 
6322 		remove_hist_vars(hist_data);
6323 
6324 		unregister_field_var_hists(hist_data);
6325 
6326 		destroy_hist_data(hist_data);
6327 	}
6328 	free_hist_pad();
6329 }
6330 
6331 static const struct event_trigger_ops event_hist_trigger_ops = {
6332 	.trigger		= event_hist_trigger,
6333 	.print			= event_hist_trigger_print,
6334 	.init			= event_hist_trigger_init,
6335 	.free			= event_hist_trigger_free,
6336 };
6337 
event_hist_trigger_named_init(struct event_trigger_data * data)6338 static int event_hist_trigger_named_init(struct event_trigger_data *data)
6339 {
6340 	data->ref++;
6341 
6342 	save_named_trigger(data->named_data->name, data);
6343 
6344 	return event_hist_trigger_init(data->named_data);
6345 }
6346 
event_hist_trigger_named_free(struct event_trigger_data * data)6347 static void event_hist_trigger_named_free(struct event_trigger_data *data)
6348 {
6349 	if (WARN_ON_ONCE(data->ref <= 0))
6350 		return;
6351 
6352 	event_hist_trigger_free(data->named_data);
6353 
6354 	data->ref--;
6355 	if (!data->ref) {
6356 		del_named_trigger(data);
6357 		trigger_data_free(data);
6358 	}
6359 }
6360 
6361 static const struct event_trigger_ops event_hist_trigger_named_ops = {
6362 	.trigger		= event_hist_trigger,
6363 	.print			= event_hist_trigger_print,
6364 	.init			= event_hist_trigger_named_init,
6365 	.free			= event_hist_trigger_named_free,
6366 };
6367 
event_hist_get_trigger_ops(char * cmd,char * param)6368 static const struct event_trigger_ops *event_hist_get_trigger_ops(char *cmd,
6369 								  char *param)
6370 {
6371 	return &event_hist_trigger_ops;
6372 }
6373 
hist_clear(struct event_trigger_data * data)6374 static void hist_clear(struct event_trigger_data *data)
6375 {
6376 	struct hist_trigger_data *hist_data = data->private_data;
6377 
6378 	if (data->name)
6379 		pause_named_trigger(data);
6380 
6381 	tracepoint_synchronize_unregister();
6382 
6383 	tracing_map_clear(hist_data->map);
6384 
6385 	if (data->name)
6386 		unpause_named_trigger(data);
6387 }
6388 
compatible_field(struct ftrace_event_field * field,struct ftrace_event_field * test_field)6389 static bool compatible_field(struct ftrace_event_field *field,
6390 			     struct ftrace_event_field *test_field)
6391 {
6392 	if (field == test_field)
6393 		return true;
6394 	if (field == NULL || test_field == NULL)
6395 		return false;
6396 	if (strcmp(field->name, test_field->name) != 0)
6397 		return false;
6398 	if (strcmp(field->type, test_field->type) != 0)
6399 		return false;
6400 	if (field->size != test_field->size)
6401 		return false;
6402 	if (field->is_signed != test_field->is_signed)
6403 		return false;
6404 
6405 	return true;
6406 }
6407 
hist_trigger_match(struct event_trigger_data * data,struct event_trigger_data * data_test,struct event_trigger_data * named_data,bool ignore_filter)6408 static bool hist_trigger_match(struct event_trigger_data *data,
6409 			       struct event_trigger_data *data_test,
6410 			       struct event_trigger_data *named_data,
6411 			       bool ignore_filter)
6412 {
6413 	struct tracing_map_sort_key *sort_key, *sort_key_test;
6414 	struct hist_trigger_data *hist_data, *hist_data_test;
6415 	struct hist_field *key_field, *key_field_test;
6416 	unsigned int i;
6417 
6418 	if (named_data && (named_data != data_test) &&
6419 	    (named_data != data_test->named_data))
6420 		return false;
6421 
6422 	if (!named_data && is_named_trigger(data_test))
6423 		return false;
6424 
6425 	hist_data = data->private_data;
6426 	hist_data_test = data_test->private_data;
6427 
6428 	if (hist_data->n_vals != hist_data_test->n_vals ||
6429 	    hist_data->n_fields != hist_data_test->n_fields ||
6430 	    hist_data->n_sort_keys != hist_data_test->n_sort_keys)
6431 		return false;
6432 
6433 	if (!ignore_filter) {
6434 		if ((data->filter_str && !data_test->filter_str) ||
6435 		   (!data->filter_str && data_test->filter_str))
6436 			return false;
6437 	}
6438 
6439 	for_each_hist_field(i, hist_data) {
6440 		key_field = hist_data->fields[i];
6441 		key_field_test = hist_data_test->fields[i];
6442 
6443 		if (key_field->flags != key_field_test->flags)
6444 			return false;
6445 		if (!compatible_field(key_field->field, key_field_test->field))
6446 			return false;
6447 		if (key_field->offset != key_field_test->offset)
6448 			return false;
6449 		if (key_field->size != key_field_test->size)
6450 			return false;
6451 		if (key_field->is_signed != key_field_test->is_signed)
6452 			return false;
6453 		if (!!key_field->var.name != !!key_field_test->var.name)
6454 			return false;
6455 		if (key_field->var.name &&
6456 		    strcmp(key_field->var.name, key_field_test->var.name) != 0)
6457 			return false;
6458 	}
6459 
6460 	for (i = 0; i < hist_data->n_sort_keys; i++) {
6461 		sort_key = &hist_data->sort_keys[i];
6462 		sort_key_test = &hist_data_test->sort_keys[i];
6463 
6464 		if (sort_key->field_idx != sort_key_test->field_idx ||
6465 		    sort_key->descending != sort_key_test->descending)
6466 			return false;
6467 	}
6468 
6469 	if (!ignore_filter && data->filter_str &&
6470 	    (strcmp(data->filter_str, data_test->filter_str) != 0))
6471 		return false;
6472 
6473 	if (!actions_match(hist_data, hist_data_test))
6474 		return false;
6475 
6476 	return true;
6477 }
6478 
existing_hist_update_only(char * glob,struct event_trigger_data * data,struct trace_event_file * file)6479 static bool existing_hist_update_only(char *glob,
6480 				      struct event_trigger_data *data,
6481 				      struct trace_event_file *file)
6482 {
6483 	struct hist_trigger_data *hist_data = data->private_data;
6484 	struct event_trigger_data *test, *named_data = NULL;
6485 	bool updated = false;
6486 
6487 	if (!hist_data->attrs->pause && !hist_data->attrs->cont &&
6488 	    !hist_data->attrs->clear)
6489 		goto out;
6490 
6491 	if (hist_data->attrs->name) {
6492 		named_data = find_named_trigger(hist_data->attrs->name);
6493 		if (named_data) {
6494 			if (!hist_trigger_match(data, named_data, named_data,
6495 						true))
6496 				goto out;
6497 		}
6498 	}
6499 
6500 	if (hist_data->attrs->name && !named_data)
6501 		goto out;
6502 
6503 	list_for_each_entry(test, &file->triggers, list) {
6504 		if (test->cmd_ops->trigger_type == ETT_EVENT_HIST) {
6505 			if (!hist_trigger_match(data, test, named_data, false))
6506 				continue;
6507 			if (hist_data->attrs->pause)
6508 				test->paused = true;
6509 			else if (hist_data->attrs->cont)
6510 				test->paused = false;
6511 			else if (hist_data->attrs->clear)
6512 				hist_clear(test);
6513 			updated = true;
6514 			goto out;
6515 		}
6516 	}
6517  out:
6518 	return updated;
6519 }
6520 
hist_register_trigger(char * glob,struct event_trigger_data * data,struct trace_event_file * file)6521 static int hist_register_trigger(char *glob,
6522 				 struct event_trigger_data *data,
6523 				 struct trace_event_file *file)
6524 {
6525 	struct hist_trigger_data *hist_data = data->private_data;
6526 	struct event_trigger_data *test, *named_data = NULL;
6527 	struct trace_array *tr = file->tr;
6528 	int ret = 0;
6529 
6530 	if (hist_data->attrs->name) {
6531 		named_data = find_named_trigger(hist_data->attrs->name);
6532 		if (named_data) {
6533 			if (!hist_trigger_match(data, named_data, named_data,
6534 						true)) {
6535 				hist_err(tr, HIST_ERR_NAMED_MISMATCH, errpos(hist_data->attrs->name));
6536 				ret = -EINVAL;
6537 				goto out;
6538 			}
6539 		}
6540 	}
6541 
6542 	if (hist_data->attrs->name && !named_data)
6543 		goto new;
6544 
6545 	lockdep_assert_held(&event_mutex);
6546 
6547 	list_for_each_entry(test, &file->triggers, list) {
6548 		if (test->cmd_ops->trigger_type == ETT_EVENT_HIST) {
6549 			if (hist_trigger_match(data, test, named_data, false)) {
6550 				hist_err(tr, HIST_ERR_TRIGGER_EEXIST, 0);
6551 				ret = -EEXIST;
6552 				goto out;
6553 			}
6554 		}
6555 	}
6556  new:
6557 	if (hist_data->attrs->cont || hist_data->attrs->clear) {
6558 		hist_err(tr, HIST_ERR_TRIGGER_ENOENT_CLEAR, 0);
6559 		ret = -ENOENT;
6560 		goto out;
6561 	}
6562 
6563 	if (hist_data->attrs->pause)
6564 		data->paused = true;
6565 
6566 	if (named_data) {
6567 		data->private_data = named_data->private_data;
6568 		set_named_trigger_data(data, named_data);
6569 		data->ops = &event_hist_trigger_named_ops;
6570 	}
6571 
6572 	if (data->ops->init) {
6573 		ret = data->ops->init(data);
6574 		if (ret < 0)
6575 			goto out;
6576 	}
6577 
6578 	if (hist_data->enable_timestamps) {
6579 		char *clock = hist_data->attrs->clock;
6580 
6581 		ret = tracing_set_clock(file->tr, hist_data->attrs->clock);
6582 		if (ret) {
6583 			hist_err(tr, HIST_ERR_SET_CLOCK_FAIL, errpos(clock));
6584 			goto out;
6585 		}
6586 
6587 		tracing_set_filter_buffering(file->tr, true);
6588 	}
6589 
6590 	if (named_data)
6591 		destroy_hist_data(hist_data);
6592  out:
6593 	return ret;
6594 }
6595 
hist_trigger_enable(struct event_trigger_data * data,struct trace_event_file * file)6596 static int hist_trigger_enable(struct event_trigger_data *data,
6597 			       struct trace_event_file *file)
6598 {
6599 	int ret = 0;
6600 
6601 	list_add_tail_rcu(&data->list, &file->triggers);
6602 
6603 	update_cond_flag(file);
6604 
6605 	if (trace_event_trigger_enable_disable(file, 1) < 0) {
6606 		list_del_rcu(&data->list);
6607 		update_cond_flag(file);
6608 		ret--;
6609 	}
6610 
6611 	return ret;
6612 }
6613 
have_hist_trigger_match(struct event_trigger_data * data,struct trace_event_file * file)6614 static bool have_hist_trigger_match(struct event_trigger_data *data,
6615 				    struct trace_event_file *file)
6616 {
6617 	struct hist_trigger_data *hist_data = data->private_data;
6618 	struct event_trigger_data *test, *named_data = NULL;
6619 	bool match = false;
6620 
6621 	lockdep_assert_held(&event_mutex);
6622 
6623 	if (hist_data->attrs->name)
6624 		named_data = find_named_trigger(hist_data->attrs->name);
6625 
6626 	list_for_each_entry(test, &file->triggers, list) {
6627 		if (test->cmd_ops->trigger_type == ETT_EVENT_HIST) {
6628 			if (hist_trigger_match(data, test, named_data, false)) {
6629 				match = true;
6630 				break;
6631 			}
6632 		}
6633 	}
6634 
6635 	return match;
6636 }
6637 
hist_trigger_check_refs(struct event_trigger_data * data,struct trace_event_file * file)6638 static bool hist_trigger_check_refs(struct event_trigger_data *data,
6639 				    struct trace_event_file *file)
6640 {
6641 	struct hist_trigger_data *hist_data = data->private_data;
6642 	struct event_trigger_data *test, *named_data = NULL;
6643 
6644 	lockdep_assert_held(&event_mutex);
6645 
6646 	if (hist_data->attrs->name)
6647 		named_data = find_named_trigger(hist_data->attrs->name);
6648 
6649 	list_for_each_entry(test, &file->triggers, list) {
6650 		if (test->cmd_ops->trigger_type == ETT_EVENT_HIST) {
6651 			if (!hist_trigger_match(data, test, named_data, false))
6652 				continue;
6653 			hist_data = test->private_data;
6654 			if (check_var_refs(hist_data))
6655 				return true;
6656 			break;
6657 		}
6658 	}
6659 
6660 	return false;
6661 }
6662 
hist_unregister_trigger(char * glob,struct event_trigger_data * data,struct trace_event_file * file)6663 static void hist_unregister_trigger(char *glob,
6664 				    struct event_trigger_data *data,
6665 				    struct trace_event_file *file)
6666 {
6667 	struct event_trigger_data *test = NULL, *iter, *named_data = NULL;
6668 	struct hist_trigger_data *hist_data = data->private_data;
6669 
6670 	lockdep_assert_held(&event_mutex);
6671 
6672 	if (hist_data->attrs->name)
6673 		named_data = find_named_trigger(hist_data->attrs->name);
6674 
6675 	list_for_each_entry(iter, &file->triggers, list) {
6676 		if (iter->cmd_ops->trigger_type == ETT_EVENT_HIST) {
6677 			if (!hist_trigger_match(data, iter, named_data, false))
6678 				continue;
6679 			test = iter;
6680 			list_del_rcu(&test->list);
6681 			trace_event_trigger_enable_disable(file, 0);
6682 			update_cond_flag(file);
6683 			break;
6684 		}
6685 	}
6686 
6687 	if (test && test->ops->free)
6688 		test->ops->free(test);
6689 
6690 	if (hist_data->enable_timestamps) {
6691 		if (!hist_data->remove || test)
6692 			tracing_set_filter_buffering(file->tr, false);
6693 	}
6694 }
6695 
hist_file_check_refs(struct trace_event_file * file)6696 static bool hist_file_check_refs(struct trace_event_file *file)
6697 {
6698 	struct hist_trigger_data *hist_data;
6699 	struct event_trigger_data *test;
6700 
6701 	lockdep_assert_held(&event_mutex);
6702 
6703 	list_for_each_entry(test, &file->triggers, list) {
6704 		if (test->cmd_ops->trigger_type == ETT_EVENT_HIST) {
6705 			hist_data = test->private_data;
6706 			if (check_var_refs(hist_data))
6707 				return true;
6708 		}
6709 	}
6710 
6711 	return false;
6712 }
6713 
hist_unreg_all(struct trace_event_file * file)6714 static void hist_unreg_all(struct trace_event_file *file)
6715 {
6716 	struct event_trigger_data *test, *n;
6717 	struct hist_trigger_data *hist_data;
6718 	struct synth_event *se;
6719 	const char *se_name;
6720 
6721 	lockdep_assert_held(&event_mutex);
6722 
6723 	if (hist_file_check_refs(file))
6724 		return;
6725 
6726 	list_for_each_entry_safe(test, n, &file->triggers, list) {
6727 		if (test->cmd_ops->trigger_type == ETT_EVENT_HIST) {
6728 			hist_data = test->private_data;
6729 			list_del_rcu(&test->list);
6730 			trace_event_trigger_enable_disable(file, 0);
6731 
6732 			se_name = trace_event_name(file->event_call);
6733 			se = find_synth_event(se_name);
6734 			if (se)
6735 				se->ref--;
6736 
6737 			update_cond_flag(file);
6738 			if (hist_data->enable_timestamps)
6739 				tracing_set_filter_buffering(file->tr, false);
6740 			if (test->ops->free)
6741 				test->ops->free(test);
6742 		}
6743 	}
6744 }
6745 
event_hist_trigger_parse(struct event_command * cmd_ops,struct trace_event_file * file,char * glob,char * cmd,char * param_and_filter)6746 static int event_hist_trigger_parse(struct event_command *cmd_ops,
6747 				    struct trace_event_file *file,
6748 				    char *glob, char *cmd,
6749 				    char *param_and_filter)
6750 {
6751 	unsigned int hist_trigger_bits = TRACING_MAP_BITS_DEFAULT;
6752 	struct event_trigger_data *trigger_data;
6753 	struct hist_trigger_attrs *attrs;
6754 	struct hist_trigger_data *hist_data;
6755 	char *param, *filter, *p, *start;
6756 	struct synth_event *se;
6757 	const char *se_name;
6758 	bool remove;
6759 	int ret = 0;
6760 
6761 	lockdep_assert_held(&event_mutex);
6762 
6763 	if (WARN_ON(!glob))
6764 		return -EINVAL;
6765 
6766 	if (glob[0]) {
6767 		hist_err_clear();
6768 		last_cmd_set(file, param_and_filter);
6769 	}
6770 
6771 	remove = event_trigger_check_remove(glob);
6772 
6773 	if (event_trigger_empty_param(param_and_filter))
6774 		return -EINVAL;
6775 
6776 	/*
6777 	 * separate the trigger from the filter (k:v [if filter])
6778 	 * allowing for whitespace in the trigger
6779 	 */
6780 	p = param = param_and_filter;
6781 	do {
6782 		p = strstr(p, "if");
6783 		if (!p)
6784 			break;
6785 		if (p == param_and_filter)
6786 			return -EINVAL;
6787 		if (*(p - 1) != ' ' && *(p - 1) != '\t') {
6788 			p++;
6789 			continue;
6790 		}
6791 		if (p >= param_and_filter + strlen(param_and_filter) - (sizeof("if") - 1) - 1)
6792 			return -EINVAL;
6793 		if (*(p + sizeof("if") - 1) != ' ' && *(p + sizeof("if") - 1) != '\t') {
6794 			p++;
6795 			continue;
6796 		}
6797 		break;
6798 	} while (1);
6799 
6800 	if (!p)
6801 		filter = NULL;
6802 	else {
6803 		*(p - 1) = '\0';
6804 		filter = strstrip(p);
6805 		param = strstrip(param);
6806 	}
6807 
6808 	/*
6809 	 * To simplify arithmetic expression parsing, replace occurrences of
6810 	 * '.sym-offset' modifier with '.symXoffset'
6811 	 */
6812 	start = strstr(param, ".sym-offset");
6813 	while (start) {
6814 		*(start + 4) = 'X';
6815 		start = strstr(start + 11, ".sym-offset");
6816 	}
6817 
6818 	attrs = parse_hist_trigger_attrs(file->tr, param);
6819 	if (IS_ERR(attrs))
6820 		return PTR_ERR(attrs);
6821 
6822 	if (attrs->map_bits)
6823 		hist_trigger_bits = attrs->map_bits;
6824 
6825 	hist_data = create_hist_data(hist_trigger_bits, attrs, file, remove);
6826 	if (IS_ERR(hist_data)) {
6827 		destroy_hist_trigger_attrs(attrs);
6828 		return PTR_ERR(hist_data);
6829 	}
6830 
6831 	trigger_data = trigger_data_alloc(cmd_ops, cmd, param, hist_data);
6832 	if (!trigger_data) {
6833 		ret = -ENOMEM;
6834 		goto out_free;
6835 	}
6836 
6837 	ret = event_trigger_set_filter(cmd_ops, file, filter, trigger_data);
6838 	if (ret < 0)
6839 		goto out_free;
6840 
6841 	if (remove) {
6842 		if (!have_hist_trigger_match(trigger_data, file))
6843 			goto out_free;
6844 
6845 		if (hist_trigger_check_refs(trigger_data, file)) {
6846 			ret = -EBUSY;
6847 			goto out_free;
6848 		}
6849 
6850 		event_trigger_unregister(cmd_ops, file, glob+1, trigger_data);
6851 		se_name = trace_event_name(file->event_call);
6852 		se = find_synth_event(se_name);
6853 		if (se)
6854 			se->ref--;
6855 		ret = 0;
6856 		goto out_free;
6857 	}
6858 
6859 	if (existing_hist_update_only(glob, trigger_data, file))
6860 		goto out_free;
6861 
6862 	if (!get_named_trigger_data(trigger_data)) {
6863 
6864 		ret = create_actions(hist_data);
6865 		if (ret)
6866 			goto out_free;
6867 
6868 		if (has_hist_vars(hist_data) || hist_data->n_var_refs) {
6869 			ret = save_hist_vars(hist_data);
6870 			if (ret)
6871 				goto out_free;
6872 		}
6873 
6874 		ret = tracing_map_init(hist_data->map);
6875 		if (ret)
6876 			goto out_free;
6877 	}
6878 
6879 	ret = event_trigger_register(cmd_ops, file, glob, trigger_data);
6880 	if (ret < 0)
6881 		goto out_free;
6882 
6883 	ret = hist_trigger_enable(trigger_data, file);
6884 	if (ret)
6885 		goto out_unreg;
6886 
6887 	se_name = trace_event_name(file->event_call);
6888 	se = find_synth_event(se_name);
6889 	if (se)
6890 		se->ref++;
6891  out:
6892 	if (ret == 0 && glob[0])
6893 		hist_err_clear();
6894 
6895 	return ret;
6896  out_unreg:
6897 	event_trigger_unregister(cmd_ops, file, glob+1, trigger_data);
6898  out_free:
6899 	event_trigger_reset_filter(cmd_ops, trigger_data);
6900 
6901 	remove_hist_vars(hist_data);
6902 
6903 	kfree(trigger_data);
6904 
6905 	destroy_hist_data(hist_data);
6906 	goto out;
6907 }
6908 
6909 static struct event_command trigger_hist_cmd = {
6910 	.name			= "hist",
6911 	.trigger_type		= ETT_EVENT_HIST,
6912 	.flags			= EVENT_CMD_FL_NEEDS_REC,
6913 	.parse			= event_hist_trigger_parse,
6914 	.reg			= hist_register_trigger,
6915 	.unreg			= hist_unregister_trigger,
6916 	.unreg_all		= hist_unreg_all,
6917 	.get_trigger_ops	= event_hist_get_trigger_ops,
6918 	.set_filter		= set_trigger_filter,
6919 };
6920 
register_trigger_hist_cmd(void)6921 __init int register_trigger_hist_cmd(void)
6922 {
6923 	int ret;
6924 
6925 	ret = register_event_command(&trigger_hist_cmd);
6926 	WARN_ON(ret < 0);
6927 
6928 	return ret;
6929 }
6930 
6931 static void
hist_enable_trigger(struct event_trigger_data * data,struct trace_buffer * buffer,void * rec,struct ring_buffer_event * event)6932 hist_enable_trigger(struct event_trigger_data *data,
6933 		    struct trace_buffer *buffer,  void *rec,
6934 		    struct ring_buffer_event *event)
6935 {
6936 	struct enable_trigger_data *enable_data = data->private_data;
6937 	struct event_trigger_data *test;
6938 
6939 	list_for_each_entry_rcu(test, &enable_data->file->triggers, list,
6940 				lockdep_is_held(&event_mutex)) {
6941 		if (test->cmd_ops->trigger_type == ETT_EVENT_HIST) {
6942 			if (enable_data->enable)
6943 				test->paused = false;
6944 			else
6945 				test->paused = true;
6946 		}
6947 	}
6948 }
6949 
6950 static void
hist_enable_count_trigger(struct event_trigger_data * data,struct trace_buffer * buffer,void * rec,struct ring_buffer_event * event)6951 hist_enable_count_trigger(struct event_trigger_data *data,
6952 			  struct trace_buffer *buffer,  void *rec,
6953 			  struct ring_buffer_event *event)
6954 {
6955 	if (!data->count)
6956 		return;
6957 
6958 	if (data->count != -1)
6959 		(data->count)--;
6960 
6961 	hist_enable_trigger(data, buffer, rec, event);
6962 }
6963 
6964 static const struct event_trigger_ops hist_enable_trigger_ops = {
6965 	.trigger		= hist_enable_trigger,
6966 	.print			= event_enable_trigger_print,
6967 	.init			= event_trigger_init,
6968 	.free			= event_enable_trigger_free,
6969 };
6970 
6971 static const struct event_trigger_ops hist_enable_count_trigger_ops = {
6972 	.trigger		= hist_enable_count_trigger,
6973 	.print			= event_enable_trigger_print,
6974 	.init			= event_trigger_init,
6975 	.free			= event_enable_trigger_free,
6976 };
6977 
6978 static const struct event_trigger_ops hist_disable_trigger_ops = {
6979 	.trigger		= hist_enable_trigger,
6980 	.print			= event_enable_trigger_print,
6981 	.init			= event_trigger_init,
6982 	.free			= event_enable_trigger_free,
6983 };
6984 
6985 static const struct event_trigger_ops hist_disable_count_trigger_ops = {
6986 	.trigger		= hist_enable_count_trigger,
6987 	.print			= event_enable_trigger_print,
6988 	.init			= event_trigger_init,
6989 	.free			= event_enable_trigger_free,
6990 };
6991 
6992 static const struct event_trigger_ops *
hist_enable_get_trigger_ops(char * cmd,char * param)6993 hist_enable_get_trigger_ops(char *cmd, char *param)
6994 {
6995 	const struct event_trigger_ops *ops;
6996 	bool enable;
6997 
6998 	enable = (strcmp(cmd, ENABLE_HIST_STR) == 0);
6999 
7000 	if (enable)
7001 		ops = param ? &hist_enable_count_trigger_ops :
7002 			&hist_enable_trigger_ops;
7003 	else
7004 		ops = param ? &hist_disable_count_trigger_ops :
7005 			&hist_disable_trigger_ops;
7006 
7007 	return ops;
7008 }
7009 
hist_enable_unreg_all(struct trace_event_file * file)7010 static void hist_enable_unreg_all(struct trace_event_file *file)
7011 {
7012 	struct event_trigger_data *test, *n;
7013 
7014 	list_for_each_entry_safe(test, n, &file->triggers, list) {
7015 		if (test->cmd_ops->trigger_type == ETT_HIST_ENABLE) {
7016 			list_del_rcu(&test->list);
7017 			update_cond_flag(file);
7018 			trace_event_trigger_enable_disable(file, 0);
7019 			if (test->ops->free)
7020 				test->ops->free(test);
7021 		}
7022 	}
7023 }
7024 
7025 static struct event_command trigger_hist_enable_cmd = {
7026 	.name			= ENABLE_HIST_STR,
7027 	.trigger_type		= ETT_HIST_ENABLE,
7028 	.parse			= event_enable_trigger_parse,
7029 	.reg			= event_enable_register_trigger,
7030 	.unreg			= event_enable_unregister_trigger,
7031 	.unreg_all		= hist_enable_unreg_all,
7032 	.get_trigger_ops	= hist_enable_get_trigger_ops,
7033 	.set_filter		= set_trigger_filter,
7034 };
7035 
7036 static struct event_command trigger_hist_disable_cmd = {
7037 	.name			= DISABLE_HIST_STR,
7038 	.trigger_type		= ETT_HIST_ENABLE,
7039 	.parse			= event_enable_trigger_parse,
7040 	.reg			= event_enable_register_trigger,
7041 	.unreg			= event_enable_unregister_trigger,
7042 	.unreg_all		= hist_enable_unreg_all,
7043 	.get_trigger_ops	= hist_enable_get_trigger_ops,
7044 	.set_filter		= set_trigger_filter,
7045 };
7046 
unregister_trigger_hist_enable_disable_cmds(void)7047 static __init void unregister_trigger_hist_enable_disable_cmds(void)
7048 {
7049 	unregister_event_command(&trigger_hist_enable_cmd);
7050 	unregister_event_command(&trigger_hist_disable_cmd);
7051 }
7052 
register_trigger_hist_enable_disable_cmds(void)7053 __init int register_trigger_hist_enable_disable_cmds(void)
7054 {
7055 	int ret;
7056 
7057 	ret = register_event_command(&trigger_hist_enable_cmd);
7058 	if (WARN_ON(ret < 0))
7059 		return ret;
7060 	ret = register_event_command(&trigger_hist_disable_cmd);
7061 	if (WARN_ON(ret < 0))
7062 		unregister_trigger_hist_enable_disable_cmds();
7063 
7064 	return ret;
7065 }
7066