xref: /linux/net/sched/act_police.c (revision 8f7e001e0325de63a42f23342ac3b8139150c5cf)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  * net/sched/act_police.c	Input police filter
4  *
5  * Authors:	Alexey Kuznetsov, <kuznet@ms2.inr.ac.ru>
6  * 		J Hadi Salim (action changes)
7  */
8 
9 #include <linux/module.h>
10 #include <linux/types.h>
11 #include <linux/kernel.h>
12 #include <linux/string.h>
13 #include <linux/errno.h>
14 #include <linux/skbuff.h>
15 #include <linux/rtnetlink.h>
16 #include <linux/init.h>
17 #include <linux/slab.h>
18 #include <net/act_api.h>
19 #include <net/netlink.h>
20 #include <net/pkt_cls.h>
21 #include <net/tc_act/tc_police.h>
22 
23 /* Each policer is serialized by its individual spinlock */
24 
25 static unsigned int police_net_id;
26 static struct tc_action_ops act_police_ops;
27 
28 static int tcf_police_walker(struct net *net, struct sk_buff *skb,
29 				 struct netlink_callback *cb, int type,
30 				 const struct tc_action_ops *ops,
31 				 struct netlink_ext_ack *extack)
32 {
33 	struct tc_action_net *tn = net_generic(net, police_net_id);
34 
35 	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
36 }
37 
38 static const struct nla_policy police_policy[TCA_POLICE_MAX + 1] = {
39 	[TCA_POLICE_RATE]	= { .len = TC_RTAB_SIZE },
40 	[TCA_POLICE_PEAKRATE]	= { .len = TC_RTAB_SIZE },
41 	[TCA_POLICE_AVRATE]	= { .type = NLA_U32 },
42 	[TCA_POLICE_RESULT]	= { .type = NLA_U32 },
43 	[TCA_POLICE_RATE64]     = { .type = NLA_U64 },
44 	[TCA_POLICE_PEAKRATE64] = { .type = NLA_U64 },
45 	[TCA_POLICE_PKTRATE64]  = { .type = NLA_U64, .min = 1 },
46 	[TCA_POLICE_PKTBURST64] = { .type = NLA_U64, .min = 1 },
47 };
48 
49 static int tcf_police_init(struct net *net, struct nlattr *nla,
50 			       struct nlattr *est, struct tc_action **a,
51 			       struct tcf_proto *tp, u32 flags,
52 			       struct netlink_ext_ack *extack)
53 {
54 	int ret = 0, tcfp_result = TC_ACT_OK, err, size;
55 	bool bind = flags & TCA_ACT_FLAGS_BIND;
56 	struct nlattr *tb[TCA_POLICE_MAX + 1];
57 	struct tcf_chain *goto_ch = NULL;
58 	struct tc_police *parm;
59 	struct tcf_police *police;
60 	struct qdisc_rate_table *R_tab = NULL, *P_tab = NULL;
61 	struct tc_action_net *tn = net_generic(net, police_net_id);
62 	struct tcf_police_params *new;
63 	bool exists = false;
64 	u32 index;
65 	u64 rate64, prate64;
66 	u64 pps, ppsburst;
67 
68 	if (nla == NULL)
69 		return -EINVAL;
70 
71 	err = nla_parse_nested_deprecated(tb, TCA_POLICE_MAX, nla,
72 					  police_policy, NULL);
73 	if (err < 0)
74 		return err;
75 
76 	if (tb[TCA_POLICE_TBF] == NULL)
77 		return -EINVAL;
78 	size = nla_len(tb[TCA_POLICE_TBF]);
79 	if (size != sizeof(*parm) && size != sizeof(struct tc_police_compat))
80 		return -EINVAL;
81 
82 	parm = nla_data(tb[TCA_POLICE_TBF]);
83 	index = parm->index;
84 	err = tcf_idr_check_alloc(tn, &index, a, bind);
85 	if (err < 0)
86 		return err;
87 	exists = err;
88 	if (exists && bind)
89 		return 0;
90 
91 	if (!exists) {
92 		ret = tcf_idr_create(tn, index, NULL, a,
93 				     &act_police_ops, bind, true, flags);
94 		if (ret) {
95 			tcf_idr_cleanup(tn, index);
96 			return ret;
97 		}
98 		ret = ACT_P_CREATED;
99 		spin_lock_init(&(to_police(*a)->tcfp_lock));
100 	} else if (!(flags & TCA_ACT_FLAGS_REPLACE)) {
101 		tcf_idr_release(*a, bind);
102 		return -EEXIST;
103 	}
104 	err = tcf_action_check_ctrlact(parm->action, tp, &goto_ch, extack);
105 	if (err < 0)
106 		goto release_idr;
107 
108 	police = to_police(*a);
109 	if (parm->rate.rate) {
110 		err = -ENOMEM;
111 		R_tab = qdisc_get_rtab(&parm->rate, tb[TCA_POLICE_RATE], NULL);
112 		if (R_tab == NULL)
113 			goto failure;
114 
115 		if (parm->peakrate.rate) {
116 			P_tab = qdisc_get_rtab(&parm->peakrate,
117 					       tb[TCA_POLICE_PEAKRATE], NULL);
118 			if (P_tab == NULL)
119 				goto failure;
120 		}
121 	}
122 
123 	if (est) {
124 		err = gen_replace_estimator(&police->tcf_bstats,
125 					    police->common.cpu_bstats,
126 					    &police->tcf_rate_est,
127 					    &police->tcf_lock,
128 					    false, est);
129 		if (err)
130 			goto failure;
131 	} else if (tb[TCA_POLICE_AVRATE] &&
132 		   (ret == ACT_P_CREATED ||
133 		    !gen_estimator_active(&police->tcf_rate_est))) {
134 		err = -EINVAL;
135 		goto failure;
136 	}
137 
138 	if (tb[TCA_POLICE_RESULT]) {
139 		tcfp_result = nla_get_u32(tb[TCA_POLICE_RESULT]);
140 		if (TC_ACT_EXT_CMP(tcfp_result, TC_ACT_GOTO_CHAIN)) {
141 			NL_SET_ERR_MSG(extack,
142 				       "goto chain not allowed on fallback");
143 			err = -EINVAL;
144 			goto failure;
145 		}
146 	}
147 
148 	if ((tb[TCA_POLICE_PKTRATE64] && !tb[TCA_POLICE_PKTBURST64]) ||
149 	    (!tb[TCA_POLICE_PKTRATE64] && tb[TCA_POLICE_PKTBURST64])) {
150 		NL_SET_ERR_MSG(extack,
151 			       "Both or neither packet-per-second burst and rate must be provided");
152 		err = -EINVAL;
153 		goto failure;
154 	}
155 
156 	if (tb[TCA_POLICE_PKTRATE64] && R_tab) {
157 		NL_SET_ERR_MSG(extack,
158 			       "packet-per-second and byte-per-second rate limits not allowed in same action");
159 		err = -EINVAL;
160 		goto failure;
161 	}
162 
163 	new = kzalloc(sizeof(*new), GFP_KERNEL);
164 	if (unlikely(!new)) {
165 		err = -ENOMEM;
166 		goto failure;
167 	}
168 
169 	/* No failure allowed after this point */
170 	new->tcfp_result = tcfp_result;
171 	new->tcfp_mtu = parm->mtu;
172 	if (!new->tcfp_mtu) {
173 		new->tcfp_mtu = ~0;
174 		if (R_tab)
175 			new->tcfp_mtu = 255 << R_tab->rate.cell_log;
176 	}
177 	if (R_tab) {
178 		new->rate_present = true;
179 		rate64 = tb[TCA_POLICE_RATE64] ?
180 			 nla_get_u64(tb[TCA_POLICE_RATE64]) : 0;
181 		psched_ratecfg_precompute(&new->rate, &R_tab->rate, rate64);
182 		qdisc_put_rtab(R_tab);
183 	} else {
184 		new->rate_present = false;
185 	}
186 	if (P_tab) {
187 		new->peak_present = true;
188 		prate64 = tb[TCA_POLICE_PEAKRATE64] ?
189 			  nla_get_u64(tb[TCA_POLICE_PEAKRATE64]) : 0;
190 		psched_ratecfg_precompute(&new->peak, &P_tab->rate, prate64);
191 		qdisc_put_rtab(P_tab);
192 	} else {
193 		new->peak_present = false;
194 	}
195 
196 	new->tcfp_burst = PSCHED_TICKS2NS(parm->burst);
197 	if (new->peak_present)
198 		new->tcfp_mtu_ptoks = (s64)psched_l2t_ns(&new->peak,
199 							 new->tcfp_mtu);
200 
201 	if (tb[TCA_POLICE_AVRATE])
202 		new->tcfp_ewma_rate = nla_get_u32(tb[TCA_POLICE_AVRATE]);
203 
204 	if (tb[TCA_POLICE_PKTRATE64]) {
205 		pps = nla_get_u64(tb[TCA_POLICE_PKTRATE64]);
206 		ppsburst = nla_get_u64(tb[TCA_POLICE_PKTBURST64]);
207 		new->pps_present = true;
208 		new->tcfp_pkt_burst = PSCHED_TICKS2NS(ppsburst);
209 		psched_ppscfg_precompute(&new->ppsrate, pps);
210 	}
211 
212 	spin_lock_bh(&police->tcf_lock);
213 	spin_lock_bh(&police->tcfp_lock);
214 	police->tcfp_t_c = ktime_get_ns();
215 	police->tcfp_toks = new->tcfp_burst;
216 	if (new->peak_present)
217 		police->tcfp_ptoks = new->tcfp_mtu_ptoks;
218 	spin_unlock_bh(&police->tcfp_lock);
219 	goto_ch = tcf_action_set_ctrlact(*a, parm->action, goto_ch);
220 	new = rcu_replace_pointer(police->params,
221 				  new,
222 				  lockdep_is_held(&police->tcf_lock));
223 	spin_unlock_bh(&police->tcf_lock);
224 
225 	if (goto_ch)
226 		tcf_chain_put_by_act(goto_ch);
227 	if (new)
228 		kfree_rcu(new, rcu);
229 
230 	return ret;
231 
232 failure:
233 	qdisc_put_rtab(P_tab);
234 	qdisc_put_rtab(R_tab);
235 	if (goto_ch)
236 		tcf_chain_put_by_act(goto_ch);
237 release_idr:
238 	tcf_idr_release(*a, bind);
239 	return err;
240 }
241 
242 static bool tcf_police_mtu_check(struct sk_buff *skb, u32 limit)
243 {
244 	u32 len;
245 
246 	if (skb_is_gso(skb))
247 		return skb_gso_validate_mac_len(skb, limit);
248 
249 	len = qdisc_pkt_len(skb);
250 	if (skb_at_tc_ingress(skb))
251 		len += skb->mac_len;
252 
253 	return len <= limit;
254 }
255 
256 static int tcf_police_act(struct sk_buff *skb, const struct tc_action *a,
257 			  struct tcf_result *res)
258 {
259 	struct tcf_police *police = to_police(a);
260 	s64 now, toks, ppstoks = 0, ptoks = 0;
261 	struct tcf_police_params *p;
262 	int ret;
263 
264 	tcf_lastuse_update(&police->tcf_tm);
265 	bstats_update(this_cpu_ptr(police->common.cpu_bstats), skb);
266 
267 	ret = READ_ONCE(police->tcf_action);
268 	p = rcu_dereference_bh(police->params);
269 
270 	if (p->tcfp_ewma_rate) {
271 		struct gnet_stats_rate_est64 sample;
272 
273 		if (!gen_estimator_read(&police->tcf_rate_est, &sample) ||
274 		    sample.bps >= p->tcfp_ewma_rate)
275 			goto inc_overlimits;
276 	}
277 
278 	if (tcf_police_mtu_check(skb, p->tcfp_mtu)) {
279 		if (!p->rate_present && !p->pps_present) {
280 			ret = p->tcfp_result;
281 			goto end;
282 		}
283 
284 		now = ktime_get_ns();
285 		spin_lock_bh(&police->tcfp_lock);
286 		toks = min_t(s64, now - police->tcfp_t_c, p->tcfp_burst);
287 		if (p->peak_present) {
288 			ptoks = toks + police->tcfp_ptoks;
289 			if (ptoks > p->tcfp_mtu_ptoks)
290 				ptoks = p->tcfp_mtu_ptoks;
291 			ptoks -= (s64)psched_l2t_ns(&p->peak,
292 						    qdisc_pkt_len(skb));
293 		}
294 		if (p->rate_present) {
295 			toks += police->tcfp_toks;
296 			if (toks > p->tcfp_burst)
297 				toks = p->tcfp_burst;
298 			toks -= (s64)psched_l2t_ns(&p->rate, qdisc_pkt_len(skb));
299 		} else if (p->pps_present) {
300 			ppstoks = min_t(s64, now - police->tcfp_t_c, p->tcfp_pkt_burst);
301 			ppstoks += police->tcfp_pkttoks;
302 			if (ppstoks > p->tcfp_pkt_burst)
303 				ppstoks = p->tcfp_pkt_burst;
304 			ppstoks -= (s64)psched_pkt2t_ns(&p->ppsrate, 1);
305 		}
306 		if ((toks | ptoks | ppstoks) >= 0) {
307 			police->tcfp_t_c = now;
308 			police->tcfp_toks = toks;
309 			police->tcfp_ptoks = ptoks;
310 			police->tcfp_pkttoks = ppstoks;
311 			spin_unlock_bh(&police->tcfp_lock);
312 			ret = p->tcfp_result;
313 			goto inc_drops;
314 		}
315 		spin_unlock_bh(&police->tcfp_lock);
316 	}
317 
318 inc_overlimits:
319 	qstats_overlimit_inc(this_cpu_ptr(police->common.cpu_qstats));
320 inc_drops:
321 	if (ret == TC_ACT_SHOT)
322 		qstats_drop_inc(this_cpu_ptr(police->common.cpu_qstats));
323 end:
324 	return ret;
325 }
326 
327 static void tcf_police_cleanup(struct tc_action *a)
328 {
329 	struct tcf_police *police = to_police(a);
330 	struct tcf_police_params *p;
331 
332 	p = rcu_dereference_protected(police->params, 1);
333 	if (p)
334 		kfree_rcu(p, rcu);
335 }
336 
337 static void tcf_police_stats_update(struct tc_action *a,
338 				    u64 bytes, u64 packets, u64 drops,
339 				    u64 lastuse, bool hw)
340 {
341 	struct tcf_police *police = to_police(a);
342 	struct tcf_t *tm = &police->tcf_tm;
343 
344 	tcf_action_update_stats(a, bytes, packets, drops, hw);
345 	tm->lastuse = max_t(u64, tm->lastuse, lastuse);
346 }
347 
348 static int tcf_police_dump(struct sk_buff *skb, struct tc_action *a,
349 			       int bind, int ref)
350 {
351 	unsigned char *b = skb_tail_pointer(skb);
352 	struct tcf_police *police = to_police(a);
353 	struct tcf_police_params *p;
354 	struct tc_police opt = {
355 		.index = police->tcf_index,
356 		.refcnt = refcount_read(&police->tcf_refcnt) - ref,
357 		.bindcnt = atomic_read(&police->tcf_bindcnt) - bind,
358 	};
359 	struct tcf_t t;
360 
361 	spin_lock_bh(&police->tcf_lock);
362 	opt.action = police->tcf_action;
363 	p = rcu_dereference_protected(police->params,
364 				      lockdep_is_held(&police->tcf_lock));
365 	opt.mtu = p->tcfp_mtu;
366 	opt.burst = PSCHED_NS2TICKS(p->tcfp_burst);
367 	if (p->rate_present) {
368 		psched_ratecfg_getrate(&opt.rate, &p->rate);
369 		if ((police->params->rate.rate_bytes_ps >= (1ULL << 32)) &&
370 		    nla_put_u64_64bit(skb, TCA_POLICE_RATE64,
371 				      police->params->rate.rate_bytes_ps,
372 				      TCA_POLICE_PAD))
373 			goto nla_put_failure;
374 	}
375 	if (p->peak_present) {
376 		psched_ratecfg_getrate(&opt.peakrate, &p->peak);
377 		if ((police->params->peak.rate_bytes_ps >= (1ULL << 32)) &&
378 		    nla_put_u64_64bit(skb, TCA_POLICE_PEAKRATE64,
379 				      police->params->peak.rate_bytes_ps,
380 				      TCA_POLICE_PAD))
381 			goto nla_put_failure;
382 	}
383 	if (p->pps_present) {
384 		if (nla_put_u64_64bit(skb, TCA_POLICE_PKTRATE64,
385 				      police->params->ppsrate.rate_pkts_ps,
386 				      TCA_POLICE_PAD))
387 			goto nla_put_failure;
388 		if (nla_put_u64_64bit(skb, TCA_POLICE_PKTBURST64,
389 				      PSCHED_NS2TICKS(p->tcfp_pkt_burst),
390 				      TCA_POLICE_PAD))
391 			goto nla_put_failure;
392 	}
393 	if (nla_put(skb, TCA_POLICE_TBF, sizeof(opt), &opt))
394 		goto nla_put_failure;
395 	if (p->tcfp_result &&
396 	    nla_put_u32(skb, TCA_POLICE_RESULT, p->tcfp_result))
397 		goto nla_put_failure;
398 	if (p->tcfp_ewma_rate &&
399 	    nla_put_u32(skb, TCA_POLICE_AVRATE, p->tcfp_ewma_rate))
400 		goto nla_put_failure;
401 
402 	tcf_tm_dump(&t, &police->tcf_tm);
403 	if (nla_put_64bit(skb, TCA_POLICE_TM, sizeof(t), &t, TCA_POLICE_PAD))
404 		goto nla_put_failure;
405 	spin_unlock_bh(&police->tcf_lock);
406 
407 	return skb->len;
408 
409 nla_put_failure:
410 	spin_unlock_bh(&police->tcf_lock);
411 	nlmsg_trim(skb, b);
412 	return -1;
413 }
414 
415 static int tcf_police_search(struct net *net, struct tc_action **a, u32 index)
416 {
417 	struct tc_action_net *tn = net_generic(net, police_net_id);
418 
419 	return tcf_idr_search(tn, a, index);
420 }
421 
422 static int tcf_police_act_to_flow_act(int tc_act, u32 *extval,
423 				      struct netlink_ext_ack *extack)
424 {
425 	int act_id = -EOPNOTSUPP;
426 
427 	if (!TC_ACT_EXT_OPCODE(tc_act)) {
428 		if (tc_act == TC_ACT_OK)
429 			act_id = FLOW_ACTION_ACCEPT;
430 		else if (tc_act ==  TC_ACT_SHOT)
431 			act_id = FLOW_ACTION_DROP;
432 		else if (tc_act == TC_ACT_PIPE)
433 			act_id = FLOW_ACTION_PIPE;
434 		else if (tc_act == TC_ACT_RECLASSIFY)
435 			NL_SET_ERR_MSG_MOD(extack, "Offload not supported when conform/exceed action is \"reclassify\"");
436 		else
437 			NL_SET_ERR_MSG_MOD(extack, "Unsupported conform/exceed action offload");
438 	} else if (TC_ACT_EXT_CMP(tc_act, TC_ACT_GOTO_CHAIN)) {
439 		act_id = FLOW_ACTION_GOTO;
440 		*extval = tc_act & TC_ACT_EXT_VAL_MASK;
441 	} else if (TC_ACT_EXT_CMP(tc_act, TC_ACT_JUMP)) {
442 		act_id = FLOW_ACTION_JUMP;
443 		*extval = tc_act & TC_ACT_EXT_VAL_MASK;
444 	} else if (tc_act == TC_ACT_UNSPEC) {
445 		act_id = FLOW_ACTION_CONTINUE;
446 	} else {
447 		NL_SET_ERR_MSG_MOD(extack, "Unsupported conform/exceed action offload");
448 	}
449 
450 	return act_id;
451 }
452 
453 static int tcf_police_offload_act_setup(struct tc_action *act, void *entry_data,
454 					u32 *index_inc, bool bind,
455 					struct netlink_ext_ack *extack)
456 {
457 	if (bind) {
458 		struct flow_action_entry *entry = entry_data;
459 		struct tcf_police *police = to_police(act);
460 		struct tcf_police_params *p;
461 		int act_id;
462 
463 		p = rcu_dereference_protected(police->params,
464 					      lockdep_is_held(&police->tcf_lock));
465 
466 		entry->id = FLOW_ACTION_POLICE;
467 		entry->police.burst = tcf_police_burst(act);
468 		entry->police.rate_bytes_ps =
469 			tcf_police_rate_bytes_ps(act);
470 		entry->police.peakrate_bytes_ps = tcf_police_peakrate_bytes_ps(act);
471 		entry->police.avrate = tcf_police_tcfp_ewma_rate(act);
472 		entry->police.overhead = tcf_police_rate_overhead(act);
473 		entry->police.burst_pkt = tcf_police_burst_pkt(act);
474 		entry->police.rate_pkt_ps =
475 			tcf_police_rate_pkt_ps(act);
476 		entry->police.mtu = tcf_police_tcfp_mtu(act);
477 
478 		act_id = tcf_police_act_to_flow_act(police->tcf_action,
479 						    &entry->police.exceed.extval,
480 						    extack);
481 		if (act_id < 0)
482 			return act_id;
483 
484 		entry->police.exceed.act_id = act_id;
485 
486 		act_id = tcf_police_act_to_flow_act(p->tcfp_result,
487 						    &entry->police.notexceed.extval,
488 						    extack);
489 		if (act_id < 0)
490 			return act_id;
491 
492 		entry->police.notexceed.act_id = act_id;
493 
494 		*index_inc = 1;
495 	} else {
496 		struct flow_offload_action *fl_action = entry_data;
497 
498 		fl_action->id = FLOW_ACTION_POLICE;
499 	}
500 
501 	return 0;
502 }
503 
504 MODULE_AUTHOR("Alexey Kuznetsov");
505 MODULE_DESCRIPTION("Policing actions");
506 MODULE_LICENSE("GPL");
507 
508 static struct tc_action_ops act_police_ops = {
509 	.kind		=	"police",
510 	.id		=	TCA_ID_POLICE,
511 	.owner		=	THIS_MODULE,
512 	.stats_update	=	tcf_police_stats_update,
513 	.act		=	tcf_police_act,
514 	.dump		=	tcf_police_dump,
515 	.init		=	tcf_police_init,
516 	.walk		=	tcf_police_walker,
517 	.lookup		=	tcf_police_search,
518 	.cleanup	=	tcf_police_cleanup,
519 	.offload_act_setup =	tcf_police_offload_act_setup,
520 	.size		=	sizeof(struct tcf_police),
521 };
522 
523 static __net_init int police_init_net(struct net *net)
524 {
525 	struct tc_action_net *tn = net_generic(net, police_net_id);
526 
527 	return tc_action_net_init(net, tn, &act_police_ops);
528 }
529 
530 static void __net_exit police_exit_net(struct list_head *net_list)
531 {
532 	tc_action_net_exit(net_list, police_net_id);
533 }
534 
535 static struct pernet_operations police_net_ops = {
536 	.init = police_init_net,
537 	.exit_batch = police_exit_net,
538 	.id   = &police_net_id,
539 	.size = sizeof(struct tc_action_net),
540 };
541 
542 static int __init police_init_module(void)
543 {
544 	return tcf_register_action(&act_police_ops, &police_net_ops);
545 }
546 
547 static void __exit police_cleanup_module(void)
548 {
549 	tcf_unregister_action(&act_police_ops, &police_net_ops);
550 }
551 
552 module_init(police_init_module);
553 module_exit(police_cleanup_module);
554