xref: /linux/drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_matchall.c (revision 4b132aacb0768ac1e652cf517097ea6f237214b9)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /* Copyright (C) 2019 Chelsio Communications.  All rights reserved. */
3 
4 #include "cxgb4.h"
5 #include "cxgb4_tc_matchall.h"
6 #include "sched.h"
7 #include "cxgb4_uld.h"
8 #include "cxgb4_filter.h"
9 #include "cxgb4_tc_flower.h"
10 
11 static int cxgb4_policer_validate(const struct flow_action *action,
12 				  const struct flow_action_entry *act,
13 				  struct netlink_ext_ack *extack)
14 {
15 	if (act->police.exceed.act_id != FLOW_ACTION_DROP) {
16 		NL_SET_ERR_MSG_MOD(extack,
17 				   "Offload not supported when exceed action is not drop");
18 		return -EOPNOTSUPP;
19 	}
20 
21 	if (act->police.notexceed.act_id != FLOW_ACTION_PIPE &&
22 	    act->police.notexceed.act_id != FLOW_ACTION_ACCEPT) {
23 		NL_SET_ERR_MSG_MOD(extack,
24 				   "Offload not supported when conform action is not pipe or ok");
25 		return -EOPNOTSUPP;
26 	}
27 
28 	if (act->police.notexceed.act_id == FLOW_ACTION_ACCEPT &&
29 	    !flow_action_is_last_entry(action, act)) {
30 		NL_SET_ERR_MSG_MOD(extack,
31 				   "Offload not supported when conform action is ok, but action is not last");
32 		return -EOPNOTSUPP;
33 	}
34 
35 	if (act->police.peakrate_bytes_ps ||
36 	    act->police.avrate || act->police.overhead) {
37 		NL_SET_ERR_MSG_MOD(extack,
38 				   "Offload not supported when peakrate/avrate/overhead is configured");
39 		return -EOPNOTSUPP;
40 	}
41 
42 	if (act->police.rate_pkt_ps) {
43 		NL_SET_ERR_MSG_MOD(extack,
44 				   "QoS offload not support packets per second");
45 		return -EOPNOTSUPP;
46 	}
47 
48 	return 0;
49 }
50 
51 static int cxgb4_matchall_egress_validate(struct net_device *dev,
52 					  struct tc_cls_matchall_offload *cls)
53 {
54 	struct netlink_ext_ack *extack = cls->common.extack;
55 	struct flow_action *actions = &cls->rule->action;
56 	struct port_info *pi = netdev2pinfo(dev);
57 	struct flow_action_entry *entry;
58 	struct ch_sched_queue qe;
59 	struct sched_class *e;
60 	u64 max_link_rate;
61 	u32 i, speed;
62 	int ret;
63 
64 	if (!flow_action_has_entries(actions)) {
65 		NL_SET_ERR_MSG_MOD(extack,
66 				   "Egress MATCHALL offload needs at least 1 policing action");
67 		return -EINVAL;
68 	} else if (!flow_offload_has_one_action(actions)) {
69 		NL_SET_ERR_MSG_MOD(extack,
70 				   "Egress MATCHALL offload only supports 1 policing action");
71 		return -EINVAL;
72 	} else if (pi->tc_block_shared) {
73 		NL_SET_ERR_MSG_MOD(extack,
74 				   "Egress MATCHALL offload not supported with shared blocks");
75 		return -EINVAL;
76 	}
77 
78 	ret = t4_get_link_params(pi, NULL, &speed, NULL);
79 	if (ret) {
80 		NL_SET_ERR_MSG_MOD(extack,
81 				   "Failed to get max speed supported by the link");
82 		return -EINVAL;
83 	}
84 
85 	/* Convert from Mbps to bps */
86 	max_link_rate = (u64)speed * 1000 * 1000;
87 
88 	flow_action_for_each(i, entry, actions) {
89 		switch (entry->id) {
90 		case FLOW_ACTION_POLICE:
91 			ret = cxgb4_policer_validate(actions, entry, extack);
92 			if (ret)
93 				return ret;
94 
95 			/* Convert bytes per second to bits per second */
96 			if (entry->police.rate_bytes_ps * 8 > max_link_rate) {
97 				NL_SET_ERR_MSG_MOD(extack,
98 						   "Specified policing max rate is larger than underlying link speed");
99 				return -ERANGE;
100 			}
101 			break;
102 		default:
103 			NL_SET_ERR_MSG_MOD(extack,
104 					   "Only policing action supported with Egress MATCHALL offload");
105 			return -EOPNOTSUPP;
106 		}
107 	}
108 
109 	for (i = 0; i < pi->nqsets; i++) {
110 		memset(&qe, 0, sizeof(qe));
111 		qe.queue = i;
112 
113 		e = cxgb4_sched_queue_lookup(dev, &qe);
114 		if (e && e->info.u.params.level != SCHED_CLASS_LEVEL_CH_RL) {
115 			NL_SET_ERR_MSG_MOD(extack,
116 					   "Some queues are already bound to different class");
117 			return -EBUSY;
118 		}
119 	}
120 
121 	return 0;
122 }
123 
124 static int cxgb4_matchall_tc_bind_queues(struct net_device *dev, u32 tc)
125 {
126 	struct port_info *pi = netdev2pinfo(dev);
127 	struct ch_sched_queue qe;
128 	int ret;
129 	u32 i;
130 
131 	for (i = 0; i < pi->nqsets; i++) {
132 		qe.queue = i;
133 		qe.class = tc;
134 		ret = cxgb4_sched_class_bind(dev, &qe, SCHED_QUEUE);
135 		if (ret)
136 			goto out_free;
137 	}
138 
139 	return 0;
140 
141 out_free:
142 	while (i--) {
143 		qe.queue = i;
144 		qe.class = SCHED_CLS_NONE;
145 		cxgb4_sched_class_unbind(dev, &qe, SCHED_QUEUE);
146 	}
147 
148 	return ret;
149 }
150 
151 static void cxgb4_matchall_tc_unbind_queues(struct net_device *dev)
152 {
153 	struct port_info *pi = netdev2pinfo(dev);
154 	struct ch_sched_queue qe;
155 	u32 i;
156 
157 	for (i = 0; i < pi->nqsets; i++) {
158 		qe.queue = i;
159 		qe.class = SCHED_CLS_NONE;
160 		cxgb4_sched_class_unbind(dev, &qe, SCHED_QUEUE);
161 	}
162 }
163 
164 static int cxgb4_matchall_alloc_tc(struct net_device *dev,
165 				   struct tc_cls_matchall_offload *cls)
166 {
167 	struct ch_sched_params p = {
168 		.type = SCHED_CLASS_TYPE_PACKET,
169 		.u.params.level = SCHED_CLASS_LEVEL_CH_RL,
170 		.u.params.mode = SCHED_CLASS_MODE_CLASS,
171 		.u.params.rateunit = SCHED_CLASS_RATEUNIT_BITS,
172 		.u.params.ratemode = SCHED_CLASS_RATEMODE_ABS,
173 		.u.params.class = SCHED_CLS_NONE,
174 		.u.params.minrate = 0,
175 		.u.params.weight = 0,
176 		.u.params.pktsize = dev->mtu,
177 	};
178 	struct netlink_ext_ack *extack = cls->common.extack;
179 	struct cxgb4_tc_port_matchall *tc_port_matchall;
180 	struct port_info *pi = netdev2pinfo(dev);
181 	struct adapter *adap = netdev2adap(dev);
182 	struct flow_action_entry *entry;
183 	struct sched_class *e;
184 	int ret;
185 	u32 i;
186 
187 	tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id];
188 
189 	flow_action_for_each(i, entry, &cls->rule->action)
190 		if (entry->id == FLOW_ACTION_POLICE)
191 			break;
192 
193 	ret = cxgb4_policer_validate(&cls->rule->action, entry, extack);
194 	if (ret)
195 		return ret;
196 
197 	/* Convert from bytes per second to Kbps */
198 	p.u.params.maxrate = div_u64(entry->police.rate_bytes_ps * 8, 1000);
199 	p.u.params.channel = pi->tx_chan;
200 	e = cxgb4_sched_class_alloc(dev, &p);
201 	if (!e) {
202 		NL_SET_ERR_MSG_MOD(extack,
203 				   "No free traffic class available for policing action");
204 		return -ENOMEM;
205 	}
206 
207 	ret = cxgb4_matchall_tc_bind_queues(dev, e->idx);
208 	if (ret) {
209 		NL_SET_ERR_MSG_MOD(extack,
210 				   "Could not bind queues to traffic class");
211 		goto out_free;
212 	}
213 
214 	tc_port_matchall->egress.hwtc = e->idx;
215 	tc_port_matchall->egress.cookie = cls->cookie;
216 	tc_port_matchall->egress.state = CXGB4_MATCHALL_STATE_ENABLED;
217 	return 0;
218 
219 out_free:
220 	cxgb4_sched_class_free(dev, e->idx);
221 	return ret;
222 }
223 
224 static void cxgb4_matchall_free_tc(struct net_device *dev)
225 {
226 	struct cxgb4_tc_port_matchall *tc_port_matchall;
227 	struct port_info *pi = netdev2pinfo(dev);
228 	struct adapter *adap = netdev2adap(dev);
229 
230 	tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id];
231 	cxgb4_matchall_tc_unbind_queues(dev);
232 	cxgb4_sched_class_free(dev, tc_port_matchall->egress.hwtc);
233 
234 	tc_port_matchall->egress.hwtc = SCHED_CLS_NONE;
235 	tc_port_matchall->egress.cookie = 0;
236 	tc_port_matchall->egress.state = CXGB4_MATCHALL_STATE_DISABLED;
237 }
238 
239 static int cxgb4_matchall_mirror_alloc(struct net_device *dev,
240 				       struct tc_cls_matchall_offload *cls)
241 {
242 	struct netlink_ext_ack *extack = cls->common.extack;
243 	struct cxgb4_tc_port_matchall *tc_port_matchall;
244 	struct port_info *pi = netdev2pinfo(dev);
245 	struct adapter *adap = netdev2adap(dev);
246 	struct flow_action_entry *act;
247 	int ret;
248 	u32 i;
249 
250 	tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id];
251 	flow_action_for_each(i, act, &cls->rule->action) {
252 		if (act->id == FLOW_ACTION_MIRRED) {
253 			ret = cxgb4_port_mirror_alloc(dev);
254 			if (ret) {
255 				NL_SET_ERR_MSG_MOD(extack,
256 						   "Couldn't allocate mirror");
257 				return ret;
258 			}
259 
260 			tc_port_matchall->ingress.viid_mirror = pi->viid_mirror;
261 			break;
262 		}
263 	}
264 
265 	return 0;
266 }
267 
268 static void cxgb4_matchall_mirror_free(struct net_device *dev)
269 {
270 	struct cxgb4_tc_port_matchall *tc_port_matchall;
271 	struct port_info *pi = netdev2pinfo(dev);
272 	struct adapter *adap = netdev2adap(dev);
273 
274 	tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id];
275 	if (!tc_port_matchall->ingress.viid_mirror)
276 		return;
277 
278 	cxgb4_port_mirror_free(dev);
279 	tc_port_matchall->ingress.viid_mirror = 0;
280 }
281 
282 static int cxgb4_matchall_del_filter(struct net_device *dev, u8 filter_type)
283 {
284 	struct cxgb4_tc_port_matchall *tc_port_matchall;
285 	struct port_info *pi = netdev2pinfo(dev);
286 	struct adapter *adap = netdev2adap(dev);
287 	int ret;
288 
289 	tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id];
290 	ret = cxgb4_del_filter(dev, tc_port_matchall->ingress.tid[filter_type],
291 			       &tc_port_matchall->ingress.fs[filter_type]);
292 	if (ret)
293 		return ret;
294 
295 	tc_port_matchall->ingress.tid[filter_type] = 0;
296 	return 0;
297 }
298 
299 static int cxgb4_matchall_add_filter(struct net_device *dev,
300 				     struct tc_cls_matchall_offload *cls,
301 				     u8 filter_type)
302 {
303 	struct netlink_ext_ack *extack = cls->common.extack;
304 	struct cxgb4_tc_port_matchall *tc_port_matchall;
305 	struct port_info *pi = netdev2pinfo(dev);
306 	struct adapter *adap = netdev2adap(dev);
307 	struct ch_filter_specification *fs;
308 	int ret, fidx;
309 
310 	/* Get a free filter entry TID, where we can insert this new
311 	 * rule. Only insert rule if its prio doesn't conflict with
312 	 * existing rules.
313 	 */
314 	fidx = cxgb4_get_free_ftid(dev, filter_type ? PF_INET6 : PF_INET,
315 				   false, cls->common.prio);
316 	if (fidx < 0) {
317 		NL_SET_ERR_MSG_MOD(extack,
318 				   "No free LETCAM index available");
319 		return -ENOMEM;
320 	}
321 
322 	tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id];
323 	fs = &tc_port_matchall->ingress.fs[filter_type];
324 	memset(fs, 0, sizeof(*fs));
325 
326 	if (fidx < adap->tids.nhpftids)
327 		fs->prio = 1;
328 	fs->tc_prio = cls->common.prio;
329 	fs->tc_cookie = cls->cookie;
330 	fs->type = filter_type;
331 	fs->hitcnts = 1;
332 
333 	fs->val.pfvf_vld = 1;
334 	fs->val.pf = adap->pf;
335 	fs->val.vf = pi->vin;
336 
337 	cxgb4_process_flow_actions(dev, &cls->rule->action, fs);
338 
339 	ret = cxgb4_set_filter(dev, fidx, fs);
340 	if (ret)
341 		return ret;
342 
343 	tc_port_matchall->ingress.tid[filter_type] = fidx;
344 	return 0;
345 }
346 
347 static int cxgb4_matchall_alloc_filter(struct net_device *dev,
348 				       struct tc_cls_matchall_offload *cls)
349 {
350 	struct cxgb4_tc_port_matchall *tc_port_matchall;
351 	struct port_info *pi = netdev2pinfo(dev);
352 	struct adapter *adap = netdev2adap(dev);
353 	int ret, i;
354 
355 	tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id];
356 
357 	ret = cxgb4_matchall_mirror_alloc(dev, cls);
358 	if (ret)
359 		return ret;
360 
361 	for (i = 0; i < CXGB4_FILTER_TYPE_MAX; i++) {
362 		ret = cxgb4_matchall_add_filter(dev, cls, i);
363 		if (ret)
364 			goto out_free;
365 	}
366 
367 	tc_port_matchall->ingress.state = CXGB4_MATCHALL_STATE_ENABLED;
368 	return 0;
369 
370 out_free:
371 	while (i-- > 0)
372 		cxgb4_matchall_del_filter(dev, i);
373 
374 	cxgb4_matchall_mirror_free(dev);
375 	return ret;
376 }
377 
378 static int cxgb4_matchall_free_filter(struct net_device *dev)
379 {
380 	struct cxgb4_tc_port_matchall *tc_port_matchall;
381 	struct port_info *pi = netdev2pinfo(dev);
382 	struct adapter *adap = netdev2adap(dev);
383 	int ret;
384 	u8 i;
385 
386 	tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id];
387 
388 	for (i = 0; i < CXGB4_FILTER_TYPE_MAX; i++) {
389 		ret = cxgb4_matchall_del_filter(dev, i);
390 		if (ret)
391 			return ret;
392 	}
393 
394 	cxgb4_matchall_mirror_free(dev);
395 
396 	tc_port_matchall->ingress.packets = 0;
397 	tc_port_matchall->ingress.bytes = 0;
398 	tc_port_matchall->ingress.last_used = 0;
399 	tc_port_matchall->ingress.state = CXGB4_MATCHALL_STATE_DISABLED;
400 	return 0;
401 }
402 
403 int cxgb4_tc_matchall_replace(struct net_device *dev,
404 			      struct tc_cls_matchall_offload *cls_matchall,
405 			      bool ingress)
406 {
407 	struct netlink_ext_ack *extack = cls_matchall->common.extack;
408 	struct cxgb4_tc_port_matchall *tc_port_matchall;
409 	struct port_info *pi = netdev2pinfo(dev);
410 	struct adapter *adap = netdev2adap(dev);
411 	int ret;
412 
413 	tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id];
414 	if (ingress) {
415 		if (tc_port_matchall->ingress.state ==
416 		    CXGB4_MATCHALL_STATE_ENABLED) {
417 			NL_SET_ERR_MSG_MOD(extack,
418 					   "Only 1 Ingress MATCHALL can be offloaded");
419 			return -ENOMEM;
420 		}
421 
422 		ret = cxgb4_validate_flow_actions(dev,
423 						  &cls_matchall->rule->action,
424 						  extack, 1);
425 		if (ret)
426 			return ret;
427 
428 		return cxgb4_matchall_alloc_filter(dev, cls_matchall);
429 	}
430 
431 	if (tc_port_matchall->egress.state == CXGB4_MATCHALL_STATE_ENABLED) {
432 		NL_SET_ERR_MSG_MOD(extack,
433 				   "Only 1 Egress MATCHALL can be offloaded");
434 		return -ENOMEM;
435 	}
436 
437 	ret = cxgb4_matchall_egress_validate(dev, cls_matchall);
438 	if (ret)
439 		return ret;
440 
441 	return cxgb4_matchall_alloc_tc(dev, cls_matchall);
442 }
443 
444 int cxgb4_tc_matchall_destroy(struct net_device *dev,
445 			      struct tc_cls_matchall_offload *cls_matchall,
446 			      bool ingress)
447 {
448 	struct cxgb4_tc_port_matchall *tc_port_matchall;
449 	struct port_info *pi = netdev2pinfo(dev);
450 	struct adapter *adap = netdev2adap(dev);
451 
452 	tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id];
453 	if (ingress) {
454 		/* All the filter types of this matchall rule save the
455 		 * same cookie. So, checking for the first one is
456 		 * enough.
457 		 */
458 		if (cls_matchall->cookie !=
459 		    tc_port_matchall->ingress.fs[0].tc_cookie)
460 			return -ENOENT;
461 
462 		return cxgb4_matchall_free_filter(dev);
463 	}
464 
465 	if (cls_matchall->cookie != tc_port_matchall->egress.cookie)
466 		return -ENOENT;
467 
468 	cxgb4_matchall_free_tc(dev);
469 	return 0;
470 }
471 
472 int cxgb4_tc_matchall_stats(struct net_device *dev,
473 			    struct tc_cls_matchall_offload *cls_matchall)
474 {
475 	u64 tmp_packets, tmp_bytes, packets = 0, bytes = 0;
476 	struct cxgb4_tc_port_matchall *tc_port_matchall;
477 	struct cxgb4_matchall_ingress_entry *ingress;
478 	struct port_info *pi = netdev2pinfo(dev);
479 	struct adapter *adap = netdev2adap(dev);
480 	int ret;
481 	u8 i;
482 
483 	tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id];
484 	if (tc_port_matchall->ingress.state == CXGB4_MATCHALL_STATE_DISABLED)
485 		return -ENOENT;
486 
487 	ingress = &tc_port_matchall->ingress;
488 	for (i = 0; i < CXGB4_FILTER_TYPE_MAX; i++) {
489 		ret = cxgb4_get_filter_counters(dev, ingress->tid[i],
490 						&tmp_packets, &tmp_bytes,
491 						ingress->fs[i].hash);
492 		if (ret)
493 			return ret;
494 
495 		packets += tmp_packets;
496 		bytes += tmp_bytes;
497 	}
498 
499 	if (tc_port_matchall->ingress.packets != packets) {
500 		flow_stats_update(&cls_matchall->stats,
501 				  bytes - tc_port_matchall->ingress.bytes,
502 				  packets - tc_port_matchall->ingress.packets,
503 				  0, tc_port_matchall->ingress.last_used,
504 				  FLOW_ACTION_HW_STATS_IMMEDIATE);
505 
506 		tc_port_matchall->ingress.packets = packets;
507 		tc_port_matchall->ingress.bytes = bytes;
508 		tc_port_matchall->ingress.last_used = jiffies;
509 	}
510 
511 	return 0;
512 }
513 
514 static void cxgb4_matchall_disable_offload(struct net_device *dev)
515 {
516 	struct cxgb4_tc_port_matchall *tc_port_matchall;
517 	struct port_info *pi = netdev2pinfo(dev);
518 	struct adapter *adap = netdev2adap(dev);
519 
520 	tc_port_matchall = &adap->tc_matchall->port_matchall[pi->port_id];
521 	if (tc_port_matchall->egress.state == CXGB4_MATCHALL_STATE_ENABLED)
522 		cxgb4_matchall_free_tc(dev);
523 
524 	if (tc_port_matchall->ingress.state == CXGB4_MATCHALL_STATE_ENABLED)
525 		cxgb4_matchall_free_filter(dev);
526 }
527 
528 int cxgb4_init_tc_matchall(struct adapter *adap)
529 {
530 	struct cxgb4_tc_port_matchall *tc_port_matchall;
531 	struct cxgb4_tc_matchall *tc_matchall;
532 	int ret;
533 
534 	tc_matchall = kzalloc(sizeof(*tc_matchall), GFP_KERNEL);
535 	if (!tc_matchall)
536 		return -ENOMEM;
537 
538 	tc_port_matchall = kcalloc(adap->params.nports,
539 				   sizeof(*tc_port_matchall),
540 				   GFP_KERNEL);
541 	if (!tc_port_matchall) {
542 		ret = -ENOMEM;
543 		goto out_free_matchall;
544 	}
545 
546 	tc_matchall->port_matchall = tc_port_matchall;
547 	adap->tc_matchall = tc_matchall;
548 	return 0;
549 
550 out_free_matchall:
551 	kfree(tc_matchall);
552 	return ret;
553 }
554 
555 void cxgb4_cleanup_tc_matchall(struct adapter *adap)
556 {
557 	u8 i;
558 
559 	if (adap->tc_matchall) {
560 		if (adap->tc_matchall->port_matchall) {
561 			for (i = 0; i < adap->params.nports; i++) {
562 				struct net_device *dev = adap->port[i];
563 
564 				if (dev)
565 					cxgb4_matchall_disable_offload(dev);
566 			}
567 			kfree(adap->tc_matchall->port_matchall);
568 		}
569 		kfree(adap->tc_matchall);
570 	}
571 }
572