xref: /linux/net/core/net-procfs.c (revision 7effbd18e496a954540fb6319db0d87d28a70496)
1 // SPDX-License-Identifier: GPL-2.0
2 #include <linux/netdevice.h>
3 #include <linux/proc_fs.h>
4 #include <linux/seq_file.h>
5 #include <net/wext.h>
6 
7 #include "dev.h"
8 
9 #define BUCKET_SPACE (32 - NETDEV_HASHBITS - 1)
10 
11 #define get_bucket(x) ((x) >> BUCKET_SPACE)
12 #define get_offset(x) ((x) & ((1 << BUCKET_SPACE) - 1))
13 #define set_bucket_offset(b, o) ((b) << BUCKET_SPACE | (o))
14 
15 static inline struct net_device *dev_from_same_bucket(struct seq_file *seq, loff_t *pos)
16 {
17 	struct net *net = seq_file_net(seq);
18 	struct net_device *dev;
19 	struct hlist_head *h;
20 	unsigned int count = 0, offset = get_offset(*pos);
21 
22 	h = &net->dev_index_head[get_bucket(*pos)];
23 	hlist_for_each_entry_rcu(dev, h, index_hlist) {
24 		if (++count == offset)
25 			return dev;
26 	}
27 
28 	return NULL;
29 }
30 
31 static inline struct net_device *dev_from_bucket(struct seq_file *seq, loff_t *pos)
32 {
33 	struct net_device *dev;
34 	unsigned int bucket;
35 
36 	do {
37 		dev = dev_from_same_bucket(seq, pos);
38 		if (dev)
39 			return dev;
40 
41 		bucket = get_bucket(*pos) + 1;
42 		*pos = set_bucket_offset(bucket, 1);
43 	} while (bucket < NETDEV_HASHENTRIES);
44 
45 	return NULL;
46 }
47 
48 /*
49  *	This is invoked by the /proc filesystem handler to display a device
50  *	in detail.
51  */
52 static void *dev_seq_start(struct seq_file *seq, loff_t *pos)
53 	__acquires(RCU)
54 {
55 	rcu_read_lock();
56 	if (!*pos)
57 		return SEQ_START_TOKEN;
58 
59 	if (get_bucket(*pos) >= NETDEV_HASHENTRIES)
60 		return NULL;
61 
62 	return dev_from_bucket(seq, pos);
63 }
64 
65 static void *dev_seq_next(struct seq_file *seq, void *v, loff_t *pos)
66 {
67 	++*pos;
68 	return dev_from_bucket(seq, pos);
69 }
70 
71 static void dev_seq_stop(struct seq_file *seq, void *v)
72 	__releases(RCU)
73 {
74 	rcu_read_unlock();
75 }
76 
77 static void dev_seq_printf_stats(struct seq_file *seq, struct net_device *dev)
78 {
79 	struct rtnl_link_stats64 temp;
80 	const struct rtnl_link_stats64 *stats = dev_get_stats(dev, &temp);
81 
82 	seq_printf(seq, "%6s: %7llu %7llu %4llu %4llu %4llu %5llu %10llu %9llu "
83 		   "%8llu %7llu %4llu %4llu %4llu %5llu %7llu %10llu\n",
84 		   dev->name, stats->rx_bytes, stats->rx_packets,
85 		   stats->rx_errors,
86 		   stats->rx_dropped + stats->rx_missed_errors,
87 		   stats->rx_fifo_errors,
88 		   stats->rx_length_errors + stats->rx_over_errors +
89 		    stats->rx_crc_errors + stats->rx_frame_errors,
90 		   stats->rx_compressed, stats->multicast,
91 		   stats->tx_bytes, stats->tx_packets,
92 		   stats->tx_errors, stats->tx_dropped,
93 		   stats->tx_fifo_errors, stats->collisions,
94 		   stats->tx_carrier_errors +
95 		    stats->tx_aborted_errors +
96 		    stats->tx_window_errors +
97 		    stats->tx_heartbeat_errors,
98 		   stats->tx_compressed);
99 }
100 
101 /*
102  *	Called from the PROCfs module. This now uses the new arbitrary sized
103  *	/proc/net interface to create /proc/net/dev
104  */
105 static int dev_seq_show(struct seq_file *seq, void *v)
106 {
107 	if (v == SEQ_START_TOKEN)
108 		seq_puts(seq, "Inter-|   Receive                            "
109 			      "                    |  Transmit\n"
110 			      " face |bytes    packets errs drop fifo frame "
111 			      "compressed multicast|bytes    packets errs "
112 			      "drop fifo colls carrier compressed\n");
113 	else
114 		dev_seq_printf_stats(seq, v);
115 	return 0;
116 }
117 
118 static u32 softnet_backlog_len(struct softnet_data *sd)
119 {
120 	return skb_queue_len_lockless(&sd->input_pkt_queue) +
121 	       skb_queue_len_lockless(&sd->process_queue);
122 }
123 
124 static struct softnet_data *softnet_get_online(loff_t *pos)
125 {
126 	struct softnet_data *sd = NULL;
127 
128 	while (*pos < nr_cpu_ids)
129 		if (cpu_online(*pos)) {
130 			sd = &per_cpu(softnet_data, *pos);
131 			break;
132 		} else
133 			++*pos;
134 	return sd;
135 }
136 
137 static void *softnet_seq_start(struct seq_file *seq, loff_t *pos)
138 {
139 	return softnet_get_online(pos);
140 }
141 
142 static void *softnet_seq_next(struct seq_file *seq, void *v, loff_t *pos)
143 {
144 	++*pos;
145 	return softnet_get_online(pos);
146 }
147 
148 static void softnet_seq_stop(struct seq_file *seq, void *v)
149 {
150 }
151 
152 static int softnet_seq_show(struct seq_file *seq, void *v)
153 {
154 	struct softnet_data *sd = v;
155 	unsigned int flow_limit_count = 0;
156 
157 #ifdef CONFIG_NET_FLOW_LIMIT
158 	struct sd_flow_limit *fl;
159 
160 	rcu_read_lock();
161 	fl = rcu_dereference(sd->flow_limit);
162 	if (fl)
163 		flow_limit_count = fl->count;
164 	rcu_read_unlock();
165 #endif
166 
167 	/* the index is the CPU id owing this sd. Since offline CPUs are not
168 	 * displayed, it would be othrwise not trivial for the user-space
169 	 * mapping the data a specific CPU
170 	 */
171 	seq_printf(seq,
172 		   "%08x %08x %08x %08x %08x %08x %08x %08x %08x %08x %08x %08x %08x\n",
173 		   sd->processed, sd->dropped, sd->time_squeeze, 0,
174 		   0, 0, 0, 0, /* was fastroute */
175 		   0,	/* was cpu_collision */
176 		   sd->received_rps, flow_limit_count,
177 		   softnet_backlog_len(sd), (int)seq->index);
178 	return 0;
179 }
180 
181 static const struct seq_operations dev_seq_ops = {
182 	.start = dev_seq_start,
183 	.next  = dev_seq_next,
184 	.stop  = dev_seq_stop,
185 	.show  = dev_seq_show,
186 };
187 
188 static const struct seq_operations softnet_seq_ops = {
189 	.start = softnet_seq_start,
190 	.next  = softnet_seq_next,
191 	.stop  = softnet_seq_stop,
192 	.show  = softnet_seq_show,
193 };
194 
195 static void *ptype_get_idx(struct seq_file *seq, loff_t pos)
196 {
197 	struct list_head *ptype_list = NULL;
198 	struct packet_type *pt = NULL;
199 	struct net_device *dev;
200 	loff_t i = 0;
201 	int t;
202 
203 	for_each_netdev_rcu(seq_file_net(seq), dev) {
204 		ptype_list = &dev->ptype_all;
205 		list_for_each_entry_rcu(pt, ptype_list, list) {
206 			if (i == pos)
207 				return pt;
208 			++i;
209 		}
210 	}
211 
212 	list_for_each_entry_rcu(pt, &ptype_all, list) {
213 		if (i == pos)
214 			return pt;
215 		++i;
216 	}
217 
218 	for (t = 0; t < PTYPE_HASH_SIZE; t++) {
219 		list_for_each_entry_rcu(pt, &ptype_base[t], list) {
220 			if (i == pos)
221 				return pt;
222 			++i;
223 		}
224 	}
225 	return NULL;
226 }
227 
228 static void *ptype_seq_start(struct seq_file *seq, loff_t *pos)
229 	__acquires(RCU)
230 {
231 	rcu_read_lock();
232 	return *pos ? ptype_get_idx(seq, *pos - 1) : SEQ_START_TOKEN;
233 }
234 
235 static void *ptype_seq_next(struct seq_file *seq, void *v, loff_t *pos)
236 {
237 	struct net_device *dev;
238 	struct packet_type *pt;
239 	struct list_head *nxt;
240 	int hash;
241 
242 	++*pos;
243 	if (v == SEQ_START_TOKEN)
244 		return ptype_get_idx(seq, 0);
245 
246 	pt = v;
247 	nxt = pt->list.next;
248 	if (pt->dev) {
249 		if (nxt != &pt->dev->ptype_all)
250 			goto found;
251 
252 		dev = pt->dev;
253 		for_each_netdev_continue_rcu(seq_file_net(seq), dev) {
254 			if (!list_empty(&dev->ptype_all)) {
255 				nxt = dev->ptype_all.next;
256 				goto found;
257 			}
258 		}
259 
260 		nxt = ptype_all.next;
261 		goto ptype_all;
262 	}
263 
264 	if (pt->type == htons(ETH_P_ALL)) {
265 ptype_all:
266 		if (nxt != &ptype_all)
267 			goto found;
268 		hash = 0;
269 		nxt = ptype_base[0].next;
270 	} else
271 		hash = ntohs(pt->type) & PTYPE_HASH_MASK;
272 
273 	while (nxt == &ptype_base[hash]) {
274 		if (++hash >= PTYPE_HASH_SIZE)
275 			return NULL;
276 		nxt = ptype_base[hash].next;
277 	}
278 found:
279 	return list_entry(nxt, struct packet_type, list);
280 }
281 
282 static void ptype_seq_stop(struct seq_file *seq, void *v)
283 	__releases(RCU)
284 {
285 	rcu_read_unlock();
286 }
287 
288 static int ptype_seq_show(struct seq_file *seq, void *v)
289 {
290 	struct packet_type *pt = v;
291 
292 	if (v == SEQ_START_TOKEN)
293 		seq_puts(seq, "Type Device      Function\n");
294 	else if ((!pt->af_packet_net || net_eq(pt->af_packet_net, seq_file_net(seq))) &&
295 		 (!pt->dev || net_eq(dev_net(pt->dev), seq_file_net(seq)))) {
296 		if (pt->type == htons(ETH_P_ALL))
297 			seq_puts(seq, "ALL ");
298 		else
299 			seq_printf(seq, "%04x", ntohs(pt->type));
300 
301 		seq_printf(seq, " %-8s %ps\n",
302 			   pt->dev ? pt->dev->name : "", pt->func);
303 	}
304 
305 	return 0;
306 }
307 
308 static const struct seq_operations ptype_seq_ops = {
309 	.start = ptype_seq_start,
310 	.next  = ptype_seq_next,
311 	.stop  = ptype_seq_stop,
312 	.show  = ptype_seq_show,
313 };
314 
315 static int __net_init dev_proc_net_init(struct net *net)
316 {
317 	int rc = -ENOMEM;
318 
319 	if (!proc_create_net("dev", 0444, net->proc_net, &dev_seq_ops,
320 			sizeof(struct seq_net_private)))
321 		goto out;
322 	if (!proc_create_seq("softnet_stat", 0444, net->proc_net,
323 			 &softnet_seq_ops))
324 		goto out_dev;
325 	if (!proc_create_net("ptype", 0444, net->proc_net, &ptype_seq_ops,
326 			sizeof(struct seq_net_private)))
327 		goto out_softnet;
328 
329 	if (wext_proc_init(net))
330 		goto out_ptype;
331 	rc = 0;
332 out:
333 	return rc;
334 out_ptype:
335 	remove_proc_entry("ptype", net->proc_net);
336 out_softnet:
337 	remove_proc_entry("softnet_stat", net->proc_net);
338 out_dev:
339 	remove_proc_entry("dev", net->proc_net);
340 	goto out;
341 }
342 
343 static void __net_exit dev_proc_net_exit(struct net *net)
344 {
345 	wext_proc_exit(net);
346 
347 	remove_proc_entry("ptype", net->proc_net);
348 	remove_proc_entry("softnet_stat", net->proc_net);
349 	remove_proc_entry("dev", net->proc_net);
350 }
351 
352 static struct pernet_operations __net_initdata dev_proc_ops = {
353 	.init = dev_proc_net_init,
354 	.exit = dev_proc_net_exit,
355 };
356 
357 static int dev_mc_seq_show(struct seq_file *seq, void *v)
358 {
359 	struct netdev_hw_addr *ha;
360 	struct net_device *dev = v;
361 
362 	if (v == SEQ_START_TOKEN)
363 		return 0;
364 
365 	netif_addr_lock_bh(dev);
366 	netdev_for_each_mc_addr(ha, dev) {
367 		seq_printf(seq, "%-4d %-15s %-5d %-5d %*phN\n",
368 			   dev->ifindex, dev->name,
369 			   ha->refcount, ha->global_use,
370 			   (int)dev->addr_len, ha->addr);
371 	}
372 	netif_addr_unlock_bh(dev);
373 	return 0;
374 }
375 
376 static const struct seq_operations dev_mc_seq_ops = {
377 	.start = dev_seq_start,
378 	.next  = dev_seq_next,
379 	.stop  = dev_seq_stop,
380 	.show  = dev_mc_seq_show,
381 };
382 
383 static int __net_init dev_mc_net_init(struct net *net)
384 {
385 	if (!proc_create_net("dev_mcast", 0, net->proc_net, &dev_mc_seq_ops,
386 			sizeof(struct seq_net_private)))
387 		return -ENOMEM;
388 	return 0;
389 }
390 
391 static void __net_exit dev_mc_net_exit(struct net *net)
392 {
393 	remove_proc_entry("dev_mcast", net->proc_net);
394 }
395 
396 static struct pernet_operations __net_initdata dev_mc_net_ops = {
397 	.init = dev_mc_net_init,
398 	.exit = dev_mc_net_exit,
399 };
400 
401 int __init dev_proc_init(void)
402 {
403 	int ret = register_pernet_subsys(&dev_proc_ops);
404 	if (!ret)
405 		return register_pernet_subsys(&dev_mc_net_ops);
406 	return ret;
407 }
408