xref: /linux/drivers/infiniband/hw/qib/qib_sysfs.c (revision 26fbb4c8c7c3ee9a4c3b4de555a8587b5a19154e)
1 /*
2  * Copyright (c) 2012 Intel Corporation.  All rights reserved.
3  * Copyright (c) 2006 - 2012 QLogic Corporation. All rights reserved.
4  * Copyright (c) 2006 PathScale, Inc. All rights reserved.
5  *
6  * This software is available to you under a choice of one of two
7  * licenses.  You may choose to be licensed under the terms of the GNU
8  * General Public License (GPL) Version 2, available from the file
9  * COPYING in the main directory of this source tree, or the
10  * OpenIB.org BSD license below:
11  *
12  *     Redistribution and use in source and binary forms, with or
13  *     without modification, are permitted provided that the following
14  *     conditions are met:
15  *
16  *      - Redistributions of source code must retain the above
17  *        copyright notice, this list of conditions and the following
18  *        disclaimer.
19  *
20  *      - Redistributions in binary form must reproduce the above
21  *        copyright notice, this list of conditions and the following
22  *        disclaimer in the documentation and/or other materials
23  *        provided with the distribution.
24  *
25  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
26  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
27  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
28  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
29  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
30  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
31  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
32  * SOFTWARE.
33  */
34 #include <linux/ctype.h>
35 
36 #include "qib.h"
37 #include "qib_mad.h"
38 
39 /* start of per-port functions */
40 /*
41  * Get/Set heartbeat enable. OR of 1=enabled, 2=auto
42  */
43 static ssize_t show_hrtbt_enb(struct qib_pportdata *ppd, char *buf)
44 {
45 	struct qib_devdata *dd = ppd->dd;
46 
47 	return sysfs_emit(buf, "%d\n", dd->f_get_ib_cfg(ppd, QIB_IB_CFG_HRTBT));
48 }
49 
50 static ssize_t store_hrtbt_enb(struct qib_pportdata *ppd, const char *buf,
51 			       size_t count)
52 {
53 	struct qib_devdata *dd = ppd->dd;
54 	int ret;
55 	u16 val;
56 
57 	ret = kstrtou16(buf, 0, &val);
58 	if (ret) {
59 		qib_dev_err(dd, "attempt to set invalid Heartbeat enable\n");
60 		return ret;
61 	}
62 
63 	/*
64 	 * Set the "intentional" heartbeat enable per either of
65 	 * "Enable" and "Auto", as these are normally set together.
66 	 * This bit is consulted when leaving loopback mode,
67 	 * because entering loopback mode overrides it and automatically
68 	 * disables heartbeat.
69 	 */
70 	ret = dd->f_set_ib_cfg(ppd, QIB_IB_CFG_HRTBT, val);
71 	return ret < 0 ? ret : count;
72 }
73 
74 static ssize_t store_loopback(struct qib_pportdata *ppd, const char *buf,
75 			      size_t count)
76 {
77 	struct qib_devdata *dd = ppd->dd;
78 	int ret = count, r;
79 
80 	r = dd->f_set_ib_loopback(ppd, buf);
81 	if (r < 0)
82 		ret = r;
83 
84 	return ret;
85 }
86 
87 static ssize_t store_led_override(struct qib_pportdata *ppd, const char *buf,
88 				  size_t count)
89 {
90 	struct qib_devdata *dd = ppd->dd;
91 	int ret;
92 	u16 val;
93 
94 	ret = kstrtou16(buf, 0, &val);
95 	if (ret) {
96 		qib_dev_err(dd, "attempt to set invalid LED override\n");
97 		return ret;
98 	}
99 
100 	qib_set_led_override(ppd, val);
101 	return count;
102 }
103 
104 static ssize_t show_status(struct qib_pportdata *ppd, char *buf)
105 {
106 	if (!ppd->statusp)
107 		return -EINVAL;
108 
109 	return sysfs_emit(buf, "0x%llx\n", (unsigned long long)*(ppd->statusp));
110 }
111 
112 /*
113  * For userland compatibility, these offsets must remain fixed.
114  * They are strings for QIB_STATUS_*
115  */
116 static const char * const qib_status_str[] = {
117 	"Initted",
118 	"",
119 	"",
120 	"",
121 	"",
122 	"Present",
123 	"IB_link_up",
124 	"IB_configured",
125 	"",
126 	"Fatal_Hardware_Error",
127 	NULL,
128 };
129 
130 static ssize_t show_status_str(struct qib_pportdata *ppd, char *buf)
131 {
132 	int i, any;
133 	u64 s;
134 	ssize_t ret;
135 
136 	if (!ppd->statusp) {
137 		ret = -EINVAL;
138 		goto bail;
139 	}
140 
141 	s = *(ppd->statusp);
142 	*buf = '\0';
143 	for (any = i = 0; s && qib_status_str[i]; i++) {
144 		if (s & 1) {
145 			/* if overflow */
146 			if (any && strlcat(buf, " ", PAGE_SIZE) >= PAGE_SIZE)
147 				break;
148 			if (strlcat(buf, qib_status_str[i], PAGE_SIZE) >=
149 					PAGE_SIZE)
150 				break;
151 			any = 1;
152 		}
153 		s >>= 1;
154 	}
155 	if (any)
156 		strlcat(buf, "\n", PAGE_SIZE);
157 
158 	ret = strlen(buf);
159 
160 bail:
161 	return ret;
162 }
163 
164 /* end of per-port functions */
165 
166 /*
167  * Start of per-port file structures and support code
168  * Because we are fitting into other infrastructure, we have to supply the
169  * full set of kobject/sysfs_ops structures and routines.
170  */
171 #define QIB_PORT_ATTR(name, mode, show, store) \
172 	static struct qib_port_attr qib_port_attr_##name = \
173 		__ATTR(name, mode, show, store)
174 
175 struct qib_port_attr {
176 	struct attribute attr;
177 	ssize_t (*show)(struct qib_pportdata *, char *);
178 	ssize_t (*store)(struct qib_pportdata *, const char *, size_t);
179 };
180 
181 QIB_PORT_ATTR(loopback, S_IWUSR, NULL, store_loopback);
182 QIB_PORT_ATTR(led_override, S_IWUSR, NULL, store_led_override);
183 QIB_PORT_ATTR(hrtbt_enable, S_IWUSR | S_IRUGO, show_hrtbt_enb,
184 	      store_hrtbt_enb);
185 QIB_PORT_ATTR(status, S_IRUGO, show_status, NULL);
186 QIB_PORT_ATTR(status_str, S_IRUGO, show_status_str, NULL);
187 
188 static struct attribute *port_default_attributes[] = {
189 	&qib_port_attr_loopback.attr,
190 	&qib_port_attr_led_override.attr,
191 	&qib_port_attr_hrtbt_enable.attr,
192 	&qib_port_attr_status.attr,
193 	&qib_port_attr_status_str.attr,
194 	NULL
195 };
196 
197 /*
198  * Start of per-port congestion control structures and support code
199  */
200 
201 /*
202  * Congestion control table size followed by table entries
203  */
204 static ssize_t read_cc_table_bin(struct file *filp, struct kobject *kobj,
205 		struct bin_attribute *bin_attr,
206 		char *buf, loff_t pos, size_t count)
207 {
208 	int ret;
209 	struct qib_pportdata *ppd =
210 		container_of(kobj, struct qib_pportdata, pport_cc_kobj);
211 
212 	if (!qib_cc_table_size || !ppd->ccti_entries_shadow)
213 		return -EINVAL;
214 
215 	ret = ppd->total_cct_entry * sizeof(struct ib_cc_table_entry_shadow)
216 		 + sizeof(__be16);
217 
218 	if (pos > ret)
219 		return -EINVAL;
220 
221 	if (count > ret - pos)
222 		count = ret - pos;
223 
224 	if (!count)
225 		return count;
226 
227 	spin_lock(&ppd->cc_shadow_lock);
228 	memcpy(buf, ppd->ccti_entries_shadow, count);
229 	spin_unlock(&ppd->cc_shadow_lock);
230 
231 	return count;
232 }
233 
234 static void qib_port_release(struct kobject *kobj)
235 {
236 	/* nothing to do since memory is freed by qib_free_devdata() */
237 }
238 
239 static struct kobj_type qib_port_cc_ktype = {
240 	.release = qib_port_release,
241 };
242 
243 static const struct bin_attribute cc_table_bin_attr = {
244 	.attr = {.name = "cc_table_bin", .mode = 0444},
245 	.read = read_cc_table_bin,
246 	.size = PAGE_SIZE,
247 };
248 
249 /*
250  * Congestion settings: port control, control map and an array of 16
251  * entries for the congestion entries - increase, timer, event log
252  * trigger threshold and the minimum injection rate delay.
253  */
254 static ssize_t read_cc_setting_bin(struct file *filp, struct kobject *kobj,
255 		struct bin_attribute *bin_attr,
256 		char *buf, loff_t pos, size_t count)
257 {
258 	int ret;
259 	struct qib_pportdata *ppd =
260 		container_of(kobj, struct qib_pportdata, pport_cc_kobj);
261 
262 	if (!qib_cc_table_size || !ppd->congestion_entries_shadow)
263 		return -EINVAL;
264 
265 	ret = sizeof(struct ib_cc_congestion_setting_attr_shadow);
266 
267 	if (pos > ret)
268 		return -EINVAL;
269 	if (count > ret - pos)
270 		count = ret - pos;
271 
272 	if (!count)
273 		return count;
274 
275 	spin_lock(&ppd->cc_shadow_lock);
276 	memcpy(buf, ppd->congestion_entries_shadow, count);
277 	spin_unlock(&ppd->cc_shadow_lock);
278 
279 	return count;
280 }
281 
282 static const struct bin_attribute cc_setting_bin_attr = {
283 	.attr = {.name = "cc_settings_bin", .mode = 0444},
284 	.read = read_cc_setting_bin,
285 	.size = PAGE_SIZE,
286 };
287 
288 
289 static ssize_t qib_portattr_show(struct kobject *kobj,
290 	struct attribute *attr, char *buf)
291 {
292 	struct qib_port_attr *pattr =
293 		container_of(attr, struct qib_port_attr, attr);
294 	struct qib_pportdata *ppd =
295 		container_of(kobj, struct qib_pportdata, pport_kobj);
296 
297 	if (!pattr->show)
298 		return -EIO;
299 
300 	return pattr->show(ppd, buf);
301 }
302 
303 static ssize_t qib_portattr_store(struct kobject *kobj,
304 	struct attribute *attr, const char *buf, size_t len)
305 {
306 	struct qib_port_attr *pattr =
307 		container_of(attr, struct qib_port_attr, attr);
308 	struct qib_pportdata *ppd =
309 		container_of(kobj, struct qib_pportdata, pport_kobj);
310 
311 	if (!pattr->store)
312 		return -EIO;
313 
314 	return pattr->store(ppd, buf, len);
315 }
316 
317 
318 static const struct sysfs_ops qib_port_ops = {
319 	.show = qib_portattr_show,
320 	.store = qib_portattr_store,
321 };
322 
323 static struct kobj_type qib_port_ktype = {
324 	.release = qib_port_release,
325 	.sysfs_ops = &qib_port_ops,
326 	.default_attrs = port_default_attributes
327 };
328 
329 /* Start sl2vl */
330 
331 #define QIB_SL2VL_ATTR(N) \
332 	static struct qib_sl2vl_attr qib_sl2vl_attr_##N = { \
333 		.attr = { .name = __stringify(N), .mode = 0444 }, \
334 		.sl = N \
335 	}
336 
337 struct qib_sl2vl_attr {
338 	struct attribute attr;
339 	int sl;
340 };
341 
342 QIB_SL2VL_ATTR(0);
343 QIB_SL2VL_ATTR(1);
344 QIB_SL2VL_ATTR(2);
345 QIB_SL2VL_ATTR(3);
346 QIB_SL2VL_ATTR(4);
347 QIB_SL2VL_ATTR(5);
348 QIB_SL2VL_ATTR(6);
349 QIB_SL2VL_ATTR(7);
350 QIB_SL2VL_ATTR(8);
351 QIB_SL2VL_ATTR(9);
352 QIB_SL2VL_ATTR(10);
353 QIB_SL2VL_ATTR(11);
354 QIB_SL2VL_ATTR(12);
355 QIB_SL2VL_ATTR(13);
356 QIB_SL2VL_ATTR(14);
357 QIB_SL2VL_ATTR(15);
358 
359 static struct attribute *sl2vl_default_attributes[] = {
360 	&qib_sl2vl_attr_0.attr,
361 	&qib_sl2vl_attr_1.attr,
362 	&qib_sl2vl_attr_2.attr,
363 	&qib_sl2vl_attr_3.attr,
364 	&qib_sl2vl_attr_4.attr,
365 	&qib_sl2vl_attr_5.attr,
366 	&qib_sl2vl_attr_6.attr,
367 	&qib_sl2vl_attr_7.attr,
368 	&qib_sl2vl_attr_8.attr,
369 	&qib_sl2vl_attr_9.attr,
370 	&qib_sl2vl_attr_10.attr,
371 	&qib_sl2vl_attr_11.attr,
372 	&qib_sl2vl_attr_12.attr,
373 	&qib_sl2vl_attr_13.attr,
374 	&qib_sl2vl_attr_14.attr,
375 	&qib_sl2vl_attr_15.attr,
376 	NULL
377 };
378 
379 static ssize_t sl2vl_attr_show(struct kobject *kobj, struct attribute *attr,
380 			       char *buf)
381 {
382 	struct qib_sl2vl_attr *sattr =
383 		container_of(attr, struct qib_sl2vl_attr, attr);
384 	struct qib_pportdata *ppd =
385 		container_of(kobj, struct qib_pportdata, sl2vl_kobj);
386 	struct qib_ibport *qibp = &ppd->ibport_data;
387 
388 	return sysfs_emit(buf, "%u\n", qibp->sl_to_vl[sattr->sl]);
389 }
390 
391 static const struct sysfs_ops qib_sl2vl_ops = {
392 	.show = sl2vl_attr_show,
393 };
394 
395 static struct kobj_type qib_sl2vl_ktype = {
396 	.release = qib_port_release,
397 	.sysfs_ops = &qib_sl2vl_ops,
398 	.default_attrs = sl2vl_default_attributes
399 };
400 
401 /* End sl2vl */
402 
403 /* Start diag_counters */
404 
405 #define QIB_DIAGC_ATTR(N) \
406 	static struct qib_diagc_attr qib_diagc_attr_##N = { \
407 		.attr = { .name = __stringify(N), .mode = 0664 }, \
408 		.counter = offsetof(struct qib_ibport, rvp.n_##N) \
409 	}
410 
411 #define QIB_DIAGC_ATTR_PER_CPU(N) \
412 	static struct qib_diagc_attr qib_diagc_attr_##N = { \
413 		.attr = { .name = __stringify(N), .mode = 0664 }, \
414 		.counter = offsetof(struct qib_ibport, rvp.z_##N) \
415 	}
416 
417 struct qib_diagc_attr {
418 	struct attribute attr;
419 	size_t counter;
420 };
421 
422 QIB_DIAGC_ATTR_PER_CPU(rc_acks);
423 QIB_DIAGC_ATTR_PER_CPU(rc_qacks);
424 QIB_DIAGC_ATTR_PER_CPU(rc_delayed_comp);
425 
426 QIB_DIAGC_ATTR(rc_resends);
427 QIB_DIAGC_ATTR(seq_naks);
428 QIB_DIAGC_ATTR(rdma_seq);
429 QIB_DIAGC_ATTR(rnr_naks);
430 QIB_DIAGC_ATTR(other_naks);
431 QIB_DIAGC_ATTR(rc_timeouts);
432 QIB_DIAGC_ATTR(loop_pkts);
433 QIB_DIAGC_ATTR(pkt_drops);
434 QIB_DIAGC_ATTR(dmawait);
435 QIB_DIAGC_ATTR(unaligned);
436 QIB_DIAGC_ATTR(rc_dupreq);
437 QIB_DIAGC_ATTR(rc_seqnak);
438 QIB_DIAGC_ATTR(rc_crwaits);
439 
440 static struct attribute *diagc_default_attributes[] = {
441 	&qib_diagc_attr_rc_resends.attr,
442 	&qib_diagc_attr_rc_acks.attr,
443 	&qib_diagc_attr_rc_qacks.attr,
444 	&qib_diagc_attr_rc_delayed_comp.attr,
445 	&qib_diagc_attr_seq_naks.attr,
446 	&qib_diagc_attr_rdma_seq.attr,
447 	&qib_diagc_attr_rnr_naks.attr,
448 	&qib_diagc_attr_other_naks.attr,
449 	&qib_diagc_attr_rc_timeouts.attr,
450 	&qib_diagc_attr_loop_pkts.attr,
451 	&qib_diagc_attr_pkt_drops.attr,
452 	&qib_diagc_attr_dmawait.attr,
453 	&qib_diagc_attr_unaligned.attr,
454 	&qib_diagc_attr_rc_dupreq.attr,
455 	&qib_diagc_attr_rc_seqnak.attr,
456 	&qib_diagc_attr_rc_crwaits.attr,
457 	NULL
458 };
459 
460 static u64 get_all_cpu_total(u64 __percpu *cntr)
461 {
462 	int cpu;
463 	u64 counter = 0;
464 
465 	for_each_possible_cpu(cpu)
466 		counter += *per_cpu_ptr(cntr, cpu);
467 	return counter;
468 }
469 
470 #define def_write_per_cpu(cntr) \
471 static void write_per_cpu_##cntr(struct qib_pportdata *ppd, u32 data)	\
472 {									\
473 	struct qib_devdata *dd = ppd->dd;				\
474 	struct qib_ibport *qibp = &ppd->ibport_data;			\
475 	/*  A write can only zero the counter */			\
476 	if (data == 0)							\
477 		qibp->rvp.z_##cntr = get_all_cpu_total(qibp->rvp.cntr); \
478 	else								\
479 		qib_dev_err(dd, "Per CPU cntrs can only be zeroed");	\
480 }
481 
482 def_write_per_cpu(rc_acks)
483 def_write_per_cpu(rc_qacks)
484 def_write_per_cpu(rc_delayed_comp)
485 
486 #define READ_PER_CPU_CNTR(cntr) (get_all_cpu_total(qibp->rvp.cntr) - \
487 							qibp->rvp.z_##cntr)
488 
489 static ssize_t diagc_attr_show(struct kobject *kobj, struct attribute *attr,
490 			       char *buf)
491 {
492 	struct qib_diagc_attr *dattr =
493 		container_of(attr, struct qib_diagc_attr, attr);
494 	struct qib_pportdata *ppd =
495 		container_of(kobj, struct qib_pportdata, diagc_kobj);
496 	struct qib_ibport *qibp = &ppd->ibport_data;
497 	u64 val;
498 
499 	if (!strncmp(dattr->attr.name, "rc_acks", 7))
500 		val = READ_PER_CPU_CNTR(rc_acks);
501 	else if (!strncmp(dattr->attr.name, "rc_qacks", 8))
502 		val = READ_PER_CPU_CNTR(rc_qacks);
503 	else if (!strncmp(dattr->attr.name, "rc_delayed_comp", 15))
504 		val = READ_PER_CPU_CNTR(rc_delayed_comp);
505 	else
506 		val = *(u32 *)((char *)qibp + dattr->counter);
507 
508 	return sysfs_emit(buf, "%llu\n", val);
509 }
510 
511 static ssize_t diagc_attr_store(struct kobject *kobj, struct attribute *attr,
512 				const char *buf, size_t size)
513 {
514 	struct qib_diagc_attr *dattr =
515 		container_of(attr, struct qib_diagc_attr, attr);
516 	struct qib_pportdata *ppd =
517 		container_of(kobj, struct qib_pportdata, diagc_kobj);
518 	struct qib_ibport *qibp = &ppd->ibport_data;
519 	u32 val;
520 	int ret;
521 
522 	ret = kstrtou32(buf, 0, &val);
523 	if (ret)
524 		return ret;
525 
526 	if (!strncmp(dattr->attr.name, "rc_acks", 7))
527 		write_per_cpu_rc_acks(ppd, val);
528 	else if (!strncmp(dattr->attr.name, "rc_qacks", 8))
529 		write_per_cpu_rc_qacks(ppd, val);
530 	else if (!strncmp(dattr->attr.name, "rc_delayed_comp", 15))
531 		write_per_cpu_rc_delayed_comp(ppd, val);
532 	else
533 		*(u32 *)((char *)qibp + dattr->counter) = val;
534 	return size;
535 }
536 
537 static const struct sysfs_ops qib_diagc_ops = {
538 	.show = diagc_attr_show,
539 	.store = diagc_attr_store,
540 };
541 
542 static struct kobj_type qib_diagc_ktype = {
543 	.release = qib_port_release,
544 	.sysfs_ops = &qib_diagc_ops,
545 	.default_attrs = diagc_default_attributes
546 };
547 
548 /* End diag_counters */
549 
550 /* end of per-port file structures and support code */
551 
552 /*
553  * Start of per-unit (or driver, in some cases, but replicated
554  * per unit) functions (these get a device *)
555  */
556 static ssize_t hw_rev_show(struct device *device, struct device_attribute *attr,
557 			   char *buf)
558 {
559 	struct qib_ibdev *dev =
560 		rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev);
561 
562 	return sysfs_emit(buf, "%x\n", dd_from_dev(dev)->minrev);
563 }
564 static DEVICE_ATTR_RO(hw_rev);
565 
566 static ssize_t hca_type_show(struct device *device,
567 			     struct device_attribute *attr, char *buf)
568 {
569 	struct qib_ibdev *dev =
570 		rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev);
571 	struct qib_devdata *dd = dd_from_dev(dev);
572 
573 	if (!dd->boardname)
574 		return -EINVAL;
575 	return sysfs_emit(buf, "%s\n", dd->boardname);
576 }
577 static DEVICE_ATTR_RO(hca_type);
578 static DEVICE_ATTR(board_id, 0444, hca_type_show, NULL);
579 
580 static ssize_t version_show(struct device *device,
581 			    struct device_attribute *attr, char *buf)
582 {
583 	/* The string printed here is already newline-terminated. */
584 	return sysfs_emit(buf, "%s", (char *)ib_qib_version);
585 }
586 static DEVICE_ATTR_RO(version);
587 
588 static ssize_t boardversion_show(struct device *device,
589 				 struct device_attribute *attr, char *buf)
590 {
591 	struct qib_ibdev *dev =
592 		rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev);
593 	struct qib_devdata *dd = dd_from_dev(dev);
594 
595 	/* The string printed here is already newline-terminated. */
596 	return sysfs_emit(buf, "%s", dd->boardversion);
597 }
598 static DEVICE_ATTR_RO(boardversion);
599 
600 static ssize_t localbus_info_show(struct device *device,
601 				  struct device_attribute *attr, char *buf)
602 {
603 	struct qib_ibdev *dev =
604 		rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev);
605 	struct qib_devdata *dd = dd_from_dev(dev);
606 
607 	/* The string printed here is already newline-terminated. */
608 	return sysfs_emit(buf, "%s", dd->lbus_info);
609 }
610 static DEVICE_ATTR_RO(localbus_info);
611 
612 static ssize_t nctxts_show(struct device *device,
613 			   struct device_attribute *attr, char *buf)
614 {
615 	struct qib_ibdev *dev =
616 		rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev);
617 	struct qib_devdata *dd = dd_from_dev(dev);
618 
619 	/* Return the number of user ports (contexts) available. */
620 	/* The calculation below deals with a special case where
621 	 * cfgctxts is set to 1 on a single-port board. */
622 	return sysfs_emit(buf, "%u\n",
623 			  (dd->first_user_ctxt > dd->cfgctxts) ?
624 				  0 :
625 				  (dd->cfgctxts - dd->first_user_ctxt));
626 }
627 static DEVICE_ATTR_RO(nctxts);
628 
629 static ssize_t nfreectxts_show(struct device *device,
630 			       struct device_attribute *attr, char *buf)
631 {
632 	struct qib_ibdev *dev =
633 		rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev);
634 	struct qib_devdata *dd = dd_from_dev(dev);
635 
636 	/* Return the number of free user ports (contexts) available. */
637 	return sysfs_emit(buf, "%u\n", dd->freectxts);
638 }
639 static DEVICE_ATTR_RO(nfreectxts);
640 
641 static ssize_t serial_show(struct device *device, struct device_attribute *attr,
642 			   char *buf)
643 {
644 	struct qib_ibdev *dev =
645 		rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev);
646 	struct qib_devdata *dd = dd_from_dev(dev);
647 	const u8 *end = memchr(dd->serial, 0, ARRAY_SIZE(dd->serial));
648 	int size = end ? end - dd->serial : ARRAY_SIZE(dd->serial);
649 
650 	return sysfs_emit(buf, ".%*s\n", size, dd->serial);
651 }
652 static DEVICE_ATTR_RO(serial);
653 
654 static ssize_t chip_reset_store(struct device *device,
655 				struct device_attribute *attr, const char *buf,
656 				size_t count)
657 {
658 	struct qib_ibdev *dev =
659 		rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev);
660 	struct qib_devdata *dd = dd_from_dev(dev);
661 	int ret;
662 
663 	if (count < 5 || memcmp(buf, "reset", 5) || !dd->diag_client) {
664 		ret = -EINVAL;
665 		goto bail;
666 	}
667 
668 	ret = qib_reset_device(dd->unit);
669 bail:
670 	return ret < 0 ? ret : count;
671 }
672 static DEVICE_ATTR_WO(chip_reset);
673 
674 /*
675  * Dump tempsense regs. in decimal, to ease shell-scripts.
676  */
677 static ssize_t tempsense_show(struct device *device,
678 			      struct device_attribute *attr, char *buf)
679 {
680 	struct qib_ibdev *dev =
681 		rdma_device_to_drv_device(device, struct qib_ibdev, rdi.ibdev);
682 	struct qib_devdata *dd = dd_from_dev(dev);
683 	int i;
684 	u8 regvals[8];
685 
686 	for (i = 0; i < 8; i++) {
687 		int ret;
688 
689 		if (i == 6)
690 			continue;
691 		ret = dd->f_tempsense_rd(dd, i);
692 		if (ret < 0)
693 			return ret;	/* return error on bad read */
694 		regvals[i] = ret;
695 	}
696 	return sysfs_emit(buf, "%d %d %02X %02X %d %d\n",
697 			  (signed char)regvals[0],
698 			  (signed char)regvals[1],
699 			  regvals[2],
700 			  regvals[3],
701 			  (signed char)regvals[5],
702 			  (signed char)regvals[7]);
703 }
704 static DEVICE_ATTR_RO(tempsense);
705 
706 /*
707  * end of per-unit (or driver, in some cases, but replicated
708  * per unit) functions
709  */
710 
711 /* start of per-unit file structures and support code */
712 static struct attribute *qib_attributes[] = {
713 	&dev_attr_hw_rev.attr,
714 	&dev_attr_hca_type.attr,
715 	&dev_attr_board_id.attr,
716 	&dev_attr_version.attr,
717 	&dev_attr_nctxts.attr,
718 	&dev_attr_nfreectxts.attr,
719 	&dev_attr_serial.attr,
720 	&dev_attr_boardversion.attr,
721 	&dev_attr_tempsense.attr,
722 	&dev_attr_localbus_info.attr,
723 	&dev_attr_chip_reset.attr,
724 	NULL,
725 };
726 
727 const struct attribute_group qib_attr_group = {
728 	.attrs = qib_attributes,
729 };
730 
731 int qib_create_port_files(struct ib_device *ibdev, u8 port_num,
732 			  struct kobject *kobj)
733 {
734 	struct qib_pportdata *ppd;
735 	struct qib_devdata *dd = dd_from_ibdev(ibdev);
736 	int ret;
737 
738 	if (!port_num || port_num > dd->num_pports) {
739 		qib_dev_err(dd,
740 			"Skipping infiniband class with invalid port %u\n",
741 			port_num);
742 		ret = -ENODEV;
743 		goto bail;
744 	}
745 	ppd = &dd->pport[port_num - 1];
746 
747 	ret = kobject_init_and_add(&ppd->pport_kobj, &qib_port_ktype, kobj,
748 				   "linkcontrol");
749 	if (ret) {
750 		qib_dev_err(dd,
751 			"Skipping linkcontrol sysfs info, (err %d) port %u\n",
752 			ret, port_num);
753 		goto bail_link;
754 	}
755 	kobject_uevent(&ppd->pport_kobj, KOBJ_ADD);
756 
757 	ret = kobject_init_and_add(&ppd->sl2vl_kobj, &qib_sl2vl_ktype, kobj,
758 				   "sl2vl");
759 	if (ret) {
760 		qib_dev_err(dd,
761 			"Skipping sl2vl sysfs info, (err %d) port %u\n",
762 			ret, port_num);
763 		goto bail_sl;
764 	}
765 	kobject_uevent(&ppd->sl2vl_kobj, KOBJ_ADD);
766 
767 	ret = kobject_init_and_add(&ppd->diagc_kobj, &qib_diagc_ktype, kobj,
768 				   "diag_counters");
769 	if (ret) {
770 		qib_dev_err(dd,
771 			"Skipping diag_counters sysfs info, (err %d) port %u\n",
772 			ret, port_num);
773 		goto bail_diagc;
774 	}
775 	kobject_uevent(&ppd->diagc_kobj, KOBJ_ADD);
776 
777 	if (!qib_cc_table_size || !ppd->congestion_entries_shadow)
778 		return 0;
779 
780 	ret = kobject_init_and_add(&ppd->pport_cc_kobj, &qib_port_cc_ktype,
781 				kobj, "CCMgtA");
782 	if (ret) {
783 		qib_dev_err(dd,
784 		 "Skipping Congestion Control sysfs info, (err %d) port %u\n",
785 		 ret, port_num);
786 		goto bail_cc;
787 	}
788 
789 	kobject_uevent(&ppd->pport_cc_kobj, KOBJ_ADD);
790 
791 	ret = sysfs_create_bin_file(&ppd->pport_cc_kobj,
792 				&cc_setting_bin_attr);
793 	if (ret) {
794 		qib_dev_err(dd,
795 		 "Skipping Congestion Control setting sysfs info, (err %d) port %u\n",
796 		 ret, port_num);
797 		goto bail_cc;
798 	}
799 
800 	ret = sysfs_create_bin_file(&ppd->pport_cc_kobj,
801 				&cc_table_bin_attr);
802 	if (ret) {
803 		qib_dev_err(dd,
804 		 "Skipping Congestion Control table sysfs info, (err %d) port %u\n",
805 		 ret, port_num);
806 		goto bail_cc_entry_bin;
807 	}
808 
809 	qib_devinfo(dd->pcidev,
810 		"IB%u: Congestion Control Agent enabled for port %d\n",
811 		dd->unit, port_num);
812 
813 	return 0;
814 
815 bail_cc_entry_bin:
816 	sysfs_remove_bin_file(&ppd->pport_cc_kobj, &cc_setting_bin_attr);
817 bail_cc:
818 	kobject_put(&ppd->pport_cc_kobj);
819 bail_diagc:
820 	kobject_put(&ppd->diagc_kobj);
821 bail_sl:
822 	kobject_put(&ppd->sl2vl_kobj);
823 bail_link:
824 	kobject_put(&ppd->pport_kobj);
825 bail:
826 	return ret;
827 }
828 
829 /*
830  * Unregister and remove our files in /sys/class/infiniband.
831  */
832 void qib_verbs_unregister_sysfs(struct qib_devdata *dd)
833 {
834 	struct qib_pportdata *ppd;
835 	int i;
836 
837 	for (i = 0; i < dd->num_pports; i++) {
838 		ppd = &dd->pport[i];
839 		if (qib_cc_table_size &&
840 			ppd->congestion_entries_shadow) {
841 			sysfs_remove_bin_file(&ppd->pport_cc_kobj,
842 				&cc_setting_bin_attr);
843 			sysfs_remove_bin_file(&ppd->pport_cc_kobj,
844 				&cc_table_bin_attr);
845 			kobject_put(&ppd->pport_cc_kobj);
846 		}
847 		kobject_put(&ppd->diagc_kobj);
848 		kobject_put(&ppd->sl2vl_kobj);
849 		kobject_put(&ppd->pport_kobj);
850 	}
851 }
852