xref: /linux/include/rdma/ib_mad.h (revision f3956ebb3bf06ab2266ad5ee2214aed46405810c)
1 /* SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB */
2 /*
3  * Copyright (c) 2004 Mellanox Technologies Ltd.  All rights reserved.
4  * Copyright (c) 2004 Infinicon Corporation.  All rights reserved.
5  * Copyright (c) 2004 Intel Corporation.  All rights reserved.
6  * Copyright (c) 2004 Topspin Corporation.  All rights reserved.
7  * Copyright (c) 2004-2006 Voltaire Corporation.  All rights reserved.
8  */
9 
10 #ifndef IB_MAD_H
11 #define IB_MAD_H
12 
13 #include <linux/list.h>
14 
15 #include <rdma/ib_verbs.h>
16 #include <uapi/rdma/ib_user_mad.h>
17 
18 /* Management base versions */
19 #define IB_MGMT_BASE_VERSION			1
20 #define OPA_MGMT_BASE_VERSION			0x80
21 
22 #define OPA_SM_CLASS_VERSION			0x80
23 
24 /* Management classes */
25 #define IB_MGMT_CLASS_SUBN_LID_ROUTED		0x01
26 #define IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE	0x81
27 #define IB_MGMT_CLASS_SUBN_ADM			0x03
28 #define IB_MGMT_CLASS_PERF_MGMT			0x04
29 #define IB_MGMT_CLASS_BM			0x05
30 #define IB_MGMT_CLASS_DEVICE_MGMT		0x06
31 #define IB_MGMT_CLASS_CM			0x07
32 #define IB_MGMT_CLASS_SNMP			0x08
33 #define IB_MGMT_CLASS_DEVICE_ADM		0x10
34 #define IB_MGMT_CLASS_BOOT_MGMT			0x11
35 #define IB_MGMT_CLASS_BIS			0x12
36 #define IB_MGMT_CLASS_CONG_MGMT			0x21
37 #define IB_MGMT_CLASS_VENDOR_RANGE2_START	0x30
38 #define IB_MGMT_CLASS_VENDOR_RANGE2_END		0x4F
39 
40 #define	IB_OPENIB_OUI				(0x001405)
41 
42 /* Management methods */
43 #define IB_MGMT_METHOD_GET			0x01
44 #define IB_MGMT_METHOD_SET			0x02
45 #define IB_MGMT_METHOD_GET_RESP			0x81
46 #define IB_MGMT_METHOD_SEND			0x03
47 #define IB_MGMT_METHOD_TRAP			0x05
48 #define IB_MGMT_METHOD_REPORT			0x06
49 #define IB_MGMT_METHOD_REPORT_RESP		0x86
50 #define IB_MGMT_METHOD_TRAP_REPRESS		0x07
51 
52 #define IB_MGMT_METHOD_RESP			0x80
53 #define IB_BM_ATTR_MOD_RESP			cpu_to_be32(1)
54 
55 #define IB_MGMT_MAX_METHODS			128
56 
57 /* MAD Status field bit masks */
58 #define IB_MGMT_MAD_STATUS_SUCCESS			0x0000
59 #define IB_MGMT_MAD_STATUS_BUSY				0x0001
60 #define IB_MGMT_MAD_STATUS_REDIRECT_REQD		0x0002
61 #define IB_MGMT_MAD_STATUS_BAD_VERSION			0x0004
62 #define IB_MGMT_MAD_STATUS_UNSUPPORTED_METHOD		0x0008
63 #define IB_MGMT_MAD_STATUS_UNSUPPORTED_METHOD_ATTRIB	0x000c
64 #define IB_MGMT_MAD_STATUS_INVALID_ATTRIB_VALUE		0x001c
65 
66 /* RMPP information */
67 #define IB_MGMT_RMPP_VERSION			1
68 
69 #define IB_MGMT_RMPP_TYPE_DATA			1
70 #define IB_MGMT_RMPP_TYPE_ACK			2
71 #define IB_MGMT_RMPP_TYPE_STOP			3
72 #define IB_MGMT_RMPP_TYPE_ABORT			4
73 
74 #define IB_MGMT_RMPP_FLAG_ACTIVE		1
75 #define IB_MGMT_RMPP_FLAG_FIRST			(1<<1)
76 #define IB_MGMT_RMPP_FLAG_LAST			(1<<2)
77 
78 #define IB_MGMT_RMPP_NO_RESPTIME		0x1F
79 
80 #define	IB_MGMT_RMPP_STATUS_SUCCESS		0
81 #define	IB_MGMT_RMPP_STATUS_RESX		1
82 #define	IB_MGMT_RMPP_STATUS_ABORT_MIN		118
83 #define	IB_MGMT_RMPP_STATUS_T2L			118
84 #define	IB_MGMT_RMPP_STATUS_BAD_LEN		119
85 #define	IB_MGMT_RMPP_STATUS_BAD_SEG		120
86 #define	IB_MGMT_RMPP_STATUS_BADT		121
87 #define	IB_MGMT_RMPP_STATUS_W2S			122
88 #define	IB_MGMT_RMPP_STATUS_S2B			123
89 #define	IB_MGMT_RMPP_STATUS_BAD_STATUS		124
90 #define	IB_MGMT_RMPP_STATUS_UNV			125
91 #define	IB_MGMT_RMPP_STATUS_TMR			126
92 #define	IB_MGMT_RMPP_STATUS_UNSPEC		127
93 #define	IB_MGMT_RMPP_STATUS_ABORT_MAX		127
94 
95 #define IB_QP0		0
96 #define IB_QP1		cpu_to_be32(1)
97 #define IB_QP1_QKEY	0x80010000
98 #define IB_QP_SET_QKEY	0x80000000
99 
100 #define IB_DEFAULT_PKEY_PARTIAL 0x7FFF
101 #define IB_DEFAULT_PKEY_FULL	0xFFFF
102 
103 /*
104  * Generic trap/notice types
105  */
106 #define IB_NOTICE_TYPE_FATAL	0x80
107 #define IB_NOTICE_TYPE_URGENT	0x81
108 #define IB_NOTICE_TYPE_SECURITY	0x82
109 #define IB_NOTICE_TYPE_SM	0x83
110 #define IB_NOTICE_TYPE_INFO	0x84
111 
112 /*
113  * Generic trap/notice producers
114  */
115 #define IB_NOTICE_PROD_CA		cpu_to_be16(1)
116 #define IB_NOTICE_PROD_SWITCH		cpu_to_be16(2)
117 #define IB_NOTICE_PROD_ROUTER		cpu_to_be16(3)
118 #define IB_NOTICE_PROD_CLASS_MGR	cpu_to_be16(4)
119 
120 enum {
121 	IB_MGMT_MAD_HDR = 24,
122 	IB_MGMT_MAD_DATA = 232,
123 	IB_MGMT_RMPP_HDR = 36,
124 	IB_MGMT_RMPP_DATA = 220,
125 	IB_MGMT_VENDOR_HDR = 40,
126 	IB_MGMT_VENDOR_DATA = 216,
127 	IB_MGMT_SA_HDR = 56,
128 	IB_MGMT_SA_DATA = 200,
129 	IB_MGMT_DEVICE_HDR = 64,
130 	IB_MGMT_DEVICE_DATA = 192,
131 	IB_MGMT_MAD_SIZE = IB_MGMT_MAD_HDR + IB_MGMT_MAD_DATA,
132 	OPA_MGMT_MAD_DATA = 2024,
133 	OPA_MGMT_RMPP_DATA = 2012,
134 	OPA_MGMT_MAD_SIZE = IB_MGMT_MAD_HDR + OPA_MGMT_MAD_DATA,
135 };
136 
137 struct ib_mad_hdr {
138 	u8	base_version;
139 	u8	mgmt_class;
140 	u8	class_version;
141 	u8	method;
142 	__be16	status;
143 	__be16	class_specific;
144 	__be64	tid;
145 	__be16	attr_id;
146 	__be16	resv;
147 	__be32	attr_mod;
148 };
149 
150 struct ib_rmpp_hdr {
151 	u8	rmpp_version;
152 	u8	rmpp_type;
153 	u8	rmpp_rtime_flags;
154 	u8	rmpp_status;
155 	__be32	seg_num;
156 	__be32	paylen_newwin;
157 };
158 
159 typedef u64 __bitwise ib_sa_comp_mask;
160 
161 #define IB_SA_COMP_MASK(n) ((__force ib_sa_comp_mask) cpu_to_be64(1ull << (n)))
162 
163 /*
164  * ib_sa_hdr and ib_sa_mad structures must be packed because they have
165  * 64-bit fields that are only 32-bit aligned. 64-bit architectures will
166  * lay them out wrong otherwise.  (And unfortunately they are sent on
167  * the wire so we can't change the layout)
168  */
169 struct ib_sa_hdr {
170 	__be64			sm_key;
171 	__be16			attr_offset;
172 	__be16			reserved;
173 	ib_sa_comp_mask		comp_mask;
174 } __packed;
175 
176 struct ib_mad {
177 	struct ib_mad_hdr	mad_hdr;
178 	u8			data[IB_MGMT_MAD_DATA];
179 };
180 
181 struct opa_mad {
182 	struct ib_mad_hdr	mad_hdr;
183 	u8			data[OPA_MGMT_MAD_DATA];
184 };
185 
186 struct ib_rmpp_mad {
187 	struct ib_mad_hdr	mad_hdr;
188 	struct ib_rmpp_hdr	rmpp_hdr;
189 	u8			data[IB_MGMT_RMPP_DATA];
190 };
191 
192 struct opa_rmpp_mad {
193 	struct ib_mad_hdr	mad_hdr;
194 	struct ib_rmpp_hdr	rmpp_hdr;
195 	u8			data[OPA_MGMT_RMPP_DATA];
196 };
197 
198 struct ib_sa_mad {
199 	struct ib_mad_hdr	mad_hdr;
200 	struct ib_rmpp_hdr	rmpp_hdr;
201 	struct ib_sa_hdr	sa_hdr;
202 	u8			data[IB_MGMT_SA_DATA];
203 } __packed;
204 
205 struct ib_vendor_mad {
206 	struct ib_mad_hdr	mad_hdr;
207 	struct ib_rmpp_hdr	rmpp_hdr;
208 	u8			reserved;
209 	u8			oui[3];
210 	u8			data[IB_MGMT_VENDOR_DATA];
211 };
212 
213 #define IB_MGMT_CLASSPORTINFO_ATTR_ID	cpu_to_be16(0x0001)
214 
215 #define IB_CLASS_PORT_INFO_RESP_TIME_MASK	0x1F
216 #define IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE 5
217 
218 struct ib_class_port_info {
219 	u8			base_version;
220 	u8			class_version;
221 	__be16			capability_mask;
222 	  /* 27 bits for cap_mask2, 5 bits for resp_time */
223 	__be32			cap_mask2_resp_time;
224 	u8			redirect_gid[16];
225 	__be32			redirect_tcslfl;
226 	__be16			redirect_lid;
227 	__be16			redirect_pkey;
228 	__be32			redirect_qp;
229 	__be32			redirect_qkey;
230 	u8			trap_gid[16];
231 	__be32			trap_tcslfl;
232 	__be16			trap_lid;
233 	__be16			trap_pkey;
234 	__be32			trap_hlqp;
235 	__be32			trap_qkey;
236 };
237 
238 /* PortInfo CapabilityMask */
239 enum ib_port_capability_mask_bits {
240 	IB_PORT_SM = 1 << 1,
241 	IB_PORT_NOTICE_SUP = 1 << 2,
242 	IB_PORT_TRAP_SUP = 1 << 3,
243 	IB_PORT_OPT_IPD_SUP = 1 << 4,
244 	IB_PORT_AUTO_MIGR_SUP = 1 << 5,
245 	IB_PORT_SL_MAP_SUP = 1 << 6,
246 	IB_PORT_MKEY_NVRAM = 1 << 7,
247 	IB_PORT_PKEY_NVRAM = 1 << 8,
248 	IB_PORT_LED_INFO_SUP = 1 << 9,
249 	IB_PORT_SM_DISABLED = 1 << 10,
250 	IB_PORT_SYS_IMAGE_GUID_SUP = 1 << 11,
251 	IB_PORT_PKEY_SW_EXT_PORT_TRAP_SUP = 1 << 12,
252 	IB_PORT_EXTENDED_SPEEDS_SUP = 1 << 14,
253 	IB_PORT_CAP_MASK2_SUP = 1 << 15,
254 	IB_PORT_CM_SUP = 1 << 16,
255 	IB_PORT_SNMP_TUNNEL_SUP = 1 << 17,
256 	IB_PORT_REINIT_SUP = 1 << 18,
257 	IB_PORT_DEVICE_MGMT_SUP = 1 << 19,
258 	IB_PORT_VENDOR_CLASS_SUP = 1 << 20,
259 	IB_PORT_DR_NOTICE_SUP = 1 << 21,
260 	IB_PORT_CAP_MASK_NOTICE_SUP = 1 << 22,
261 	IB_PORT_BOOT_MGMT_SUP = 1 << 23,
262 	IB_PORT_LINK_LATENCY_SUP = 1 << 24,
263 	IB_PORT_CLIENT_REG_SUP = 1 << 25,
264 	IB_PORT_OTHER_LOCAL_CHANGES_SUP = 1 << 26,
265 	IB_PORT_LINK_SPEED_WIDTH_TABLE_SUP = 1 << 27,
266 	IB_PORT_VENDOR_SPECIFIC_MADS_TABLE_SUP = 1 << 28,
267 	IB_PORT_MCAST_PKEY_TRAP_SUPPRESSION_SUP = 1 << 29,
268 	IB_PORT_MCAST_FDB_TOP_SUP = 1 << 30,
269 	IB_PORT_HIERARCHY_INFO_SUP = 1ULL << 31,
270 };
271 
272 enum ib_port_capability_mask2_bits {
273 	IB_PORT_SET_NODE_DESC_SUP		= 1 << 0,
274 	IB_PORT_EX_PORT_INFO_EX_SUP		= 1 << 1,
275 	IB_PORT_VIRT_SUP			= 1 << 2,
276 	IB_PORT_SWITCH_PORT_STATE_TABLE_SUP	= 1 << 3,
277 	IB_PORT_LINK_WIDTH_2X_SUP		= 1 << 4,
278 	IB_PORT_LINK_SPEED_HDR_SUP		= 1 << 5,
279 };
280 
281 #define OPA_CLASS_PORT_INFO_PR_SUPPORT BIT(26)
282 
283 struct opa_class_port_info {
284 	u8 base_version;
285 	u8 class_version;
286 	__be16 cap_mask;
287 	__be32 cap_mask2_resp_time;
288 
289 	u8 redirect_gid[16];
290 	__be32 redirect_tc_fl;
291 	__be32 redirect_lid;
292 	__be32 redirect_sl_qp;
293 	__be32 redirect_qkey;
294 
295 	u8 trap_gid[16];
296 	__be32 trap_tc_fl;
297 	__be32 trap_lid;
298 	__be32 trap_hl_qp;
299 	__be32 trap_qkey;
300 
301 	__be16 trap_pkey;
302 	__be16 redirect_pkey;
303 
304 	u8 trap_sl_rsvd;
305 	u8 reserved[3];
306 } __packed;
307 
308 /**
309  * ib_get_cpi_resp_time - Returns the resp_time value from
310  * cap_mask2_resp_time in ib_class_port_info.
311  * @cpi: A struct ib_class_port_info mad.
312  */
313 static inline u8 ib_get_cpi_resp_time(struct ib_class_port_info *cpi)
314 {
315 	return (u8)(be32_to_cpu(cpi->cap_mask2_resp_time) &
316 		    IB_CLASS_PORT_INFO_RESP_TIME_MASK);
317 }
318 
319 /**
320  * ib_set_cpi_resptime - Sets the response time in an
321  * ib_class_port_info mad.
322  * @cpi: A struct ib_class_port_info.
323  * @rtime: The response time to set.
324  */
325 static inline void ib_set_cpi_resp_time(struct ib_class_port_info *cpi,
326 					u8 rtime)
327 {
328 	cpi->cap_mask2_resp_time =
329 		(cpi->cap_mask2_resp_time &
330 		 cpu_to_be32(~IB_CLASS_PORT_INFO_RESP_TIME_MASK)) |
331 		cpu_to_be32(rtime & IB_CLASS_PORT_INFO_RESP_TIME_MASK);
332 }
333 
334 /**
335  * ib_get_cpi_capmask2 - Returns the capmask2 value from
336  * cap_mask2_resp_time in ib_class_port_info.
337  * @cpi: A struct ib_class_port_info mad.
338  */
339 static inline u32 ib_get_cpi_capmask2(struct ib_class_port_info *cpi)
340 {
341 	return (be32_to_cpu(cpi->cap_mask2_resp_time) >>
342 		IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE);
343 }
344 
345 /**
346  * ib_set_cpi_capmask2 - Sets the capmask2 in an
347  * ib_class_port_info mad.
348  * @cpi: A struct ib_class_port_info.
349  * @capmask2: The capmask2 to set.
350  */
351 static inline void ib_set_cpi_capmask2(struct ib_class_port_info *cpi,
352 				       u32 capmask2)
353 {
354 	cpi->cap_mask2_resp_time =
355 		(cpi->cap_mask2_resp_time &
356 		 cpu_to_be32(IB_CLASS_PORT_INFO_RESP_TIME_MASK)) |
357 		cpu_to_be32(capmask2 <<
358 			    IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE);
359 }
360 
361 /**
362  * opa_get_cpi_capmask2 - Returns the capmask2 value from
363  * cap_mask2_resp_time in ib_class_port_info.
364  * @cpi: A struct opa_class_port_info mad.
365  */
366 static inline u32 opa_get_cpi_capmask2(struct opa_class_port_info *cpi)
367 {
368 	return (be32_to_cpu(cpi->cap_mask2_resp_time) >>
369 		IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE);
370 }
371 
372 struct ib_mad_notice_attr {
373 	u8 generic_type;
374 	u8 prod_type_msb;
375 	__be16 prod_type_lsb;
376 	__be16 trap_num;
377 	__be16 issuer_lid;
378 	__be16 toggle_count;
379 
380 	union {
381 		struct {
382 			u8	details[54];
383 		} raw_data;
384 
385 		struct {
386 			__be16	reserved;
387 			__be16	lid;		/* where violation happened */
388 			u8	port_num;	/* where violation happened */
389 		} __packed ntc_129_131;
390 
391 		struct {
392 			__be16	reserved;
393 			__be16	lid;		/* LID where change occurred */
394 			u8	reserved2;
395 			u8	local_changes;	/* low bit - local changes */
396 			__be32	new_cap_mask;	/* new capability mask */
397 			u8	reserved3;
398 			u8	change_flags;	/* low 3 bits only */
399 		} __packed ntc_144;
400 
401 		struct {
402 			__be16	reserved;
403 			__be16	lid;		/* lid where sys guid changed */
404 			__be16	reserved2;
405 			__be64	new_sys_guid;
406 		} __packed ntc_145;
407 
408 		struct {
409 			__be16	reserved;
410 			__be16	lid;
411 			__be16	dr_slid;
412 			u8	method;
413 			u8	reserved2;
414 			__be16	attr_id;
415 			__be32	attr_mod;
416 			__be64	mkey;
417 			u8	reserved3;
418 			u8	dr_trunc_hop;
419 			u8	dr_rtn_path[30];
420 		} __packed ntc_256;
421 
422 		struct {
423 			__be16		reserved;
424 			__be16		lid1;
425 			__be16		lid2;
426 			__be32		key;
427 			__be32		sl_qp1;	/* SL: high 4 bits */
428 			__be32		qp2;	/* high 8 bits reserved */
429 			union ib_gid	gid1;
430 			union ib_gid	gid2;
431 		} __packed ntc_257_258;
432 
433 	} details;
434 };
435 
436 /**
437  * ib_mad_send_buf - MAD data buffer and work request for sends.
438  * @next: A pointer used to chain together MADs for posting.
439  * @mad: References an allocated MAD data buffer for MADs that do not have
440  *   RMPP active.  For MADs using RMPP, references the common and management
441  *   class specific headers.
442  * @mad_agent: MAD agent that allocated the buffer.
443  * @ah: The address handle to use when sending the MAD.
444  * @context: User-controlled context fields.
445  * @hdr_len: Indicates the size of the data header of the MAD.  This length
446  *   includes the common MAD, RMPP, and class specific headers.
447  * @data_len: Indicates the total size of user-transferred data.
448  * @seg_count: The number of RMPP segments allocated for this send.
449  * @seg_size: Size of the data in each RMPP segment.  This does not include
450  *   class specific headers.
451  * @seg_rmpp_size: Size of each RMPP segment including the class specific
452  *   headers.
453  * @timeout_ms: Time to wait for a response.
454  * @retries: Number of times to retry a request for a response.  For MADs
455  *   using RMPP, this applies per window.  On completion, returns the number
456  *   of retries needed to complete the transfer.
457  *
458  * Users are responsible for initializing the MAD buffer itself, with the
459  * exception of any RMPP header.  Additional segment buffer space allocated
460  * beyond data_len is padding.
461  */
462 struct ib_mad_send_buf {
463 	struct ib_mad_send_buf	*next;
464 	void			*mad;
465 	struct ib_mad_agent	*mad_agent;
466 	struct ib_ah		*ah;
467 	void			*context[2];
468 	int			hdr_len;
469 	int			data_len;
470 	int			seg_count;
471 	int			seg_size;
472 	int			seg_rmpp_size;
473 	int			timeout_ms;
474 	int			retries;
475 };
476 
477 /**
478  * ib_response_mad - Returns if the specified MAD has been generated in
479  *   response to a sent request or trap.
480  */
481 int ib_response_mad(const struct ib_mad_hdr *hdr);
482 
483 /**
484  * ib_get_rmpp_resptime - Returns the RMPP response time.
485  * @rmpp_hdr: An RMPP header.
486  */
487 static inline u8 ib_get_rmpp_resptime(struct ib_rmpp_hdr *rmpp_hdr)
488 {
489 	return rmpp_hdr->rmpp_rtime_flags >> 3;
490 }
491 
492 /**
493  * ib_get_rmpp_flags - Returns the RMPP flags.
494  * @rmpp_hdr: An RMPP header.
495  */
496 static inline u8 ib_get_rmpp_flags(struct ib_rmpp_hdr *rmpp_hdr)
497 {
498 	return rmpp_hdr->rmpp_rtime_flags & 0x7;
499 }
500 
501 /**
502  * ib_set_rmpp_resptime - Sets the response time in an RMPP header.
503  * @rmpp_hdr: An RMPP header.
504  * @rtime: The response time to set.
505  */
506 static inline void ib_set_rmpp_resptime(struct ib_rmpp_hdr *rmpp_hdr, u8 rtime)
507 {
508 	rmpp_hdr->rmpp_rtime_flags = ib_get_rmpp_flags(rmpp_hdr) | (rtime << 3);
509 }
510 
511 /**
512  * ib_set_rmpp_flags - Sets the flags in an RMPP header.
513  * @rmpp_hdr: An RMPP header.
514  * @flags: The flags to set.
515  */
516 static inline void ib_set_rmpp_flags(struct ib_rmpp_hdr *rmpp_hdr, u8 flags)
517 {
518 	rmpp_hdr->rmpp_rtime_flags = (rmpp_hdr->rmpp_rtime_flags & 0xF8) |
519 				     (flags & 0x7);
520 }
521 
522 struct ib_mad_agent;
523 struct ib_mad_send_wc;
524 struct ib_mad_recv_wc;
525 
526 /**
527  * ib_mad_send_handler - callback handler for a sent MAD.
528  * @mad_agent: MAD agent that sent the MAD.
529  * @mad_send_wc: Send work completion information on the sent MAD.
530  */
531 typedef void (*ib_mad_send_handler)(struct ib_mad_agent *mad_agent,
532 				    struct ib_mad_send_wc *mad_send_wc);
533 
534 /**
535  * ib_mad_recv_handler - callback handler for a received MAD.
536  * @mad_agent: MAD agent requesting the received MAD.
537  * @send_buf: Send buffer if found, else NULL
538  * @mad_recv_wc: Received work completion information on the received MAD.
539  *
540  * MADs received in response to a send request operation will be handed to
541  * the user before the send operation completes.  All data buffers given
542  * to registered agents through this routine are owned by the receiving
543  * client.
544  */
545 typedef void (*ib_mad_recv_handler)(struct ib_mad_agent *mad_agent,
546 				    struct ib_mad_send_buf *send_buf,
547 				    struct ib_mad_recv_wc *mad_recv_wc);
548 
549 /**
550  * ib_mad_agent - Used to track MAD registration with the access layer.
551  * @device: Reference to device registration is on.
552  * @qp: Reference to QP used for sending and receiving MADs.
553  * @mr: Memory region for system memory usable for DMA.
554  * @recv_handler: Callback handler for a received MAD.
555  * @send_handler: Callback handler for a sent MAD.
556  * @context: User-specified context associated with this registration.
557  * @hi_tid: Access layer assigned transaction ID for this client.
558  *   Unsolicited MADs sent by this client will have the upper 32-bits
559  *   of their TID set to this value.
560  * @flags: registration flags
561  * @port_num: Port number on which QP is registered
562  * @rmpp_version: If set, indicates the RMPP version used by this agent.
563  */
564 enum {
565 	IB_MAD_USER_RMPP = IB_USER_MAD_USER_RMPP,
566 };
567 struct ib_mad_agent {
568 	struct ib_device	*device;
569 	struct ib_qp		*qp;
570 	ib_mad_recv_handler	recv_handler;
571 	ib_mad_send_handler	send_handler;
572 	void			*context;
573 	u32			hi_tid;
574 	u32			flags;
575 	void			*security;
576 	struct list_head	mad_agent_sec_list;
577 	u8			port_num;
578 	u8			rmpp_version;
579 	bool			smp_allowed;
580 };
581 
582 /**
583  * ib_mad_send_wc - MAD send completion information.
584  * @send_buf: Send MAD data buffer associated with the send MAD request.
585  * @status: Completion status.
586  * @vendor_err: Optional vendor error information returned with a failed
587  *   request.
588  */
589 struct ib_mad_send_wc {
590 	struct ib_mad_send_buf	*send_buf;
591 	enum ib_wc_status	status;
592 	u32			vendor_err;
593 };
594 
595 /**
596  * ib_mad_recv_buf - received MAD buffer information.
597  * @list: Reference to next data buffer for a received RMPP MAD.
598  * @grh: References a data buffer containing the global route header.
599  *   The data refereced by this buffer is only valid if the GRH is
600  *   valid.
601  * @mad: References the start of the received MAD.
602  */
603 struct ib_mad_recv_buf {
604 	struct list_head	list;
605 	struct ib_grh		*grh;
606 	union {
607 		struct ib_mad	*mad;
608 		struct opa_mad	*opa_mad;
609 	};
610 };
611 
612 /**
613  * ib_mad_recv_wc - received MAD information.
614  * @wc: Completion information for the received data.
615  * @recv_buf: Specifies the location of the received data buffer(s).
616  * @rmpp_list: Specifies a list of RMPP reassembled received MAD buffers.
617  * @mad_len: The length of the received MAD, without duplicated headers.
618  * @mad_seg_size: The size of individual MAD segments
619  *
620  * For received response, the wr_id contains a pointer to the ib_mad_send_buf
621  *   for the corresponding send request.
622  */
623 struct ib_mad_recv_wc {
624 	struct ib_wc		*wc;
625 	struct ib_mad_recv_buf	recv_buf;
626 	struct list_head	rmpp_list;
627 	int			mad_len;
628 	size_t			mad_seg_size;
629 };
630 
631 /**
632  * ib_mad_reg_req - MAD registration request
633  * @mgmt_class: Indicates which management class of MADs should be receive
634  *   by the caller.  This field is only required if the user wishes to
635  *   receive unsolicited MADs, otherwise it should be 0.
636  * @mgmt_class_version: Indicates which version of MADs for the given
637  *   management class to receive.
638  * @oui: Indicates IEEE OUI when mgmt_class is a vendor class
639  *   in the range from 0x30 to 0x4f. Otherwise not used.
640  * @method_mask: The caller will receive unsolicited MADs for any method
641  *   where @method_mask = 1.
642  *
643  */
644 struct ib_mad_reg_req {
645 	u8	mgmt_class;
646 	u8	mgmt_class_version;
647 	u8	oui[3];
648 	DECLARE_BITMAP(method_mask, IB_MGMT_MAX_METHODS);
649 };
650 
651 /**
652  * ib_register_mad_agent - Register to send/receive MADs.
653  * @device: The device to register with.
654  * @port_num: The port on the specified device to use.
655  * @qp_type: Specifies which QP to access.  Must be either
656  *   IB_QPT_SMI or IB_QPT_GSI.
657  * @mad_reg_req: Specifies which unsolicited MADs should be received
658  *   by the caller.  This parameter may be NULL if the caller only
659  *   wishes to receive solicited responses.
660  * @rmpp_version: If set, indicates that the client will send
661  *   and receive MADs that contain the RMPP header for the given version.
662  *   If set to 0, indicates that RMPP is not used by this client.
663  * @send_handler: The completion callback routine invoked after a send
664  *   request has completed.
665  * @recv_handler: The completion callback routine invoked for a received
666  *   MAD.
667  * @context: User specified context associated with the registration.
668  * @registration_flags: Registration flags to set for this agent
669  */
670 struct ib_mad_agent *ib_register_mad_agent(struct ib_device *device,
671 					   u32 port_num,
672 					   enum ib_qp_type qp_type,
673 					   struct ib_mad_reg_req *mad_reg_req,
674 					   u8 rmpp_version,
675 					   ib_mad_send_handler send_handler,
676 					   ib_mad_recv_handler recv_handler,
677 					   void *context,
678 					   u32 registration_flags);
679 /**
680  * ib_unregister_mad_agent - Unregisters a client from using MAD services.
681  * @mad_agent: Corresponding MAD registration request to deregister.
682  *
683  * After invoking this routine, MAD services are no longer usable by the
684  * client on the associated QP.
685  */
686 void ib_unregister_mad_agent(struct ib_mad_agent *mad_agent);
687 
688 /**
689  * ib_post_send_mad - Posts MAD(s) to the send queue of the QP associated
690  *   with the registered client.
691  * @send_buf: Specifies the information needed to send the MAD(s).
692  * @bad_send_buf: Specifies the MAD on which an error was encountered.  This
693  *   parameter is optional if only a single MAD is posted.
694  *
695  * Sent MADs are not guaranteed to complete in the order that they were posted.
696  *
697  * If the MAD requires RMPP, the data buffer should contain a single copy
698  * of the common MAD, RMPP, and class specific headers, followed by the class
699  * defined data.  If the class defined data would not divide evenly into
700  * RMPP segments, then space must be allocated at the end of the referenced
701  * buffer for any required padding.  To indicate the amount of class defined
702  * data being transferred, the paylen_newwin field in the RMPP header should
703  * be set to the size of the class specific header plus the amount of class
704  * defined data being transferred.  The paylen_newwin field should be
705  * specified in network-byte order.
706  */
707 int ib_post_send_mad(struct ib_mad_send_buf *send_buf,
708 		     struct ib_mad_send_buf **bad_send_buf);
709 
710 
711 /**
712  * ib_free_recv_mad - Returns data buffers used to receive a MAD.
713  * @mad_recv_wc: Work completion information for a received MAD.
714  *
715  * Clients receiving MADs through their ib_mad_recv_handler must call this
716  * routine to return the work completion buffers to the access layer.
717  */
718 void ib_free_recv_mad(struct ib_mad_recv_wc *mad_recv_wc);
719 
720 /**
721  * ib_modify_mad - Modifies an outstanding send MAD operation.
722  * @send_buf: Indicates the MAD to modify.
723  * @timeout_ms: New timeout value for sent MAD.
724  *
725  * This call will reset the timeout value for a sent MAD to the specified
726  * value.
727  */
728 int ib_modify_mad(struct ib_mad_send_buf *send_buf, u32 timeout_ms);
729 
730 /**
731  * ib_cancel_mad - Cancels an outstanding send MAD operation.
732  * @send_buf: Indicates the MAD to cancel.
733  *
734  * MADs will be returned to the user through the corresponding
735  * ib_mad_send_handler.
736  */
737 static inline void ib_cancel_mad(struct ib_mad_send_buf *send_buf)
738 {
739 	ib_modify_mad(send_buf, 0);
740 }
741 
742 /**
743  * ib_create_send_mad - Allocate and initialize a data buffer and work request
744  *   for sending a MAD.
745  * @mad_agent: Specifies the registered MAD service to associate with the MAD.
746  * @remote_qpn: Specifies the QPN of the receiving node.
747  * @pkey_index: Specifies which PKey the MAD will be sent using.  This field
748  *   is valid only if the remote_qpn is QP 1.
749  * @rmpp_active: Indicates if the send will enable RMPP.
750  * @hdr_len: Indicates the size of the data header of the MAD.  This length
751  *   should include the common MAD header, RMPP header, plus any class
752  *   specific header.
753  * @data_len: Indicates the size of any user-transferred data.  The call will
754  *   automatically adjust the allocated buffer size to account for any
755  *   additional padding that may be necessary.
756  * @gfp_mask: GFP mask used for the memory allocation.
757  * @base_version: Base Version of this MAD
758  *
759  * This routine allocates a MAD for sending.  The returned MAD send buffer
760  * will reference a data buffer usable for sending a MAD, along
761  * with an initialized work request structure.  Users may modify the returned
762  * MAD data buffer before posting the send.
763  *
764  * The returned MAD header, class specific headers, and any padding will be
765  * cleared.  Users are responsible for initializing the common MAD header,
766  * any class specific header, and MAD data area.
767  * If @rmpp_active is set, the RMPP header will be initialized for sending.
768  */
769 struct ib_mad_send_buf *ib_create_send_mad(struct ib_mad_agent *mad_agent,
770 					   u32 remote_qpn, u16 pkey_index,
771 					   int rmpp_active,
772 					   int hdr_len, int data_len,
773 					   gfp_t gfp_mask,
774 					   u8 base_version);
775 
776 /**
777  * ib_is_mad_class_rmpp - returns whether given management class
778  * supports RMPP.
779  * @mgmt_class: management class
780  *
781  * This routine returns whether the management class supports RMPP.
782  */
783 int ib_is_mad_class_rmpp(u8 mgmt_class);
784 
785 /**
786  * ib_get_mad_data_offset - returns the data offset for a given
787  * management class.
788  * @mgmt_class: management class
789  *
790  * This routine returns the data offset in the MAD for the management
791  * class requested.
792  */
793 int ib_get_mad_data_offset(u8 mgmt_class);
794 
795 /**
796  * ib_get_rmpp_segment - returns the data buffer for a given RMPP segment.
797  * @send_buf: Previously allocated send data buffer.
798  * @seg_num: number of segment to return
799  *
800  * This routine returns a pointer to the data buffer of an RMPP MAD.
801  * Users must provide synchronization to @send_buf around this call.
802  */
803 void *ib_get_rmpp_segment(struct ib_mad_send_buf *send_buf, int seg_num);
804 
805 /**
806  * ib_free_send_mad - Returns data buffers used to send a MAD.
807  * @send_buf: Previously allocated send data buffer.
808  */
809 void ib_free_send_mad(struct ib_mad_send_buf *send_buf);
810 
811 /**
812  * ib_mad_kernel_rmpp_agent - Returns if the agent is performing RMPP.
813  * @agent: the agent in question
814  * @return: true if agent is performing rmpp, false otherwise.
815  */
816 int ib_mad_kernel_rmpp_agent(const struct ib_mad_agent *agent);
817 
818 #endif /* IB_MAD_H */
819