1 /* 2 * Copyright (c) 2004 Mellanox Technologies Ltd. All rights reserved. 3 * Copyright (c) 2004 Infinicon Corporation. All rights reserved. 4 * Copyright (c) 2004 Intel Corporation. All rights reserved. 5 * Copyright (c) 2004 Topspin Corporation. All rights reserved. 6 * Copyright (c) 2004-2006 Voltaire Corporation. All rights reserved. 7 * 8 * This software is available to you under a choice of one of two 9 * licenses. You may choose to be licensed under the terms of the GNU 10 * General Public License (GPL) Version 2, available from the file 11 * COPYING in the main directory of this source tree, or the 12 * OpenIB.org BSD license below: 13 * 14 * Redistribution and use in source and binary forms, with or 15 * without modification, are permitted provided that the following 16 * conditions are met: 17 * 18 * - Redistributions of source code must retain the above 19 * copyright notice, this list of conditions and the following 20 * disclaimer. 21 * 22 * - Redistributions in binary form must reproduce the above 23 * copyright notice, this list of conditions and the following 24 * disclaimer in the documentation and/or other materials 25 * provided with the distribution. 26 * 27 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 28 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 29 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 30 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS 31 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 32 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 33 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 34 * SOFTWARE. 35 */ 36 37 #if !defined(IB_MAD_H) 38 #define IB_MAD_H 39 40 #include <linux/list.h> 41 42 #include <rdma/ib_verbs.h> 43 #include <uapi/rdma/ib_user_mad.h> 44 45 /* Management base versions */ 46 #define IB_MGMT_BASE_VERSION 1 47 #define OPA_MGMT_BASE_VERSION 0x80 48 49 #define OPA_SM_CLASS_VERSION 0x80 50 51 /* Management classes */ 52 #define IB_MGMT_CLASS_SUBN_LID_ROUTED 0x01 53 #define IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE 0x81 54 #define IB_MGMT_CLASS_SUBN_ADM 0x03 55 #define IB_MGMT_CLASS_PERF_MGMT 0x04 56 #define IB_MGMT_CLASS_BM 0x05 57 #define IB_MGMT_CLASS_DEVICE_MGMT 0x06 58 #define IB_MGMT_CLASS_CM 0x07 59 #define IB_MGMT_CLASS_SNMP 0x08 60 #define IB_MGMT_CLASS_DEVICE_ADM 0x10 61 #define IB_MGMT_CLASS_BOOT_MGMT 0x11 62 #define IB_MGMT_CLASS_BIS 0x12 63 #define IB_MGMT_CLASS_CONG_MGMT 0x21 64 #define IB_MGMT_CLASS_VENDOR_RANGE2_START 0x30 65 #define IB_MGMT_CLASS_VENDOR_RANGE2_END 0x4F 66 67 #define IB_OPENIB_OUI (0x001405) 68 69 /* Management methods */ 70 #define IB_MGMT_METHOD_GET 0x01 71 #define IB_MGMT_METHOD_SET 0x02 72 #define IB_MGMT_METHOD_GET_RESP 0x81 73 #define IB_MGMT_METHOD_SEND 0x03 74 #define IB_MGMT_METHOD_TRAP 0x05 75 #define IB_MGMT_METHOD_REPORT 0x06 76 #define IB_MGMT_METHOD_REPORT_RESP 0x86 77 #define IB_MGMT_METHOD_TRAP_REPRESS 0x07 78 79 #define IB_MGMT_METHOD_RESP 0x80 80 #define IB_BM_ATTR_MOD_RESP cpu_to_be32(1) 81 82 #define IB_MGMT_MAX_METHODS 128 83 84 /* MAD Status field bit masks */ 85 #define IB_MGMT_MAD_STATUS_SUCCESS 0x0000 86 #define IB_MGMT_MAD_STATUS_BUSY 0x0001 87 #define IB_MGMT_MAD_STATUS_REDIRECT_REQD 0x0002 88 #define IB_MGMT_MAD_STATUS_BAD_VERSION 0x0004 89 #define IB_MGMT_MAD_STATUS_UNSUPPORTED_METHOD 0x0008 90 #define IB_MGMT_MAD_STATUS_UNSUPPORTED_METHOD_ATTRIB 0x000c 91 #define IB_MGMT_MAD_STATUS_INVALID_ATTRIB_VALUE 0x001c 92 93 /* RMPP information */ 94 #define IB_MGMT_RMPP_VERSION 1 95 96 #define IB_MGMT_RMPP_TYPE_DATA 1 97 #define IB_MGMT_RMPP_TYPE_ACK 2 98 #define IB_MGMT_RMPP_TYPE_STOP 3 99 #define IB_MGMT_RMPP_TYPE_ABORT 4 100 101 #define IB_MGMT_RMPP_FLAG_ACTIVE 1 102 #define IB_MGMT_RMPP_FLAG_FIRST (1<<1) 103 #define IB_MGMT_RMPP_FLAG_LAST (1<<2) 104 105 #define IB_MGMT_RMPP_NO_RESPTIME 0x1F 106 107 #define IB_MGMT_RMPP_STATUS_SUCCESS 0 108 #define IB_MGMT_RMPP_STATUS_RESX 1 109 #define IB_MGMT_RMPP_STATUS_ABORT_MIN 118 110 #define IB_MGMT_RMPP_STATUS_T2L 118 111 #define IB_MGMT_RMPP_STATUS_BAD_LEN 119 112 #define IB_MGMT_RMPP_STATUS_BAD_SEG 120 113 #define IB_MGMT_RMPP_STATUS_BADT 121 114 #define IB_MGMT_RMPP_STATUS_W2S 122 115 #define IB_MGMT_RMPP_STATUS_S2B 123 116 #define IB_MGMT_RMPP_STATUS_BAD_STATUS 124 117 #define IB_MGMT_RMPP_STATUS_UNV 125 118 #define IB_MGMT_RMPP_STATUS_TMR 126 119 #define IB_MGMT_RMPP_STATUS_UNSPEC 127 120 #define IB_MGMT_RMPP_STATUS_ABORT_MAX 127 121 122 #define IB_QP0 0 123 #define IB_QP1 cpu_to_be32(1) 124 #define IB_QP1_QKEY 0x80010000 125 #define IB_QP_SET_QKEY 0x80000000 126 127 #define IB_DEFAULT_PKEY_PARTIAL 0x7FFF 128 #define IB_DEFAULT_PKEY_FULL 0xFFFF 129 130 /* 131 * Generic trap/notice types 132 */ 133 #define IB_NOTICE_TYPE_FATAL 0x80 134 #define IB_NOTICE_TYPE_URGENT 0x81 135 #define IB_NOTICE_TYPE_SECURITY 0x82 136 #define IB_NOTICE_TYPE_SM 0x83 137 #define IB_NOTICE_TYPE_INFO 0x84 138 139 /* 140 * Generic trap/notice producers 141 */ 142 #define IB_NOTICE_PROD_CA cpu_to_be16(1) 143 #define IB_NOTICE_PROD_SWITCH cpu_to_be16(2) 144 #define IB_NOTICE_PROD_ROUTER cpu_to_be16(3) 145 #define IB_NOTICE_PROD_CLASS_MGR cpu_to_be16(4) 146 147 enum { 148 IB_MGMT_MAD_HDR = 24, 149 IB_MGMT_MAD_DATA = 232, 150 IB_MGMT_RMPP_HDR = 36, 151 IB_MGMT_RMPP_DATA = 220, 152 IB_MGMT_VENDOR_HDR = 40, 153 IB_MGMT_VENDOR_DATA = 216, 154 IB_MGMT_SA_HDR = 56, 155 IB_MGMT_SA_DATA = 200, 156 IB_MGMT_DEVICE_HDR = 64, 157 IB_MGMT_DEVICE_DATA = 192, 158 IB_MGMT_MAD_SIZE = IB_MGMT_MAD_HDR + IB_MGMT_MAD_DATA, 159 OPA_MGMT_MAD_DATA = 2024, 160 OPA_MGMT_RMPP_DATA = 2012, 161 OPA_MGMT_MAD_SIZE = IB_MGMT_MAD_HDR + OPA_MGMT_MAD_DATA, 162 }; 163 164 struct ib_mad_hdr { 165 u8 base_version; 166 u8 mgmt_class; 167 u8 class_version; 168 u8 method; 169 __be16 status; 170 __be16 class_specific; 171 __be64 tid; 172 __be16 attr_id; 173 __be16 resv; 174 __be32 attr_mod; 175 }; 176 177 struct ib_rmpp_hdr { 178 u8 rmpp_version; 179 u8 rmpp_type; 180 u8 rmpp_rtime_flags; 181 u8 rmpp_status; 182 __be32 seg_num; 183 __be32 paylen_newwin; 184 }; 185 186 typedef u64 __bitwise ib_sa_comp_mask; 187 188 #define IB_SA_COMP_MASK(n) ((__force ib_sa_comp_mask) cpu_to_be64(1ull << (n))) 189 190 /* 191 * ib_sa_hdr and ib_sa_mad structures must be packed because they have 192 * 64-bit fields that are only 32-bit aligned. 64-bit architectures will 193 * lay them out wrong otherwise. (And unfortunately they are sent on 194 * the wire so we can't change the layout) 195 */ 196 struct ib_sa_hdr { 197 __be64 sm_key; 198 __be16 attr_offset; 199 __be16 reserved; 200 ib_sa_comp_mask comp_mask; 201 } __packed; 202 203 struct ib_mad { 204 struct ib_mad_hdr mad_hdr; 205 u8 data[IB_MGMT_MAD_DATA]; 206 }; 207 208 struct opa_mad { 209 struct ib_mad_hdr mad_hdr; 210 u8 data[OPA_MGMT_MAD_DATA]; 211 }; 212 213 struct ib_rmpp_mad { 214 struct ib_mad_hdr mad_hdr; 215 struct ib_rmpp_hdr rmpp_hdr; 216 u8 data[IB_MGMT_RMPP_DATA]; 217 }; 218 219 struct opa_rmpp_mad { 220 struct ib_mad_hdr mad_hdr; 221 struct ib_rmpp_hdr rmpp_hdr; 222 u8 data[OPA_MGMT_RMPP_DATA]; 223 }; 224 225 struct ib_sa_mad { 226 struct ib_mad_hdr mad_hdr; 227 struct ib_rmpp_hdr rmpp_hdr; 228 struct ib_sa_hdr sa_hdr; 229 u8 data[IB_MGMT_SA_DATA]; 230 } __packed; 231 232 struct ib_vendor_mad { 233 struct ib_mad_hdr mad_hdr; 234 struct ib_rmpp_hdr rmpp_hdr; 235 u8 reserved; 236 u8 oui[3]; 237 u8 data[IB_MGMT_VENDOR_DATA]; 238 }; 239 240 #define IB_MGMT_CLASSPORTINFO_ATTR_ID cpu_to_be16(0x0001) 241 242 #define IB_CLASS_PORT_INFO_RESP_TIME_MASK 0x1F 243 #define IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE 5 244 245 struct ib_class_port_info { 246 u8 base_version; 247 u8 class_version; 248 __be16 capability_mask; 249 /* 27 bits for cap_mask2, 5 bits for resp_time */ 250 __be32 cap_mask2_resp_time; 251 u8 redirect_gid[16]; 252 __be32 redirect_tcslfl; 253 __be16 redirect_lid; 254 __be16 redirect_pkey; 255 __be32 redirect_qp; 256 __be32 redirect_qkey; 257 u8 trap_gid[16]; 258 __be32 trap_tcslfl; 259 __be16 trap_lid; 260 __be16 trap_pkey; 261 __be32 trap_hlqp; 262 __be32 trap_qkey; 263 }; 264 265 /* PortInfo CapabilityMask */ 266 enum ib_port_capability_mask_bits { 267 IB_PORT_SM = 1 << 1, 268 IB_PORT_NOTICE_SUP = 1 << 2, 269 IB_PORT_TRAP_SUP = 1 << 3, 270 IB_PORT_OPT_IPD_SUP = 1 << 4, 271 IB_PORT_AUTO_MIGR_SUP = 1 << 5, 272 IB_PORT_SL_MAP_SUP = 1 << 6, 273 IB_PORT_MKEY_NVRAM = 1 << 7, 274 IB_PORT_PKEY_NVRAM = 1 << 8, 275 IB_PORT_LED_INFO_SUP = 1 << 9, 276 IB_PORT_SM_DISABLED = 1 << 10, 277 IB_PORT_SYS_IMAGE_GUID_SUP = 1 << 11, 278 IB_PORT_PKEY_SW_EXT_PORT_TRAP_SUP = 1 << 12, 279 IB_PORT_EXTENDED_SPEEDS_SUP = 1 << 14, 280 IB_PORT_CAP_MASK2_SUP = 1 << 15, 281 IB_PORT_CM_SUP = 1 << 16, 282 IB_PORT_SNMP_TUNNEL_SUP = 1 << 17, 283 IB_PORT_REINIT_SUP = 1 << 18, 284 IB_PORT_DEVICE_MGMT_SUP = 1 << 19, 285 IB_PORT_VENDOR_CLASS_SUP = 1 << 20, 286 IB_PORT_DR_NOTICE_SUP = 1 << 21, 287 IB_PORT_CAP_MASK_NOTICE_SUP = 1 << 22, 288 IB_PORT_BOOT_MGMT_SUP = 1 << 23, 289 IB_PORT_LINK_LATENCY_SUP = 1 << 24, 290 IB_PORT_CLIENT_REG_SUP = 1 << 25, 291 IB_PORT_OTHER_LOCAL_CHANGES_SUP = 1 << 26, 292 IB_PORT_LINK_SPEED_WIDTH_TABLE_SUP = 1 << 27, 293 IB_PORT_VENDOR_SPECIFIC_MADS_TABLE_SUP = 1 << 28, 294 IB_PORT_MCAST_PKEY_TRAP_SUPPRESSION_SUP = 1 << 29, 295 IB_PORT_MCAST_FDB_TOP_SUP = 1 << 30, 296 IB_PORT_HIERARCHY_INFO_SUP = 1ULL << 31, 297 }; 298 299 enum ib_port_capability_mask2_bits { 300 IB_PORT_SET_NODE_DESC_SUP = 1 << 0, 301 IB_PORT_EX_PORT_INFO_EX_SUP = 1 << 1, 302 IB_PORT_VIRT_SUP = 1 << 2, 303 IB_PORT_SWITCH_PORT_STATE_TABLE_SUP = 1 << 3, 304 IB_PORT_LINK_WIDTH_2X_SUP = 1 << 4, 305 IB_PORT_LINK_SPEED_HDR_SUP = 1 << 5, 306 }; 307 308 #define OPA_CLASS_PORT_INFO_PR_SUPPORT BIT(26) 309 310 struct opa_class_port_info { 311 u8 base_version; 312 u8 class_version; 313 __be16 cap_mask; 314 __be32 cap_mask2_resp_time; 315 316 u8 redirect_gid[16]; 317 __be32 redirect_tc_fl; 318 __be32 redirect_lid; 319 __be32 redirect_sl_qp; 320 __be32 redirect_qkey; 321 322 u8 trap_gid[16]; 323 __be32 trap_tc_fl; 324 __be32 trap_lid; 325 __be32 trap_hl_qp; 326 __be32 trap_qkey; 327 328 __be16 trap_pkey; 329 __be16 redirect_pkey; 330 331 u8 trap_sl_rsvd; 332 u8 reserved[3]; 333 } __packed; 334 335 /** 336 * ib_get_cpi_resp_time - Returns the resp_time value from 337 * cap_mask2_resp_time in ib_class_port_info. 338 * @cpi: A struct ib_class_port_info mad. 339 */ 340 static inline u8 ib_get_cpi_resp_time(struct ib_class_port_info *cpi) 341 { 342 return (u8)(be32_to_cpu(cpi->cap_mask2_resp_time) & 343 IB_CLASS_PORT_INFO_RESP_TIME_MASK); 344 } 345 346 /** 347 * ib_set_cpi_resptime - Sets the response time in an 348 * ib_class_port_info mad. 349 * @cpi: A struct ib_class_port_info. 350 * @rtime: The response time to set. 351 */ 352 static inline void ib_set_cpi_resp_time(struct ib_class_port_info *cpi, 353 u8 rtime) 354 { 355 cpi->cap_mask2_resp_time = 356 (cpi->cap_mask2_resp_time & 357 cpu_to_be32(~IB_CLASS_PORT_INFO_RESP_TIME_MASK)) | 358 cpu_to_be32(rtime & IB_CLASS_PORT_INFO_RESP_TIME_MASK); 359 } 360 361 /** 362 * ib_get_cpi_capmask2 - Returns the capmask2 value from 363 * cap_mask2_resp_time in ib_class_port_info. 364 * @cpi: A struct ib_class_port_info mad. 365 */ 366 static inline u32 ib_get_cpi_capmask2(struct ib_class_port_info *cpi) 367 { 368 return (be32_to_cpu(cpi->cap_mask2_resp_time) >> 369 IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE); 370 } 371 372 /** 373 * ib_set_cpi_capmask2 - Sets the capmask2 in an 374 * ib_class_port_info mad. 375 * @cpi: A struct ib_class_port_info. 376 * @capmask2: The capmask2 to set. 377 */ 378 static inline void ib_set_cpi_capmask2(struct ib_class_port_info *cpi, 379 u32 capmask2) 380 { 381 cpi->cap_mask2_resp_time = 382 (cpi->cap_mask2_resp_time & 383 cpu_to_be32(IB_CLASS_PORT_INFO_RESP_TIME_MASK)) | 384 cpu_to_be32(capmask2 << 385 IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE); 386 } 387 388 /** 389 * opa_get_cpi_capmask2 - Returns the capmask2 value from 390 * cap_mask2_resp_time in ib_class_port_info. 391 * @cpi: A struct opa_class_port_info mad. 392 */ 393 static inline u32 opa_get_cpi_capmask2(struct opa_class_port_info *cpi) 394 { 395 return (be32_to_cpu(cpi->cap_mask2_resp_time) >> 396 IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE); 397 } 398 399 struct ib_mad_notice_attr { 400 u8 generic_type; 401 u8 prod_type_msb; 402 __be16 prod_type_lsb; 403 __be16 trap_num; 404 __be16 issuer_lid; 405 __be16 toggle_count; 406 407 union { 408 struct { 409 u8 details[54]; 410 } raw_data; 411 412 struct { 413 __be16 reserved; 414 __be16 lid; /* where violation happened */ 415 u8 port_num; /* where violation happened */ 416 } __packed ntc_129_131; 417 418 struct { 419 __be16 reserved; 420 __be16 lid; /* LID where change occurred */ 421 u8 reserved2; 422 u8 local_changes; /* low bit - local changes */ 423 __be32 new_cap_mask; /* new capability mask */ 424 u8 reserved3; 425 u8 change_flags; /* low 3 bits only */ 426 } __packed ntc_144; 427 428 struct { 429 __be16 reserved; 430 __be16 lid; /* lid where sys guid changed */ 431 __be16 reserved2; 432 __be64 new_sys_guid; 433 } __packed ntc_145; 434 435 struct { 436 __be16 reserved; 437 __be16 lid; 438 __be16 dr_slid; 439 u8 method; 440 u8 reserved2; 441 __be16 attr_id; 442 __be32 attr_mod; 443 __be64 mkey; 444 u8 reserved3; 445 u8 dr_trunc_hop; 446 u8 dr_rtn_path[30]; 447 } __packed ntc_256; 448 449 struct { 450 __be16 reserved; 451 __be16 lid1; 452 __be16 lid2; 453 __be32 key; 454 __be32 sl_qp1; /* SL: high 4 bits */ 455 __be32 qp2; /* high 8 bits reserved */ 456 union ib_gid gid1; 457 union ib_gid gid2; 458 } __packed ntc_257_258; 459 460 } details; 461 }; 462 463 /** 464 * ib_mad_send_buf - MAD data buffer and work request for sends. 465 * @next: A pointer used to chain together MADs for posting. 466 * @mad: References an allocated MAD data buffer for MADs that do not have 467 * RMPP active. For MADs using RMPP, references the common and management 468 * class specific headers. 469 * @mad_agent: MAD agent that allocated the buffer. 470 * @ah: The address handle to use when sending the MAD. 471 * @context: User-controlled context fields. 472 * @hdr_len: Indicates the size of the data header of the MAD. This length 473 * includes the common MAD, RMPP, and class specific headers. 474 * @data_len: Indicates the total size of user-transferred data. 475 * @seg_count: The number of RMPP segments allocated for this send. 476 * @seg_size: Size of the data in each RMPP segment. This does not include 477 * class specific headers. 478 * @seg_rmpp_size: Size of each RMPP segment including the class specific 479 * headers. 480 * @timeout_ms: Time to wait for a response. 481 * @retries: Number of times to retry a request for a response. For MADs 482 * using RMPP, this applies per window. On completion, returns the number 483 * of retries needed to complete the transfer. 484 * 485 * Users are responsible for initializing the MAD buffer itself, with the 486 * exception of any RMPP header. Additional segment buffer space allocated 487 * beyond data_len is padding. 488 */ 489 struct ib_mad_send_buf { 490 struct ib_mad_send_buf *next; 491 void *mad; 492 struct ib_mad_agent *mad_agent; 493 struct ib_ah *ah; 494 void *context[2]; 495 int hdr_len; 496 int data_len; 497 int seg_count; 498 int seg_size; 499 int seg_rmpp_size; 500 int timeout_ms; 501 int retries; 502 }; 503 504 /** 505 * ib_response_mad - Returns if the specified MAD has been generated in 506 * response to a sent request or trap. 507 */ 508 int ib_response_mad(const struct ib_mad_hdr *hdr); 509 510 /** 511 * ib_get_rmpp_resptime - Returns the RMPP response time. 512 * @rmpp_hdr: An RMPP header. 513 */ 514 static inline u8 ib_get_rmpp_resptime(struct ib_rmpp_hdr *rmpp_hdr) 515 { 516 return rmpp_hdr->rmpp_rtime_flags >> 3; 517 } 518 519 /** 520 * ib_get_rmpp_flags - Returns the RMPP flags. 521 * @rmpp_hdr: An RMPP header. 522 */ 523 static inline u8 ib_get_rmpp_flags(struct ib_rmpp_hdr *rmpp_hdr) 524 { 525 return rmpp_hdr->rmpp_rtime_flags & 0x7; 526 } 527 528 /** 529 * ib_set_rmpp_resptime - Sets the response time in an RMPP header. 530 * @rmpp_hdr: An RMPP header. 531 * @rtime: The response time to set. 532 */ 533 static inline void ib_set_rmpp_resptime(struct ib_rmpp_hdr *rmpp_hdr, u8 rtime) 534 { 535 rmpp_hdr->rmpp_rtime_flags = ib_get_rmpp_flags(rmpp_hdr) | (rtime << 3); 536 } 537 538 /** 539 * ib_set_rmpp_flags - Sets the flags in an RMPP header. 540 * @rmpp_hdr: An RMPP header. 541 * @flags: The flags to set. 542 */ 543 static inline void ib_set_rmpp_flags(struct ib_rmpp_hdr *rmpp_hdr, u8 flags) 544 { 545 rmpp_hdr->rmpp_rtime_flags = (rmpp_hdr->rmpp_rtime_flags & 0xF8) | 546 (flags & 0x7); 547 } 548 549 struct ib_mad_agent; 550 struct ib_mad_send_wc; 551 struct ib_mad_recv_wc; 552 553 /** 554 * ib_mad_send_handler - callback handler for a sent MAD. 555 * @mad_agent: MAD agent that sent the MAD. 556 * @mad_send_wc: Send work completion information on the sent MAD. 557 */ 558 typedef void (*ib_mad_send_handler)(struct ib_mad_agent *mad_agent, 559 struct ib_mad_send_wc *mad_send_wc); 560 561 /** 562 * ib_mad_snoop_handler - Callback handler for snooping sent MADs. 563 * @mad_agent: MAD agent that snooped the MAD. 564 * @send_buf: send MAD data buffer. 565 * @mad_send_wc: Work completion information on the sent MAD. Valid 566 * only for snooping that occurs on a send completion. 567 * 568 * Clients snooping MADs should not modify data referenced by the @send_buf 569 * or @mad_send_wc. 570 */ 571 typedef void (*ib_mad_snoop_handler)(struct ib_mad_agent *mad_agent, 572 struct ib_mad_send_buf *send_buf, 573 struct ib_mad_send_wc *mad_send_wc); 574 575 /** 576 * ib_mad_recv_handler - callback handler for a received MAD. 577 * @mad_agent: MAD agent requesting the received MAD. 578 * @send_buf: Send buffer if found, else NULL 579 * @mad_recv_wc: Received work completion information on the received MAD. 580 * 581 * MADs received in response to a send request operation will be handed to 582 * the user before the send operation completes. All data buffers given 583 * to registered agents through this routine are owned by the receiving 584 * client, except for snooping agents. Clients snooping MADs should not 585 * modify the data referenced by @mad_recv_wc. 586 */ 587 typedef void (*ib_mad_recv_handler)(struct ib_mad_agent *mad_agent, 588 struct ib_mad_send_buf *send_buf, 589 struct ib_mad_recv_wc *mad_recv_wc); 590 591 /** 592 * ib_mad_agent - Used to track MAD registration with the access layer. 593 * @device: Reference to device registration is on. 594 * @qp: Reference to QP used for sending and receiving MADs. 595 * @mr: Memory region for system memory usable for DMA. 596 * @recv_handler: Callback handler for a received MAD. 597 * @send_handler: Callback handler for a sent MAD. 598 * @snoop_handler: Callback handler for snooped sent MADs. 599 * @context: User-specified context associated with this registration. 600 * @hi_tid: Access layer assigned transaction ID for this client. 601 * Unsolicited MADs sent by this client will have the upper 32-bits 602 * of their TID set to this value. 603 * @flags: registration flags 604 * @port_num: Port number on which QP is registered 605 * @rmpp_version: If set, indicates the RMPP version used by this agent. 606 */ 607 enum { 608 IB_MAD_USER_RMPP = IB_USER_MAD_USER_RMPP, 609 }; 610 struct ib_mad_agent { 611 struct ib_device *device; 612 struct ib_qp *qp; 613 ib_mad_recv_handler recv_handler; 614 ib_mad_send_handler send_handler; 615 ib_mad_snoop_handler snoop_handler; 616 void *context; 617 u32 hi_tid; 618 u32 flags; 619 void *security; 620 struct list_head mad_agent_sec_list; 621 u8 port_num; 622 u8 rmpp_version; 623 bool smp_allowed; 624 }; 625 626 /** 627 * ib_mad_send_wc - MAD send completion information. 628 * @send_buf: Send MAD data buffer associated with the send MAD request. 629 * @status: Completion status. 630 * @vendor_err: Optional vendor error information returned with a failed 631 * request. 632 */ 633 struct ib_mad_send_wc { 634 struct ib_mad_send_buf *send_buf; 635 enum ib_wc_status status; 636 u32 vendor_err; 637 }; 638 639 /** 640 * ib_mad_recv_buf - received MAD buffer information. 641 * @list: Reference to next data buffer for a received RMPP MAD. 642 * @grh: References a data buffer containing the global route header. 643 * The data refereced by this buffer is only valid if the GRH is 644 * valid. 645 * @mad: References the start of the received MAD. 646 */ 647 struct ib_mad_recv_buf { 648 struct list_head list; 649 struct ib_grh *grh; 650 union { 651 struct ib_mad *mad; 652 struct opa_mad *opa_mad; 653 }; 654 }; 655 656 /** 657 * ib_mad_recv_wc - received MAD information. 658 * @wc: Completion information for the received data. 659 * @recv_buf: Specifies the location of the received data buffer(s). 660 * @rmpp_list: Specifies a list of RMPP reassembled received MAD buffers. 661 * @mad_len: The length of the received MAD, without duplicated headers. 662 * @mad_seg_size: The size of individual MAD segments 663 * 664 * For received response, the wr_id contains a pointer to the ib_mad_send_buf 665 * for the corresponding send request. 666 */ 667 struct ib_mad_recv_wc { 668 struct ib_wc *wc; 669 struct ib_mad_recv_buf recv_buf; 670 struct list_head rmpp_list; 671 int mad_len; 672 size_t mad_seg_size; 673 }; 674 675 /** 676 * ib_mad_reg_req - MAD registration request 677 * @mgmt_class: Indicates which management class of MADs should be receive 678 * by the caller. This field is only required if the user wishes to 679 * receive unsolicited MADs, otherwise it should be 0. 680 * @mgmt_class_version: Indicates which version of MADs for the given 681 * management class to receive. 682 * @oui: Indicates IEEE OUI when mgmt_class is a vendor class 683 * in the range from 0x30 to 0x4f. Otherwise not used. 684 * @method_mask: The caller will receive unsolicited MADs for any method 685 * where @method_mask = 1. 686 * 687 */ 688 struct ib_mad_reg_req { 689 u8 mgmt_class; 690 u8 mgmt_class_version; 691 u8 oui[3]; 692 DECLARE_BITMAP(method_mask, IB_MGMT_MAX_METHODS); 693 }; 694 695 /** 696 * ib_register_mad_agent - Register to send/receive MADs. 697 * @device: The device to register with. 698 * @port_num: The port on the specified device to use. 699 * @qp_type: Specifies which QP to access. Must be either 700 * IB_QPT_SMI or IB_QPT_GSI. 701 * @mad_reg_req: Specifies which unsolicited MADs should be received 702 * by the caller. This parameter may be NULL if the caller only 703 * wishes to receive solicited responses. 704 * @rmpp_version: If set, indicates that the client will send 705 * and receive MADs that contain the RMPP header for the given version. 706 * If set to 0, indicates that RMPP is not used by this client. 707 * @send_handler: The completion callback routine invoked after a send 708 * request has completed. 709 * @recv_handler: The completion callback routine invoked for a received 710 * MAD. 711 * @context: User specified context associated with the registration. 712 * @registration_flags: Registration flags to set for this agent 713 */ 714 struct ib_mad_agent *ib_register_mad_agent(struct ib_device *device, 715 u8 port_num, 716 enum ib_qp_type qp_type, 717 struct ib_mad_reg_req *mad_reg_req, 718 u8 rmpp_version, 719 ib_mad_send_handler send_handler, 720 ib_mad_recv_handler recv_handler, 721 void *context, 722 u32 registration_flags); 723 724 enum ib_mad_snoop_flags { 725 /*IB_MAD_SNOOP_POSTED_SENDS = 1,*/ 726 /*IB_MAD_SNOOP_RMPP_SENDS = (1<<1),*/ 727 IB_MAD_SNOOP_SEND_COMPLETIONS = (1<<2), 728 /*IB_MAD_SNOOP_RMPP_SEND_COMPLETIONS = (1<<3),*/ 729 IB_MAD_SNOOP_RECVS = (1<<4) 730 /*IB_MAD_SNOOP_RMPP_RECVS = (1<<5),*/ 731 /*IB_MAD_SNOOP_REDIRECTED_QPS = (1<<6)*/ 732 }; 733 734 /** 735 * ib_register_mad_snoop - Register to snoop sent and received MADs. 736 * @device: The device to register with. 737 * @port_num: The port on the specified device to use. 738 * @qp_type: Specifies which QP traffic to snoop. Must be either 739 * IB_QPT_SMI or IB_QPT_GSI. 740 * @mad_snoop_flags: Specifies information where snooping occurs. 741 * @send_handler: The callback routine invoked for a snooped send. 742 * @recv_handler: The callback routine invoked for a snooped receive. 743 * @context: User specified context associated with the registration. 744 */ 745 struct ib_mad_agent *ib_register_mad_snoop(struct ib_device *device, 746 u8 port_num, 747 enum ib_qp_type qp_type, 748 int mad_snoop_flags, 749 ib_mad_snoop_handler snoop_handler, 750 ib_mad_recv_handler recv_handler, 751 void *context); 752 753 /** 754 * ib_unregister_mad_agent - Unregisters a client from using MAD services. 755 * @mad_agent: Corresponding MAD registration request to deregister. 756 * 757 * After invoking this routine, MAD services are no longer usable by the 758 * client on the associated QP. 759 */ 760 void ib_unregister_mad_agent(struct ib_mad_agent *mad_agent); 761 762 /** 763 * ib_post_send_mad - Posts MAD(s) to the send queue of the QP associated 764 * with the registered client. 765 * @send_buf: Specifies the information needed to send the MAD(s). 766 * @bad_send_buf: Specifies the MAD on which an error was encountered. This 767 * parameter is optional if only a single MAD is posted. 768 * 769 * Sent MADs are not guaranteed to complete in the order that they were posted. 770 * 771 * If the MAD requires RMPP, the data buffer should contain a single copy 772 * of the common MAD, RMPP, and class specific headers, followed by the class 773 * defined data. If the class defined data would not divide evenly into 774 * RMPP segments, then space must be allocated at the end of the referenced 775 * buffer for any required padding. To indicate the amount of class defined 776 * data being transferred, the paylen_newwin field in the RMPP header should 777 * be set to the size of the class specific header plus the amount of class 778 * defined data being transferred. The paylen_newwin field should be 779 * specified in network-byte order. 780 */ 781 int ib_post_send_mad(struct ib_mad_send_buf *send_buf, 782 struct ib_mad_send_buf **bad_send_buf); 783 784 785 /** 786 * ib_free_recv_mad - Returns data buffers used to receive a MAD. 787 * @mad_recv_wc: Work completion information for a received MAD. 788 * 789 * Clients receiving MADs through their ib_mad_recv_handler must call this 790 * routine to return the work completion buffers to the access layer. 791 */ 792 void ib_free_recv_mad(struct ib_mad_recv_wc *mad_recv_wc); 793 794 /** 795 * ib_cancel_mad - Cancels an outstanding send MAD operation. 796 * @mad_agent: Specifies the registration associated with sent MAD. 797 * @send_buf: Indicates the MAD to cancel. 798 * 799 * MADs will be returned to the user through the corresponding 800 * ib_mad_send_handler. 801 */ 802 void ib_cancel_mad(struct ib_mad_agent *mad_agent, 803 struct ib_mad_send_buf *send_buf); 804 805 /** 806 * ib_modify_mad - Modifies an outstanding send MAD operation. 807 * @mad_agent: Specifies the registration associated with sent MAD. 808 * @send_buf: Indicates the MAD to modify. 809 * @timeout_ms: New timeout value for sent MAD. 810 * 811 * This call will reset the timeout value for a sent MAD to the specified 812 * value. 813 */ 814 int ib_modify_mad(struct ib_mad_agent *mad_agent, 815 struct ib_mad_send_buf *send_buf, u32 timeout_ms); 816 817 /** 818 * ib_redirect_mad_qp - Registers a QP for MAD services. 819 * @qp: Reference to a QP that requires MAD services. 820 * @rmpp_version: If set, indicates that the client will send 821 * and receive MADs that contain the RMPP header for the given version. 822 * If set to 0, indicates that RMPP is not used by this client. 823 * @send_handler: The completion callback routine invoked after a send 824 * request has completed. 825 * @recv_handler: The completion callback routine invoked for a received 826 * MAD. 827 * @context: User specified context associated with the registration. 828 * 829 * Use of this call allows clients to use MAD services, such as RMPP, 830 * on user-owned QPs. After calling this routine, users may send 831 * MADs on the specified QP by calling ib_mad_post_send. 832 */ 833 struct ib_mad_agent *ib_redirect_mad_qp(struct ib_qp *qp, 834 u8 rmpp_version, 835 ib_mad_send_handler send_handler, 836 ib_mad_recv_handler recv_handler, 837 void *context); 838 839 /** 840 * ib_process_mad_wc - Processes a work completion associated with a 841 * MAD sent or received on a redirected QP. 842 * @mad_agent: Specifies the registered MAD service using the redirected QP. 843 * @wc: References a work completion associated with a sent or received 844 * MAD segment. 845 * 846 * This routine is used to complete or continue processing on a MAD request. 847 * If the work completion is associated with a send operation, calling 848 * this routine is required to continue an RMPP transfer or to wait for a 849 * corresponding response, if it is a request. If the work completion is 850 * associated with a receive operation, calling this routine is required to 851 * process an inbound or outbound RMPP transfer, or to match a response MAD 852 * with its corresponding request. 853 */ 854 int ib_process_mad_wc(struct ib_mad_agent *mad_agent, 855 struct ib_wc *wc); 856 857 /** 858 * ib_create_send_mad - Allocate and initialize a data buffer and work request 859 * for sending a MAD. 860 * @mad_agent: Specifies the registered MAD service to associate with the MAD. 861 * @remote_qpn: Specifies the QPN of the receiving node. 862 * @pkey_index: Specifies which PKey the MAD will be sent using. This field 863 * is valid only if the remote_qpn is QP 1. 864 * @rmpp_active: Indicates if the send will enable RMPP. 865 * @hdr_len: Indicates the size of the data header of the MAD. This length 866 * should include the common MAD header, RMPP header, plus any class 867 * specific header. 868 * @data_len: Indicates the size of any user-transferred data. The call will 869 * automatically adjust the allocated buffer size to account for any 870 * additional padding that may be necessary. 871 * @gfp_mask: GFP mask used for the memory allocation. 872 * @base_version: Base Version of this MAD 873 * 874 * This routine allocates a MAD for sending. The returned MAD send buffer 875 * will reference a data buffer usable for sending a MAD, along 876 * with an initialized work request structure. Users may modify the returned 877 * MAD data buffer before posting the send. 878 * 879 * The returned MAD header, class specific headers, and any padding will be 880 * cleared. Users are responsible for initializing the common MAD header, 881 * any class specific header, and MAD data area. 882 * If @rmpp_active is set, the RMPP header will be initialized for sending. 883 */ 884 struct ib_mad_send_buf *ib_create_send_mad(struct ib_mad_agent *mad_agent, 885 u32 remote_qpn, u16 pkey_index, 886 int rmpp_active, 887 int hdr_len, int data_len, 888 gfp_t gfp_mask, 889 u8 base_version); 890 891 /** 892 * ib_is_mad_class_rmpp - returns whether given management class 893 * supports RMPP. 894 * @mgmt_class: management class 895 * 896 * This routine returns whether the management class supports RMPP. 897 */ 898 int ib_is_mad_class_rmpp(u8 mgmt_class); 899 900 /** 901 * ib_get_mad_data_offset - returns the data offset for a given 902 * management class. 903 * @mgmt_class: management class 904 * 905 * This routine returns the data offset in the MAD for the management 906 * class requested. 907 */ 908 int ib_get_mad_data_offset(u8 mgmt_class); 909 910 /** 911 * ib_get_rmpp_segment - returns the data buffer for a given RMPP segment. 912 * @send_buf: Previously allocated send data buffer. 913 * @seg_num: number of segment to return 914 * 915 * This routine returns a pointer to the data buffer of an RMPP MAD. 916 * Users must provide synchronization to @send_buf around this call. 917 */ 918 void *ib_get_rmpp_segment(struct ib_mad_send_buf *send_buf, int seg_num); 919 920 /** 921 * ib_free_send_mad - Returns data buffers used to send a MAD. 922 * @send_buf: Previously allocated send data buffer. 923 */ 924 void ib_free_send_mad(struct ib_mad_send_buf *send_buf); 925 926 /** 927 * ib_mad_kernel_rmpp_agent - Returns if the agent is performing RMPP. 928 * @agent: the agent in question 929 * @return: true if agent is performing rmpp, false otherwise. 930 */ 931 int ib_mad_kernel_rmpp_agent(const struct ib_mad_agent *agent); 932 933 #endif /* IB_MAD_H */ 934