/linux/drivers/infiniband/hw/mthca/ |
H A D | mthca_mad.c | 114 if ((mad->mad_hdr.mgmt_class == IB_MGMT_CLASS_SUBN_LID_ROUTED || in smp_snoop() 115 mad->mad_hdr.mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) && in smp_snoop() 153 if ((mad->mad_hdr.mgmt_class == IB_MGMT_CLASS_SUBN_LID_ROUTED || in node_desc_override() 154 mad->mad_hdr.mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) && in node_desc_override() 168 int qpn = mad->mad_hdr.mgmt_class != IB_MGMT_CLASS_SUBN_LID_ROUTED; in forward_trap() 221 if (in->mad_hdr.mgmt_class == IB_MGMT_CLASS_SUBN_LID_ROUTED || in mthca_process_mad() 222 in->mad_hdr.mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) { in mthca_process_mad() 236 } else if (in->mad_hdr.mgmt_class == IB_MGMT_CLASS_PERF_MGMT || in mthca_process_mad() 237 in->mad_hdr.mgmt_class == MTHCA_VENDOR_CLASS1 || in mthca_process_mad() 238 in->mad_hdr.mgmt_class == MTHCA_VENDOR_CLASS2) { in mthca_process_mad() [all …]
|
/linux/drivers/infiniband/core/ |
H A D | mad.c | 104 u8 mgmt_class); 144 static inline u8 convert_mgmt_class(u8 mgmt_class) in convert_mgmt_class() argument 147 return mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE ? in convert_mgmt_class() 148 0 : mgmt_class; in convert_mgmt_class() 163 static int vendor_class_index(u8 mgmt_class) in vendor_class_index() argument 165 return mgmt_class - IB_MGMT_CLASS_VENDOR_RANGE2_START; in vendor_class_index() 168 static int is_vendor_class(u8 mgmt_class) in is_vendor_class() argument 170 if ((mgmt_class < IB_MGMT_CLASS_VENDOR_RANGE2_START) || in is_vendor_class() 171 (mgmt_class > IB_MGMT_CLASS_VENDOR_RANGE2_END)) in is_vendor_class() 208 ((hdr->mgmt_class == IB_MGMT_CLASS_BM) && in ib_response_mad() [all …]
|
H A D | user_mad.c | 341 offset = ib_get_mad_data_offset(recv_buf->mad->mad_hdr.mgmt_class); in copy_recv_mad() 485 (hdr->mgmt_class != sent_hdr->mgmt_class)) in is_duplicate() 578 hdr_len = ib_get_mad_data_offset(rmpp_mad_hdr->mad_hdr.mgmt_class); in ib_umad_write() 580 if (ib_is_mad_class_rmpp(rmpp_mad_hdr->mad_hdr.mgmt_class) in ib_umad_write() 636 && ib_is_mad_class_rmpp(rmpp_mad_hdr->mad_hdr.mgmt_class) in ib_umad_write() 734 if (ureq.mgmt_class) { in ib_umad_reg_agent() 736 req.mgmt_class = ureq.mgmt_class; in ib_umad_reg_agent() 754 ureq.mgmt_class ? &req : NULL, in ib_umad_reg_agent() 847 if (ureq.mgmt_class) { in ib_umad_reg_agent2() 849 req.mgmt_class = ureq.mgmt_class; in ib_umad_reg_agent2() [all …]
|
H A D | mad_rmpp.c | 67 u8 mgmt_class; member 136 hdr_len = ib_get_mad_data_offset(recv_wc->recv_buf.mad->mad_hdr.mgmt_class); in ack_recv() 163 hdr_len = ib_get_mad_data_offset(recv_wc->recv_buf.mad->mad_hdr.mgmt_class); in alloc_response_msg() 312 rmpp_recv->mgmt_class = mad_hdr->mgmt_class; in create_rmpp_recv() 333 rmpp_recv->mgmt_class == mad_hdr->mgmt_class && in find_rmpp_recv() 436 hdr_size = ib_get_mad_data_offset(rmpp_mad->mad_hdr.mgmt_class); in get_mad_len() 857 rmpp_recv->mgmt_class != mad_hdr->mgmt_class || in init_newwin()
|
H A D | sysfs.c | 578 in_mad->mad_hdr.mgmt_class = IB_MGMT_CLASS_PERF_MGMT; in get_perf_mad()
|
H A D | sa_query.c | 1327 mad->mad_hdr.mgmt_class = IB_MGMT_CLASS_SUBN_ADM; in init_mad()
|
/linux/include/trace/events/ |
H A D | ib_mad.h | 30 __field(u8, mgmt_class) 68 __entry->mgmt_class = 69 ((struct ib_mad_hdr *)wr->send_buf.mad)->mgmt_class; 96 __entry->base_version, __entry->mgmt_class, 127 __field(u8, mgmt_class) 155 __entry->mgmt_class = 156 ((struct ib_mad_hdr *)wr->send_buf.mad)->mgmt_class; 177 __entry->base_version, __entry->mgmt_class, 190 __field(u8, mgmt_class) 214 __entry->mgmt_class = mad_hdr->mgmt_class; [all …]
|
H A D | ib_umad.h | 39 __field(u8, mgmt_class) 75 __entry->mgmt_class = mad_hdr->mgmt_class; 100 __entry->base_version, __entry->mgmt_class,
|
/linux/samples/bpf/ |
H A D | ibumad_kern.c | 64 u8 mgmt_class; member 81 u8 class = ctx->mgmt_class; in on_ib_umad_read_recv() 101 u8 class = ctx->mgmt_class; in on_ib_umad_read_send() 121 u8 class = ctx->mgmt_class; in on_ib_umad_write()
|
/linux/include/rdma/ |
H A D | opa_smi.h | 24 u8 mgmt_class; member 102 if (smp->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) in opa_get_smp_data() 110 if (smp->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) in opa_get_smp_data_size() 118 if (smp->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) in opa_get_smp_header_size()
|
H A D | ib_mad.h | 139 u8 mgmt_class; member 648 u8 mgmt_class; member 786 int ib_is_mad_class_rmpp(u8 mgmt_class); 796 int ib_get_mad_data_offset(u8 mgmt_class);
|
H A D | ib_smi.h | 20 u8 mgmt_class; member 154 mad->mgmt_class = IB_MGMT_CLASS_SUBN_LID_ROUTED; in ib_init_query_mad()
|
/linux/include/uapi/rdma/ |
H A D | ib_user_mad.h | 189 __u8 mgmt_class; member 224 __u8 mgmt_class; member
|
/linux/drivers/infiniband/hw/mlx4/ |
H A D | mad.c | 231 if ((mad->mad_hdr.mgmt_class == IB_MGMT_CLASS_SUBN_LID_ROUTED || in smp_snoop() 232 mad->mad_hdr.mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) && in smp_snoop() 372 if ((mad->mad_hdr.mgmt_class == IB_MGMT_CLASS_SUBN_LID_ROUTED || in node_desc_override() 373 mad->mad_hdr.mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) && in node_desc_override() 386 int qpn = mad->mad_hdr.mgmt_class != IB_MGMT_CLASS_SUBN_LID_ROUTED; in forward_trap() 707 if (mad->mad_hdr.mgmt_class != IB_MGMT_CLASS_CM) { in mlx4_ib_demux_mad() 770 switch (mad->mad_hdr.mgmt_class) { in mlx4_ib_demux_mad() 780 slave, mad->mad_hdr.mgmt_class, in mlx4_ib_demux_mad() 803 "for slave:%d\n", mad->mad_hdr.mgmt_class, slave); in mlx4_ib_demux_mad() 837 if (in_mad->mad_hdr.mgmt_class == IB_MGMT_CLASS_SUBN_LID_ROUTED || in ib_process_mad() [all …]
|
H A D | mcg.c | 287 mad.mad_hdr.mgmt_class = IB_MGMT_CLASS_SUBN_ADM; in send_leave_to_wire() 328 mad.mad_hdr.mgmt_class = IB_MGMT_CLASS_SUBN_ADM; in send_reply_to_slave()
|
/linux/drivers/infiniband/hw/qib/ |
H A D | qib_mad.h | 121 u8 mgmt_class; member
|
H A D | qib_mad.c | 47 if (smp->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) in reply() 59 if (smp->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) in reply_failure() 94 smp->mgmt_class = IB_MGMT_CLASS_SUBN_LID_ROUTED; in qib_send_trap() 184 if (smp->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) { in qib_bad_mkey() 2372 switch (in->mad_hdr.mgmt_class) { in qib_process_mad()
|
H A D | qib_ud.c | 500 smp->mgmt_class != IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) in qib_ud_rcv()
|
/linux/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_ah.c | 261 if (in->mad_hdr.mgmt_class == IB_MGMT_CLASS_PERF_MGMT) { in ocrdma_process_mad()
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | ud.c | 746 if (smp->mgmt_class != IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE && in opa_smp_check() 747 smp->mgmt_class != IB_MGMT_CLASS_SUBN_LID_ROUTED) { in opa_smp_check() 911 smp->mgmt_class != IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) in hfi1_ud_rcv()
|
H A D | mad.c | 46 if (smp->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) in reply() 312 smp->mgmt_class = IB_MGMT_CLASS_SUBN_LID_ROUTED; in send_trap() 455 if (mad->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) { in bad_mkey() 4359 if (smp->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) { in is_local_mad() 4797 switch (in_mad->mad_hdr.mgmt_class) { in hfi1_process_opa_mad() 4838 switch (in_mad->mad_hdr.mgmt_class) { in hfi1_process_ib_mad()
|
/linux/drivers/infiniband/ulp/opa_vnic/ |
H A D | opa_vnic_vema.c | 809 trap_mad->mad_hdr.mgmt_class = OPA_MGMT_CLASS_INTEL_EMA; in opa_vnic_vema_send_trap() 925 .mgmt_class = OPA_MGMT_CLASS_INTEL_EMA, in vema_register()
|
/linux/drivers/net/ethernet/mellanox/mlx4/ |
H A D | cmd.c | 919 smp->mgmt_class == IB_MGMT_CLASS_SUBN_LID_ROUTED && in mlx4_MAD_IFC_wrapper() 1010 if (!(smp->mgmt_class == IB_MGMT_CLASS_SUBN_LID_ROUTED && in mlx4_MAD_IFC_wrapper() 1013 slave, smp->mgmt_class, smp->method, in mlx4_MAD_IFC_wrapper()
|
H A D | port.c | 2052 inmad->mgmt_class = 0x1; in mlx4_get_module_id() 2170 inmad->mgmt_class = 0x1; in mlx4_get_module_info()
|
/linux/include/linux/mlx4/ |
H A D | device.h | 1006 u8 mgmt_class; member
|