/linux/include/rdma/ |
H A D | ib_sa.h | 135 __be16 slid; member 153 __be32 slid; member 223 (be32_to_cpu(opa->opa.slid) >= in path_conv_opa_to_ib() 231 = OPA_MAKE_ID(be32_to_cpu(opa->opa.slid)); in path_conv_opa_to_ib() 236 ib->ib.slid = 0; in path_conv_opa_to_ib() 239 ib->ib.slid = htons(ntohl(opa->opa.slid)); in path_conv_opa_to_ib() 248 __be32 slid, dlid; in path_conv_ib_to_opa() local 252 slid = htonl(opa_get_lid_from_gid(&ib->sgid)); in path_conv_ib_to_opa() 255 slid = htonl(ntohs(ib->ib.slid)); in path_conv_ib_to_opa() 258 opa->opa.slid = slid; in path_conv_ib_to_opa() [all …]
|
H A D | opa_addr.h | 58 static inline bool opa_is_extended_lid(__be32 dlid, __be32 slid) in opa_is_extended_lid() argument 62 (be32_to_cpu(slid) >= in opa_is_extended_lid()
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | ud.c | 71 u32 slid; in ud_loopback() local 75 slid = ppd->lid | (rdma_ah_get_path_bits(ah_attr) & in ud_loopback() 79 slid, false))) { in ud_loopback() 83 slid, rdma_ah_get_dlid(ah_attr)); in ud_loopback() 207 wc.slid = (ppd->lid | (rdma_ah_get_path_bits(ah_attr) & in ud_loopback() 210 if (wc.slid == 0 && sqp->ibqp.qp_type == IB_QPT_GSI) in ud_loopback() 211 wc.slid = be16_to_cpu(IB_LID_PERMISSIVE); in ud_loopback() 265 u16 len, slid, dlid, pkey; in hfi1_make_ud_req_9B() local 311 slid = be16_to_cpu(IB_LID_PERMISSIVE); in hfi1_make_ud_req_9B() 318 slid = lid; in hfi1_make_ud_req_9B() [all …]
|
H A D | ruc.c | 34 u32 slid = packet->slid; in hfi1_ruc_check_hdr() local 63 sc5, slid))) { in hfi1_ruc_check_hdr() 65 slid, dlid); in hfi1_ruc_check_hdr() 69 if (slid != rdma_ah_get_dlid(&qp->alt_ah_attr) || in hfi1_ruc_check_hdr() 100 sc5, slid))) { in hfi1_ruc_check_hdr() 102 slid, dlid); in hfi1_ruc_check_hdr() 106 if ((slid != rdma_ah_get_dlid(&qp->remote_ah_attr)) || in hfi1_ruc_check_hdr() 240 u32 slid; in hfi1_make_ruc_header_16B() local 291 slid = be32_to_cpu(OPA_LID_PERMISSIVE); in hfi1_make_ruc_header_16B() 293 slid = ppd->lid | in hfi1_make_ruc_header_16B() [all …]
|
H A D | trace_ibhdrs.h | 80 u16 *len, u32 *dlid, u32 *slid); 89 u32 *dlid, u32 *slid); 95 u16 len, u16 pkey, u32 dlid, u32 slid); 144 __field(u32, slid) 170 &__entry->slid); 196 &__entry->slid); 240 __entry->slid), 303 __field(u32, slid) 330 &__entry->slid); 361 &__entry->slid); [all …]
|
H A D | trace.c | 157 u16 *len, u32 *dlid, u32 *slid) in hfi1_trace_parse_9b_hdr() argument 165 *slid = ib_get_slid(hdr); in hfi1_trace_parse_9b_hdr() 172 u32 *dlid, u32 *slid) in hfi1_trace_parse_16b_hdr() argument 184 *slid = hfi1_16B_get_slid(hdr); in hfi1_trace_parse_16b_hdr() 195 u16 len, u16 pkey, u32 dlid, u32 slid) in hfi1_trace_fmt_lrh() argument 199 trace_seq_printf(p, LRH_PRN, len, sc, dlid, slid); in hfi1_trace_fmt_lrh()
|
H A D | ipoib_tx.c | 267 u16 slid; in hfi1_ipoib_build_ib_tx_headers() local 305 slid = be16_to_cpu(IB_LID_PERMISSIVE); in hfi1_ipoib_build_ib_tx_headers() 312 slid = lid; in hfi1_ipoib_build_ib_tx_headers() 314 slid = be16_to_cpu(IB_LID_PERMISSIVE); in hfi1_ipoib_build_ib_tx_headers() 323 hfi1_make_ib_hdr(&sdma_hdr->hdr.ibh, lrh0, dwords, dlid, slid); in hfi1_ipoib_build_ib_tx_headers()
|
H A D | driver.c | 410 u32 rlid, slid, dlid = 0; in hfi1_process_ecn_slowpath() local 420 slid = hfi1_16B_get_slid(pkt->hdr); in hfi1_process_ecn_slowpath() 431 slid = ib_get_slid(pkt->hdr); in hfi1_process_ecn_slowpath() 442 rlid = slid; in hfi1_process_ecn_slowpath() 448 rlid = slid; in hfi1_process_ecn_slowpath() 1449 if ((!packet->slid) || (!packet->dlid)) in hfi1_bypass_ingress_pkt_check() 1503 packet->slid = ib_get_slid(hdr); in hfi1_setup_9B_packet() 1584 packet->slid = hfi1_16B_get_slid(packet->hdr); in hfi1_setup_bypass_packet()
|
H A D | verbs.c | 450 packet->slid, true); in hfi1_do_pkey_check() 1116 int egress_pkey_check(struct hfi1_pportdata *ppd, u32 slid, u16 pkey, in egress_pkey_check() argument 1160 dd->err_info_xmit_constraint.slid = slid; in egress_pkey_check() 1221 u32 slid; in hfi1_verbs_send() local 1234 slid = hfi1_16B_get_slid(hdr); in hfi1_verbs_send() 1244 slid = ib_get_slid(hdr); in hfi1_verbs_send() 1254 ret = egress_pkey_check(dd->pport, slid, pkey, in hfi1_verbs_send()
|
H A D | user_sdma.c | 293 u32 slid; in hfi1_user_sdma_process_request() local 428 slid = be16_to_cpu(req->hdr.lrh[3]); in hfi1_user_sdma_process_request() 429 if (egress_pkey_check(dd->pport, slid, pkey, sc, PKEY_CHECK_INVALID)) { in hfi1_user_sdma_process_request()
|
H A D | mad.c | 2516 __be32 slid; member 2524 __be32 slid; member 3405 rsp->port_xmit_constraint_ei.slid = in pma_get_opa_errorinfo() 3406 cpu_to_be32(dd->err_info_xmit_constraint.slid); in pma_get_opa_errorinfo() 3412 rsp->port_rcv_constraint_ei.slid = in pma_get_opa_errorinfo() 3413 cpu_to_be32(dd->err_info_rcv_constraint.slid); in pma_get_opa_errorinfo() 4365 return (in_wc->slid == ppd->lid); in is_local_mad() 4408 ingress_pkey_table_fail(ppd, pkey, in_wc->slid); in opa_local_smp_check()
|
/linux/drivers/infiniband/core/ |
H A D | uverbs_marshall.c | 143 dst->slid = htons(ntohl(sa_path_get_slid(src))); in __ib_copy_path_rec_to_user() 178 u32 slid, dlid; in ib_copy_path_rec_from_user() local 184 slid = opa_get_lid_from_gid((union ib_gid *)src->sgid); in ib_copy_path_rec_from_user() 188 slid = ntohs(src->slid); in ib_copy_path_rec_from_user() 195 sa_path_set_slid(dst, slid); in ib_copy_path_rec_from_user()
|
H A D | mad_rmpp.c | 66 u32 slid; member 311 rmpp_recv->slid = mad_recv_wc->wc->slid; in create_rmpp_recv() 332 rmpp_recv->slid == mad_recv_wc->wc->slid && in find_rmpp_recv() 865 if (rmpp_recv->slid == rdma_ah_get_dlid(&ah_attr)) { in init_newwin()
|
/linux/drivers/infiniband/hw/mthca/ |
H A D | mthca_mad.c | 205 u16 slid = in_wc ? ib_lid_cpu16(in_wc->slid) : be16_to_cpu(IB_LID_PERMISSIVE); in mthca_process_mad() local 210 if (in->mad_hdr.method == IB_MGMT_METHOD_TRAP && !slid) { in mthca_process_mad()
|
/linux/include/trace/events/ |
H A D | ib_mad.h | 204 __field(u32, slid) 222 __entry->slid = wc->slid; 242 __entry->slid, __entry->src_qp, __entry->sl
|
/linux/include/uapi/rdma/ |
H A D | ib_user_sa.h | 59 __be16 slid; member
|
H A D | vmw_pvrdma-abi.h | 301 __u16 slid; member
|
/linux/drivers/infiniband/ulp/opa_vnic/ |
H A D | opa_vnic_encap.c | 73 static inline void opa_vnic_make_header(u8 *hdr, u32 slid, u32 dlid, u16 len, in opa_vnic_make_header() argument 85 h[0] |= (slid & OPA_16B_LID_MASK); in opa_vnic_make_header() 86 h[2] |= ((slid >> (20 - OPA_16B_SLID_HIGH_SHFT)) & OPA_16B_SLID_MASK); in opa_vnic_make_header()
|
/linux/tools/testing/selftests/bpf/progs/ |
H A D | btf_dump_test_case_padding.c | 183 int slid; member
|
/linux/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_cq.c | 366 wc->slid = cqe->slid; in pvrdma_poll_one()
|
/linux/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 597 wc->slid = 0; in use_tunnel_data() 603 wc->slid = be16_to_cpu(hdr->tun.slid_mac_47_32); in use_tunnel_data() 857 wc->slid = 0; in mlx4_ib_poll_one() 869 wc->slid = be16_to_cpu(cqe->rlid); in mlx4_ib_poll_one()
|
H A D | mad.c | 172 in_modifier |= ib_lid_cpu16(in_wc->slid) << 16; in mlx4_MAD_IFC() 646 tun_mad->hdr.slid_mac_47_32 = ib_lid_be16(wc->slid); in mlx4_ib_send_to_slave() 826 u16 slid, prev_lid = 0; in ib_process_mad() local 830 slid = in_wc ? ib_lid_cpu16(in_wc->slid) : be16_to_cpu(IB_LID_PERMISSIVE); in ib_process_mad() 832 if (in_mad->mad_hdr.method == IB_MGMT_METHOD_TRAP && slid == 0) { in ib_process_mad()
|
/linux/drivers/infiniband/sw/rdmavt/ |
H A D | cq.c | 91 uqueue[head].slid = ib_lid_cpu16(entry->slid); in rvt_cq_enter()
|
/linux/drivers/infiniband/hw/qib/ |
H A D | qib_ud.c | 208 wc.slid = ppd->lid | (rdma_ah_get_path_bits(ah_attr) & in qib_ud_loopback() 568 wc.slid = be16_to_cpu(hdr->lrh[3]); in qib_ud_rcv()
|
/linux/drivers/infiniband/hw/mlx5/ |
H A D | cq.c | 243 wc->slid = be16_to_cpu(cqe->slid); in handle_responder() 248 wc->slid = 0; in handle_responder()
|