| /linux/Documentation/translations/zh_CN/infiniband/ |
| H A D | user_mad.rst | 57 要的缓冲区长度被设置在mad.length中。 61 struct ib_user_mad *mad; 62 mad = malloc(sizeof *mad + 256); 63 ret = read(fd, mad, sizeof *mad + 256); 64 if (ret != sizeof mad + 256) { 66 free(mad); 71 struct ib_user_mad *mad; 72 mad = malloc(sizeof *mad + 256); 73 ret = read(fd, mad, sizeof *mad + 256); 75 length = mad.length; [all …]
|
| /linux/drivers/infiniband/hw/mlx4/ |
| H A D | cm.c | 85 static void set_local_comm_id(struct ib_mad *mad, u32 cm_id) in set_local_comm_id() argument 87 if (mad->mad_hdr.attr_id == CM_SIDR_REQ_ATTR_ID) { in set_local_comm_id() 89 (struct cm_sidr_generic_msg *)mad; in set_local_comm_id() 91 } else if (mad->mad_hdr.attr_id == CM_SIDR_REP_ATTR_ID) { in set_local_comm_id() 95 struct cm_generic_msg *msg = (struct cm_generic_msg *)mad; in set_local_comm_id() 100 static u32 get_local_comm_id(struct ib_mad *mad) in get_local_comm_id() argument 102 if (mad->mad_hdr.attr_id == CM_SIDR_REQ_ATTR_ID) { in get_local_comm_id() 104 (struct cm_sidr_generic_msg *)mad; in get_local_comm_id() 106 } else if (mad->mad_hdr.attr_id == CM_SIDR_REP_ATTR_ID) { in get_local_comm_id() 110 struct cm_generic_msg *msg = (struct cm_generic_msg *)mad; in get_local_comm_id() [all …]
|
| H A D | mcg.c | 209 static int send_mad_to_wire(struct mlx4_ib_demux_ctx *ctx, struct ib_mad *mad) in send_mad_to_wire() argument 225 &ah_attr, NULL, 0xffff, mad); in send_mad_to_wire() 229 struct ib_mad *mad) in send_mad_to_slave() argument 249 return mlx4_ib_send_to_slave(dev, slave, ctx->port, IB_QPT_GSI, &wc, NULL, mad); in send_mad_to_slave() 254 struct ib_sa_mad mad; in send_join_to_wire() local 255 struct ib_sa_mcmember_data *sa_mad_data = (struct ib_sa_mcmember_data *)&mad.data; in send_join_to_wire() 259 memcpy(&mad, sa_mad, sizeof mad); in send_join_to_wire() 265 mad.mad_hdr.tid = mlx4_ib_get_new_demux_tid(group->demux); in send_join_to_wire() 266 group->last_req_tid = mad.mad_hdr.tid; /* keep it for later validation */ in send_join_to_wire() 268 ret = send_mad_to_wire(group->demux, (struct ib_mad *)&mad); in send_join_to_wire() [all …]
|
| H A D | Makefile | 4 mlx4_ib-y := ah.o cq.o doorbell.o mad.o main.o mr.o qp.o srq.o mcg.o cm.o alias_GUID.o sysfs.o
|
| /linux/drivers/iio/adc/ |
| H A D | mt6360-adc.c | 66 static int mt6360_adc_read_channel(struct mt6360_adc_data *mad, int channel, int *val) in mt6360_adc_read_channel() argument 74 mutex_lock(&mad->adc_lock); in mt6360_adc_read_channel() 77 ret = regmap_update_bits(mad->regmap, MT6360_REG_PMUADCRPT1, MT6360_PREFERCH_MASK, in mt6360_adc_read_channel() 83 ret = regmap_raw_write(mad->regmap, MT6360_REG_PMUADCCFG, &adc_enable, sizeof(adc_enable)); in mt6360_adc_read_channel() 87 predict_end_t = ktime_add_ms(mad->last_off_timestamps[channel], 2 * ADC_WAIT_TIME_MS); in mt6360_adc_read_channel() 101 ret = regmap_raw_read(mad->regmap, MT6360_REG_PMUADCRPT1, rpt, sizeof(rpt)); in mt6360_adc_read_channel() 133 regmap_raw_write(mad->regmap, MT6360_REG_PMUADCCFG, &adc_enable, sizeof(adc_enable)); in mt6360_adc_read_channel() 134 mad->last_off_timestamps[channel] = ktime_get(); in mt6360_adc_read_channel() 136 regmap_update_bits(mad->regmap, MT6360_REG_PMUADCRPT1, MT6360_PREFERCH_MASK, in mt6360_adc_read_channel() 139 mutex_unlock(&mad->adc_lock); in mt6360_adc_read_channel() [all …]
|
| /linux/Documentation/infiniband/ |
| H A D | user_mad.rst | 48 buffer needed is set in mad.length. 52 struct ib_user_mad *mad; 53 mad = malloc(sizeof *mad + 256); 54 ret = read(fd, mad, sizeof *mad + 256); 55 if (ret != sizeof mad + 256) { 57 free(mad); 62 struct ib_user_mad *mad; 63 mad = malloc(sizeof *mad + 256); 64 ret = read(fd, mad, sizeof *mad + 256); 66 length = mad.length; [all …]
|
| /linux/drivers/infiniband/hw/mthca/ |
| H A D | mthca_mad.c | 109 const struct ib_mad *mad, in smp_snoop() argument 114 if ((mad->mad_hdr.mgmt_class == IB_MGMT_CLASS_SUBN_LID_ROUTED || in smp_snoop() 115 mad->mad_hdr.mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) && in smp_snoop() 116 mad->mad_hdr.method == IB_MGMT_METHOD_SET) { in smp_snoop() 117 if (mad->mad_hdr.attr_id == IB_SMP_ATTR_PORT_INFO) { in smp_snoop() 119 (struct ib_port_info *) ((struct ib_smp *) mad)->data; in smp_snoop() 141 if (mad->mad_hdr.attr_id == IB_SMP_ATTR_PKEY_TABLE) { in smp_snoop() 151 struct ib_mad *mad) in node_desc_override() argument 153 if ((mad->mad_hdr.mgmt_class == IB_MGMT_CLASS_SUBN_LID_ROUTED || in node_desc_override() 154 mad->mad_hdr.mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) && in node_desc_override() [all …]
|
| /linux/drivers/infiniband/core/ |
| H A D | user_mad.c | 134 struct ib_user_mad mad; member 197 for (packet->mad.hdr.id = 0; in queue_packet() 198 packet->mad.hdr.id < IB_UMAD_MAX_AGENTS; in queue_packet() 199 packet->mad.hdr.id++) in queue_packet() 200 if (agent == __get_agent(file, packet->mad.hdr.id)) { in queue_packet() 233 packet->mad.hdr.status = ETIMEDOUT; in send_handler() 257 packet->mad.hdr.status = 0; in recv_handler() 258 packet->mad.hdr.length = hdr_size(file) + mad_recv_wc->mad_len; in recv_handler() 259 packet->mad.hdr.qpn = cpu_to_be32(mad_recv_wc->wc->src_qp); in recv_handler() 265 packet->mad.hdr.lid = ib_lid_be16(0xFFFF & in recv_handler() [all …]
|
| H A D | cm.c | 1583 req_msg = (struct cm_req_msg *)msg->mad; in ib_send_cm_req() 1621 rcv_msg = (struct cm_rej_msg *) mad_recv_wc->recv_buf.mad; in cm_issue_rej() 1622 rej_msg = (struct cm_rej_msg *) msg->mad; in cm_issue_rej() 1825 req_msg = (struct cm_req_msg *)work->mad_recv_wc->recv_buf.mad; in cm_format_req_event() 1976 cm_format_mra((struct cm_mra_msg *) msg->mad, cm_id_priv, in cm_dup_req_handler() 1982 cm_format_rej((struct cm_rej_msg *)msg->mad, cm_id_priv, in cm_dup_req_handler() 2008 req_msg = (struct cm_req_msg *)work->mad_recv_wc->recv_buf.mad; in cm_match_req() 2102 req_msg = (struct cm_req_msg *)work->mad_recv_wc->recv_buf.mad; in cm_req_handler() 2318 rep_msg = (struct cm_rep_msg *) msg->mad; in ib_send_cm_rep() 2393 cm_format_rtu((struct cm_rtu_msg *) msg->mad, cm_id_priv, in ib_send_cm_rtu() [all …]
|
| /linux/drivers/scsi/ibmvscsi/ |
| H A D | ibmvfc.c | 224 struct ibmvfc_mad_common *mad = &evt->iu.mad_common; in ibmvfc_trc_start() local 244 entry->op_code = be32_to_cpu(mad->opcode); in ibmvfc_trc_start() 260 struct ibmvfc_mad_common *mad = &evt->xfer_iu->mad_common; in ibmvfc_trc_end() local 285 entry->op_code = be32_to_cpu(mad->opcode); in ibmvfc_trc_end() 286 entry->u.end.status = be16_to_cpu(mad->status); in ibmvfc_trc_end() 2189 struct ibmvfc_passthru_mad *mad; in ibmvfc_bsg_request() local 2273 mad = &evt->iu.passthru; in ibmvfc_bsg_request() 2275 memset(mad, 0, sizeof(*mad)); in ibmvfc_bsg_request() 2276 mad->common.version = cpu_to_be32(1); in ibmvfc_bsg_request() 2277 mad->common.opcode = cpu_to_be32(IBMVFC_PASSTHRU); in ibmvfc_bsg_request() [all …]
|
| H A D | ibmvscsi.c | 1240 if (evt_struct->xfer_iu->mad.capabilities.common.status) { in capabilities_rsp() 1242 evt_struct->xfer_iu->mad.capabilities.common.status); in capabilities_rsp() 1279 req = &evt_struct->iu.mad.capabilities; in send_mad_capabilities() 1335 u16 status = be16_to_cpu(evt_struct->xfer_iu->mad.fast_fail.common.status); in fast_fail_rsp() 1370 fast_fail_mad = &evt_struct->iu.mad.fast_fail; in enable_fast_fail() 1392 if (evt_struct->xfer_iu->mad.adapter_info.common.status) { in adapter_info_rsp() 1394 evt_struct->xfer_iu->mad.adapter_info.common.status); in adapter_info_rsp() 1449 req = &evt_struct->iu.mad.adapter_info; in send_mad_adapter_info()
|
| /linux/include/trace/events/ |
| H A D | ib_mad.h | 67 ((struct ib_mad_hdr *)wr->send_buf.mad)->base_version; 69 ((struct ib_mad_hdr *)wr->send_buf.mad)->mgmt_class; 71 ((struct ib_mad_hdr *)wr->send_buf.mad)->class_version; 73 ((struct ib_mad_hdr *)wr->send_buf.mad)->method; 75 ((struct ib_mad_hdr *)wr->send_buf.mad)->status; 77 ((struct ib_mad_hdr *)wr->send_buf.mad)->class_specific; 78 __entry->tid = ((struct ib_mad_hdr *)wr->send_buf.mad)->tid; 80 ((struct ib_mad_hdr *)wr->send_buf.mad)->attr_id; 82 ((struct ib_mad_hdr *)wr->send_buf.mad)->attr_mod; 154 ((struct ib_mad_hdr *)wr->send_buf.mad)->base_version; [all …]
|
| /linux/include/rdma/ |
| H A D | ib_smi.h | 151 static inline void ib_init_query_mad(struct ib_smp *mad) in ib_init_query_mad() argument 153 mad->base_version = IB_MGMT_BASE_VERSION; in ib_init_query_mad() 154 mad->mgmt_class = IB_MGMT_CLASS_SUBN_LID_ROUTED; in ib_init_query_mad() 155 mad->class_version = 1; in ib_init_query_mad() 156 mad->method = IB_MGMT_METHOD_GET; in ib_init_query_mad()
|
| /linux/drivers/scsi/ibmvscsi_tgt/ |
| H A D | ibmvscsi_tgt.c | 1504 struct viosrp_adapter_info *mad = &vio_iu(iue)->mad.adapter_info; in ibmvscsis_adapter_info() local 1510 mad->common.status = cpu_to_be16(VIOSRP_MAD_SUCCESS); in ibmvscsis_adapter_info() 1512 if (be16_to_cpu(mad->common.length) > sizeof(*info)) { in ibmvscsis_adapter_info() 1513 mad->common.status = cpu_to_be16(VIOSRP_MAD_FAILED); in ibmvscsis_adapter_info() 1522 mad->common.status = cpu_to_be16(VIOSRP_MAD_FAILED); in ibmvscsis_adapter_info() 1527 rc = h_copy_rdma(be16_to_cpu(mad->common.length), in ibmvscsis_adapter_info() 1529 be64_to_cpu(mad->buffer), in ibmvscsis_adapter_info() 1540 be64_to_cpu(mad->buffer), vscsi->flags, flag_bits); in ibmvscsis_adapter_info() 1575 be64_to_cpu(mad->buffer)); in ibmvscsis_adapter_info() 1615 struct viosrp_capabilities *mad = &vio_iu(iue)->mad.capabilities; in ibmvscsis_cap_mad() local [all …]
|
| /linux/drivers/infiniband/ulp/opa_vnic/ |
| H A D | opa_vnic_vema.c | 630 if (!mad_wc || !mad_wc->recv_buf.mad) in vema_recv() 647 vema_mad = rsp->mad; in vema_recv() 648 memcpy(vema_mad, mad_wc->recv_buf.mad, IB_MGMT_VENDOR_HDR); in vema_recv() 655 switch (mad_wc->recv_buf.mad->mad_hdr.method) { in vema_recv() 657 vema_get(port, (struct opa_vnic_vema_mad *)mad_wc->recv_buf.mad, in vema_recv() 661 vema_set(port, (struct opa_vnic_vema_mad *)mad_wc->recv_buf.mad, in vema_recv() 807 trap_mad = send_buf->mad; in opa_vnic_vema_send_trap()
|
| /linux/net/802/ |
| H A D | mrp.c | 240 struct rb_node *parent = app->mad.rb_node; in mrp_attr_lookup() 260 struct rb_node *parent = NULL, **p = &app->mad.rb_node; in mrp_attr_create() 286 rb_insert_color(&attr->node, &app->mad); in mrp_attr_create() 292 rb_erase(&attr->node, &app->mad); in mrp_attr_destroy() 301 for (node = rb_first(&app->mad); in mrp_attr_destroy_all() 584 for (node = rb_first(&app->mad); in mrp_mad_event() 879 app->mad = RB_ROOT; in mrp_init_applicant()
|
| /linux/drivers/infiniband/sw/rdmavt/ |
| H A D | Makefile | 11 rdmavt-y := vt.o ah.o cq.o mad.o mcast.o mmap.o mr.o pd.o qp.o \
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | Makefile | 28 mad.o \
|
| H A D | mad.c | 310 smp = send_buf->mad; in send_trap() 438 static void bad_mkey(struct hfi1_ibport *ibp, struct ib_mad_hdr *mad, in bad_mkey() argument 451 trap->data.ntc_256.method = mad->method; in bad_mkey() 452 trap->data.ntc_256.attr_id = mad->attr_id; in bad_mkey() 453 trap->data.ntc_256.attr_mod = mad->attr_mod; in bad_mkey() 455 if (mad->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) { in bad_mkey() 650 static int check_mkey(struct hfi1_ibport *ibp, struct ib_mad_hdr *mad, in check_mkey() argument 671 (mad->method == IB_MGMT_METHOD_GET || in check_mkey() 672 mad->method == IB_MGMT_METHOD_SET || in check_mkey() 673 mad->method == IB_MGMT_METHOD_TRAP_REPRESS)) in check_mkey() [all …]
|
| /linux/include/net/ |
| H A D | mrp.h | 125 struct rb_root mad; member
|
| /linux/include/scsi/ |
| H A D | viosrp.h | 202 union mad_iu mad; member
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | cmd.c | 206 data = MLX5_ADDR_OF(mad_ifc_in, in, mad); in mlx5_cmd_mad_ifc() 207 memcpy(data, inb, MLX5_FLD_SZ_BYTES(mad_ifc_in, mad)); in mlx5_cmd_mad_ifc()
|
| /linux/drivers/ata/ |
| H A D | sata_inic162x.c | 221 __le32 mad; /* Physical Memory Address */ member 471 prd->mad = cpu_to_le32(sg_dma_address(sg)); in inic_fill_sg() 525 prd->mad = cpu_to_le32(pp->pkt_dma + in inic_qc_prep()
|
| /linux/arch/s390/tools/ |
| H A D | opcodes.txt | 1239 ed3e mad RXF_FRRDF
|
| /linux/include/linux/mlx5/ |
| H A D | mlx5_ifc.h | 8258 u8 mad[256][0x8]; member
|