| /linux/drivers/infiniband/hw/mthca/ |
| H A D | mthca_main.c | 136 static int mthca_tune_pci(struct mthca_dev *mdev) in mthca_tune_pci() argument 142 if (pci_find_capability(mdev->pdev, PCI_CAP_ID_PCIX)) { in mthca_tune_pci() 143 if (pcix_set_mmrbc(mdev->pdev, pcix_get_max_mmrbc(mdev->pdev))) { in mthca_tune_pci() 144 mthca_err(mdev, "Couldn't set PCI-X max read count, " in mthca_tune_pci() 148 } else if (!(mdev->mthca_flags & MTHCA_FLAG_PCIE)) in mthca_tune_pci() 149 mthca_info(mdev, "No PCI-X capability, not setting RBC.\n"); in mthca_tune_pci() 151 if (pci_is_pcie(mdev->pdev)) { in mthca_tune_pci() 152 if (pcie_set_readrq(mdev->pdev, 4096)) { in mthca_tune_pci() 153 mthca_err(mdev, "Couldn't write PCI Express read request, " in mthca_tune_pci() 157 } else if (mdev->mthca_flags & MTHCA_FLAG_PCIE) in mthca_tune_pci() [all …]
|
| /linux/drivers/staging/most/video/ |
| H A D | video.c | 51 struct most_video_dev *mdev; member 63 static inline bool data_ready(struct most_video_dev *mdev) in data_ready() argument 65 return !list_empty(&mdev->pending_mbos); in data_ready() 68 static inline struct mbo *get_top_mbo(struct most_video_dev *mdev) in get_top_mbo() argument 70 return list_first_entry(&mdev->pending_mbos, struct mbo, list); in get_top_mbo() 77 struct most_video_dev *mdev = video_drvdata(filp); in comp_vdev_open() local 91 if (!atomic_inc_and_test(&mdev->access_ref)) { in comp_vdev_open() 92 v4l2_err(&mdev->v4l2_dev, "too many clients\n"); in comp_vdev_open() 97 fh->mdev = mdev; in comp_vdev_open() 101 ret = most_start_channel(mdev->iface, mdev->ch_idx, &comp); in comp_vdev_open() [all …]
|
| /linux/drivers/sh/maple/ |
| H A D | maple.c | 139 struct maple_device *mdev; in maple_release_device() local 142 mdev = to_maple_dev(dev); in maple_release_device() 143 mq = mdev->mq; in maple_release_device() 146 kfree(mdev); in maple_release_device() 151 * @mdev: maple device 157 int maple_add_packet(struct maple_device *mdev, u32 function, u32 command, in maple_add_packet() argument 172 mdev->mq->command = command; in maple_add_packet() 173 mdev->mq->length = length; in maple_add_packet() 176 mdev->mq->sendbuf = sendbuf; in maple_add_packet() 179 list_add_tail(&mdev->mq->list, &maple_waitq); in maple_add_packet() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
| H A D | clock.c | 83 struct mlx5_core_dev *mdev; member 91 struct mlx5_core_dev *mdev; member 92 struct mutex lock; /* protect mdev and used in PTP callbacks */ 113 return clock_priv(clock)->mdev; in mlx5_clock_mdev_get() 132 static bool mlx5_real_time_mode(struct mlx5_core_dev *mdev) in mlx5_real_time_mode() argument 134 return (mlx5_is_real_time_rq(mdev) || mlx5_is_real_time_sq(mdev)); in mlx5_real_time_mode() 137 static bool mlx5_npps_real_time_supported(struct mlx5_core_dev *mdev) in mlx5_npps_real_time_supported() argument 139 return (mlx5_real_time_mode(mdev) && in mlx5_npps_real_time_supported() 140 MLX5_CAP_MCAM_FEATURE(mdev, npps_period) && in mlx5_npps_real_time_supported() 141 MLX5_CAP_MCAM_FEATURE(mdev, out_pulse_duration_ns)); in mlx5_npps_real_time_supported() [all …]
|
| H A D | ipsec_fs_roce.c | 59 static bool ipsec_fs_create_alias_supported_one(struct mlx5_core_dev *mdev) in ipsec_fs_create_alias_supported_one() argument 61 u64 obj_allowed = MLX5_CAP_GEN_2_64(mdev, allowed_object_for_other_vhca_access); in ipsec_fs_create_alias_supported_one() 62 u32 obj_supp = MLX5_CAP_GEN_2(mdev, cross_vhca_object_to_object_supported); in ipsec_fs_create_alias_supported_one() 74 static bool ipsec_fs_create_alias_supported(struct mlx5_core_dev *mdev, in ipsec_fs_create_alias_supported() argument 77 if (ipsec_fs_create_alias_supported_one(mdev) && in ipsec_fs_create_alias_supported() 131 ipsec_fs_roce_rx_rule_setup(struct mlx5_core_dev *mdev, in ipsec_fs_roce_rx_rule_setup() argument 135 bool is_mpv_slave = mlx5_core_is_mp_slave(mdev); in ipsec_fs_roce_rx_rule_setup() 159 mlx5_core_err(mdev, "Fail to add RX RoCE IPsec rule err=%d\n", in ipsec_fs_roce_rx_rule_setup() 172 mlx5_core_err(mdev, "Fail to add RX RoCE IPsec miss rule err=%d\n", in ipsec_fs_roce_rx_rule_setup() 191 mlx5_core_err(mdev, "Fail to add RX RoCE IPsec rule for alias err=%d\n", in ipsec_fs_roce_rx_rule_setup() [all …]
|
| /linux/drivers/most/ |
| H A D | most_usb.c | 77 struct most_dev *mdev; member 238 struct most_dev *mdev = to_mdev(iface); in hdm_poison_channel() local 243 dev_warn(&mdev->usb_device->dev, "Channel ID out of range.\n"); in hdm_poison_channel() 247 lock = mdev->channel_lock + channel; in hdm_poison_channel() 249 mdev->is_channel_healthy[channel] = false; in hdm_poison_channel() 252 cancel_work_sync(&mdev->clear_work[channel].ws); in hdm_poison_channel() 254 mutex_lock(&mdev->io_mutex); in hdm_poison_channel() 255 usb_kill_anchored_urbs(&mdev->busy_urbs[channel]); in hdm_poison_channel() 256 if (mdev->padding_active[channel]) in hdm_poison_channel() 257 mdev->padding_active[channel] = false; in hdm_poison_channel() [all …]
|
| /linux/drivers/dma/ |
| H A D | altera-msgdma.c | 206 * @mdev: Pointer to the Altera mSGDMA device structure 210 static struct msgdma_sw_desc *msgdma_get_descriptor(struct msgdma_device *mdev) in msgdma_get_descriptor() argument 215 spin_lock_irqsave(&mdev->lock, flags); in msgdma_get_descriptor() 216 desc = list_first_entry(&mdev->free_list, struct msgdma_sw_desc, node); in msgdma_get_descriptor() 218 spin_unlock_irqrestore(&mdev->lock, flags); in msgdma_get_descriptor() 227 * @mdev: Pointer to the Altera mSGDMA device structure 230 static void msgdma_free_descriptor(struct msgdma_device *mdev, in msgdma_free_descriptor() argument 235 mdev->desc_free_cnt++; in msgdma_free_descriptor() 236 list_move_tail(&desc->node, &mdev->free_list); in msgdma_free_descriptor() 238 mdev->desc_free_cnt++; in msgdma_free_descriptor() [all …]
|
| /linux/sound/core/seq/oss/ |
| H A D | seq_oss_midi.c | 59 …t send_midi_event(struct seq_oss_devinfo *dp, struct snd_seq_event *ev, struct seq_oss_midi *mdev); 93 struct seq_oss_midi *mdev; in get_mdev() local 96 mdev = midi_devs[dev]; in get_mdev() 97 if (mdev) in get_mdev() 98 snd_use_lock_use(&mdev->use_lock); in get_mdev() 99 return mdev; in get_mdev() 109 struct seq_oss_midi *mdev; in find_slot() local 113 mdev = midi_devs[i]; in find_slot() 114 if (mdev && mdev->client == client && mdev->port == port) { in find_slot() 116 snd_use_lock_use(&mdev->use_lock); in find_slot() [all …]
|
| /linux/drivers/virtio/ |
| H A D | virtio_pci_modern_dev.c | 10 * @mdev: the modern virtio-pci device 22 vp_modern_map_capability(struct virtio_pci_modern_device *mdev, int off, in vp_modern_map_capability() argument 26 struct pci_dev *dev = mdev->pci_dev; in vp_modern_map_capability() 40 if (bar >= PCI_STD_NUM_BARS || !(mdev->modern_bars & (1 << bar))) { in vp_modern_map_capability() 219 * @mdev: the modern virtio-pci device 223 int vp_modern_probe(struct virtio_pci_modern_device *mdev) in vp_modern_probe() argument 225 struct pci_dev *pci_dev = mdev->pci_dev; in vp_modern_probe() 233 if (mdev->device_id_check) { in vp_modern_probe() 234 devid = mdev->device_id_check(pci_dev); in vp_modern_probe() 237 mdev->id.device = devid; in vp_modern_probe() [all …]
|
| /linux/drivers/mtd/maps/ |
| H A D | vmu-flash.c | 24 struct maple_device *mdev; member 60 struct maple_device *mdev; in ofs_to_block() local 66 mdev = mpart->mdev; in ofs_to_block() 67 card = maple_get_drvdata(mdev); in ofs_to_block() 91 struct maple_device *mdev; in vmu_blockread() local 94 mdev = mq->dev; in vmu_blockread() 95 card = maple_get_drvdata(mdev); in vmu_blockread() 114 struct maple_device *mdev; in maple_vmu_read_block() local 121 mdev = mpart->mdev; in maple_vmu_read_block() 123 card = maple_get_drvdata(mdev); in maple_vmu_read_block() [all …]
|
| /linux/net/mctp/ |
| H A D | device.c | 32 struct mctp_dev *mdev = rcu_dereference(dev->mctp_ptr); in __mctp_dev_get() local 34 /* RCU guarantees that any mdev is still live. in __mctp_dev_get() 37 if (mdev) in __mctp_dev_get() 38 if (!refcount_inc_not_zero(&mdev->refs)) in __mctp_dev_get() 40 return mdev; in __mctp_dev_get() 61 struct mctp_dev *mdev, mctp_eid_t eid, in mctp_fill_addrinfo() argument 78 hdr->ifa_index = mdev->dev->ifindex; in mctp_fill_addrinfo() 95 static int mctp_dump_dev_addrinfo(struct mctp_dev *mdev, struct sk_buff *skb, in mctp_dump_dev_addrinfo() argument 104 for (; mcb->a_idx < mdev->num_addrs; mcb->a_idx++) { in mctp_dump_dev_addrinfo() 105 rc = mctp_fill_addrinfo(skb, mdev, mdev->addrs[mcb->a_idx], in mctp_dump_dev_addrinfo() [all …]
|
| /linux/drivers/crypto/intel/qat/qat_common/ |
| H A D | qat_mig_dev.c | 15 struct qat_mig_dev *mdev; in qat_vfmig_create() local 27 mdev = kmalloc_obj(*mdev); in qat_vfmig_create() 28 if (!mdev) in qat_vfmig_create() 31 mdev->vf_id = vf_id; in qat_vfmig_create() 32 mdev->parent_accel_dev = accel_dev; in qat_vfmig_create() 34 return mdev; in qat_vfmig_create() 38 int qat_vfmig_init(struct qat_mig_dev *mdev) in qat_vfmig_init() argument 40 struct adf_accel_dev *accel_dev = mdev->parent_accel_dev; in qat_vfmig_init() 42 return GET_VFMIG_OPS(accel_dev)->init(mdev); in qat_vfmig_init() 46 void qat_vfmig_cleanup(struct qat_mig_dev *mdev) in qat_vfmig_cleanup() argument [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | en_clock.c | 43 struct mlx4_en_dev *mdev = in mlx4_en_read_clock() local 45 struct mlx4_dev *dev = mdev->dev; in mlx4_en_read_clock() 61 u64 mlx4_en_get_hwtstamp(struct mlx4_en_dev *mdev, u64 timestamp) in mlx4_en_get_hwtstamp() argument 67 seq = read_seqbegin(&mdev->clock_lock); in mlx4_en_get_hwtstamp() 68 nsec = timecounter_cyc2time(&mdev->clock, timestamp); in mlx4_en_get_hwtstamp() 69 } while (read_seqretry(&mdev->clock_lock, seq)); in mlx4_en_get_hwtstamp() 74 void mlx4_en_fill_hwtstamps(struct mlx4_en_dev *mdev, in mlx4_en_fill_hwtstamps() argument 79 hwts->hwtstamp = mlx4_en_get_hwtstamp(mdev, timestamp); in mlx4_en_fill_hwtstamps() 84 * @mdev: board private structure 88 void mlx4_en_remove_timestamp(struct mlx4_en_dev *mdev) in mlx4_en_remove_timestamp() argument [all …]
|
| H A D | en_main.c | 96 level, DRV_NAME, dev_name(&priv->mdev->pdev->dev), in en_print() 117 if (mlx4_is_mfunc(priv->mdev->dev) && in mlx4_en_update_loopback_state() 124 if (mlx4_is_mfunc(priv->mdev->dev) || priv->validate_loopback) in mlx4_en_update_loopback_state() 127 mutex_lock(&priv->mdev->state_lock); in mlx4_en_update_loopback_state() 128 if ((priv->mdev->dev->caps.flags2 & in mlx4_en_update_loopback_state() 145 mlx4_warn(priv->mdev, "failed to change mcast loopback\n"); in mlx4_en_update_loopback_state() 147 mutex_unlock(&priv->mdev->state_lock); in mlx4_en_update_loopback_state() 150 static void mlx4_en_get_profile(struct mlx4_en_dev *mdev) in mlx4_en_get_profile() argument 152 struct mlx4_en_profile *params = &mdev->profile; in mlx4_en_get_profile() 160 if (params->udp_rss && !(mdev->dev->caps.flags in mlx4_en_get_profile() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | en_common.c | 41 void mlx5e_mkey_set_relaxed_ordering(struct mlx5_core_dev *mdev, void *mkc) in mlx5e_mkey_set_relaxed_ordering() argument 43 bool ro_write = MLX5_CAP_GEN(mdev, relaxed_ordering_write); in mlx5e_mkey_set_relaxed_ordering() 44 bool ro_read = MLX5_CAP_GEN(mdev, relaxed_ordering_read) || in mlx5e_mkey_set_relaxed_ordering() 45 (pcie_relaxed_ordering_enabled(mdev->pdev) && in mlx5e_mkey_set_relaxed_ordering() 46 MLX5_CAP_GEN(mdev, relaxed_ordering_read_pci_enabled)); in mlx5e_mkey_set_relaxed_ordering() 52 int mlx5e_create_mkey(struct mlx5_core_dev *mdev, u32 pdn, u32 *mkey) in mlx5e_create_mkey() argument 67 mlx5e_mkey_set_relaxed_ordering(mdev, mkc); in mlx5e_create_mkey() 72 err = mlx5_core_create_mkey(mdev, mkey, in, inlen); in mlx5e_create_mkey() 78 int mlx5e_create_tis(struct mlx5_core_dev *mdev, void *in, u32 *tisn) in mlx5e_create_tis() argument 82 MLX5_SET(tisc, tisc, transport_domain, mdev->mlx5e_res.hw_objs.td.tdn); in mlx5e_create_tis() [all …]
|
| H A D | wc.c | 29 struct mlx5_core_dev *mdev; member 49 static int mlx5_wc_create_cqwq(struct mlx5_core_dev *mdev, void *cqc, in mlx5_wc_create_cqwq() argument 57 err = mlx5_cqwq_create(mdev, ¶m, cqc, &cq->wq, &cq->wq_ctrl); in mlx5_wc_create_cqwq() 71 cq->mdev = mdev; in mlx5_wc_create_cqwq() 79 struct mlx5_core_dev *mdev = cq->mdev; in create_wc_cq() local 84 err = mlx5_comp_eqn_get(mdev, 0, &eqn); in create_wc_cq() 103 MLX5_SET(cqc, cqc, uar_page, mdev->priv.bfreg.up->index); in create_wc_cq() 108 err = mlx5_core_create_cq(mdev, mcq, in, inlen, out, sizeof(out)); in create_wc_cq() 115 static int mlx5_wc_create_cq(struct mlx5_core_dev *mdev, struct mlx5_wc_cq *cq) in mlx5_wc_create_cq() argument 125 MLX5_SET(cqc, cqc, uar_page, mdev->priv.bfreg.up->index); in mlx5_wc_create_cq() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/sf/dev/ |
| H A D | driver.c | 28 struct mlx5_core_dev *mdev; in mlx5_sf_dev_probe() local 36 mdev = devlink_priv(devlink); in mlx5_sf_dev_probe() 37 mdev->device = &adev->dev; in mlx5_sf_dev_probe() 38 mdev->pdev = sf_dev->parent_mdev->pdev; in mlx5_sf_dev_probe() 39 mdev->bar_addr = sf_dev->bar_base_addr; in mlx5_sf_dev_probe() 40 mdev->iseg_base = sf_dev->bar_base_addr; in mlx5_sf_dev_probe() 41 mdev->coredev_type = MLX5_COREDEV_SF; in mlx5_sf_dev_probe() 42 mdev->priv.parent_mdev = sf_dev->parent_mdev; in mlx5_sf_dev_probe() 43 mdev->priv.adev_idx = adev->id; in mlx5_sf_dev_probe() 44 sf_dev->mdev = mdev; in mlx5_sf_dev_probe() [all …]
|
| /linux/drivers/dma/qcom/ |
| H A D | hidma_mgmt_sys.c | 14 struct hidma_mgmt_dev *mdev; member 22 int (*get)(struct hidma_mgmt_dev *mdev); 23 int (*set)(struct hidma_mgmt_dev *mdev, u64 val); 27 static int get_##name(struct hidma_mgmt_dev *mdev) \ 29 return mdev->name; \ 31 static int set_##name(struct hidma_mgmt_dev *mdev, u64 val) \ 36 tmp = mdev->name; \ 37 mdev->name = val; \ 38 rc = hidma_mgmt_setup(mdev); \ 40 mdev->name = tmp; \ [all …]
|
| /linux/drivers/media/mc/ |
| H A D | mc-device.c | 81 static struct media_entity *find_entity(struct media_device *mdev, u32 id) in find_entity() argument 88 media_device_for_each_entity(entity, mdev) { in find_entity() 98 static long media_device_enum_entities(struct media_device *mdev, void *arg) in media_device_enum_entities() argument 103 ent = find_entity(mdev, entd->id); in media_device_enum_entities() 151 static long media_device_enum_links(struct media_device *mdev, void *arg) in media_device_enum_links() argument 156 entity = find_entity(mdev, links->entity); in media_device_enum_links() 200 static long media_device_setup_link(struct media_device *mdev, void *arg) in media_device_setup_link() argument 209 source = find_entity(mdev, linkd->source.entity); in media_device_setup_link() 210 sink = find_entity(mdev, linkd->sink.entity); in media_device_setup_link() 230 static long media_device_get_topology(struct media_device *mdev, void *arg) in media_device_get_topology() argument [all …]
|
| /linux/drivers/vfio/mdev/ |
| H A D | mdev_core.c | 13 #include <linux/mdev.h> 27 static void mdev_device_remove_common(struct mdev_device *mdev) in mdev_device_remove_common() argument 29 struct mdev_parent *parent = mdev->type->parent; in mdev_device_remove_common() 31 mdev_remove_sysfs_files(mdev); in mdev_device_remove_common() 32 device_del(&mdev->dev); in mdev_device_remove_common() 35 put_device(&mdev->dev); in mdev_device_remove_common() 49 * @mdev_driver: Device driver to bind to the newly created mdev 50 * @types: Array of supported mdev types 53 * Registers the @parent stucture as a parent for mdev types and thus mdev 83 dev_info(dev, "MDEV: Registered\n"); in mdev_register_parent() [all …]
|
| /linux/drivers/mailbox/ |
| H A D | mailbox-sti.c | 36 #define MBOX_BASE(mdev, inst) ((mdev)->base + ((inst) * 4)) argument 76 * @mdev: Pointer to parent Mailbox device 81 struct sti_mbox_device *mdev; member 89 struct sti_mbox_device *mdev = chan_info->mdev; in sti_mbox_channel_is_enabled() local 93 return mdev->enabled[instance] & BIT(channel); in sti_mbox_channel_is_enabled() 122 struct sti_mbox_device *mdev = chan_info->mdev; in sti_mbox_enable_channel() local 126 void __iomem *base = MBOX_BASE(mdev, instance); in sti_mbox_enable_channel() 128 spin_lock_irqsave(&mdev->lock, flags); in sti_mbox_enable_channel() 129 mdev->enabled[instance] |= BIT(channel); in sti_mbox_enable_channel() 131 spin_unlock_irqrestore(&mdev->lock, flags); in sti_mbox_enable_channel() [all …]
|
| /linux/drivers/net/ethernet/marvell/octeontx2/af/ |
| H A D | mbox.c | 23 struct otx2_mbox_dev *mdev = &mbox->dev[devid]; in __otx2_mbox_reset() local 25 void *hw_mbase = mdev->hwbase; in __otx2_mbox_reset() 30 mdev->msg_size = 0; in __otx2_mbox_reset() 31 mdev->rsp_size = 0; in __otx2_mbox_reset() 43 struct otx2_mbox_dev *mdev = &mbox->dev[devid]; in otx2_mbox_reset() local 45 spin_lock(&mdev->mbox_lock); in otx2_mbox_reset() 47 spin_unlock(&mdev->mbox_lock); in otx2_mbox_reset() 227 struct otx2_mbox_dev *mdev; in otx2_mbox_init() local 237 mdev = &mbox->dev[devid]; in otx2_mbox_init() 238 mdev->mbase = mbox->hwbase + (devid * MBOX_SIZE); in otx2_mbox_init() [all …]
|
| /linux/drivers/vdpa/virtio_pci/ |
| H A D | vp_vdpa.c | 36 struct virtio_pci_modern_device *mdev; member 48 struct virtio_pci_modern_device *mdev; member 61 return vp_vdpa->mdev; in vdpa_to_mdev() 66 return vp_vdpa->mdev; in vp_vdpa_to_mdev() 78 struct virtio_pci_modern_device *mdev = vdpa_to_mdev(vdpa); in vp_vdpa_set_driver_features() local 80 vp_modern_set_features(mdev, features); in vp_vdpa_set_driver_features() 87 struct virtio_pci_modern_device *mdev = vdpa_to_mdev(vdpa); in vp_vdpa_get_driver_features() local 89 return vp_modern_get_driver_features(mdev); in vp_vdpa_get_driver_features() 94 struct virtio_pci_modern_device *mdev = vdpa_to_mdev(vdpa); in vp_vdpa_get_status() local 96 return vp_modern_get_status(mdev); in vp_vdpa_get_status() [all …]
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | main.c | 113 int port_type_cap = MLX5_CAP_GEN(dev->mdev, port_type); in mlx5_ib_port_link_layer() 211 struct mlx5_core_dev *mdev; in mlx5_netdev_event() local 215 mdev = mlx5_ib_get_native_port_mdev(ibdev, port_num, NULL); in mlx5_netdev_event() 216 if (!mdev) in mlx5_netdev_event() 230 if (ndev->dev.parent == mdev->device) in mlx5_netdev_event() 249 !mlx5_core_mp_enabled(mdev)) in mlx5_netdev_event() 252 if (mlx5_lag_is_roce(mdev) || mlx5_lag_is_sriov(mdev)) { in mlx5_netdev_event() 255 if(mlx5_lag_is_roce(mdev)) in mlx5_netdev_event() 317 struct mlx5_core_dev *mdev = NULL; in mlx5_ib_get_native_port_mdev() local 325 return ibdev->mdev; in mlx5_ib_get_native_port_mdev() [all …]
|
| /linux/drivers/infiniband/hw/mana/ |
| H A D | qp.c | 96 struct mana_ib_dev *mdev = in mana_ib_create_qp_rss() local 119 ibdev_dbg(&mdev->ib_dev, in mana_ib_create_qp_rss() 125 if (attr->cap.max_recv_wr > mdev->adapter_caps.max_qp_wr) { in mana_ib_create_qp_rss() 126 ibdev_dbg(&mdev->ib_dev, in mana_ib_create_qp_rss() 133 ibdev_dbg(&mdev->ib_dev, in mana_ib_create_qp_rss() 141 ibdev_dbg(&mdev->ib_dev, in mana_ib_create_qp_rss() 148 ibdev_dbg(&mdev->ib_dev, in mana_ib_create_qp_rss() 158 ibdev_dbg(&mdev->ib_dev, "Invalid port %u in creating qp\n", in mana_ib_create_qp_rss() 164 ibdev_dbg(&mdev->ib_dev, "rx_hash_function %d port %d\n", in mana_ib_create_qp_rss() 209 ibdev_dbg(&mdev->ib_dev, in mana_ib_create_qp_rss() [all …]
|