/linux/net/openvswitch/ |
H A D | vport-netdev.c | 31 struct vport *vport; in netdev_port_receive() local 33 vport = ovs_netdev_get_vport(skb->dev); in netdev_port_receive() 34 if (unlikely(!vport)) in netdev_port_receive() 50 ovs_vport_receive(vport, skb, skb_tunnel_info(skb)); in netdev_port_receive() 70 struct vport *local; in get_dpdev() 76 struct vport *ovs_netdev_link(struct vport *vport, const char *name) in ovs_netdev_link() argument 80 vport->dev = dev_get_by_name(ovs_dp_get_net(vport->dp), name); in ovs_netdev_link() 81 if (!vport->dev) { in ovs_netdev_link() 88 if (strcmp(name, ovs_vport_name(vport))) { in ovs_netdev_link() 92 netdev_tracker_alloc(vport->dev, &vport->dev_tracker, GFP_KERNEL); in ovs_netdev_link() [all …]
|
H A D | vport.c | 95 struct vport *ovs_vport_locate(const struct net *net, const char *name) in ovs_vport_locate() 98 struct vport *vport; in ovs_vport_locate() local 100 hlist_for_each_entry_rcu(vport, bucket, hash_node, in ovs_vport_locate() 102 if (!strcmp(name, ovs_vport_name(vport)) && in ovs_vport_locate() 103 net_eq(ovs_dp_get_net(vport->dp), net)) in ovs_vport_locate() 104 return vport; in ovs_vport_locate() 122 struct vport *ovs_vport_alloc(int priv_size, const struct vport_ops *ops, in ovs_vport_alloc() 125 struct vport *vport; in ovs_vport_alloc() local 129 alloc_size = sizeof(struct vport); in ovs_vport_alloc() 135 vport = kzalloc(alloc_size, GFP_KERNEL); in ovs_vport_alloc() [all …]
|
H A D | vport.h | 20 struct vport; 28 struct vport *ovs_vport_add(const struct vport_parms *); 29 void ovs_vport_del(struct vport *); 31 struct vport *ovs_vport_locate(const struct net *net, const char *name); 33 void ovs_vport_get_stats(struct vport *, struct ovs_vport_stats *); 35 int ovs_vport_get_upcall_stats(struct vport *vport, struct sk_buff *skb); 37 int ovs_vport_set_options(struct vport *, struct nlattr *options); 38 int ovs_vport_get_options(const struct vport *, struct sk_buff *); 40 int ovs_vport_set_upcall_portids(struct vport *, const struct nlattr *pids); 41 int ovs_vport_get_upcall_portids(const struct vport *, struct sk_buff *); [all …]
|
H A D | vport-internal_dev.c | 22 struct vport *vport; member 42 err = ovs_vport_receive(internal_dev_priv(netdev)->vport, skb, NULL); in internal_dev_xmit() 78 struct vport *vport = ovs_internal_dev_get_vport(dev); in internal_dev_destructor() local 80 ovs_vport_free(vport); in internal_dev_destructor() 123 static struct vport *internal_dev_create(const struct vport_parms *parms) in internal_dev_create() 125 struct vport *vport; in internal_dev_create() local 130 vport = ovs_vport_alloc(0, &ovs_internal_vport_ops, parms); in internal_dev_create() 131 if (IS_ERR(vport)) { in internal_dev_create() 132 err = PTR_ERR(vport); in internal_dev_create() 138 vport->dev = dev; in internal_dev_create() [all …]
|
H A D | vport-geneve.c | 36 static inline struct geneve_port *geneve_vport(const struct vport *vport) in geneve_vport() argument 38 return vport_priv(vport); in geneve_vport() 41 static int geneve_get_options(const struct vport *vport, in geneve_get_options() argument 44 struct geneve_port *geneve_port = geneve_vport(vport); in geneve_get_options() 51 static struct vport *geneve_tnl_create(const struct vport_parms *parms) in geneve_tnl_create() 57 struct vport *vport; in geneve_tnl_create() local 76 vport = ovs_vport_alloc(sizeof(struct geneve_port), in geneve_tnl_create() 78 if (IS_ERR(vport)) in geneve_tnl_create() 79 return vport; in geneve_tnl_create() 81 geneve_port = geneve_vport(vport); in geneve_tnl_create() [all …]
|
H A D | vport-vxlan.c | 22 static int vxlan_get_options(const struct vport *vport, struct sk_buff *skb) in vxlan_get_options() argument 24 struct vxlan_dev *vxlan = netdev_priv(vport->dev); in vxlan_get_options() 51 static int vxlan_configure_exts(struct vport *vport, struct nlattr *attr, in vxlan_configure_exts() argument 71 static struct vport *vxlan_tnl_create(const struct vport_parms *parms) in vxlan_tnl_create() 76 struct vport *vport; in vxlan_tnl_create() local 100 vport = ovs_vport_alloc(0, &ovs_vxlan_netdev_vport_ops, parms); in vxlan_tnl_create() 101 if (IS_ERR(vport)) in vxlan_tnl_create() 102 return vport; in vxlan_tnl_create() 106 err = vxlan_configure_exts(vport, a, &conf); in vxlan_tnl_create() 108 ovs_vport_free(vport); in vxlan_tnl_create() [all …]
|
H A D | vport-gre.c | 39 static struct vport *gre_tnl_create(const struct vport_parms *parms) in gre_tnl_create() 43 struct vport *vport; in gre_tnl_create() local 46 vport = ovs_vport_alloc(0, &ovs_gre_vport_ops, parms); in gre_tnl_create() 47 if (IS_ERR(vport)) in gre_tnl_create() 48 return vport; in gre_tnl_create() 54 ovs_vport_free(vport); in gre_tnl_create() 62 ovs_vport_free(vport); in gre_tnl_create() 67 return vport; in gre_tnl_create() 70 static struct vport *gre_create(const struct vport_parms *parms) in gre_create() 72 struct vport *vport; in gre_create() local [all …]
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/esw/acl/ |
H A D | ingress_ofld.c | 10 acl_ingress_ofld_setup(struct mlx5_eswitch *esw, struct mlx5_vport *vport); 14 const struct mlx5_vport *vport) in esw_acl_ingress_prio_tag_enabled() argument 17 mlx5_eswitch_is_vf_vport(esw, vport->vport)); in esw_acl_ingress_prio_tag_enabled() 21 struct mlx5_vport *vport) in esw_acl_ingress_prio_tag_create() argument 46 if (vport->ingress.offloads.modify_metadata_rule) { in esw_acl_ingress_prio_tag_create() 48 flow_act.modify_hdr = vport->ingress.offloads.modify_metadata; in esw_acl_ingress_prio_tag_create() 51 vport->ingress.allow_rule = mlx5_add_flow_rules(vport->ingress.acl, spec, in esw_acl_ingress_prio_tag_create() 53 if (IS_ERR(vport->ingress.allow_rule)) { in esw_acl_ingress_prio_tag_create() 54 err = PTR_ERR(vport->ingress.allow_rule); in esw_acl_ingress_prio_tag_create() 57 vport->vport, err); in esw_acl_ingress_prio_tag_create() [all …]
|
H A D | egress_ofld.c | 9 static void esw_acl_egress_ofld_fwd2vport_destroy(struct mlx5_vport *vport) in esw_acl_egress_ofld_fwd2vport_destroy() argument 11 if (!vport->egress.offloads.fwd_rule) in esw_acl_egress_ofld_fwd2vport_destroy() 14 mlx5_del_flow_rules(vport->egress.offloads.fwd_rule); in esw_acl_egress_ofld_fwd2vport_destroy() 15 vport->egress.offloads.fwd_rule = NULL; in esw_acl_egress_ofld_fwd2vport_destroy() 18 void esw_acl_egress_ofld_bounce_rule_destroy(struct mlx5_vport *vport, int rule_index) in esw_acl_egress_ofld_bounce_rule_destroy() argument 21 xa_load(&vport->egress.offloads.bounce_rules, rule_index); in esw_acl_egress_ofld_bounce_rule_destroy() 27 xa_erase(&vport->egress.offloads.bounce_rules, rule_index); in esw_acl_egress_ofld_bounce_rule_destroy() 30 static void esw_acl_egress_ofld_bounce_rules_destroy(struct mlx5_vport *vport) in esw_acl_egress_ofld_bounce_rules_destroy() argument 35 xa_for_each(&vport->egress.offloads.bounce_rules, i, bounce_rule) { in esw_acl_egress_ofld_bounce_rules_destroy() 37 xa_erase(&vport->egress.offloads.bounce_rules, i); in esw_acl_egress_ofld_bounce_rules_destroy() [all …]
|
H A D | helper.c | 9 esw_acl_table_create(struct mlx5_eswitch *esw, struct mlx5_vport *vport, int ns, int size) in esw_acl_table_create() argument 26 vport_num = vport->vport; in esw_acl_table_create() 30 root_ns = mlx5_get_flow_vport_acl_namespace(dev, ns, vport->index); in esw_acl_table_create() 50 struct mlx5_vport *vport, in esw_egress_acl_vlan_create() argument 58 if (vport->egress.allowed_vlan) in esw_egress_acl_vlan_create() 72 vport->egress.allowed_vlan = in esw_egress_acl_vlan_create() 73 mlx5_add_flow_rules(vport->egress.acl, spec, in esw_egress_acl_vlan_create() 75 if (IS_ERR(vport->egress.allowed_vlan)) { in esw_egress_acl_vlan_create() 76 err = PTR_ERR(vport->egress.allowed_vlan); in esw_egress_acl_vlan_create() 79 vport->vport, err); in esw_egress_acl_vlan_create() [all …]
|
/linux/drivers/scsi/lpfc/ |
H A D | lpfc_vport.c | 53 inline void lpfc_vport_set_state(struct lpfc_vport *vport, in lpfc_vport_set_state() argument 56 struct fc_vport *fc_vport = vport->fc_vport; in lpfc_vport_set_state() 76 vport->port_state = LPFC_VPORT_FAILED; in lpfc_vport_set_state() 79 vport->port_state = LPFC_VPORT_UNKNOWN; in lpfc_vport_set_state() 118 lpfc_vport_sparm(struct lpfc_hba *phba, struct lpfc_vport *vport) in lpfc_vport_sparm() argument 131 rc = lpfc_read_sparam(phba, pmb, vport->vpi); in lpfc_vport_sparm() 146 pmb->vport = vport; in lpfc_vport_sparm() 150 lpfc_printf_vlog(vport, KERN_ERR, LOG_TRACE_EVENT, in lpfc_vport_sparm() 158 lpfc_printf_vlog(vport, KERN_ERR, LOG_TRACE_EVENT, in lpfc_vport_sparm() 170 memcpy(&vport->fc_sparam, mp->virt, sizeof (struct serv_parm)); in lpfc_vport_sparm() [all …]
|
H A D | lpfc_vmid.c | 41 * @vport: The virtual port for which this call is being executed. 47 struct lpfc_vmid *lpfc_get_vmid_from_hashtable(struct lpfc_vport *vport, in lpfc_get_vmid_from_hashtable() argument 52 hash_for_each_possible(vport->hash_table, vmp, hnode, hash) { in lpfc_get_vmid_from_hashtable() 61 * @vport: The virtual port for which this call is being executed. 69 lpfc_put_vmid_in_hashtable(struct lpfc_vport *vport, u32 hash, in lpfc_put_vmid_in_hashtable() argument 72 hash_add(vport->hash_table, &vmp->hnode, hash); in lpfc_put_vmid_in_hashtable() 102 * @vport: The virtual port for which this call is being executed. 107 static void lpfc_vmid_update_entry(struct lpfc_vport *vport, in lpfc_vmid_update_entry() argument 114 if (vport->phba->pport->vmid_flag & LPFC_VMID_TYPE_PRIO) in lpfc_vmid_update_entry() 116 else if (vport in lpfc_vmid_update_entry() 129 lpfc_vmid_assign_cs_ctl(struct lpfc_vport * vport,struct lpfc_vmid * vmid) lpfc_vmid_assign_cs_ctl() argument 158 lpfc_vmid_get_appid(struct lpfc_vport * vport,char * uuid,enum dma_data_direction iodir,union lpfc_vmid_io_tag * tag) lpfc_vmid_get_appid() argument 298 lpfc_reinit_vmid(struct lpfc_vport * vport) lpfc_reinit_vmid() argument [all...] |
H A D | lpfc_ct.c | 157 struct lpfc_vport *vport = ndlp->vport; in lpfc_ct_reject_event() local 158 struct lpfc_hba *phba = vport->phba; in lpfc_ct_reject_event() 241 cmdiocbq->vport = vport; in lpfc_ct_reject_event() 267 lpfc_printf_vlog(vport, KERN_ERR, LOG_ELS, in lpfc_ct_reject_event() 269 rc, vport->fc_flag); in lpfc_ct_reject_event() 286 struct lpfc_vport *vport = ctiocbq->vport; in lpfc_ct_handle_mibreq() local 294 lpfc_vlog_msg(vport, KERN_WARNING, LOG_ELS, in lpfc_ct_handle_mibreq() 301 if (test_bit(FC_UNLOADING, &vport->load_flag)) in lpfc_ct_handle_mibreq() 304 ndlp = lpfc_findnode_did(vport, did); in lpfc_ct_handle_mibreq() 306 lpfc_vlog_msg(vport, KERN_WARNING, LOG_ELS, in lpfc_ct_handle_mibreq() [all …]
|
H A D | lpfc_hbadisc.c | 71 static void lpfc_disc_flush_list(struct lpfc_vport *vport); 123 if (!ndlp->vport) { in lpfc_rport_invalid() 137 struct lpfc_vport *vport; in lpfc_terminate_rport_io() local 144 vport = ndlp->vport; in lpfc_terminate_rport_io() 145 lpfc_debugfs_disc_trc(vport, LPFC_DISC_TRC_RPORT, in lpfc_terminate_rport_io() 150 lpfc_sli_abort_iocb(vport, ndlp->nlp_sid, 0, LPFC_CTX_TGT); in lpfc_terminate_rport_io() 160 struct lpfc_vport *vport; in lpfc_dev_loss_tmo_callbk() local 170 vport = ndlp->vport; in lpfc_dev_loss_tmo_callbk() 171 phba = vport->phba; in lpfc_dev_loss_tmo_callbk() 173 lpfc_debugfs_disc_trc(vport, LPFC_DISC_TRC_RPORT, in lpfc_dev_loss_tmo_callbk() [all …]
|
H A D | lpfc_els.c | 54 static void lpfc_fabric_abort_vport(struct lpfc_vport *vport); 55 static int lpfc_issue_els_fdisc(struct lpfc_vport *vport, 67 static void lpfc_init_cs_ctl_bitmap(struct lpfc_vport *vport); 68 static void lpfc_vmid_set_cs_ctl_range(struct lpfc_vport *vport, u32 min, u32 max); 69 static void lpfc_vmid_put_cs_ctl(struct lpfc_vport *vport, u32 ctcl_vmid); 94 lpfc_els_chk_latt(struct lpfc_vport *vport) in lpfc_els_chk_latt() argument 96 struct lpfc_hba *phba = vport->phba; in lpfc_els_chk_latt() 99 if (vport->port_state >= LPFC_VPORT_READY || in lpfc_els_chk_latt() 112 lpfc_printf_vlog(vport, KERN_ERR, LOG_TRACE_EVENT, in lpfc_els_chk_latt() 123 set_bit(FC_ABORT_DISCOVERY, &vport->fc_flag); in lpfc_els_chk_latt() [all …]
|
H A D | lpfc_nvme.c | 87 struct lpfc_vport *vport; in lpfc_nvme_create_queue() local 95 vport = lport->vport; in lpfc_nvme_create_queue() 97 if (!vport || test_bit(FC_UNLOADING, &vport->load_flag) || in lpfc_nvme_create_queue() 98 test_bit(HBA_IOQ_FLUSH, &vport->phba->hba_flag)) in lpfc_nvme_create_queue() 121 lpfc_printf_vlog(vport, KERN_INFO, LOG_NVME, in lpfc_nvme_create_queue() 149 struct lpfc_vport *vport; in lpfc_nvme_delete_queue() local 155 vport = lport->vport; in lpfc_nvme_delete_queue() 157 lpfc_printf_vlog(vport, KERN_INFO, LOG_NVME, in lpfc_nvme_delete_queue() 168 lpfc_printf_vlog(lport->vport, KERN_INFO, LOG_NVME, in lpfc_nvme_localport_delete() 173 if (lport->vport->localport) in lpfc_nvme_localport_delete() [all …]
|
/linux/drivers/net/ethernet/intel/idpf/ |
H A D | idpf_lib.c | 466 static int idpf_del_mac_filter(struct idpf_vport *vport, in idpf_del_mac_filter() argument 489 err = idpf_add_del_mac_filters(vport, np, false, async); in idpf_del_mac_filter() 545 static int idpf_add_mac_filter(struct idpf_vport *vport, in idpf_add_mac_filter() argument 558 err = idpf_add_del_mac_filters(vport, np, true, async); in idpf_add_mac_filter() 569 static void idpf_del_all_mac_filters(struct idpf_vport *vport) in idpf_del_all_mac_filters() argument 574 vport_config = vport->adapter->vport_config[vport->idx]; in idpf_del_all_mac_filters() 593 static void idpf_restore_mac_filters(struct idpf_vport *vport) in idpf_restore_mac_filters() argument 598 vport_config = vport->adapter->vport_config[vport->idx]; in idpf_restore_mac_filters() 606 idpf_add_del_mac_filters(vport, netdev_priv(vport->netdev), in idpf_restore_mac_filters() 617 static void idpf_remove_mac_filters(struct idpf_vport *vport) in idpf_remove_mac_filters() argument [all …]
|
H A D | idpf_virtchnl.h | 19 int idpf_get_reg_intr_vecs(struct idpf_vport *vport, 21 int idpf_queue_reg_init(struct idpf_vport *vport); 22 int idpf_vport_queue_ids_init(struct idpf_vport *vport); 28 void idpf_vport_init(struct idpf_vport *vport, struct idpf_vport_max_q *max_q); 29 u32 idpf_get_vport_id(struct idpf_vport *vport); 32 int idpf_send_destroy_vport_msg(struct idpf_vport *vport); 33 int idpf_send_enable_vport_msg(struct idpf_vport *vport); 34 int idpf_send_disable_vport_msg(struct idpf_vport *vport); 36 int idpf_vport_adjust_qs(struct idpf_vport *vport); 41 int idpf_send_add_queues_msg(const struct idpf_vport *vport, u16 num_tx_q, [all …]
|
H A D | idpf_txrx.c | 171 static void idpf_tx_desc_rel_all(struct idpf_vport *vport) in idpf_tx_desc_rel_all() argument 175 if (!vport->txq_grps) in idpf_tx_desc_rel_all() 178 for (i = 0; i < vport->num_txq_grp; i++) { in idpf_tx_desc_rel_all() 179 struct idpf_txq_group *txq_grp = &vport->txq_grps[i]; in idpf_tx_desc_rel_all() 184 if (idpf_is_queue_model_split(vport->txq_model)) in idpf_tx_desc_rel_all() 242 static int idpf_tx_desc_alloc(const struct idpf_vport *vport, in idpf_tx_desc_alloc() argument 284 static int idpf_compl_desc_alloc(const struct idpf_vport *vport, in idpf_compl_desc_alloc() argument 308 static int idpf_tx_desc_alloc_all(struct idpf_vport *vport) in idpf_tx_desc_alloc_all() argument 316 for (i = 0; i < vport->num_txq_grp; i++) { in idpf_tx_desc_alloc_all() 317 for (j = 0; j < vport->txq_grps[i].num_txq; j++) { in idpf_tx_desc_alloc_all() [all …]
|
H A D | idpf_ethtool.c | 17 struct idpf_vport *vport; in idpf_get_rxnfc() local 20 vport = idpf_netdev_to_vport(netdev); in idpf_get_rxnfc() 24 cmd->data = vport->num_rxq; in idpf_get_rxnfc() 136 struct idpf_vport *vport; in idpf_set_rxfh() local 141 vport = idpf_netdev_to_vport(netdev); in idpf_set_rxfh() 143 adapter = vport->adapter; in idpf_set_rxfh() 150 rss_data = &adapter->vport_config[vport->idx]->user_config.rss_data; in idpf_set_rxfh() 168 err = idpf_config_rss(vport); in idpf_set_rxfh() 227 struct idpf_vport *vport; in idpf_set_channels() local 239 vport = idpf_netdev_to_vport(netdev); in idpf_set_channels() [all …]
|
H A D | idpf_virtchnl.c | 134 struct idpf_vport *vport; in idpf_handle_event_link() local 136 vport = idpf_vid_to_vport(adapter, le32_to_cpu(v2e->vport_id)); in idpf_handle_event_link() 137 if (!vport) { in idpf_handle_event_link() 142 np = netdev_priv(vport->netdev); in idpf_handle_event_link() 146 if (vport->link_up == v2e->link_status) in idpf_handle_event_link() 149 vport->link_up = v2e->link_status; in idpf_handle_event_link() 154 if (vport->link_up) { in idpf_handle_event_link() 155 netif_tx_start_all_queues(vport->netdev); in idpf_handle_event_link() 156 netif_carrier_on(vport->netdev); in idpf_handle_event_link() 158 netif_tx_stop_all_queues(vport->netdev); in idpf_handle_event_link() [all …]
|
/linux/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
H A D | hclge_mbx.c | 31 static int hclge_gen_resp_to_vf(struct hclge_vport *vport, in hclge_gen_resp_to_vf() argument 36 struct hclge_dev *hdev = vport->back; in hclge_gen_resp_to_vf() 91 static int hclge_send_mbx_msg(struct hclge_vport *vport, u8 *msg, u16 msg_len, in hclge_send_mbx_msg() argument 95 struct hclge_dev *hdev = vport->back; in hclge_send_mbx_msg() 127 int hclge_inform_vf_reset(struct hclge_vport *vport, u16 reset_type) in hclge_inform_vf_reset() argument 132 dest_vfid = (u8)vport->vport_id; in hclge_inform_vf_reset() 136 return hclge_send_mbx_msg(vport, (u8 *)&msg_data, sizeof(msg_data), in hclge_inform_vf_reset() 140 int hclge_inform_reset_assert_to_vf(struct hclge_vport *vport) in hclge_inform_reset_assert_to_vf() argument 142 struct hclge_dev *hdev = vport->back; in hclge_inform_reset_assert_to_vf() 154 return hclge_inform_vf_reset(vport, reset_type); in hclge_inform_reset_assert_to_vf() [all …]
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | eswitch.c | 63 u16 vport; member 109 struct mlx5_vport *vport; in mlx5_eswitch_get_vport() local 114 vport = xa_load(&esw->vports, vport_num); in mlx5_eswitch_get_vport() 115 if (!vport) { in mlx5_eswitch_get_vport() 119 return vport; in mlx5_eswitch_get_vport() 122 static int arm_vport_context_events_cmd(struct mlx5_core_dev *dev, u16 vport, in arm_vport_context_events_cmd() argument 131 MLX5_SET(modify_nic_vport_context_in, in, vport_number, vport); in arm_vport_context_events_cmd() 132 if (vport || mlx5_core_is_ecpf(dev)) in arm_vport_context_events_cmd() 153 int mlx5_eswitch_modify_esw_vport_context(struct mlx5_core_dev *dev, u16 vport, in mlx5_eswitch_modify_esw_vport_context() argument 158 MLX5_SET(modify_esw_vport_context_in, in, vport_number, vport); in mlx5_eswitch_modify_esw_vport_context() [all …]
|
/linux/sound/isa/ |
H A D | sc6000.c | 183 static int sc6000_wait_data(char __iomem *vport) in sc6000_wait_data() argument 189 val = ioread8(vport + DSP_DATAVAIL); in sc6000_wait_data() 198 static int sc6000_read(char __iomem *vport) in sc6000_read() argument 200 if (sc6000_wait_data(vport)) in sc6000_read() 203 return ioread8(vport + DSP_READ); in sc6000_read() 207 static int sc6000_write(struct device *devptr, char __iomem *vport, int cmd) in sc6000_write() argument 213 val = ioread8(vport + DSP_STATUS); in sc6000_write() 218 iowrite8(cmd, vport + DSP_COMMAND); in sc6000_write() 230 char __iomem *vport, int command, in sc6000_dsp_get_answer() argument 235 if (sc6000_write(devptr, vport, command)) { in sc6000_dsp_get_answer() [all …]
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/esw/ |
H A D | qos.c | 92 struct mlx5_vport *vport; member 104 void mlx5_esw_qos_vport_qos_free(struct mlx5_vport *vport) in mlx5_esw_qos_vport_qos_free() argument 106 kfree(vport->qos.sched_node); in mlx5_esw_qos_vport_qos_free() 107 memset(&vport->qos, 0, sizeof(vport->qos)); in mlx5_esw_qos_vport_qos_free() 110 u32 mlx5_esw_qos_vport_get_sched_elem_ix(const struct mlx5_vport *vport) in mlx5_esw_qos_vport_get_sched_elem_ix() argument 112 if (!vport->qos.sched_node) in mlx5_esw_qos_vport_get_sched_elem_ix() 115 return vport->qos.sched_node->ix; in mlx5_esw_qos_vport_get_sched_elem_ix() 119 mlx5_esw_qos_vport_get_parent(const struct mlx5_vport *vport) in mlx5_esw_qos_vport_get_parent() argument 121 if (!vport->qos.sched_node) in mlx5_esw_qos_vport_get_parent() 124 return vport->qos.sched_node->parent; in mlx5_esw_qos_vport_get_parent() [all …]
|