Lines Matching refs:sriov
281 if (!dev->sriov.is_going_down) in smp_snoop()
296 !dev->sriov.is_going_down) { in smp_snoop()
434 if (dev->sriov.demux[port - 1].guid_cache[i] == guid) in mlx4_ib_find_real_gid()
505 tun_ctx = dev->sriov.demux[port-1].tun[slave]; in mlx4_ib_send_to_slave()
1052 if (mlx4_is_master(dev->dev) && !dev->sriov.is_going_down) in handle_lid_change_event()
1063 if (!dev->sriov.is_going_down) { in handle_client_rereg_event()
1064 mlx4_ib_mcg_port_cleanup(&dev->sriov.demux[port_num - 1], 0); in handle_client_rereg_event()
1192 (long long)atomic64_read(&dev->sriov.demux[port - 1].subnet_prefix), in handle_port_mgmt_change_event()
1194 atomic64_set(&dev->sriov.demux[port - 1].subnet_prefix, in handle_port_mgmt_change_event()
1211 if (mlx4_is_master(dev->dev) && !dev->sriov.is_going_down) in handle_port_mgmt_change_event()
1219 else if (!dev->sriov.is_going_down) { in handle_port_mgmt_change_event()
1267 spin_lock_irqsave(&dev->sriov.going_down_lock, flags); in mlx4_ib_tunnel_comp_handler()
1268 if (!dev->sriov.is_going_down && ctx->state == DEMUX_PV_STATE_ACTIVE) in mlx4_ib_tunnel_comp_handler()
1270 spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags); in mlx4_ib_tunnel_comp_handler()
1341 sqp_ctx = dev->sriov.sqps[port-1]; in mlx4_ib_send_to_wire()
1934 if (dev->sriov.demux[port - 1].tun[slave]) { in free_pv_object()
1935 kfree(dev->sriov.demux[port - 1].tun[slave]); in free_pv_object()
1936 dev->sriov.demux[port - 1].tun[slave] = NULL; in free_pv_object()
2010 ctx->wq = to_mdev(ibdev)->sriov.demux[port - 1].wq; in create_pv_resources()
2081 clean_vf_mcast(&dev->sriov.demux[port - 1], slave); in mlx4_ib_tunnels_update()
2085 dev->sriov.sqps[port - 1], 1); in mlx4_ib_tunnels_update()
2088 dev->sriov.demux[port - 1].tun[slave], 1); in mlx4_ib_tunnels_update()
2094 dev->sriov.demux[port - 1].tun[slave]); in mlx4_ib_tunnels_update()
2099 dev->sriov.sqps[port - 1]); in mlx4_ib_tunnels_update()
2250 dev->sriov.is_going_down = 0; in mlx4_ib_init_sriov()
2251 spin_lock_init(&dev->sriov.going_down_lock); in mlx4_ib_init_sriov()
2286 dev->sriov.demux[i].guid_cache[0] = gid.global.interface_id; in mlx4_ib_init_sriov()
2287 atomic64_set(&dev->sriov.demux[i].subnet_prefix, in mlx4_ib_init_sriov()
2290 &dev->sriov.sqps[i]); in mlx4_ib_init_sriov()
2293 err = mlx4_ib_alloc_demux_ctx(dev, &dev->sriov.demux[i], i + 1); in mlx4_ib_init_sriov()
2305 mlx4_ib_free_demux_ctx(&dev->sriov.demux[i]); in mlx4_ib_init_sriov()
2326 spin_lock_irqsave(&dev->sriov.going_down_lock, flags); in mlx4_ib_close_sriov()
2327 dev->sriov.is_going_down = 1; in mlx4_ib_close_sriov()
2328 spin_unlock_irqrestore(&dev->sriov.going_down_lock, flags); in mlx4_ib_close_sriov()
2331 flush_workqueue(dev->sriov.demux[i].ud_wq); in mlx4_ib_close_sriov()
2332 mlx4_ib_free_sqp_ctx(dev->sriov.sqps[i]); in mlx4_ib_close_sriov()
2333 kfree(dev->sriov.sqps[i]); in mlx4_ib_close_sriov()
2334 dev->sriov.sqps[i] = NULL; in mlx4_ib_close_sriov()
2335 mlx4_ib_free_demux_ctx(&dev->sriov.demux[i]); in mlx4_ib_close_sriov()