Home
last modified time | relevance | path

Searched refs:sriov_info (Results 1 – 12 of 12) sorted by relevance

/linux/drivers/net/ethernet/huawei/hinic/
H A Dhinic_sriov.c520 return &nic_dev->sriov_info; in hinic_get_sriov_info_by_pcidev()
601 struct hinic_sriov_info *sriov_info; in hinic_ndo_get_vf_config() local
603 sriov_info = &nic_dev->sriov_info; in hinic_ndo_get_vf_config()
604 if (vf >= sriov_info->num_vfs) in hinic_ndo_get_vf_config()
607 hinic_get_vf_config(sriov_info->hwdev, OS_VF_ID_TO_HW(vf), ivi); in hinic_ndo_get_vf_config()
645 struct hinic_sriov_info *sriov_info; in hinic_ndo_set_vf_mac() local
648 sriov_info = &nic_dev->sriov_info; in hinic_ndo_set_vf_mac()
649 if (!is_valid_ether_addr(mac) || vf >= sriov_info->num_vfs) in hinic_ndo_set_vf_mac()
652 err = hinic_set_vf_mac(sriov_info->hwdev, OS_VF_ID_TO_HW(vf), mac); in hinic_ndo_set_vf_mac()
760 dev_err(&nic_dev->sriov_info.pdev->dev, "Failed to delete vf %d old vlan %d\n", in set_hw_vf_vlan()
[all …]
H A Dhinic_dev.h111 struct hinic_sriov_info sriov_info; member
H A Dhinic_main.c1216 nic_dev->sriov_info.hwdev = hwdev; in nic_dev_init()
1217 nic_dev->sriov_info.pdev = pdev; in nic_dev_init()
1398 struct hinic_sriov_info *sriov_info = &nic_dev->sriov_info; in wait_sriov_cfg_complete() local
1401 set_bit(HINIC_FUNC_REMOVE, &sriov_info->state); in wait_sriov_cfg_complete()
1405 if (!test_bit(HINIC_SRIOV_ENABLE, &sriov_info->state) && in wait_sriov_cfg_complete()
1406 !test_bit(HINIC_SRIOV_DISABLE, &sriov_info->state)) in wait_sriov_cfg_complete()
/linux/drivers/net/ethernet/cavium/liquidio/
H A Dcn23xx_pf_device.c158 (oct->sriov_info.trs << CN23XX_PKT_MAC_CTL_RINFO_TRS_BIT_POS); in cn23xx_setup_global_mac_regs()
160 temp = oct->sriov_info.rings_per_vf & 0xff; in cn23xx_setup_global_mac_regs()
164 temp = oct->sriov_info.max_vfs & 0xff; in cn23xx_setup_global_mac_regs()
183 srn = oct->sriov_info.pf_srn; in cn23xx_reset_io_queues()
184 ern = srn + oct->sriov_info.num_pf_rings; in cn23xx_reset_io_queues()
242 srn = oct->sriov_info.pf_srn; in cn23xx_pf_setup_global_input_regs()
243 ern = srn + oct->sriov_info.num_pf_rings; in cn23xx_pf_setup_global_input_regs()
257 if (q_no < oct->sriov_info.pf_srn) { in cn23xx_pf_setup_global_input_regs()
258 vf_num = q_no / oct->sriov_info.rings_per_vf; in cn23xx_pf_setup_global_input_regs()
314 srn = oct->sriov_info.pf_srn; in cn23xx_pf_setup_global_output_regs()
[all …]
H A Dlio_main.c172 int adjusted_q_no = q_no + oct->sriov_info.pf_srn; in octeon_droq_bh()
732 max_vfs = oct->sriov_info.max_vfs; in disable_all_vf_links()
800 vfs_mask1 = READ_ONCE(oct->sriov_info.vf_drv_loaded_mask); in liquidio_watchdog()
801 vfs_mask2 = READ_ONCE(other_oct->sriov_info.vf_drv_loaded_mask); in liquidio_watchdog()
1094 if (oct->sriov_info.sriov_enabled) in octeon_destroy_resources()
1366 oct->sriov_info.max_vfs); in octeon_chip_specific_setup()
2807 if (vfidx < 0 || vfidx >= oct->sriov_info.max_vfs) in __liquidio_set_vf_mac()
2828 oct->sriov_info.vf_macaddr[vfidx] = nctrl.udd[0]; in __liquidio_set_vf_mac()
2843 if (vfidx < 0 || vfidx >= oct->sriov_info.num_vfs_alloced) in liquidio_set_vf_mac()
2867 if (vfidx < 0 || vfidx >= oct->sriov_info.num_vfs_alloced) { in liquidio_set_vf_spoofchk()
[all …]
H A Dcn23xx_vf_device.c108 if (cn23xx_vf_reset_io_queues(oct, oct->sriov_info.rings_per_vf)) in cn23xx_vf_setup_global_input_regs()
111 for (q_no = 0; q_no < (oct->sriov_info.rings_per_vf); q_no++) { in cn23xx_vf_setup_global_input_regs()
156 for (q_no = 0; q_no < (oct->sriov_info.rings_per_vf); q_no++) { in cn23xx_vf_setup_global_output_regs()
645 if (oct->sriov_info.rings_per_vf > rings_per_vf) { in cn23xx_setup_octeon_vf_device()
648 oct->sriov_info.rings_per_vf, rings_per_vf, in cn23xx_setup_octeon_vf_device()
650 oct->sriov_info.rings_per_vf = rings_per_vf; in cn23xx_setup_octeon_vf_device()
658 oct->sriov_info.rings_per_vf = in cn23xx_setup_octeon_vf_device()
661 oct->sriov_info.rings_per_vf = rings_per_vf; in cn23xx_setup_octeon_vf_device()
H A Dlio_ethtool.c514 if (oct->sriov_info.sriov_enabled) { in lio_ethtool_get_channels()
623 if (oct->sriov_info.sriov_enabled) { in lio_ethtool_set_channels()
1021 if_cfg.s.num_iqueues = oct->sriov_info.num_pf_rings; in lio_23xx_reconfigure_queue_count()
1022 if_cfg.s.num_oqueues = oct->sriov_info.num_pf_rings; in lio_23xx_reconfigure_queue_count()
1023 if_cfg.s.base_queue = oct->sriov_info.pf_srn; in lio_23xx_reconfigure_queue_count()
1149 if ((OCTEON_CN23XX_PF(oct)) && !oct->sriov_info.sriov_enabled) in lio_reset_queues()
1168 !oct->sriov_info.sriov_enabled) { in lio_reset_queues()
1169 oct->sriov_info.num_pf_rings = num_qs; in lio_reset_queues()
1176 num_qs = oct->sriov_info.num_pf_rings; in lio_reset_queues()
1196 if (OCTEON_CN23XX_PF(oct) && !oct->sriov_info.sriov_enabled) { in lio_reset_queues()
[all …]
H A Dlio_vf_main.c1970 if_cfg.s.num_iqueues = octeon_dev->sriov_info.rings_per_vf; in setup_nic_devices()
1971 if_cfg.s.num_oqueues = octeon_dev->sriov_info.rings_per_vf; in setup_nic_devices()
2349 if (octeon_allocate_ioq_vector(oct, oct->sriov_info.rings_per_vf)) { in octeon_device_init()
2356 oct->sriov_info.rings_per_vf); in octeon_device_init()
2359 if (octeon_setup_interrupt(oct, oct->sriov_info.rings_per_vf)) in octeon_device_init()
H A Docteon_mailbox.c263 pcie_flr(oct->sriov_info.dpiring_to_vfpcidev_lut[mbox->q_no]); in octeon_mbox_process_cmd()
H A Docteon_device.h566 struct octeon_sriov_info sriov_info; member
H A Docteon_device.c855 ioq_vector->ioq_num = i + oct->sriov_info.pf_srn; in octeon_allocate_ioq_vector()
968 for (q_no = 0; q_no < oct->sriov_info.rings_per_vf; q_no++) { in octeon_set_io_queues_off()
/linux/drivers/gpu/drm/xe/
H A Dxe_debugfs.c102 static int sriov_info(struct seq_file *m, void *data) in sriov_info() function
187 { .name = "sriov_info", .show = sriov_info, },