| /linux/drivers/net/ethernet/marvell/octeontx2/af/cn20k/ |
| H A D | mbox_init.c | 347 void cn20k_rvu_disable_afvf_intr(struct rvu *rvu, int vfs) in cn20k_rvu_disable_afvf_intr() argument 349 rvupf_write64(rvu, RVU_MBOX_PF_VFPF_INT_ENA_W1CX(0), INTR_MASK(vfs)); in cn20k_rvu_disable_afvf_intr() 350 rvupf_write64(rvu, RVU_MBOX_PF_VFPF1_INT_ENA_W1CX(0), INTR_MASK(vfs)); in cn20k_rvu_disable_afvf_intr() 351 rvupf_write64(rvu, RVU_PF_VFFLR_INT_ENA_W1CX(0), INTR_MASK(vfs)); in cn20k_rvu_disable_afvf_intr() 352 rvupf_write64(rvu, RVU_PF_VFME_INT_ENA_W1CX(0), INTR_MASK(vfs)); in cn20k_rvu_disable_afvf_intr() 354 if (vfs <= 64) in cn20k_rvu_disable_afvf_intr() 357 rvupf_write64(rvu, RVU_MBOX_PF_VFPF_INT_ENA_W1CX(1), INTR_MASK(vfs - 64)); in cn20k_rvu_disable_afvf_intr() 358 rvupf_write64(rvu, RVU_MBOX_PF_VFPF1_INT_ENA_W1CX(1), INTR_MASK(vfs - 64)); in cn20k_rvu_disable_afvf_intr() 359 rvupf_write64(rvu, RVU_PF_VFFLR_INT_ENA_W1CX(1), INTR_MASK(vfs - 64)); in cn20k_rvu_disable_afvf_intr() 360 rvupf_write64(rvu, RVU_PF_VFME_INT_ENA_W1CX(1), INTR_MASK(vfs - 64)); in cn20k_rvu_disable_afvf_intr() [all …]
|
| H A D | api.h | 30 void cn20k_rvu_enable_afvf_intr(struct rvu *rvu, int vfs); 31 void cn20k_rvu_disable_afvf_intr(struct rvu *rvu, int vfs);
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_sriov.c | 26 struct ice_vfs *vfs = &pf->vfs; in ice_free_vf_entries() local 35 lockdep_assert_held(&vfs->table_lock); in ice_free_vf_entries() 37 hash_for_each_safe(vfs->table, bkt, tmp, vf, entry) { in ice_free_vf_entries() 134 struct ice_vfs *vfs = &pf->vfs; in ice_free_vfs() local 154 mutex_lock(&vfs->table_lock); in ice_free_vfs() 181 vfs->num_qps_per = 0; in ice_free_vfs() 184 mutex_unlock(&vfs->table_lock); in ice_free_vfs() 377 lockdep_assert_held(&pf->vfs.table_lock); in ice_set_per_vf_res() 423 pf->vfs.num_qps_per = min_t(int, num_txq, num_rxq); in ice_set_per_vf_res() 424 pf->vfs.num_msix_per = num_msix_per_vf; in ice_set_per_vf_res() [all …]
|
| H A D | ice_vf_lib.c | 30 hash_for_each_possible_rcu(pf->vfs.table, vf, entry, vf_id) { in ice_get_vf_by_id() 94 return !hash_empty(pf->vfs.table); in ice_has_vfs() 759 mutex_lock(&pf->vfs.table_lock); in ice_reset_all_vfs() 767 mutex_unlock(&pf->vfs.table_lock); in ice_reset_all_vfs() 818 mutex_unlock(&pf->vfs.table_lock); in ice_reset_all_vfs() 999 struct ice_vfs *vfs; in ice_initialize_vf_entry() local 1001 vfs = &pf->vfs; in ice_initialize_vf_entry() 1009 vf->num_msix = vfs->num_msix_per; in ice_initialize_vf_entry() 1010 vf->num_vf_qs = vfs->num_qps_per; in ice_initialize_vf_entry()
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_sriov_pf_service.c | 92 xe->sriov.pf.vfs[vfid].version.major = major; in pf_connect() 93 xe->sriov.pf.vfs[vfid].version.minor = minor; in pf_connect() 100 xe->sriov.pf.vfs[vfid].version.major = 0; in pf_disconnect() 101 xe->sriov.pf.vfs[vfid].version.minor = 0; in pf_disconnect() 121 return major == xe->sriov.pf.vfs[vfid].version.major && in xe_sriov_pf_service_is_negotiated() 122 minor <= xe->sriov.pf.vfs[vfid].version.minor; in xe_sriov_pf_service_is_negotiated() 206 version = &xe->sriov.pf.vfs[n].version; in xe_sriov_pf_service_print_versions()
|
| H A D | xe_gt_sriov_pf_monitor.c | 32 gt->sriov.pf.vfs[vfid].monitor.guc.events[e] = 0; in xe_gt_sriov_pf_monitor_flr() 41 gt->sriov.pf.vfs[vfid].monitor.guc.events[e]++; in pf_update_event_counter() 127 data = >->sriov.pf.vfs[n].monitor; in xe_gt_sriov_pf_monitor_print_events()
|
| H A D | xe_sriov_packet.c | 20 return &xe->sriov.pf.vfs[vfid].migration.lock; in pf_migration_mutex() 29 return &xe->sriov.pf.vfs[vfid].migration.pending; in pf_pick_pending() 39 return &xe->sriov.pf.vfs[vfid].migration.descriptor; in pf_pick_descriptor() 48 return &xe->sriov.pf.vfs[vfid].migration.trailer; in pf_pick_trailer()
|
| H A D | xe_sriov_pf_sysfs.c | 523 xe_assert(xe, !xe->sriov.pf.vfs[n].kobj); in pf_setup_tree() 524 xe->sriov.pf.vfs[n].kobj = kobj; in pf_setup_tree() 539 struct kobject *kobj = xe->sriov.pf.vfs[PFID].kobj; in pf_link_pf_device() 609 err = sysfs_create_link(xe->sriov.pf.vfs[VFID(n)].kobj, in xe_sriov_pf_sysfs_link_vfs() 636 sysfs_remove_link(xe->sriov.pf.vfs[VFID(n)].kobj, "device"); in xe_sriov_pf_sysfs_unlink_vfs()
|
| /linux/drivers/net/ethernet/amd/pds_core/ |
| H A D | main.c | 149 pdsc->vfs = kzalloc_objs(struct pdsc_vf, num_vfs); in pdsc_sriov_configure() 150 if (!pdsc->vfs) in pdsc_sriov_configure() 167 kfree(pdsc->vfs); in pdsc_sriov_configure() 168 pdsc->vfs = NULL; in pdsc_sriov_configure() 191 pf->vfs[vf->vf_id].vf = vf; in pdsc_init_vf() 193 &pf->vfs[vf->vf_id].padev); in pdsc_init_vf() 426 pdsc_auxbus_dev_del(pdsc, pf, &pf->vfs[pdsc->vf_id].padev); in pdsc_remove() 427 pf->vfs[pdsc->vf_id].vf = NULL; in pdsc_remove() 493 &pf->vfs[pdsc->vf_id].padev); in pdsc_reset_prepare() 541 &pf->vfs[pdsc->vf_id].padev); in pdsc_reset_done()
|
| /linux/tools/testing/selftests/powerpc/eeh/ |
| H A D | eeh-functions.sh | 200 vfs=0 226 vfs="$((vfs + 1))" 230 test "$vfs" != 0
|
| /linux/drivers/net/ethernet/marvell/octeontx2/af/ |
| H A D | rvu.c | 2728 int vfs = rvu->vfs; in rvu_mbox_intr_handler() local 2735 if (vfs > 64) { in rvu_mbox_intr_handler() 2739 rvu_queue_work(&rvu->afvf_wq_info, 64, vfs, intr); in rvu_mbox_intr_handler() 2740 vfs = 64; in rvu_mbox_intr_handler() 2748 rvu_queue_work(&rvu->afvf_wq_info, 0, vfs, intr); in rvu_mbox_intr_handler() 2942 if (rvu->vfs > 64) in rvu_flr_intr_handler() 2943 rvu_afvf_queue_flr_work(rvu, 64, rvu->vfs - 64); in rvu_flr_intr_handler() 3307 int vfs = rvu->vfs; in rvu_disable_afvf_intr() local 3310 return cn20k_rvu_disable_afvf_intr(rvu, vfs); in rvu_disable_afvf_intr() 3312 rvupf_write64(rvu, RVU_PF_VFPF_MBOX_INT_ENA_W1CX(0), INTR_MASK(vfs)); in rvu_disable_afvf_intr() [all …]
|
| H A D | rvu_cn10k.c | 28 int pfs, vfs; in lmtst_map_table_ops() local 32 vfs = 1 << (cfg & 0xF); in lmtst_map_table_ops() 35 lmt_map_base = ioremap_wc(tbl_base, pfs * vfs * LMT_MAPTBL_ENTRY_SIZE); in lmtst_map_table_ops()
|
| /linux/drivers/net/ethernet/pensando/ionic/ |
| H A D | ionic_bus_pci.c | 118 if (!ionic->vfs) in ionic_vf_dealloc_locked() 122 v = &ionic->vfs[i]; in ionic_vf_dealloc_locked() 133 kfree(ionic->vfs); in ionic_vf_dealloc_locked() 134 ionic->vfs = NULL; in ionic_vf_dealloc_locked() 154 ionic->vfs = kzalloc_objs(struct ionic_vf, num_vfs); in ionic_vf_alloc() 155 if (!ionic->vfs) { in ionic_vf_alloc() 161 v = &ionic->vfs[i]; in ionic_vf_alloc()
|
| H A D | ionic_lif.c | 2350 if (vf >= pci_num_vf(ionic->pdev) || !ionic->vfs) { in ionic_get_vf_config() 2353 struct ionic_vf *vfdata = &ionic->vfs[vf]; in ionic_get_vf_config() 2382 if (vf >= pci_num_vf(ionic->pdev) || !ionic->vfs) { in ionic_get_vf_stats() 2386 vs = &ionic->vfs[vf].stats; in ionic_get_vf_stats() 2421 if (vf >= pci_num_vf(ionic->pdev) || !ionic->vfs) { in ionic_set_vf_mac() 2430 ether_addr_copy(ionic->vfs[vf].macaddr, mac); in ionic_set_vf_mac() 2460 if (vf >= pci_num_vf(ionic->pdev) || !ionic->vfs) { in ionic_set_vf_vlan() 2469 ionic->vfs[vf].vlanid = cpu_to_le16(vlan); in ionic_set_vf_vlan() 2493 if (vf >= pci_num_vf(ionic->pdev) || !ionic->vfs) { in ionic_set_vf_rate() 2502 ionic->vfs[vf].maxrate = cpu_to_le32(tx_max); in ionic_set_vf_rate() [all …]
|
| /linux/drivers/iio/dac/ |
| H A D | rohm-bd79703.c | 44 int vfs; member 64 *val = data->vfs / 1000; in bd79703_read_raw() 200 data->vfs = ret; in bd79703_probe()
|
| /linux/fs/fuse/ |
| H A D | virtio_fs.c | 193 struct virtio_fs *vfs = container_of(kobj, struct virtio_fs, kobj); in virtio_fs_ktype_release() local 195 kfree(vfs->mq_map); in virtio_fs_ktype_release() 196 kfree(vfs->vqs); in virtio_fs_ktype_release() 197 kfree(vfs); in virtio_fs_ktype_release() 291 struct virtio_fs *vfs = fiq->priv; in virtio_fs_fiq_release() local 293 virtio_fs_put(vfs); in virtio_fs_fiq_release() 1609 struct virtio_fs *vfs = fc->iq.priv; in virtio_fs_conn_destroy() local 1610 struct virtio_fs_vq *fsvq = &vfs->vqs[VQ_HIPRIO]; in virtio_fs_conn_destroy() 1622 virtio_fs_drain_all_queues(vfs); in virtio_fs_conn_destroy() 1631 virtio_fs_stop_all_queues(vfs); in virtio_fs_conn_destroy() [all …]
|
| /linux/drivers/crypto/intel/qat/qat_common/ |
| H A D | adf_dev_mgr.c | 346 int vfs = 0; in adf_get_num_dettached_vfs() local 353 vfs++; in adf_get_num_dettached_vfs() 356 return vfs; in adf_get_num_dettached_vfs()
|
| /linux/scripts/gdb/linux/ |
| H A D | proc.py | 20 from linux import vfs 184 pathname = vfs.dentry_name(mntpoint) + pathname
|
| /linux/drivers/net/ethernet/broadcom/bnx2x/ |
| H A D | bnx2x_sriov.h | 320 struct bnx2x_virtf *vfs; member 321 #define BP_VF(bp, idx) ((BP_VFDB(bp) && (bp)->vfdb->vfs) ? \ 322 &((bp)->vfdb->vfs[idx]) : NULL) 323 #define bnx2x_vf(bp, idx, var) ((bp)->vfdb->vfs[idx].var)
|
| /linux/scripts/gdb/ |
| H A D | vmlinux-gdb.py | 43 import linux.vfs
|
| /linux/tools/testing/ktest/examples/bootconfigs/ |
| H A D | boottrace.bconf | 58 ftrace_filter = "vfs*"
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | port.c | 1074 int vfs; in mlx4_get_slave_num_gids() local 1102 vfs = bitmap_weight(slaves_pport.slaves, dev->persist->num_vfs + 1) - 1; in mlx4_get_slave_num_gids() 1103 if (slave_gid <= ((MLX4_ROCE_MAX_GIDS - MLX4_ROCE_PF_GIDS) % vfs)) in mlx4_get_slave_num_gids() 1104 return ((MLX4_ROCE_MAX_GIDS - MLX4_ROCE_PF_GIDS) / vfs) + 1; in mlx4_get_slave_num_gids() 1105 return (MLX4_ROCE_MAX_GIDS - MLX4_ROCE_PF_GIDS) / vfs; in mlx4_get_slave_num_gids() 1113 int vfs; in mlx4_get_base_gid_ix() local 1140 vfs = bitmap_weight(slaves_pport.slaves, dev->persist->num_vfs + 1) - 1; in mlx4_get_base_gid_ix() 1141 if (slave_gid <= gids % vfs) in mlx4_get_base_gid_ix() 1142 return MLX4_ROCE_PF_GIDS + ((gids / vfs) + 1) * (slave_gid - 1); in mlx4_get_base_gid_ix() 1144 return MLX4_ROCE_PF_GIDS + (gids % vfs) + in mlx4_get_base_gid_ix() [all …]
|
| /linux/samples/ |
| H A D | Makefile | 34 subdir-$(CONFIG_SAMPLE_VFS) += vfs
|
| /linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
| H A D | otx2_pf.c | 82 int irq, vfs = pf->total_vfs; in otx2_disable_flr_me_intr() local 85 otx2_write64(pf, RVU_PF_VFME_INT_ENA_W1CX(0), INTR_MASK(vfs)); in otx2_disable_flr_me_intr() 90 otx2_write64(pf, RVU_PF_VFFLR_INT_ENA_W1CX(0), INTR_MASK(vfs)); in otx2_disable_flr_me_intr() 94 if (vfs <= 64) in otx2_disable_flr_me_intr() 97 otx2_write64(pf, RVU_PF_VFME_INT_ENA_W1CX(1), INTR_MASK(vfs - 64)); in otx2_disable_flr_me_intr() 101 otx2_write64(pf, RVU_PF_VFFLR_INT_ENA_W1CX(1), INTR_MASK(vfs - 64)); in otx2_disable_flr_me_intr() 553 int vfs = pf->total_vfs; in otx2_pfvf_mbox_intr_handler() local 559 if (vfs > 64) { in otx2_pfvf_mbox_intr_handler() 562 otx2_queue_vf_work(mbox, pf->mbox_pfvf_wq, 64, vfs, intr); in otx2_pfvf_mbox_intr_handler() 565 vfs = 64; in otx2_pfvf_mbox_intr_handler() [all …]
|
| /linux/drivers/net/ethernet/qlogic/qed/ |
| H A D | qed_int.c | 2059 u16 vfs = p_hwfn->cdev->p_iov_info->total_vfs; in qed_int_igu_reset_cam() local 2061 if (vfs != p_info->usage.iov_cnt) in qed_int_igu_reset_cam() 2065 p_info->usage.iov_cnt, vfs); in qed_int_igu_reset_cam() 2071 if (vfs > p_info->usage.free_cnt + in qed_int_igu_reset_cam() 2077 p_info->usage.cnt, vfs); in qed_int_igu_reset_cam() 2084 p_info->usage.iov_cnt = vfs; in qed_int_igu_reset_cam()
|