Home
last modified time | relevance | path

Searched refs:vfid (Results 1 – 25 of 62) sorted by relevance

123

/linux/drivers/gpu/drm/xe/
H A Dxe_gt_sriov_pf_control.c41 static int guc_action_vf_control_cmd(struct xe_guc *guc, u32 vfid, u32 cmd) in guc_action_vf_control_cmd() argument
47 FIELD_PREP(PF2GUC_VF_CONTROL_REQUEST_MSG_1_VFID, vfid), in guc_action_vf_control_cmd()
56 static int pf_send_vf_control_cmd(struct xe_gt *gt, unsigned int vfid, u32 cmd) in pf_send_vf_control_cmd() argument
60 xe_gt_assert(gt, vfid != PFID); in pf_send_vf_control_cmd()
62 vfid, control_cmd_to_string(cmd)); in pf_send_vf_control_cmd()
64 err = guc_action_vf_control_cmd(&gt->uc.guc, vfid, cmd); in pf_send_vf_control_cmd()
67 vfid, control_cmd_to_string(cmd), ERR_PTR(err)); in pf_send_vf_control_cmd()
71 static int pf_send_vf_pause(struct xe_gt *gt, unsigned int vfid) in pf_send_vf_pause() argument
73 return pf_send_vf_control_cmd(gt, vfid, GUC_PF_TRIGGER_VF_PAUSE); in pf_send_vf_pause()
76 static int pf_send_vf_resume(struct xe_gt *gt, unsigned int vfid) in pf_send_vf_resume() argument
[all …]
H A Dxe_gt_sriov_pf_config.h15 u64 xe_gt_sriov_pf_config_get_ggtt(struct xe_gt *gt, unsigned int vfid);
16 int xe_gt_sriov_pf_config_set_ggtt(struct xe_gt *gt, unsigned int vfid, u64 size);
18 unsigned int vfid, unsigned int num_vfs);
20 unsigned int vfid, unsigned int num_vfs, u64 size);
22 u32 xe_gt_sriov_pf_config_get_ctxs(struct xe_gt *gt, unsigned int vfid);
23 int xe_gt_sriov_pf_config_set_ctxs(struct xe_gt *gt, unsigned int vfid, u32 num_ctxs);
24 int xe_gt_sriov_pf_config_set_fair_ctxs(struct xe_gt *gt, unsigned int vfid, unsigned int num_vfs);
25 int xe_gt_sriov_pf_config_bulk_set_ctxs(struct xe_gt *gt, unsigned int vfid, unsigned int num_vfs,
28 u32 xe_gt_sriov_pf_config_get_dbs(struct xe_gt *gt, unsigned int vfid);
29 int xe_gt_sriov_pf_config_set_dbs(struct xe_gt *gt, unsigned int vfid, u32 num_dbs);
[all …]
H A Dxe_gt_sriov_pf_config.c43 static int guc_action_update_vf_cfg(struct xe_guc *guc, u32 vfid, in guc_action_update_vf_cfg() argument
48 vfid, in guc_action_update_vf_cfg()
60 static int pf_send_vf_cfg_reset(struct xe_gt *gt, u32 vfid) in pf_send_vf_cfg_reset() argument
65 ret = guc_action_update_vf_cfg(guc, vfid, 0, 0); in pf_send_vf_cfg_reset()
74 static int pf_send_vf_cfg_klvs(struct xe_gt *gt, u32 vfid, const u32 *klvs, u32 num_dwords) in pf_send_vf_cfg_klvs() argument
94 ret = guc_action_update_vf_cfg(guc, vfid, xe_bo_ggtt_addr(bo), num_dwords); in pf_send_vf_cfg_klvs()
105 static int pf_push_vf_cfg_klvs(struct xe_gt *gt, unsigned int vfid, u32 num_klvs, in pf_push_vf_cfg_klvs() argument
112 ret = pf_send_vf_cfg_klvs(gt, vfid, klvs, num_dwords); in pf_push_vf_cfg_klvs()
120 xe_sriov_function_name(vfid, name, sizeof(name)), in pf_push_vf_cfg_klvs()
135 static int pf_push_vf_cfg_u32(struct xe_gt *gt, unsigned int vfid, u16 key, u32 value) in pf_push_vf_cfg_u32() argument
[all …]
H A Dxe_gt_sriov_pf_monitor.c24 void xe_gt_sriov_pf_monitor_flr(struct xe_gt *gt, u32 vfid) in xe_gt_sriov_pf_monitor_flr() argument
29 xe_gt_sriov_pf_assert_vfid(gt, vfid); in xe_gt_sriov_pf_monitor_flr()
32 gt->sriov.pf.vfs[vfid].monitor.guc.events[e] = 0; in xe_gt_sriov_pf_monitor_flr()
35 static void pf_update_event_counter(struct xe_gt *gt, u32 vfid, in pf_update_event_counter() argument
41 gt->sriov.pf.vfs[vfid].monitor.guc.events[e]++; in pf_update_event_counter()
44 static int pf_handle_vf_threshold_event(struct xe_gt *gt, u32 vfid, u32 threshold) in pf_handle_vf_threshold_event() argument
50 xe_sriov_function_name(vfid, origin, sizeof(origin)); in pf_handle_vf_threshold_event()
60 origin, xe_gt_sriov_pf_config_get_threshold(gt, vfid, e), in pf_handle_vf_threshold_event()
63 pf_update_event_counter(gt, vfid, e); in pf_handle_vf_threshold_event()
81 u32 vfid; in xe_gt_sriov_pf_monitor_process_guc2pf() local
[all …]
H A Dxe_gt_sriov_pf_debugfs.c183 unsigned int vfid = extract_vfid(data); \
191 err = xe_gt_sriov_pf_config_set_##CONFIG(gt, vfid, val); \
200 unsigned int vfid = extract_vfid(data); \
202 *val = xe_gt_sriov_pf_config_get_##CONFIG(gt, vfid); \
238 unsigned int vfid = extract_vfid(data); in set_threshold() local
246 err = xe_gt_sriov_pf_config_set_threshold(gt, vfid, index, val); in set_threshold()
255 unsigned int vfid = extract_vfid(data); in get_threshold() local
257 *val = xe_gt_sriov_pf_config_get_threshold(gt, vfid, index); in get_threshold()
281 static void pf_add_config_attrs(struct xe_gt *gt, struct dentry *parent, unsigned int vfid) in MAKE_XE_GUC_KLV_THRESHOLDS_SET()
284 xe_gt_assert(gt, vfid == extract_vfid(parent)); in MAKE_XE_GUC_KLV_THRESHOLDS_SET()
[all …]
H A Dxe_lmtt.c257 static void lmtt_drop_pages(struct xe_lmtt *lmtt, unsigned int vfid) in lmtt_drop_pages() argument
262 pt = pd->entries[vfid]; in lmtt_drop_pages()
263 pd->entries[vfid] = NULL; in lmtt_drop_pages()
267 lmtt_write_pte(lmtt, pd, LMTT_PTE_INVALID, vfid); in lmtt_drop_pages()
316 static int lmtt_alloc_range(struct xe_lmtt *lmtt, unsigned int vfid, u64 start, u64 end) in lmtt_alloc_range() argument
325 lmtt_assert(lmtt, vfid <= lmtt->ops->lmtt_pte_num(pd->level)); in lmtt_alloc_range()
329 if (pd->entries[vfid]) in lmtt_alloc_range()
340 lmtt_write_pte(lmtt, pd, pde, vfid); in lmtt_alloc_range()
342 pd->entries[vfid] = pt; in lmtt_alloc_range()
357 static struct xe_lmtt_pt *lmtt_leaf_pt(struct xe_lmtt *lmtt, unsigned int vfid, u64 addr) in lmtt_leaf_pt() argument
[all …]
H A Dxe_gt_sriov_pf.c113 static struct xe_reg xe_reg_vf_to_pf(struct xe_reg vf_reg, unsigned int vfid, u32 stride) in xe_reg_vf_to_pf() argument
118 pf_reg.addr += stride * vfid; in xe_reg_vf_to_pf()
123 static void pf_clear_vf_scratch_regs(struct xe_gt *gt, unsigned int vfid) in pf_clear_vf_scratch_regs() argument
132 scratch = xe_reg_vf_to_pf(MED_VF_SW_FLAG(n), vfid, stride); in pf_clear_vf_scratch_regs()
138 scratch = xe_reg_vf_to_pf(VF_SW_FLAG(n), vfid, stride); in pf_clear_vf_scratch_regs()
151 void xe_gt_sriov_pf_sanitize_hw(struct xe_gt *gt, unsigned int vfid) in xe_gt_sriov_pf_sanitize_hw() argument
155 pf_clear_vf_scratch_regs(gt, vfid); in xe_gt_sriov_pf_sanitize_hw()
H A Dxe_gt_sriov_pf_control.h17 int xe_gt_sriov_pf_control_pause_vf(struct xe_gt *gt, unsigned int vfid);
18 int xe_gt_sriov_pf_control_resume_vf(struct xe_gt *gt, unsigned int vfid);
19 int xe_gt_sriov_pf_control_stop_vf(struct xe_gt *gt, unsigned int vfid);
20 int xe_gt_sriov_pf_control_trigger_flr(struct xe_gt *gt, unsigned int vfid);
H A Dxe_lmtt.h18 int xe_lmtt_prepare_pages(struct xe_lmtt *lmtt, unsigned int vfid, u64 range);
19 int xe_lmtt_populate_pages(struct xe_lmtt *lmtt, unsigned int vfid, struct xe_bo *bo, u64 offset);
20 void xe_lmtt_drop_pages(struct xe_lmtt *lmtt, unsigned int vfid);
H A Dxe_gt_sriov_pf_helpers.h23 #define xe_gt_sriov_pf_assert_vfid(gt, vfid) xe_sriov_pf_assert_vfid(gt_to_xe(gt), (vfid)) argument
H A Dxe_sriov_pf_helpers.h25 #define xe_sriov_pf_assert_vfid(xe, vfid) \ argument
26 xe_assert((xe), (vfid) <= xe_sriov_pf_get_totalvfs(xe))
H A Dxe_gt_pagefault.c28 u8 vfid; member
56 u8 vfid; member
268 pf->asid, pf->vfid, pf->pdata, upper_32_bits(pf->page_addr), in print_pagefault()
294 pf->vfid = FIELD_GET(PFD_VFID, desc->dw2); in get_pagefault()
383 reply.dw1 = FIELD_PREP(PFR_VFID, pf.vfid) | in pf_queue_work_func()
531 acc->asid, acc->vfid, acc->engine_class, acc->engine_instance, in print_acc()
614 acc->vfid = FIELD_GET(ACC_VFID, desc->dw2); in get_acc()
H A Dxe_ggtt.c743 static u64 xe_encode_vfid_pte(u16 vfid) in xe_encode_vfid_pte() argument
745 return FIELD_PREP(GGTT_PTE_VFID, vfid) | XE_PAGE_PRESENT; in xe_encode_vfid_pte()
748 static void xe_ggtt_assign_locked(struct xe_ggtt *ggtt, const struct drm_mm_node *node, u16 vfid) in xe_ggtt_assign_locked() argument
753 u64 pte = xe_encode_vfid_pte(vfid); in xe_ggtt_assign_locked()
777 void xe_ggtt_assign(const struct xe_ggtt_node *node, u16 vfid) in xe_ggtt_assign() argument
780 xe_ggtt_assign_locked(node->ggtt, &node->base, vfid); in xe_ggtt_assign()
/linux/drivers/net/ethernet/marvell/octeon_ep/
H A Doctep_ctrl_net.h262 int octep_ctrl_net_get_link_status(struct octep_device *oct, int vfid);
274 int octep_ctrl_net_set_link_status(struct octep_device *oct, int vfid, bool up,
287 int octep_ctrl_net_set_rx_state(struct octep_device *oct, int vfid, bool up,
299 int octep_ctrl_net_get_mac_addr(struct octep_device *oct, int vfid, u8 *addr);
311 int octep_ctrl_net_set_mac_addr(struct octep_device *oct, int vfid, u8 *addr,
322 int octep_ctrl_net_get_mtu(struct octep_device *oct, int vfid);
334 int octep_ctrl_net_set_mtu(struct octep_device *oct, int vfid, int mtu,
347 int octep_ctrl_net_get_if_stats(struct octep_device *oct, int vfid,
360 int octep_ctrl_net_get_link_info(struct octep_device *oct, int vfid,
374 int vfid,
[all …]
H A Doctep_ctrl_net.c44 u16 sz, int vfid) in init_send_req() argument
53 if (vfid != OCTEP_CTRL_NET_INVALID_VFID) { in init_send_req()
55 msg->hdr.s.vf_idx = vfid; in init_send_req()
127 int octep_ctrl_net_get_link_status(struct octep_device *oct, int vfid) in octep_ctrl_net_get_link_status() argument
133 init_send_req(&d.msg, (void *)req, state_sz, vfid); in octep_ctrl_net_get_link_status()
143 int octep_ctrl_net_set_link_status(struct octep_device *oct, int vfid, bool up, in octep_ctrl_net_set_link_status() argument
149 init_send_req(&d.msg, req, state_sz, vfid); in octep_ctrl_net_set_link_status()
158 int octep_ctrl_net_set_rx_state(struct octep_device *oct, int vfid, bool up, in octep_ctrl_net_set_rx_state() argument
164 init_send_req(&d.msg, req, state_sz, vfid); in octep_ctrl_net_set_rx_state()
173 int octep_ctrl_net_get_mac_addr(struct octep_device *oct, int vfid, u8 *addr) in octep_ctrl_net_get_mac_addr() argument
[all …]
/linux/include/linux/qed/
H A Dqed_iov_if.h16 int (*set_mac) (struct qed_dev *cdev, u8 *mac, int vfid);
18 int (*set_vlan) (struct qed_dev *cdev, u16 vid, int vfid);
26 int (*set_spoof) (struct qed_dev *cdev, int vfid, bool val);
28 int (*set_rate) (struct qed_dev *cdev, int vfid,
31 int (*set_trust) (struct qed_dev *cdev, int vfid, bool trust);
/linux/arch/s390/pci/
H A Dpci_iov.c48 static int zpci_iov_link_virtfn(struct pci_dev *pdev, struct pci_dev *virtfn, int vfid) in zpci_iov_link_virtfn() argument
52 rc = pci_iov_sysfs_link(pdev, virtfn, vfid); in zpci_iov_link_virtfn()
79 int i, vfid, devfn, cand_devfn; in zpci_iov_find_parent_pf() local
88 vfid = zdev->vfn - 1; in zpci_iov_find_parent_pf()
104 cand_devfn = pci_iov_virtfn_devfn(pdev, vfid); in zpci_iov_find_parent_pf()
/linux/drivers/net/ethernet/qlogic/qed/
H A Dqed_sriov.c23 static int qed_iov_bulletin_set_mac(struct qed_hwfn *p_hwfn, u8 *mac, int vfid);
306 int vfid, struct qed_ptt *p_ptt) in qed_iov_post_vf_bulletin() argument
313 p_vf = qed_iov_get_vf_info(p_hwfn, (u16)vfid, true); in qed_iov_post_vf_bulletin()
658 int vfid, bool b_fail_malicious) in _qed_iov_pf_sanity_check() argument
666 if (!qed_iov_is_valid_vfid(p_hwfn, vfid, true, b_fail_malicious)) in _qed_iov_pf_sanity_check()
672 static bool qed_iov_pf_sanity_check(struct qed_hwfn *p_hwfn, int vfid) in qed_iov_pf_sanity_check() argument
674 return _qed_iov_pf_sanity_check(p_hwfn, vfid, true); in qed_iov_pf_sanity_check()
939 u16 vfid, in qed_iov_set_link() argument
945 vfid, in qed_iov_set_link()
1332 static void qed_iov_clean_vf(struct qed_hwfn *p_hwfn, u8 vfid) in qed_iov_clean_vf() argument
[all …]
H A Dqed_cxt.c1798 enum protocol_type type, u32 *p_cid, u8 vfid) in _qed_cxt_acquire_cid() argument
1809 if (vfid >= MAX_NUM_VFS && vfid != QED_CXT_PF_CID) { in _qed_cxt_acquire_cid()
1810 DP_NOTICE(p_hwfn, "VF [%02x] is out of range\n", vfid); in _qed_cxt_acquire_cid()
1815 if (vfid == QED_CXT_PF_CID) in _qed_cxt_acquire_cid()
1818 p_map = &p_mngr->acquired_vf[type][vfid]; in _qed_cxt_acquire_cid()
1838 *p_cid, rel_cid, vfid, type); in _qed_cxt_acquire_cid()
1851 u8 vfid, in qed_cxt_test_cid_acquired() argument
1860 if (vfid == QED_CXT_PF_CID) in qed_cxt_test_cid_acquired()
1863 *pp_map = &p_mngr->acquired_vf[*p_type][vfid]; in qed_cxt_test_cid_acquired()
1873 DP_NOTICE(p_hwfn, "Invalid CID %d vfid %02x", cid, vfid); in qed_cxt_test_cid_acquired()
[all …]
/linux/drivers/crypto/marvell/octeontx/
H A Dotx_cptvf_mbox.c104 cptvf->vfid = mbx.data; in otx_cptvf_handle_mbox_intr()
105 dev_dbg(&cptvf->pdev->dev, "Received VFID %d\n", cptvf->vfid); in otx_cptvf_handle_mbox_intr()
111 cptvf->vfid, in otx_cptvf_handle_mbox_intr()
148 mbx->msg, cptvf->vfid); in cptvf_send_msg_to_pf_timeout()
H A Dotx_cptvf_main.c503 intr, cptvf->vfid); in cptvf_misc_intr_handler()
512 intr, cptvf->vfid); in cptvf_misc_intr_handler()
517 intr, cptvf->vfid); in cptvf_misc_intr_handler()
522 intr, cptvf->vfid); in cptvf_misc_intr_handler()
527 intr, cptvf->vfid); in cptvf_misc_intr_handler()
530 cptvf->vfid); in cptvf_misc_intr_handler()
585 cptvf->vfid); in cptvf_done_intr_handler()
603 cptvf->vfid); in cptvf_set_irq_affinity()
607 cpu = cptvf->vfid % num_online_cpus(); in cptvf_set_irq_affinity()
/linux/drivers/crypto/cavium/cpt/
H A Dcptvf_mbox.c34 cptvf->vfid = mbx.data; in cptvf_handle_mbox_intr()
35 dev_dbg(&cptvf->pdev->dev, "Received VFID %d\n", cptvf->vfid); in cptvf_handle_mbox_intr()
42 cptvf->vfid, ((mbx.data == SE_TYPES) ? "SE" : "AE"), in cptvf_handle_mbox_intr()
77 (mbx->msg & 0xFF), cptvf->vfid); in cptvf_send_msg_to_pf_timeout()
H A Dcptvf_main.c524 intr, cptvf->vfid); in cptvf_misc_intr_handler()
532 intr, cptvf->vfid); in cptvf_misc_intr_handler()
536 intr, cptvf->vfid); in cptvf_misc_intr_handler()
540 intr, cptvf->vfid); in cptvf_misc_intr_handler()
544 intr, cptvf->vfid); in cptvf_misc_intr_handler()
547 cptvf->vfid); in cptvf_misc_intr_handler()
602 cptvf->vfid); in cptvf_done_intr_handler()
619 cptvf->vfid); in cptvf_set_irq_affinity()
623 cpu = cptvf->vfid % num_online_cpus(); in cptvf_set_irq_affinity()
/linux/drivers/infiniband/hw/usnic/
H A Dusnic_ib_verbs.c110 resp.vfid = usnic_vnic_get_index(qp_grp->vf->vnic); in usnic_ib_fill_create_qp_resp()
670 unsigned int vfid; in usnic_ib_mmap() local
677 vfid = vma->vm_pgoff; in usnic_ib_mmap()
679 vma->vm_pgoff, PAGE_SHIFT, vfid); in usnic_ib_mmap()
684 if (usnic_vnic_get_index(vf->vnic) == vfid) { in usnic_ib_mmap()
707 usnic_err("No VF %u found\n", vfid); in usnic_ib_mmap()
/linux/drivers/net/ethernet/broadcom/bnx2x/
H A Dbnx2x_sriov.c1037 static int bnx2x_vf_domain(struct bnx2x *bp, int vfid) in bnx2x_vf_domain() argument
1044 static int bnx2x_vf_bus(struct bnx2x *bp, int vfid) in bnx2x_vf_bus() argument
1050 iov->stride * vfid) >> 8); in bnx2x_vf_bus()
1053 static int bnx2x_vf_devfn(struct bnx2x *bp, int vfid) in bnx2x_vf_devfn() argument
1058 return (dev->devfn + iov->offset + iov->stride * vfid) & 0xff; in bnx2x_vf_devfn()
1531 int vfid; in bnx2x_iov_link_update() local
1536 for_each_vf(bp, vfid) in bnx2x_iov_link_update()
1537 bnx2x_iov_link_update_vf(bp, vfid); in bnx2x_iov_link_update()
1543 int vfid; in bnx2x_iov_nic_init() local
1556 for_each_vf(bp, vfid) { in bnx2x_iov_nic_init()
[all …]

123