Home
last modified time | relevance | path

Searched refs:vfid (Results 1 – 25 of 70) sorted by relevance

123

/linux/drivers/gpu/drm/xe/
A Dxe_gt_sriov_pf_control.c479 pf_queue_vf(gt, vfid); in pf_enter_vf_pause_guc_done()
498 pf_queue_vf(gt, vfid); in pf_enter_vf_pause_send_pause()
658 pf_queue_vf(gt, vfid); in pf_enter_vf_resume_send_resume()
800 pf_queue_vf(gt, vfid); in pf_enter_vf_stop_send_stop()
926 pf_queue_vf(gt, vfid); in pf_enter_vf_flr_send_start()
976 pf_queue_vf(gt, vfid); in pf_enter_vf_flr_send_finish()
1003 pf_queue_vf(gt, vfid); in pf_enter_vf_flr_reset_mmio()
1022 pf_queue_vf(gt, vfid); in pf_enter_vf_flr_reset_data()
1042 pf_queue_vf(gt, vfid); in pf_enter_vf_flr_reset_config()
1276 u32 vfid; in xe_gt_sriov_pf_control_process_guc2pf() local
[all …]
A Dxe_gt_sriov_pf_config.h15 u64 xe_gt_sriov_pf_config_get_ggtt(struct xe_gt *gt, unsigned int vfid);
16 int xe_gt_sriov_pf_config_set_ggtt(struct xe_gt *gt, unsigned int vfid, u64 size);
18 unsigned int vfid, unsigned int num_vfs);
20 unsigned int vfid, unsigned int num_vfs, u64 size);
22 u32 xe_gt_sriov_pf_config_get_ctxs(struct xe_gt *gt, unsigned int vfid);
28 u32 xe_gt_sriov_pf_config_get_dbs(struct xe_gt *gt, unsigned int vfid);
34 u64 xe_gt_sriov_pf_config_get_lmem(struct xe_gt *gt, unsigned int vfid);
40 u32 xe_gt_sriov_pf_config_get_exec_quantum(struct xe_gt *gt, unsigned int vfid);
47 u32 xe_gt_sriov_pf_config_get_threshold(struct xe_gt *gt, unsigned int vfid,
49 int xe_gt_sriov_pf_config_set_threshold(struct xe_gt *gt, unsigned int vfid,
[all …]
A Dxe_gt_sriov_pf_config.c46 vfid, in guc_action_update_vf_cfg()
474 if (vfid) in xe_gt_sriov_pf_config_get_ggtt()
527 if (vfid) in xe_gt_sriov_pf_config_set_ggtt()
592 for (n = vfid; n < vfid + num_vfs; n++) { in xe_gt_sriov_pf_config_bulk_set_ggtt()
792 if (vfid) in xe_gt_sriov_pf_config_get_ctxs()
847 if (vfid) in xe_gt_sriov_pf_config_set_ctxs()
909 for (n = vfid; n < vfid + num_vfs; n++) { in xe_gt_sriov_pf_config_bulk_set_ctxs()
1092 if (vfid) in xe_gt_sriov_pf_config_get_dbs()
1119 if (vfid) in xe_gt_sriov_pf_config_set_dbs()
1153 for (n = vfid; n < vfid + num_vfs; n++) { in xe_gt_sriov_pf_config_bulk_set_dbs()
[all …]
A Dxe_gt_sriov_pf_monitor.c24 void xe_gt_sriov_pf_monitor_flr(struct xe_gt *gt, u32 vfid) in xe_gt_sriov_pf_monitor_flr() argument
29 xe_gt_sriov_pf_assert_vfid(gt, vfid); in xe_gt_sriov_pf_monitor_flr()
32 gt->sriov.pf.vfs[vfid].monitor.guc.events[e] = 0; in xe_gt_sriov_pf_monitor_flr()
35 static void pf_update_event_counter(struct xe_gt *gt, u32 vfid, in pf_update_event_counter() argument
41 gt->sriov.pf.vfs[vfid].monitor.guc.events[e]++; in pf_update_event_counter()
50 xe_sriov_function_name(vfid, origin, sizeof(origin)); in pf_handle_vf_threshold_event()
60 origin, xe_gt_sriov_pf_config_get_threshold(gt, vfid, e), in pf_handle_vf_threshold_event()
63 pf_update_event_counter(gt, vfid, e); in pf_handle_vf_threshold_event()
81 u32 vfid; in xe_gt_sriov_pf_monitor_process_guc2pf() local
102 if (unlikely(vfid > xe_gt_sriov_pf_get_totalvfs(gt))) in xe_gt_sriov_pf_monitor_process_guc2pf()
[all …]
A Dxe_gt_sriov_pf_debugfs.c175 unsigned int vfid = extract_vfid(data); \
183 err = xe_gt_sriov_pf_config_set_##CONFIG(gt, vfid, val); \
192 unsigned int vfid = extract_vfid(data); \
194 *val = xe_gt_sriov_pf_config_get_##CONFIG(gt, vfid); \
229 unsigned int vfid = extract_vfid(data); in set_threshold() local
246 unsigned int vfid = extract_vfid(data); in get_threshold() local
275 xe_gt_assert(gt, vfid == extract_vfid(parent)); in MAKE_XE_GUC_KLV_THRESHOLDS_SET()
310 int (*fn)(struct xe_gt *gt, unsigned int vfid);
323 unsigned int vfid = extract_vfid(parent); in control_write() local
328 xe_gt_assert(gt, vfid); in control_write()
[all …]
A Dxe_lmtt.c262 pt = pd->entries[vfid]; in lmtt_drop_pages()
263 pd->entries[vfid] = NULL; in lmtt_drop_pages()
329 if (pd->entries[vfid]) in lmtt_alloc_range()
340 lmtt_write_pte(lmtt, pd, pde, vfid); in lmtt_alloc_range()
342 pd->entries[vfid] = pt; in lmtt_alloc_range()
363 pt = pd->entries[vfid]; in lmtt_leaf_pt()
398 pt = lmtt_leaf_pt(lmtt, vfid, start); in lmtt_insert_bo()
427 lmtt_assert(lmtt, vfid); in xe_lmtt_prepare_pages()
450 lmtt_assert(lmtt, vfid); in xe_lmtt_populate_pages()
469 lmtt_assert(lmtt, vfid); in xe_lmtt_drop_pages()
[all …]
A Dxe_gt_sriov_pf_service.c85 xe_gt_sriov_pf_assert_vfid(gt, vfid); in pf_connect()
88 gt->sriov.pf.vfs[vfid].version.major = major; in pf_connect()
94 xe_gt_sriov_pf_assert_vfid(gt, vfid); in pf_disconnect()
96 gt->sriov.pf.vfs[vfid].version.major = 0; in pf_disconnect()
97 gt->sriov.pf.vfs[vfid].version.minor = 0; in pf_disconnect()
102 xe_gt_sriov_pf_assert_vfid(gt, vfid); in pf_is_negotiated()
317 pf_disconnect(gt, vfid); in xe_gt_sriov_pf_service_reset()
328 vfid, wanted_major, wanted_minor); in pf_process_handshake()
335 pf_disconnect(gt, vfid); in pf_process_handshake()
338 vfid, *major, *minor); in pf_process_handshake()
[all …]
A Dxe_gt_sriov_pf_control.h17 int xe_gt_sriov_pf_control_pause_vf(struct xe_gt *gt, unsigned int vfid);
18 int xe_gt_sriov_pf_control_resume_vf(struct xe_gt *gt, unsigned int vfid);
19 int xe_gt_sriov_pf_control_stop_vf(struct xe_gt *gt, unsigned int vfid);
20 int xe_gt_sriov_pf_control_trigger_flr(struct xe_gt *gt, unsigned int vfid);
A Dxe_lmtt.h18 int xe_lmtt_prepare_pages(struct xe_lmtt *lmtt, unsigned int vfid, u64 range);
19 int xe_lmtt_populate_pages(struct xe_lmtt *lmtt, unsigned int vfid, struct xe_bo *bo, u64 offset);
20 void xe_lmtt_drop_pages(struct xe_lmtt *lmtt, unsigned int vfid);
A Dxe_gt_sriov_pf_helpers.h23 #define xe_gt_sriov_pf_assert_vfid(gt, vfid) xe_sriov_pf_assert_vfid(gt_to_xe(gt), (vfid)) argument
A Dxe_sriov_pf_helpers.h25 #define xe_sriov_pf_assert_vfid(xe, vfid) \ argument
26 xe_assert((xe), (vfid) <= xe_sriov_pf_get_totalvfs(xe))
A Dxe_gt_pagefault.c29 u8 vfid; member
57 u8 vfid; member
269 pf->asid, pf->vfid, pf->pdata, upper_32_bits(pf->page_addr), in print_pagefault()
295 pf->vfid = FIELD_GET(PFD_VFID, desc->dw2); in get_pagefault()
384 reply.dw1 = FIELD_PREP(PFR_VFID, pf.vfid) | in pf_queue_work_func()
532 acc->asid, acc->vfid, acc->engine_class, acc->engine_instance, in print_acc()
615 acc->vfid = FIELD_GET(ACC_VFID, desc->dw2); in get_acc()
A Dxe_ggtt.c736 static u64 xe_encode_vfid_pte(u16 vfid) in xe_encode_vfid_pte() argument
738 return FIELD_PREP(GGTT_PTE_VFID, vfid) | XE_PAGE_PRESENT; in xe_encode_vfid_pte()
741 static void xe_ggtt_assign_locked(struct xe_ggtt *ggtt, const struct drm_mm_node *node, u16 vfid) in xe_ggtt_assign_locked() argument
746 u64 pte = xe_encode_vfid_pte(vfid); in xe_ggtt_assign_locked()
770 void xe_ggtt_assign(const struct xe_ggtt_node *node, u16 vfid) in xe_ggtt_assign() argument
773 xe_ggtt_assign_locked(node->ggtt, &node->base, vfid); in xe_ggtt_assign()
/linux/drivers/net/ethernet/marvell/octeon_ep/
A Doctep_ctrl_net.h262 int octep_ctrl_net_get_link_status(struct octep_device *oct, int vfid);
287 int octep_ctrl_net_set_rx_state(struct octep_device *oct, int vfid, bool up,
311 int octep_ctrl_net_set_mac_addr(struct octep_device *oct, int vfid, u8 *addr,
322 int octep_ctrl_net_get_mtu(struct octep_device *oct, int vfid);
334 int octep_ctrl_net_set_mtu(struct octep_device *oct, int vfid, int mtu,
347 int octep_ctrl_net_get_if_stats(struct octep_device *oct, int vfid,
360 int octep_ctrl_net_get_link_info(struct octep_device *oct, int vfid,
374 int vfid,
394 int octep_ctrl_net_get_info(struct octep_device *oct, int vfid,
405 int octep_ctrl_net_dev_remove(struct octep_device *oct, int vfid);
[all …]
A Doctep_ctrl_net.c44 u16 sz, int vfid) in init_send_req() argument
53 if (vfid != OCTEP_CTRL_NET_INVALID_VFID) { in init_send_req()
55 msg->hdr.s.vf_idx = vfid; in init_send_req()
149 init_send_req(&d.msg, req, state_sz, vfid); in octep_ctrl_net_set_link_status()
164 init_send_req(&d.msg, req, state_sz, vfid); in octep_ctrl_net_set_rx_state()
179 init_send_req(&d.msg, req, mac_sz, vfid); in octep_ctrl_net_get_mac_addr()
197 init_send_req(&d.msg, req, mac_sz, vfid); in octep_ctrl_net_set_mac_addr()
212 init_send_req(&d.msg, req, mtu_sz, vfid); in octep_ctrl_net_get_mtu()
229 init_send_req(&d.msg, req, mtu_sz, vfid); in octep_ctrl_net_set_mtu()
246 init_send_req(&d.msg, req, 0, vfid); in octep_ctrl_net_get_if_stats()
[all …]
/linux/include/linux/qed/
A Dqed_iov_if.h16 int (*set_mac) (struct qed_dev *cdev, u8 *mac, int vfid);
18 int (*set_vlan) (struct qed_dev *cdev, u16 vid, int vfid);
26 int (*set_spoof) (struct qed_dev *cdev, int vfid, bool val);
28 int (*set_rate) (struct qed_dev *cdev, int vfid,
31 int (*set_trust) (struct qed_dev *cdev, int vfid, bool trust);
/linux/arch/s390/pci/
A Dpci_iov.c48 static int zpci_iov_link_virtfn(struct pci_dev *pdev, struct pci_dev *virtfn, int vfid) in zpci_iov_link_virtfn() argument
52 rc = pci_iov_sysfs_link(pdev, virtfn, vfid); in zpci_iov_link_virtfn()
68 int vfid = vfn - 1; /* Linux' vfid's start at 0 vfn at 1*/ in zpci_iov_setup_virtfn() local
87 cand_devfn = pci_iov_virtfn_devfn(pdev, vfid); in zpci_iov_setup_virtfn()
89 rc = zpci_iov_link_virtfn(pdev, virtfn, vfid); in zpci_iov_setup_virtfn()
/linux/drivers/net/ethernet/hisilicon/hns3/hns3vf/
A Dhclgevf_trace.h23 __field(u8, vfid)
31 __entry->vfid = req->dest_vfid;
41 __get_str(pciname), __get_str(devname), __entry->vfid,
54 __field(u8, vfid)
63 __entry->vfid = req->mbx_src_vfid;
74 __get_str(pciname), __get_str(devname), __entry->vfid,
/linux/drivers/net/ethernet/qlogic/qed/
A Dqed_sriov.c939 u16 vfid, in qed_iov_set_link() argument
2980 u8 vfid, in qed_iov_pre_update_vport() argument
3733 vfid); in qed_iov_execute_vf_flr_cleanup()
3740 ack_vfs[vfid / 32] |= BIT((vfid % 32)); in qed_iov_execute_vf_flr_cleanup()
3790 u8 vfid; in qed_iov_mark_vf_flr() local
3797 if (BIT((vfid % 32)) & p_disabled_vfs[vfid / 32]) { in qed_iov_mark_vf_flr()
4183 vfid); in qed_iov_bulletin_set_mac()
4189 vfid); in qed_iov_bulletin_set_mac()
4254 vfid); in qed_iov_bulletin_set_udp_ports()
4524 u16 vfid, in qed_sriov_enable_qid_config() argument
[all …]
/linux/drivers/net/ethernet/hisilicon/hns3/hns3pf/
A Dhclge_trace.h24 __field(u8, vfid)
33 __entry->vfid = req->mbx_src_vfid;
44 __get_str(pciname), __get_str(devname), __entry->vfid,
57 __field(u8, vfid)
65 __entry->vfid = req->dest_vfid;
75 __get_str(pciname), __get_str(devname), __entry->vfid,
/linux/drivers/crypto/cavium/cpt/
A Dcptvf_mbox.c34 cptvf->vfid = mbx.data; in cptvf_handle_mbox_intr()
35 dev_dbg(&cptvf->pdev->dev, "Received VFID %d\n", cptvf->vfid); in cptvf_handle_mbox_intr()
42 cptvf->vfid, ((mbx.data == SE_TYPES) ? "SE" : "AE"), in cptvf_handle_mbox_intr()
77 (mbx->msg & 0xFF), cptvf->vfid); in cptvf_send_msg_to_pf_timeout()
A Dcptvf_main.c524 intr, cptvf->vfid); in cptvf_misc_intr_handler()
532 intr, cptvf->vfid); in cptvf_misc_intr_handler()
536 intr, cptvf->vfid); in cptvf_misc_intr_handler()
540 intr, cptvf->vfid); in cptvf_misc_intr_handler()
544 intr, cptvf->vfid); in cptvf_misc_intr_handler()
547 cptvf->vfid); in cptvf_misc_intr_handler()
602 cptvf->vfid); in cptvf_done_intr_handler()
619 cptvf->vfid); in cptvf_set_irq_affinity()
623 cpu = cptvf->vfid % num_online_cpus(); in cptvf_set_irq_affinity()
/linux/drivers/crypto/marvell/octeontx/
A Dotx_cptvf_mbox.c104 cptvf->vfid = mbx.data; in otx_cptvf_handle_mbox_intr()
105 dev_dbg(&cptvf->pdev->dev, "Received VFID %d\n", cptvf->vfid); in otx_cptvf_handle_mbox_intr()
111 cptvf->vfid, in otx_cptvf_handle_mbox_intr()
148 mbx->msg, cptvf->vfid); in cptvf_send_msg_to_pf_timeout()
A Dotx_cptvf_main.c503 intr, cptvf->vfid); in cptvf_misc_intr_handler()
512 intr, cptvf->vfid); in cptvf_misc_intr_handler()
517 intr, cptvf->vfid); in cptvf_misc_intr_handler()
522 intr, cptvf->vfid); in cptvf_misc_intr_handler()
527 intr, cptvf->vfid); in cptvf_misc_intr_handler()
530 cptvf->vfid); in cptvf_misc_intr_handler()
585 cptvf->vfid); in cptvf_done_intr_handler()
603 cptvf->vfid); in cptvf_set_irq_affinity()
607 cpu = cptvf->vfid % num_online_cpus(); in cptvf_set_irq_affinity()
/linux/drivers/net/ethernet/intel/ice/
A Dice_vf_mbx.h18 ice_aq_send_msg_to_vf(struct ice_hw *hw, u16 vfid, u32 v_opcode, u32 v_retval,
31 u16 __always_unused vfid, u32 __always_unused v_opcode, in ice_aq_send_msg_to_vf() argument

Completed in 65 milliseconds

123