From: Satyanarayana K V P <satyanarayana.k....@intel.com>

[ Upstream commit 8c0aff7d92e2be25717669eb65a81a89740a24f2 ]

When both PF and VF devices are enabled on the host, they
resume simultaneously during system resume.

However, the PF must finish provisioning the VF before any
VFs can successfully resume.

Establish a parent-child device link between the PF and VF
devices to ensure the correct order of resumption.

V4 -> V5:
- Added missing break in the error condition.
V3 -> V4:
- Made xe_pci_pf_get_vf_dev() as a static function and updated
  input parameter types.
- Updated xe_sriov_warn() to xe_sriov_abort() when VF device
  cannot be found.
V2 -> V3:
- Added function documentation for xe_pci_pf_get_vf_dev().
- Added assertion if not called from PF.
V1 -> V2:
- Added a helper function to get VF pci_dev.
- Updated xe_sriov_notice() to xe_sriov_warn() if vf pci_dev
  is not found.

Signed-off-by: Satyanarayana K V P <satyanarayana.k....@intel.com>
Cc: Michał Wajdeczko <michal.wajdec...@intel.com>
Cc: Michał Winiarski <michal.winiar...@intel.com>
Cc: Piotr Piórkowski <piotr.piorkow...@intel.com>
Reviewed-by: Piotr Piorkowski <piotr.piorkow...@intel.com>
Signed-off-by: Michal Wajdeczko <michal.wajdec...@intel.com>
Link: 
https://patchwork.freedesktop.org/patch/msgid/20250224102807.11065-2-satyanarayana.k....@intel.com
Signed-off-by: Sasha Levin <sas...@kernel.org>
---
 drivers/gpu/drm/xe/xe_pci_sriov.c | 51 +++++++++++++++++++++++++++++++
 1 file changed, 51 insertions(+)

diff --git a/drivers/gpu/drm/xe/xe_pci_sriov.c 
b/drivers/gpu/drm/xe/xe_pci_sriov.c
index aaceee748287e..09ee8a06fe2ed 100644
--- a/drivers/gpu/drm/xe/xe_pci_sriov.c
+++ b/drivers/gpu/drm/xe/xe_pci_sriov.c
@@ -62,6 +62,55 @@ static void pf_reset_vfs(struct xe_device *xe, unsigned int 
num_vfs)
                        xe_gt_sriov_pf_control_trigger_flr(gt, n);
 }
 
+static struct pci_dev *xe_pci_pf_get_vf_dev(struct xe_device *xe, unsigned int 
vf_id)
+{
+       struct pci_dev *pdev = to_pci_dev(xe->drm.dev);
+
+       xe_assert(xe, IS_SRIOV_PF(xe));
+
+       /* caller must use pci_dev_put() */
+       return pci_get_domain_bus_and_slot(pci_domain_nr(pdev->bus),
+                       pdev->bus->number,
+                       pci_iov_virtfn_devfn(pdev, vf_id));
+}
+
+static void pf_link_vfs(struct xe_device *xe, int num_vfs)
+{
+       struct pci_dev *pdev_pf = to_pci_dev(xe->drm.dev);
+       struct device_link *link;
+       struct pci_dev *pdev_vf;
+       unsigned int n;
+
+       /*
+        * When both PF and VF devices are enabled on the host, during system
+        * resume they are resuming in parallel.
+        *
+        * But PF has to complete the provision of VF first to allow any VFs to
+        * successfully resume.
+        *
+        * Create a parent-child device link between PF and VF devices that will
+        * enforce correct resume order.
+        */
+       for (n = 1; n <= num_vfs; n++) {
+               pdev_vf = xe_pci_pf_get_vf_dev(xe, n - 1);
+
+               /* unlikely, something weird is happening, abort */
+               if (!pdev_vf) {
+                       xe_sriov_err(xe, "Cannot find VF%u device, aborting 
link%s creation!\n",
+                                    n, str_plural(num_vfs));
+                       break;
+               }
+
+               link = device_link_add(&pdev_vf->dev, &pdev_pf->dev,
+                                      DL_FLAG_AUTOREMOVE_CONSUMER);
+               /* unlikely and harmless, continue with other VFs */
+               if (!link)
+                       xe_sriov_notice(xe, "Failed linking VF%u\n", n);
+
+               pci_dev_put(pdev_vf);
+       }
+}
+
 static int pf_enable_vfs(struct xe_device *xe, int num_vfs)
 {
        struct pci_dev *pdev = to_pci_dev(xe->drm.dev);
@@ -92,6 +141,8 @@ static int pf_enable_vfs(struct xe_device *xe, int num_vfs)
        if (err < 0)
                goto failed;
 
+       pf_link_vfs(xe, num_vfs);
+
        xe_sriov_info(xe, "Enabled %u of %u VF%s\n",
                      num_vfs, total_vfs, str_plural(total_vfs));
        return num_vfs;
-- 
2.39.5

Reply via email to