Contributors: 2
Author Tokens Token Proportion Commits Commit Proportion
Brett Creeley 2370 99.66% 5 83.33%
Shannon Nelson 8 0.34% 1 16.67%
Total 2378 6


// SPDX-License-Identifier: GPL-2.0
/* Copyright(c) 2023 Advanced Micro Devices, Inc. */

#include <linux/io.h>
#include <linux/types.h>
#include <linux/delay.h>

#include <linux/pds/pds_common.h>
#include <linux/pds/pds_core_if.h>
#include <linux/pds/pds_adminq.h>

#include "vfio_dev.h"
#include "cmds.h"

#define SUSPEND_TIMEOUT_S		5
#define SUSPEND_CHECK_INTERVAL_MS	1

static int pds_vfio_client_adminq_cmd(struct pds_vfio_pci_device *pds_vfio,
				      union pds_core_adminq_cmd *req,
				      union pds_core_adminq_comp *resp,
				      bool fast_poll)
{
	struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio);
	union pds_core_adminq_cmd cmd = {};
	struct pdsc *pdsc;
	int err;

	/* Wrap the client request */
	cmd.client_request.opcode = PDS_AQ_CMD_CLIENT_CMD;
	cmd.client_request.client_id = cpu_to_le16(pds_vfio->client_id);
	memcpy(cmd.client_request.client_cmd, req,
	       sizeof(cmd.client_request.client_cmd));

	pdsc = pdsc_get_pf_struct(pdev);
	if (IS_ERR(pdsc))
		return PTR_ERR(pdsc);

	err = pdsc_adminq_post(pdsc, &cmd, resp, fast_poll);
	if (err && err != -EAGAIN)
		dev_err(pds_vfio_to_dev(pds_vfio),
			"client admin cmd failed: %pe\n", ERR_PTR(err));

	return err;
}

int pds_vfio_register_client_cmd(struct pds_vfio_pci_device *pds_vfio)
{
	struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio);
	char devname[PDS_DEVNAME_LEN];
	struct pdsc *pdsc;
	int ci;

	snprintf(devname, sizeof(devname), "%s.%d-%u", PDS_VFIO_LM_DEV_NAME,
		 pci_domain_nr(pdev->bus),
		 PCI_DEVID(pdev->bus->number, pdev->devfn));

	pdsc = pdsc_get_pf_struct(pdev);
	if (IS_ERR(pdsc))
		return PTR_ERR(pdsc);

	ci = pds_client_register(pdsc, devname);
	if (ci < 0)
		return ci;

	pds_vfio->client_id = ci;

	return 0;
}

void pds_vfio_unregister_client_cmd(struct pds_vfio_pci_device *pds_vfio)
{
	struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio);
	struct pdsc *pdsc;
	int err;

	pdsc = pdsc_get_pf_struct(pdev);
	if (IS_ERR(pdsc))
		return;

	err = pds_client_unregister(pdsc, pds_vfio->client_id);
	if (err)
		dev_err(&pdev->dev, "unregister from DSC failed: %pe\n",
			ERR_PTR(err));

	pds_vfio->client_id = 0;
}

static int
pds_vfio_suspend_wait_device_cmd(struct pds_vfio_pci_device *pds_vfio, u8 type)
{
	union pds_core_adminq_cmd cmd = {
		.lm_suspend_status = {
			.opcode = PDS_LM_CMD_SUSPEND_STATUS,
			.vf_id = cpu_to_le16(pds_vfio->vf_id),
			.type = type,
		},
	};
	struct device *dev = pds_vfio_to_dev(pds_vfio);
	union pds_core_adminq_comp comp = {};
	unsigned long time_limit;
	unsigned long time_start;
	unsigned long time_done;
	int err;

	time_start = jiffies;
	time_limit = time_start + HZ * SUSPEND_TIMEOUT_S;
	do {
		err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, true);
		if (err != -EAGAIN)
			break;

		msleep(SUSPEND_CHECK_INTERVAL_MS);
	} while (time_before(jiffies, time_limit));

	time_done = jiffies;
	dev_dbg(dev, "%s: vf%u: Suspend comp received in %d msecs\n", __func__,
		pds_vfio->vf_id, jiffies_to_msecs(time_done - time_start));

	/* Check the results */
	if (time_after_eq(time_done, time_limit)) {
		dev_err(dev, "%s: vf%u: Suspend comp timeout\n", __func__,
			pds_vfio->vf_id);
		err = -ETIMEDOUT;
	}

	return err;
}

int pds_vfio_suspend_device_cmd(struct pds_vfio_pci_device *pds_vfio, u8 type)
{
	union pds_core_adminq_cmd cmd = {
		.lm_suspend = {
			.opcode = PDS_LM_CMD_SUSPEND,
			.vf_id = cpu_to_le16(pds_vfio->vf_id),
			.type = type,
		},
	};
	struct device *dev = pds_vfio_to_dev(pds_vfio);
	union pds_core_adminq_comp comp = {};
	int err;

	dev_dbg(dev, "vf%u: Suspend device\n", pds_vfio->vf_id);

	/*
	 * The initial suspend request to the firmware starts the device suspend
	 * operation and the firmware returns success if it's started
	 * successfully.
	 */
	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, true);
	if (err) {
		dev_err(dev, "vf%u: Suspend failed: %pe\n", pds_vfio->vf_id,
			ERR_PTR(err));
		return err;
	}

	/*
	 * The subsequent suspend status request(s) check if the firmware has
	 * completed the device suspend process.
	 */
	return pds_vfio_suspend_wait_device_cmd(pds_vfio, type);
}

int pds_vfio_resume_device_cmd(struct pds_vfio_pci_device *pds_vfio, u8 type)
{
	union pds_core_adminq_cmd cmd = {
		.lm_resume = {
			.opcode = PDS_LM_CMD_RESUME,
			.vf_id = cpu_to_le16(pds_vfio->vf_id),
			.type = type,
		},
	};
	struct device *dev = pds_vfio_to_dev(pds_vfio);
	union pds_core_adminq_comp comp = {};

	dev_dbg(dev, "vf%u: Resume device\n", pds_vfio->vf_id);

	return pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, true);
}

int pds_vfio_get_lm_state_size_cmd(struct pds_vfio_pci_device *pds_vfio, u64 *size)
{
	union pds_core_adminq_cmd cmd = {
		.lm_state_size = {
			.opcode = PDS_LM_CMD_STATE_SIZE,
			.vf_id = cpu_to_le16(pds_vfio->vf_id),
		},
	};
	struct device *dev = pds_vfio_to_dev(pds_vfio);
	union pds_core_adminq_comp comp = {};
	int err;

	dev_dbg(dev, "vf%u: Get migration status\n", pds_vfio->vf_id);

	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, false);
	if (err)
		return err;

	*size = le64_to_cpu(comp.lm_state_size.size);
	return 0;
}

static int pds_vfio_dma_map_lm_file(struct device *dev,
				    enum dma_data_direction dir,
				    struct pds_vfio_lm_file *lm_file)
{
	struct pds_lm_sg_elem *sgl, *sge;
	struct scatterlist *sg;
	dma_addr_t sgl_addr;
	size_t sgl_size;
	int err;
	int i;

	if (!lm_file)
		return -EINVAL;

	/* dma map file pages */
	err = dma_map_sgtable(dev, &lm_file->sg_table, dir, 0);
	if (err)
		return err;

	lm_file->num_sge = lm_file->sg_table.nents;

	/* alloc sgl */
	sgl_size = lm_file->num_sge * sizeof(struct pds_lm_sg_elem);
	sgl = kzalloc(sgl_size, GFP_KERNEL);
	if (!sgl) {
		err = -ENOMEM;
		goto out_unmap_sgtable;
	}

	/* fill sgl */
	sge = sgl;
	for_each_sgtable_dma_sg(&lm_file->sg_table, sg, i) {
		sge->addr = cpu_to_le64(sg_dma_address(sg));
		sge->len = cpu_to_le32(sg_dma_len(sg));
		dev_dbg(dev, "addr = %llx, len = %u\n", sge->addr, sge->len);
		sge++;
	}

	sgl_addr = dma_map_single(dev, sgl, sgl_size, DMA_TO_DEVICE);
	if (dma_mapping_error(dev, sgl_addr)) {
		err = -EIO;
		goto out_free_sgl;
	}

	lm_file->sgl = sgl;
	lm_file->sgl_addr = sgl_addr;

	return 0;

out_free_sgl:
	kfree(sgl);
out_unmap_sgtable:
	lm_file->num_sge = 0;
	dma_unmap_sgtable(dev, &lm_file->sg_table, dir, 0);
	return err;
}

static void pds_vfio_dma_unmap_lm_file(struct device *dev,
				       enum dma_data_direction dir,
				       struct pds_vfio_lm_file *lm_file)
{
	if (!lm_file)
		return;

	/* free sgl */
	if (lm_file->sgl) {
		dma_unmap_single(dev, lm_file->sgl_addr,
				 lm_file->num_sge * sizeof(*lm_file->sgl),
				 DMA_TO_DEVICE);
		kfree(lm_file->sgl);
		lm_file->sgl = NULL;
		lm_file->sgl_addr = DMA_MAPPING_ERROR;
		lm_file->num_sge = 0;
	}

	/* dma unmap file pages */
	dma_unmap_sgtable(dev, &lm_file->sg_table, dir, 0);
}

int pds_vfio_get_lm_state_cmd(struct pds_vfio_pci_device *pds_vfio)
{
	union pds_core_adminq_cmd cmd = {
		.lm_save = {
			.opcode = PDS_LM_CMD_SAVE,
			.vf_id = cpu_to_le16(pds_vfio->vf_id),
		},
	};
	struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio);
	struct device *pdsc_dev = &pci_physfn(pdev)->dev;
	union pds_core_adminq_comp comp = {};
	struct pds_vfio_lm_file *lm_file;
	int err;

	dev_dbg(&pdev->dev, "vf%u: Get migration state\n", pds_vfio->vf_id);

	lm_file = pds_vfio->save_file;

	err = pds_vfio_dma_map_lm_file(pdsc_dev, DMA_FROM_DEVICE, lm_file);
	if (err) {
		dev_err(&pdev->dev, "failed to map save migration file: %pe\n",
			ERR_PTR(err));
		return err;
	}

	cmd.lm_save.sgl_addr = cpu_to_le64(lm_file->sgl_addr);
	cmd.lm_save.num_sge = cpu_to_le32(lm_file->num_sge);

	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, false);
	if (err)
		dev_err(&pdev->dev, "failed to get migration state: %pe\n",
			ERR_PTR(err));

	pds_vfio_dma_unmap_lm_file(pdsc_dev, DMA_FROM_DEVICE, lm_file);

	return err;
}

int pds_vfio_set_lm_state_cmd(struct pds_vfio_pci_device *pds_vfio)
{
	union pds_core_adminq_cmd cmd = {
		.lm_restore = {
			.opcode = PDS_LM_CMD_RESTORE,
			.vf_id = cpu_to_le16(pds_vfio->vf_id),
		},
	};
	struct pci_dev *pdev = pds_vfio_to_pci_dev(pds_vfio);
	struct device *pdsc_dev = &pci_physfn(pdev)->dev;
	union pds_core_adminq_comp comp = {};
	struct pds_vfio_lm_file *lm_file;
	int err;

	dev_dbg(&pdev->dev, "vf%u: Set migration state\n", pds_vfio->vf_id);

	lm_file = pds_vfio->restore_file;

	err = pds_vfio_dma_map_lm_file(pdsc_dev, DMA_TO_DEVICE, lm_file);
	if (err) {
		dev_err(&pdev->dev,
			"failed to map restore migration file: %pe\n",
			ERR_PTR(err));
		return err;
	}

	cmd.lm_restore.sgl_addr = cpu_to_le64(lm_file->sgl_addr);
	cmd.lm_restore.num_sge = cpu_to_le32(lm_file->num_sge);

	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, false);
	if (err)
		dev_err(&pdev->dev, "failed to set migration state: %pe\n",
			ERR_PTR(err));

	pds_vfio_dma_unmap_lm_file(pdsc_dev, DMA_TO_DEVICE, lm_file);

	return err;
}

void pds_vfio_send_host_vf_lm_status_cmd(struct pds_vfio_pci_device *pds_vfio,
					 enum pds_lm_host_vf_status vf_status)
{
	union pds_core_adminq_cmd cmd = {
		.lm_host_vf_status = {
			.opcode = PDS_LM_CMD_HOST_VF_STATUS,
			.vf_id = cpu_to_le16(pds_vfio->vf_id),
			.status = vf_status,
		},
	};
	struct device *dev = pds_vfio_to_dev(pds_vfio);
	union pds_core_adminq_comp comp = {};
	int err;

	dev_dbg(dev, "vf%u: Set host VF LM status: %u", pds_vfio->vf_id,
		vf_status);
	if (vf_status != PDS_LM_STA_IN_PROGRESS &&
	    vf_status != PDS_LM_STA_NONE) {
		dev_warn(dev, "Invalid host VF migration status, %d\n",
			 vf_status);
		return;
	}

	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, false);
	if (err)
		dev_warn(dev, "failed to send host VF migration status: %pe\n",
			 ERR_PTR(err));
}

int pds_vfio_dirty_status_cmd(struct pds_vfio_pci_device *pds_vfio,
			      u64 regions_dma, u8 *max_regions, u8 *num_regions)
{
	union pds_core_adminq_cmd cmd = {
		.lm_dirty_status = {
			.opcode = PDS_LM_CMD_DIRTY_STATUS,
			.vf_id = cpu_to_le16(pds_vfio->vf_id),
		},
	};
	struct device *dev = pds_vfio_to_dev(pds_vfio);
	union pds_core_adminq_comp comp = {};
	int err;

	dev_dbg(dev, "vf%u: Dirty status\n", pds_vfio->vf_id);

	cmd.lm_dirty_status.regions_dma = cpu_to_le64(regions_dma);
	cmd.lm_dirty_status.max_regions = *max_regions;

	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, false);
	if (err) {
		dev_err(dev, "failed to get dirty status: %pe\n", ERR_PTR(err));
		return err;
	}

	/* only support seq_ack approach for now */
	if (!(le32_to_cpu(comp.lm_dirty_status.bmp_type_mask) &
	      BIT(PDS_LM_DIRTY_BMP_TYPE_SEQ_ACK))) {
		dev_err(dev, "Dirty bitmap tracking SEQ_ACK not supported\n");
		return -EOPNOTSUPP;
	}

	*num_regions = comp.lm_dirty_status.num_regions;
	*max_regions = comp.lm_dirty_status.max_regions;

	dev_dbg(dev,
		"Page Tracking Status command successful, max_regions: %d, num_regions: %d, bmp_type: %s\n",
		*max_regions, *num_regions, "PDS_LM_DIRTY_BMP_TYPE_SEQ_ACK");

	return 0;
}

int pds_vfio_dirty_enable_cmd(struct pds_vfio_pci_device *pds_vfio,
			      u64 regions_dma, u8 num_regions)
{
	union pds_core_adminq_cmd cmd = {
		.lm_dirty_enable = {
			.opcode = PDS_LM_CMD_DIRTY_ENABLE,
			.vf_id = cpu_to_le16(pds_vfio->vf_id),
			.regions_dma = cpu_to_le64(regions_dma),
			.bmp_type = PDS_LM_DIRTY_BMP_TYPE_SEQ_ACK,
			.num_regions = num_regions,
		},
	};
	struct device *dev = pds_vfio_to_dev(pds_vfio);
	union pds_core_adminq_comp comp = {};
	int err;

	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, false);
	if (err) {
		dev_err(dev, "failed dirty tracking enable: %pe\n",
			ERR_PTR(err));
		return err;
	}

	return 0;
}

int pds_vfio_dirty_disable_cmd(struct pds_vfio_pci_device *pds_vfio)
{
	union pds_core_adminq_cmd cmd = {
		.lm_dirty_disable = {
			.opcode = PDS_LM_CMD_DIRTY_DISABLE,
			.vf_id = cpu_to_le16(pds_vfio->vf_id),
		},
	};
	struct device *dev = pds_vfio_to_dev(pds_vfio);
	union pds_core_adminq_comp comp = {};
	int err;

	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, false);
	if (err || comp.lm_dirty_status.num_regions != 0) {
		/* in case num_regions is still non-zero after disable */
		err = err ? err : -EIO;
		dev_err(dev,
			"failed dirty tracking disable: %pe, num_regions %d\n",
			ERR_PTR(err), comp.lm_dirty_status.num_regions);
		return err;
	}

	return 0;
}

int pds_vfio_dirty_seq_ack_cmd(struct pds_vfio_pci_device *pds_vfio,
			       u64 sgl_dma, u16 num_sge, u32 offset,
			       u32 total_len, bool read_seq)
{
	const char *cmd_type_str = read_seq ? "read_seq" : "write_ack";
	union pds_core_adminq_cmd cmd = {
		.lm_dirty_seq_ack = {
			.vf_id = cpu_to_le16(pds_vfio->vf_id),
			.len_bytes = cpu_to_le32(total_len),
			.off_bytes = cpu_to_le32(offset),
			.sgl_addr = cpu_to_le64(sgl_dma),
			.num_sge = cpu_to_le16(num_sge),
		},
	};
	struct device *dev = pds_vfio_to_dev(pds_vfio);
	union pds_core_adminq_comp comp = {};
	int err;

	if (read_seq)
		cmd.lm_dirty_seq_ack.opcode = PDS_LM_CMD_DIRTY_READ_SEQ;
	else
		cmd.lm_dirty_seq_ack.opcode = PDS_LM_CMD_DIRTY_WRITE_ACK;

	err = pds_vfio_client_adminq_cmd(pds_vfio, &cmd, &comp, false);
	if (err) {
		dev_err(dev, "failed cmd Page Tracking %s: %pe\n", cmd_type_str,
			ERR_PTR(err));
		return err;
	}

	return 0;
}