Contributors: 1
Author Tokens Token Proportion Commits Commit Proportion
Vamsi Attunuru 3212 100.00% 1 100.00%
Total 3212 1


// SPDX-License-Identifier: GPL-2.0
/* Marvell Octeon CN10K DPI driver
 *
 * Copyright (C) 2024 Marvell.
 *
 */

#include <linux/bitfield.h>
#include <linux/compat.h>
#include <linux/delay.h>
#include <linux/miscdevice.h>
#include <linux/module.h>
#include <linux/pci.h>
#include <linux/irq.h>
#include <linux/interrupt.h>

#include <uapi/misc/mrvl_cn10k_dpi.h>

/* PCI device IDs */
#define PCI_DEVID_MRVL_CN10K_DPI_PF	0xA080
#define PCI_SUBDEVID_MRVL_CN10K_DPI_PF	0xB900

/* PCI BAR Number */
#define PCI_DPI_CFG_BAR	0

/* MSI-X interrupts */
#define DPI_MAX_REQQ_INT	0x20
#define DPI_MAX_CC_INT		0x40

/* MBOX MSI-X interrupt vector index */
#define DPI_MBOX_PF_VF_INT_IDX	0x75

#define DPI_MAX_IRQS (DPI_MBOX_PF_VF_INT_IDX + 1)

#define DPI_MAX_VFS	0x20

#define DPI_MAX_ENG_FIFO_SZ	0x20
#define DPI_MAX_ENG_MOLR	0x400

#define DPI_DMA_IDS_DMA_NPA_PF_FUNC(x)	FIELD_PREP(GENMASK_ULL(31, 16), x)
#define DPI_DMA_IDS_INST_STRM(x)	FIELD_PREP(GENMASK_ULL(47, 40), x)
#define DPI_DMA_IDS_DMA_STRM(x)		FIELD_PREP(GENMASK_ULL(39, 32), x)
#define DPI_DMA_ENG_EN_MOLR(x)		FIELD_PREP(GENMASK_ULL(41, 32), x)
#define DPI_EBUS_PORTX_CFG_MPS(x)	FIELD_PREP(GENMASK(6, 4), x)
#define DPI_DMA_IDS_DMA_SSO_PF_FUNC(x)	FIELD_PREP(GENMASK(15, 0), x)
#define DPI_DMA_IDS2_INST_AURA(x)	FIELD_PREP(GENMASK(19, 0), x)
#define DPI_DMA_IBUFF_CSIZE_CSIZE(x)	FIELD_PREP(GENMASK(13, 0), x)
#define DPI_EBUS_PORTX_CFG_MRRS(x)	FIELD_PREP(GENMASK(2, 0), x)
#define DPI_ENG_BUF_BLKS(x)		FIELD_PREP(GENMASK(5, 0), x)
#define DPI_DMA_CONTROL_DMA_ENB		GENMASK_ULL(53, 48)

#define DPI_DMA_CONTROL_O_MODE		BIT_ULL(14)
#define DPI_DMA_CONTROL_LDWB		BIT_ULL(32)
#define DPI_DMA_CONTROL_WQECSMODE1	BIT_ULL(37)
#define DPI_DMA_CONTROL_ZBWCSEN		BIT_ULL(39)
#define DPI_DMA_CONTROL_WQECSOFF(ofst)	(((u64)ofst) << 40)
#define DPI_DMA_CONTROL_WQECSDIS	BIT_ULL(47)
#define DPI_DMA_CONTROL_PKT_EN		BIT_ULL(56)
#define DPI_DMA_IBUFF_CSIZE_NPA_FREE	BIT(16)

#define DPI_CTL_EN			BIT_ULL(0)
#define DPI_DMA_CC_INT			BIT_ULL(0)
#define DPI_DMA_QRST			BIT_ULL(0)

#define DPI_REQQ_INT_INSTRFLT		BIT_ULL(0)
#define DPI_REQQ_INT_RDFLT		BIT_ULL(1)
#define DPI_REQQ_INT_WRFLT		BIT_ULL(2)
#define DPI_REQQ_INT_CSFLT		BIT_ULL(3)
#define DPI_REQQ_INT_INST_DBO		BIT_ULL(4)
#define DPI_REQQ_INT_INST_ADDR_NULL	BIT_ULL(5)
#define DPI_REQQ_INT_INST_FILL_INVAL	BIT_ULL(6)
#define DPI_REQQ_INT_INSTR_PSN		BIT_ULL(7)

#define DPI_REQQ_INT \
	(DPI_REQQ_INT_INSTRFLT		| \
	DPI_REQQ_INT_RDFLT		| \
	DPI_REQQ_INT_WRFLT		| \
	DPI_REQQ_INT_CSFLT		| \
	DPI_REQQ_INT_INST_DBO		| \
	DPI_REQQ_INT_INST_ADDR_NULL	| \
	DPI_REQQ_INT_INST_FILL_INVAL	| \
	DPI_REQQ_INT_INSTR_PSN)

#define DPI_PF_RAS_EBI_DAT_PSN	BIT_ULL(0)
#define DPI_PF_RAS_NCB_DAT_PSN	BIT_ULL(1)
#define DPI_PF_RAS_NCB_CMD_PSN	BIT_ULL(2)

#define DPI_PF_RAS_INT \
	(DPI_PF_RAS_EBI_DAT_PSN  | \
	 DPI_PF_RAS_NCB_DAT_PSN  | \
	 DPI_PF_RAS_NCB_CMD_PSN)

/* Message fields in word_l of DPI mailbox structure */
#define DPI_MBOX_VFID(msg)		FIELD_GET(GENMASK_ULL(7, 0), msg)
#define DPI_MBOX_CMD(msg)		FIELD_GET(GENMASK_ULL(11, 8), msg)
#define DPI_MBOX_CBUF_SIZE(msg)		FIELD_GET(GENMASK_ULL(27, 12), msg)
#define DPI_MBOX_CBUF_AURA(msg)		FIELD_GET(GENMASK_ULL(47, 28), msg)
#define DPI_MBOX_SSO_PFFUNC(msg)	FIELD_GET(GENMASK_ULL(63, 48), msg)

/* Message fields in word_h of DPI mailbox structure */
#define DPI_MBOX_NPA_PFFUNC(msg)	FIELD_GET(GENMASK_ULL(15, 0), msg)
#define DPI_MBOX_WQES_COMPL(msg)	FIELD_GET(GENMASK_ULL(16, 16), msg)
#define DPI_MBOX_WQES_OFFSET(msg)	FIELD_GET(GENMASK_ULL(23, 17), msg)

#define DPI_DMAX_IBUFF_CSIZE(x)	(0x0ULL | ((x) << 11))
#define DPI_DMAX_IDS(x)		(0x18ULL | ((x) << 11))
#define DPI_DMAX_IDS2(x)	(0x20ULL | ((x) << 11))
#define DPI_DMAX_QRST(x)	(0x30ULL | ((x) << 11))

#define DPI_CTL				0x10010ULL
#define DPI_DMA_CONTROL			0x10018ULL
#define DPI_PF_RAS			0x10308ULL
#define DPI_PF_RAS_ENA_W1C		0x10318ULL
#define DPI_MBOX_VF_PF_INT		0x16300ULL
#define DPI_MBOX_VF_PF_INT_W1S		0x16308ULL
#define DPI_MBOX_VF_PF_INT_ENA_W1C	0x16310ULL
#define DPI_MBOX_VF_PF_INT_ENA_W1S	0x16318ULL

#define DPI_DMA_ENGX_EN(x)		(0x10040ULL | ((x) << 3))
#define DPI_ENGX_BUF(x)			(0x100C0ULL | ((x) << 3))
#define DPI_EBUS_PORTX_CFG(x)		(0x10100ULL | ((x) << 3))
#define DPI_DMA_CCX_INT(x)		(0x11000ULL | ((x) << 3))
#define DPI_DMA_CCX_INT_ENA_W1C(x)	(0x11800ULL | ((x) << 3))
#define DPI_REQQX_INT(x)		(0x12C00ULL | ((x) << 5))
#define DPI_REQQX_INT_ENA_W1C(x)	(0x13800ULL | ((x) << 5))
#define DPI_MBOX_PF_VF_DATA0(x)		(0x16000ULL | ((x) << 4))
#define DPI_MBOX_PF_VF_DATA1(x)		(0x16008ULL | ((x) << 4))

#define DPI_WCTL_FIF_THR	0x17008ULL

#define DPI_EBUS_MAX_PORTS	2

#define DPI_EBUS_MRRS_MIN	128
#define DPI_EBUS_MRRS_MAX	1024
#define DPI_EBUS_MPS_MIN	128
#define DPI_EBUS_MPS_MAX	1024
#define DPI_WCTL_FIFO_THRESHOLD	0x30

#define DPI_QUEUE_OPEN		0x1
#define DPI_QUEUE_CLOSE		0x2
#define DPI_REG_DUMP		0x3
#define DPI_GET_REG_CFG		0x4
#define DPI_QUEUE_OPEN_V2	0x5

enum dpi_mbox_rsp_type {
	DPI_MBOX_TYPE_CMD,
	DPI_MBOX_TYPE_RSP_ACK,
	DPI_MBOX_TYPE_RSP_NACK,
};

struct dpivf_config {
	u32 aura;
	u16 csize;
	u16 sso_pf_func;
	u16 npa_pf_func;
};

struct dpipf_vf {
	struct dpivf_config vf_config;
	bool setup_done;
	u8 this_vfid;
};

/* DPI device mailbox */
struct dpi_mbox {
	struct work_struct work;
	/* lock to serialize mbox requests */
	struct mutex lock;
	struct dpipf *pf;
	u8 __iomem *pf_vf_data_reg;
	u8 __iomem *vf_pf_data_reg;
};

struct dpipf {
	struct miscdevice miscdev;
	void __iomem *reg_base;
	struct pci_dev *pdev;
	struct dpipf_vf vf[DPI_MAX_VFS];
	/* Mailbox to talk to VFs */
	struct dpi_mbox *mbox[DPI_MAX_VFS];
};

struct dpi_mbox_message {
	uint64_t word_l;
	uint64_t word_h;
};

static inline void dpi_reg_write(struct dpipf *dpi, u64 offset, u64 val)
{
	writeq(val, dpi->reg_base + offset);
}

static inline u64 dpi_reg_read(struct dpipf *dpi, u64 offset)
{
	return readq(dpi->reg_base + offset);
}

static void dpi_wqe_cs_offset(struct dpipf *dpi, u8 offset)
{
	u64 reg;

	reg = dpi_reg_read(dpi, DPI_DMA_CONTROL);
	reg &= ~DPI_DMA_CONTROL_WQECSDIS;
	reg |= DPI_DMA_CONTROL_ZBWCSEN | DPI_DMA_CONTROL_WQECSMODE1;
	reg |= DPI_DMA_CONTROL_WQECSOFF(offset);
	dpi_reg_write(dpi, DPI_DMA_CONTROL, reg);
}

static int dpi_queue_init(struct dpipf *dpi, struct dpipf_vf *dpivf, u8 vf)
{
	u16 sso_pf_func = dpivf->vf_config.sso_pf_func;
	u16 npa_pf_func = dpivf->vf_config.npa_pf_func;
	u16 csize = dpivf->vf_config.csize;
	u32 aura = dpivf->vf_config.aura;
	unsigned long timeout;
	u64 reg;

	dpi_reg_write(dpi, DPI_DMAX_QRST(vf), DPI_DMA_QRST);

	/* Wait for a maximum of 3 sec */
	timeout = jiffies + msecs_to_jiffies(3000);
	while (!time_after(jiffies, timeout)) {
		reg = dpi_reg_read(dpi, DPI_DMAX_QRST(vf));
		if (!(reg & DPI_DMA_QRST))
			break;

		/* Reset would take time for the request cache to drain */
		usleep_range(500, 1000);
	}

	if (reg & DPI_DMA_QRST) {
		dev_err(&dpi->pdev->dev, "Queue reset failed\n");
		return -EBUSY;
	}

	dpi_reg_write(dpi, DPI_DMAX_IDS2(vf), 0);
	dpi_reg_write(dpi, DPI_DMAX_IDS(vf), 0);

	reg = DPI_DMA_IBUFF_CSIZE_CSIZE(csize) | DPI_DMA_IBUFF_CSIZE_NPA_FREE;
	dpi_reg_write(dpi, DPI_DMAX_IBUFF_CSIZE(vf), reg);

	reg = dpi_reg_read(dpi, DPI_DMAX_IDS2(vf));
	reg |= DPI_DMA_IDS2_INST_AURA(aura);
	dpi_reg_write(dpi, DPI_DMAX_IDS2(vf), reg);

	reg = dpi_reg_read(dpi, DPI_DMAX_IDS(vf));
	reg |= DPI_DMA_IDS_DMA_NPA_PF_FUNC(npa_pf_func);
	reg |= DPI_DMA_IDS_DMA_SSO_PF_FUNC(sso_pf_func);
	reg |= DPI_DMA_IDS_DMA_STRM(vf + 1);
	reg |= DPI_DMA_IDS_INST_STRM(vf + 1);
	dpi_reg_write(dpi, DPI_DMAX_IDS(vf), reg);

	return 0;
}

static void dpi_queue_fini(struct dpipf *dpi, u8 vf)
{
	dpi_reg_write(dpi, DPI_DMAX_QRST(vf), DPI_DMA_QRST);

	/* Reset IDS and IDS2 registers */
	dpi_reg_write(dpi, DPI_DMAX_IDS2(vf), 0);
	dpi_reg_write(dpi, DPI_DMAX_IDS(vf), 0);
}

static irqreturn_t dpi_mbox_intr_handler(int irq, void *data)
{
	struct dpipf *dpi = data;
	u64 reg;
	u32 vf;

	reg = dpi_reg_read(dpi, DPI_MBOX_VF_PF_INT);
	if (reg) {
		for (vf = 0; vf < pci_num_vf(dpi->pdev); vf++) {
			if (reg & BIT_ULL(vf))
				schedule_work(&dpi->mbox[vf]->work);
		}
		dpi_reg_write(dpi, DPI_MBOX_VF_PF_INT, reg);
	}

	return IRQ_HANDLED;
}

static int queue_config(struct dpipf *dpi, struct dpipf_vf *dpivf, struct dpi_mbox_message *msg)
{
	int ret = 0;

	switch (DPI_MBOX_CMD(msg->word_l)) {
	case DPI_QUEUE_OPEN:
	case DPI_QUEUE_OPEN_V2:
		dpivf->vf_config.aura = DPI_MBOX_CBUF_AURA(msg->word_l);
		dpivf->vf_config.csize = DPI_MBOX_CMD(msg->word_l) == DPI_QUEUE_OPEN ?
					 DPI_MBOX_CBUF_SIZE(msg->word_l) >> 3 :
					 DPI_MBOX_CBUF_SIZE(msg->word_l);
		dpivf->vf_config.sso_pf_func = DPI_MBOX_SSO_PFFUNC(msg->word_l);
		dpivf->vf_config.npa_pf_func = DPI_MBOX_NPA_PFFUNC(msg->word_h);
		ret = dpi_queue_init(dpi, dpivf, DPI_MBOX_VFID(msg->word_l));
		if (!ret) {
			if (DPI_MBOX_WQES_COMPL(msg->word_h))
				dpi_wqe_cs_offset(dpi, DPI_MBOX_WQES_OFFSET(msg->word_h));
			dpivf->setup_done = true;
		}
		break;
	case DPI_QUEUE_CLOSE:
		memset(&dpivf->vf_config, 0, sizeof(struct dpivf_config));
		dpi_queue_fini(dpi, DPI_MBOX_VFID(msg->word_l));
		dpivf->setup_done = false;
		break;
	default:
		return -EINVAL;
	}

	return ret;
}

static void dpi_pfvf_mbox_work(struct work_struct *work)
{
	struct dpi_mbox *mbox = container_of(work, struct dpi_mbox, work);
	struct dpi_mbox_message msg;
	struct dpipf_vf *dpivf;
	struct dpipf *dpi;
	int vfid, ret;

	dpi = mbox->pf;
	memset(&msg, 0, sizeof(msg));

	mutex_lock(&mbox->lock);
	msg.word_l = readq(mbox->vf_pf_data_reg);
	if (msg.word_l == (u64)-1)
		goto exit;

	vfid = DPI_MBOX_VFID(msg.word_l);
	if (vfid >= pci_num_vf(dpi->pdev))
		goto exit;

	dpivf = &dpi->vf[vfid];
	msg.word_h = readq(mbox->pf_vf_data_reg);

	ret = queue_config(dpi, dpivf, &msg);
	if (ret < 0)
		writeq(DPI_MBOX_TYPE_RSP_NACK, mbox->pf_vf_data_reg);
	else
		writeq(DPI_MBOX_TYPE_RSP_ACK, mbox->pf_vf_data_reg);
exit:
	mutex_unlock(&mbox->lock);
}

/* Setup registers for a PF mailbox */
static void dpi_setup_mbox_regs(struct dpipf *dpi, int vf)
{
	struct dpi_mbox *mbox = dpi->mbox[vf];

	mbox->pf_vf_data_reg = dpi->reg_base + DPI_MBOX_PF_VF_DATA0(vf);
	mbox->vf_pf_data_reg = dpi->reg_base + DPI_MBOX_PF_VF_DATA1(vf);
}

static int dpi_pfvf_mbox_setup(struct dpipf *dpi)
{
	int vf;

	for (vf = 0; vf < DPI_MAX_VFS; vf++) {
		dpi->mbox[vf] = devm_kzalloc(&dpi->pdev->dev, sizeof(*dpi->mbox[vf]), GFP_KERNEL);

		if (!dpi->mbox[vf])
			return -ENOMEM;

		mutex_init(&dpi->mbox[vf]->lock);
		INIT_WORK(&dpi->mbox[vf]->work, dpi_pfvf_mbox_work);
		dpi->mbox[vf]->pf = dpi;
		dpi_setup_mbox_regs(dpi, vf);
	}

	return 0;
}

static void dpi_pfvf_mbox_destroy(struct dpipf *dpi)
{
	unsigned int vf;

	for (vf = 0; vf < DPI_MAX_VFS; vf++) {
		if (work_pending(&dpi->mbox[vf]->work))
			cancel_work_sync(&dpi->mbox[vf]->work);

		dpi->mbox[vf] = NULL;
	}
}

static void dpi_init(struct dpipf *dpi)
{
	unsigned int engine, port;
	u8 mrrs_val, mps_val;
	u64 reg;

	for (engine = 0; engine < DPI_MAX_ENGINES; engine++) {
		if (engine == 4 || engine == 5)
			reg = DPI_ENG_BUF_BLKS(16);
		else
			reg = DPI_ENG_BUF_BLKS(8);

		dpi_reg_write(dpi, DPI_ENGX_BUF(engine), reg);
	}

	reg = DPI_DMA_CONTROL_ZBWCSEN | DPI_DMA_CONTROL_PKT_EN | DPI_DMA_CONTROL_LDWB |
	      DPI_DMA_CONTROL_O_MODE | DPI_DMA_CONTROL_DMA_ENB;

	dpi_reg_write(dpi, DPI_DMA_CONTROL, reg);
	dpi_reg_write(dpi, DPI_CTL, DPI_CTL_EN);

	mrrs_val = 2; /* 512B */
	mps_val = 1; /* 256B */

	for (port = 0; port < DPI_EBUS_MAX_PORTS; port++) {
		reg = dpi_reg_read(dpi, DPI_EBUS_PORTX_CFG(port));
		reg &= ~(DPI_EBUS_PORTX_CFG_MRRS(7) | DPI_EBUS_PORTX_CFG_MPS(7));
		reg |= DPI_EBUS_PORTX_CFG_MPS(mps_val) | DPI_EBUS_PORTX_CFG_MRRS(mrrs_val);
		dpi_reg_write(dpi, DPI_EBUS_PORTX_CFG(port), reg);
	}

	dpi_reg_write(dpi, DPI_WCTL_FIF_THR, DPI_WCTL_FIFO_THRESHOLD);
}

static void dpi_fini(struct dpipf *dpi)
{
	unsigned int engine;

	for (engine = 0; engine < DPI_MAX_ENGINES; engine++)
		dpi_reg_write(dpi, DPI_ENGX_BUF(engine), 0);

	dpi_reg_write(dpi, DPI_DMA_CONTROL, 0);
	dpi_reg_write(dpi, DPI_CTL, 0);
}

static void dpi_free_irq_vectors(void *pdev)
{
	pci_free_irq_vectors((struct pci_dev *)pdev);
}

static int dpi_irq_init(struct dpipf *dpi)
{
	struct pci_dev *pdev = dpi->pdev;
	struct device *dev = &pdev->dev;
	int i, ret;

	/* Clear all RAS interrupts */
	dpi_reg_write(dpi, DPI_PF_RAS, DPI_PF_RAS_INT);

	/* Clear all RAS interrupt enable bits */
	dpi_reg_write(dpi, DPI_PF_RAS_ENA_W1C, DPI_PF_RAS_INT);

	for (i = 0; i < DPI_MAX_REQQ_INT; i++) {
		dpi_reg_write(dpi, DPI_REQQX_INT(i), DPI_REQQ_INT);
		dpi_reg_write(dpi, DPI_REQQX_INT_ENA_W1C(i), DPI_REQQ_INT);
	}

	for (i = 0; i < DPI_MAX_CC_INT; i++) {
		dpi_reg_write(dpi, DPI_DMA_CCX_INT(i), DPI_DMA_CC_INT);
		dpi_reg_write(dpi, DPI_DMA_CCX_INT_ENA_W1C(i), DPI_DMA_CC_INT);
	}

	ret = pci_alloc_irq_vectors(pdev, DPI_MAX_IRQS, DPI_MAX_IRQS, PCI_IRQ_MSIX);
	if (ret != DPI_MAX_IRQS) {
		dev_err(dev, "DPI: Failed to alloc %d msix irqs\n", DPI_MAX_IRQS);
		return ret;
	}

	ret = devm_add_action_or_reset(dev, dpi_free_irq_vectors, pdev);
	if (ret) {
		dev_err(dev, "DPI: Failed to add irq free action\n");
		return ret;
	}

	ret = devm_request_irq(dev, pci_irq_vector(pdev, DPI_MBOX_PF_VF_INT_IDX),
			       dpi_mbox_intr_handler, 0, "dpi-mbox", dpi);
	if (ret) {
		dev_err(dev, "DPI: request_irq failed for mbox; err=%d\n", ret);
		return ret;
	}

	dpi_reg_write(dpi, DPI_MBOX_VF_PF_INT_ENA_W1S, GENMASK_ULL(31, 0));

	return 0;
}

static int dpi_mps_mrrs_config(struct dpipf *dpi, void __user *arg)
{
	struct dpi_mps_mrrs_cfg cfg;
	u8 mrrs_val, mps_val;
	u64 reg;

	if (copy_from_user(&cfg, arg, sizeof(struct dpi_mps_mrrs_cfg)))
		return -EFAULT;

	if (cfg.max_read_req_sz < DPI_EBUS_MRRS_MIN || cfg.max_read_req_sz > DPI_EBUS_MRRS_MAX ||
	    !is_power_of_2(cfg.max_read_req_sz))
		return -EINVAL;

	if (cfg.max_payload_sz < DPI_EBUS_MPS_MIN || cfg.max_payload_sz > DPI_EBUS_MPS_MAX ||
	    !is_power_of_2(cfg.max_payload_sz))
		return -EINVAL;

	if (cfg.port >= DPI_EBUS_MAX_PORTS)
		return -EINVAL;

	 /* Make sure reserved fields are set to 0 */
	if (cfg.reserved)
		return -EINVAL;

	mrrs_val = fls(cfg.max_read_req_sz >> 8);
	mps_val = fls(cfg.max_payload_sz >> 8);

	reg = dpi_reg_read(dpi, DPI_EBUS_PORTX_CFG(cfg.port));
	reg &= ~(DPI_EBUS_PORTX_CFG_MRRS(0x7) | DPI_EBUS_PORTX_CFG_MPS(0x7));
	reg |= DPI_EBUS_PORTX_CFG_MPS(mps_val) | DPI_EBUS_PORTX_CFG_MRRS(mrrs_val);
	dpi_reg_write(dpi, DPI_EBUS_PORTX_CFG(cfg.port), reg);

	return 0;
}

static int dpi_engine_config(struct dpipf *dpi, void __user *arg)
{
	struct dpi_engine_cfg cfg;
	unsigned int engine;
	u8 *eng_buf;
	u64 reg;

	if (copy_from_user(&cfg, arg, sizeof(struct dpi_engine_cfg)))
		return -EFAULT;

	 /* Make sure reserved fields are set to 0 */
	if (cfg.reserved)
		return -EINVAL;

	eng_buf = (u8 *)&cfg.fifo_mask;

	for (engine = 0; engine < DPI_MAX_ENGINES; engine++) {
		if (eng_buf[engine] > DPI_MAX_ENG_FIFO_SZ)
			return -EINVAL;
		dpi_reg_write(dpi, DPI_ENGX_BUF(engine), eng_buf[engine]);

		if (cfg.update_molr) {
			if (cfg.molr[engine] > DPI_MAX_ENG_MOLR)
				return -EINVAL;
			reg = DPI_DMA_ENG_EN_MOLR(cfg.molr[engine]);
			dpi_reg_write(dpi, DPI_DMA_ENGX_EN(engine), reg);
		} else {
			/* Make sure unused fields are set to 0 */
			if (cfg.molr[engine])
				return -EINVAL;
		}
	}

	return 0;
}

static long dpi_dev_ioctl(struct file *fptr, unsigned int cmd, unsigned long data)
{
	void __user *arg = (void __user *)data;
	struct dpipf *dpi;
	int ret;

	dpi = container_of(fptr->private_data, struct dpipf, miscdev);

	switch (cmd) {
	case DPI_MPS_MRRS_CFG:
		ret = dpi_mps_mrrs_config(dpi, arg);
		break;
	case DPI_ENGINE_CFG:
		ret = dpi_engine_config(dpi, arg);
		break;
	default:
		ret = -ENOTTY;
		break;
	}

	return ret;
}

static const struct file_operations dpi_device_fops = {
	.owner = THIS_MODULE,
	.unlocked_ioctl = dpi_dev_ioctl,
	.compat_ioctl = compat_ptr_ioctl,
};

static int dpi_probe(struct pci_dev *pdev, const struct pci_device_id *id)
{
	struct device *dev = &pdev->dev;
	struct dpipf *dpi;
	int ret;

	dpi = devm_kzalloc(dev, sizeof(*dpi), GFP_KERNEL);
	if (!dpi)
		return -ENOMEM;

	dpi->pdev = pdev;

	ret = pcim_enable_device(pdev);
	if (ret) {
		dev_err(dev, "DPI: Failed to enable PCI device\n");
		return ret;
	}

	ret = pcim_iomap_regions(pdev, BIT(0) | BIT(4), KBUILD_MODNAME);
	if (ret) {
		dev_err(dev, "DPI: Failed to request MMIO region\n");
		return ret;
	}

	dpi->reg_base = pcim_iomap_table(pdev)[PCI_DPI_CFG_BAR];

	/* Initialize global PF registers */
	dpi_init(dpi);

	/* Setup PF-VF mailbox */
	ret = dpi_pfvf_mbox_setup(dpi);
	if (ret) {
		dev_err(dev, "DPI: Failed to setup pf-vf mbox\n");
		goto err_dpi_fini;
	}

	/* Register interrupts */
	ret = dpi_irq_init(dpi);
	if (ret) {
		dev_err(dev, "DPI: Failed to initialize irq vectors\n");
		goto err_dpi_mbox_free;
	}

	pci_set_drvdata(pdev, dpi);
	dpi->miscdev.minor = MISC_DYNAMIC_MINOR;
	dpi->miscdev.name = KBUILD_MODNAME;
	dpi->miscdev.fops = &dpi_device_fops;
	dpi->miscdev.parent = dev;

	ret = misc_register(&dpi->miscdev);
	if (ret) {
		dev_err(dev, "DPI: Failed to register misc device\n");
		goto err_dpi_mbox_free;
	}

	return 0;

err_dpi_mbox_free:
	dpi_pfvf_mbox_destroy(dpi);
err_dpi_fini:
	dpi_fini(dpi);
	return ret;
}

static void dpi_remove(struct pci_dev *pdev)
{
	struct dpipf *dpi = pci_get_drvdata(pdev);

	misc_deregister(&dpi->miscdev);
	pci_sriov_configure_simple(pdev, 0);
	dpi_pfvf_mbox_destroy(dpi);
	dpi_fini(dpi);
	pci_set_drvdata(pdev, NULL);
}

static const struct pci_device_id dpi_id_table[] = {
	{ PCI_DEVICE_SUB(PCI_VENDOR_ID_CAVIUM, PCI_DEVID_MRVL_CN10K_DPI_PF,
			 PCI_VENDOR_ID_CAVIUM, PCI_SUBDEVID_MRVL_CN10K_DPI_PF) },
	{ 0, }  /* end of table */
};

static struct pci_driver dpi_driver = {
	.name = KBUILD_MODNAME,
	.id_table = dpi_id_table,
	.probe = dpi_probe,
	.remove = dpi_remove,
	.sriov_configure = pci_sriov_configure_simple,
};

module_pci_driver(dpi_driver);
MODULE_DEVICE_TABLE(pci, dpi_id_table);
MODULE_AUTHOR("Marvell.");
MODULE_DESCRIPTION("Marvell Octeon CN10K DPI Driver");
MODULE_LICENSE("GPL");