Contributors: 1
Author Tokens Token Proportion Commits Commit Proportion
Sarah Walker 4669 100.00% 1 100.00%
Total 4669 1


/* SPDX-License-Identifier: GPL-2.0-only OR MIT */
/* Copyright (c) 2023 Imagination Technologies Ltd. */

#ifndef PVR_ROGUE_FWIF_H
#define PVR_ROGUE_FWIF_H

#include <linux/bits.h>
#include <linux/build_bug.h>
#include <linux/compiler.h>
#include <linux/kernel.h>
#include <linux/types.h>

#include "pvr_rogue_defs.h"
#include "pvr_rogue_fwif_common.h"
#include "pvr_rogue_fwif_shared.h"

/*
 ****************************************************************************
 * Logging type
 ****************************************************************************
 */
#define ROGUE_FWIF_LOG_TYPE_NONE 0x00000000U
#define ROGUE_FWIF_LOG_TYPE_TRACE 0x00000001U
#define ROGUE_FWIF_LOG_TYPE_GROUP_MAIN 0x00000002U
#define ROGUE_FWIF_LOG_TYPE_GROUP_MTS 0x00000004U
#define ROGUE_FWIF_LOG_TYPE_GROUP_CLEANUP 0x00000008U
#define ROGUE_FWIF_LOG_TYPE_GROUP_CSW 0x00000010U
#define ROGUE_FWIF_LOG_TYPE_GROUP_BIF 0x00000020U
#define ROGUE_FWIF_LOG_TYPE_GROUP_PM 0x00000040U
#define ROGUE_FWIF_LOG_TYPE_GROUP_RTD 0x00000080U
#define ROGUE_FWIF_LOG_TYPE_GROUP_SPM 0x00000100U
#define ROGUE_FWIF_LOG_TYPE_GROUP_POW 0x00000200U
#define ROGUE_FWIF_LOG_TYPE_GROUP_HWR 0x00000400U
#define ROGUE_FWIF_LOG_TYPE_GROUP_HWP 0x00000800U
#define ROGUE_FWIF_LOG_TYPE_GROUP_RPM 0x00001000U
#define ROGUE_FWIF_LOG_TYPE_GROUP_DMA 0x00002000U
#define ROGUE_FWIF_LOG_TYPE_GROUP_MISC 0x00004000U
#define ROGUE_FWIF_LOG_TYPE_GROUP_DEBUG 0x80000000U
#define ROGUE_FWIF_LOG_TYPE_GROUP_MASK 0x80007FFEU
#define ROGUE_FWIF_LOG_TYPE_MASK 0x80007FFFU

/* String used in pvrdebug -h output */
#define ROGUE_FWIF_LOG_GROUPS_STRING_LIST \
	"main,mts,cleanup,csw,bif,pm,rtd,spm,pow,hwr,hwp,rpm,dma,misc,debug"

/* Table entry to map log group strings to log type value */
struct rogue_fwif_log_group_map_entry {
	const char *log_group_name;
	u32 log_group_type;
};

/*
 ****************************************************************************
 * ROGUE FW signature checks
 ****************************************************************************
 */
#define ROGUE_FW_SIG_BUFFER_SIZE_MIN (8192)

#define ROGUE_FWIF_TIMEDIFF_ID ((0x1UL << 28) | ROGUE_CR_TIMER)

/*
 ****************************************************************************
 * Trace Buffer
 ****************************************************************************
 */

/* Default size of ROGUE_FWIF_TRACEBUF_SPACE in DWords */
#define ROGUE_FW_TRACE_BUF_DEFAULT_SIZE_IN_DWORDS 12000U
#define ROGUE_FW_TRACE_BUFFER_ASSERT_SIZE 200U
#define ROGUE_FW_THREAD_NUM 1U
#define ROGUE_FW_THREAD_MAX 2U

#define ROGUE_FW_POLL_TYPE_SET 0x80000000U

struct rogue_fwif_file_info_buf {
	char path[ROGUE_FW_TRACE_BUFFER_ASSERT_SIZE];
	char info[ROGUE_FW_TRACE_BUFFER_ASSERT_SIZE];
	u32 line_num;
	u32 padding;
} __aligned(8);

struct rogue_fwif_tracebuf_space {
	u32 trace_pointer;

	u32 trace_buffer_fw_addr;

	/* To be used by host when reading from trace buffer */
	u32 *trace_buffer;

	struct rogue_fwif_file_info_buf assert_buf;
} __aligned(8);

/* Total number of FW fault logs stored */
#define ROGUE_FWIF_FWFAULTINFO_MAX (8U)

struct rogue_fw_fault_info {
	aligned_u64 cr_timer;
	aligned_u64 os_timer;

	u32 data __aligned(8);
	u32 reserved;
	struct rogue_fwif_file_info_buf fault_buf;
} __aligned(8);

enum rogue_fwif_pow_state {
	ROGUE_FWIF_POW_OFF, /* idle and ready to full power down */
	ROGUE_FWIF_POW_ON, /* running HW commands */
	ROGUE_FWIF_POW_FORCED_IDLE, /* forced idle */
	ROGUE_FWIF_POW_IDLE, /* idle waiting for host handshake */
};

/* Firmware HWR states */
/* The HW state is ok or locked up */
#define ROGUE_FWIF_HWR_HARDWARE_OK BIT(0)
/* Tells if a HWR reset is in progress */
#define ROGUE_FWIF_HWR_RESET_IN_PROGRESS BIT(1)
/* A DM unrelated lockup has been detected */
#define ROGUE_FWIF_HWR_GENERAL_LOCKUP BIT(3)
/* At least one DM is running without being close to a lockup */
#define ROGUE_FWIF_HWR_DM_RUNNING_OK BIT(4)
/* At least one DM is close to lockup */
#define ROGUE_FWIF_HWR_DM_STALLING BIT(5)
/* The FW has faulted and needs to restart */
#define ROGUE_FWIF_HWR_FW_FAULT BIT(6)
/* The FW has requested the host to restart it */
#define ROGUE_FWIF_HWR_RESTART_REQUESTED BIT(7)

#define ROGUE_FWIF_PHR_STATE_SHIFT (8U)
/* The FW has requested the host to restart it, per PHR configuration */
#define ROGUE_FWIF_PHR_RESTART_REQUESTED ((1) << ROGUE_FWIF_PHR_STATE_SHIFT)
/* A PHR triggered GPU reset has just finished */
#define ROGUE_FWIF_PHR_RESTART_FINISHED ((2) << ROGUE_FWIF_PHR_STATE_SHIFT)
#define ROGUE_FWIF_PHR_RESTART_MASK \
	(ROGUE_FWIF_PHR_RESTART_REQUESTED | ROGUE_FWIF_PHR_RESTART_FINISHED)

#define ROGUE_FWIF_PHR_MODE_OFF (0UL)
#define ROGUE_FWIF_PHR_MODE_RD_RESET (1UL)
#define ROGUE_FWIF_PHR_MODE_FULL_RESET (2UL)

/* Firmware per-DM HWR states */
/* DM is working if all flags are cleared */
#define ROGUE_FWIF_DM_STATE_WORKING (0)
/* DM is idle and ready for HWR */
#define ROGUE_FWIF_DM_STATE_READY_FOR_HWR BIT(0)
/* DM need to skip to next cmd before resuming processing */
#define ROGUE_FWIF_DM_STATE_NEEDS_SKIP BIT(2)
/* DM need partial render cleanup before resuming processing */
#define ROGUE_FWIF_DM_STATE_NEEDS_PR_CLEANUP BIT(3)
/* DM need to increment Recovery Count once fully recovered */
#define ROGUE_FWIF_DM_STATE_NEEDS_TRACE_CLEAR BIT(4)
/* DM was identified as locking up and causing HWR */
#define ROGUE_FWIF_DM_STATE_GUILTY_LOCKUP BIT(5)
/* DM was innocently affected by another lockup which caused HWR */
#define ROGUE_FWIF_DM_STATE_INNOCENT_LOCKUP BIT(6)
/* DM was identified as over-running and causing HWR */
#define ROGUE_FWIF_DM_STATE_GUILTY_OVERRUNING BIT(7)
/* DM was innocently affected by another DM over-running which caused HWR */
#define ROGUE_FWIF_DM_STATE_INNOCENT_OVERRUNING BIT(8)
/* DM was forced into HWR as it delayed more important workloads */
#define ROGUE_FWIF_DM_STATE_HARD_CONTEXT_SWITCH BIT(9)
/* DM was forced into HWR due to an uncorrected GPU ECC error */
#define ROGUE_FWIF_DM_STATE_GPU_ECC_HWR BIT(10)

/* Firmware's connection state */
enum rogue_fwif_connection_fw_state {
	/* Firmware is offline */
	ROGUE_FW_CONNECTION_FW_OFFLINE = 0,
	/* Firmware is initialised */
	ROGUE_FW_CONNECTION_FW_READY,
	/* Firmware connection is fully established */
	ROGUE_FW_CONNECTION_FW_ACTIVE,
	/* Firmware is clearing up connection data*/
	ROGUE_FW_CONNECTION_FW_OFFLOADING,
	ROGUE_FW_CONNECTION_FW_STATE_COUNT
};

/* OS' connection state */
enum rogue_fwif_connection_os_state {
	/* OS is offline */
	ROGUE_FW_CONNECTION_OS_OFFLINE = 0,
	/* OS's KM driver is setup and waiting */
	ROGUE_FW_CONNECTION_OS_READY,
	/* OS connection is fully established */
	ROGUE_FW_CONNECTION_OS_ACTIVE,
	ROGUE_FW_CONNECTION_OS_STATE_COUNT
};

struct rogue_fwif_os_runtime_flags {
	unsigned int os_state : 3;
	unsigned int fl_ok : 1;
	unsigned int fl_grow_pending : 1;
	unsigned int isolated_os : 1;
	unsigned int reserved : 26;
};

#define PVR_SLR_LOG_ENTRIES 10
/* MAX_CLIENT_CCB_NAME not visible to this header */
#define PVR_SLR_LOG_STRLEN 30

struct rogue_fwif_slr_entry {
	aligned_u64 timestamp;
	u32 fw_ctx_addr;
	u32 num_ufos;
	char ccb_name[PVR_SLR_LOG_STRLEN];
	char padding[2];
} __aligned(8);

#define MAX_THREAD_NUM 2

/* firmware trace control data */
struct rogue_fwif_tracebuf {
	u32 log_type;
	struct rogue_fwif_tracebuf_space tracebuf[MAX_THREAD_NUM];
	/*
	 * Member initialised only when sTraceBuf is actually allocated (in
	 * ROGUETraceBufferInitOnDemandResources)
	 */
	u32 tracebuf_size_in_dwords;
	/* Compatibility and other flags */
	u32 tracebuf_flags;
} __aligned(8);

/* firmware system data shared with the Host driver */
struct rogue_fwif_sysdata {
	/* Configuration flags from host */
	u32 config_flags;
	/* Extended configuration flags from host */
	u32 config_flags_ext;
	enum rogue_fwif_pow_state pow_state;
	u32 hw_perf_ridx;
	u32 hw_perf_widx;
	u32 hw_perf_wrap_count;
	/* Constant after setup, needed in FW */
	u32 hw_perf_size;
	/* The number of times the FW drops a packet due to buffer full */
	u32 hw_perf_drop_count;

	/*
	 * ui32HWPerfUt, ui32FirstDropOrdinal, ui32LastDropOrdinal only valid
	 * when FW is built with ROGUE_HWPERF_UTILIZATION &
	 * ROGUE_HWPERF_DROP_TRACKING defined in rogue_fw_hwperf.c
	 */
	/* Buffer utilisation, high watermark of bytes in use */
	u32 hw_perf_ut;
	/* The ordinal of the first packet the FW dropped */
	u32 first_drop_ordinal;
	/* The ordinal of the last packet the FW dropped */
	u32 last_drop_ordinal;
	/* State flags for each Operating System mirrored from Fw coremem */
	struct rogue_fwif_os_runtime_flags
		os_runtime_flags_mirror[ROGUE_FW_MAX_NUM_OS];

	struct rogue_fw_fault_info fault_info[ROGUE_FWIF_FWFAULTINFO_MAX];
	u32 fw_faults;
	u32 cr_poll_addr[MAX_THREAD_NUM];
	u32 cr_poll_mask[MAX_THREAD_NUM];
	u32 cr_poll_count[MAX_THREAD_NUM];
	aligned_u64 start_idle_time;

#if defined(SUPPORT_ROGUE_FW_STATS_FRAMEWORK)
#	define ROGUE_FWIF_STATS_FRAMEWORK_LINESIZE (8)
#	define ROGUE_FWIF_STATS_FRAMEWORK_MAX \
		(2048 * ROGUE_FWIF_STATS_FRAMEWORK_LINESIZE)
	u32 fw_stats_buf[ROGUE_FWIF_STATS_FRAMEWORK_MAX] __aligned(8);
#endif
	u32 hwr_state_flags;
	u32 hwr_recovery_flags[PVR_FWIF_DM_MAX];
	/* Compatibility and other flags */
	u32 fw_sys_data_flags;
	/* Identify whether MC config is P-P or P-S */
	u32 mc_config;
} __aligned(8);

/* per-os firmware shared data */
struct rogue_fwif_osdata {
	/* Configuration flags from an OS */
	u32 fw_os_config_flags;
	/* Markers to signal that the host should perform a full sync check */
	u32 fw_sync_check_mark;
	u32 host_sync_check_mark;

	u32 forced_updates_requested;
	u8 slr_log_wp;
	struct rogue_fwif_slr_entry slr_log_first;
	struct rogue_fwif_slr_entry slr_log[PVR_SLR_LOG_ENTRIES];
	aligned_u64 last_forced_update_time;

	/* Interrupt count from Threads > */
	u32 interrupt_count[MAX_THREAD_NUM];
	u32 kccb_cmds_executed;
	u32 power_sync_fw_addr;
	/* Compatibility and other flags */
	u32 fw_os_data_flags;
	u32 padding;
} __aligned(8);

/* Firmware trace time-stamp field breakup */

/* ROGUE_CR_TIMER register read (48 bits) value*/
#define ROGUE_FWT_TIMESTAMP_TIME_SHIFT (0U)
#define ROGUE_FWT_TIMESTAMP_TIME_CLRMSK (0xFFFF000000000000ull)

/* Extra debug-info (16 bits) */
#define ROGUE_FWT_TIMESTAMP_DEBUG_INFO_SHIFT (48U)
#define ROGUE_FWT_TIMESTAMP_DEBUG_INFO_CLRMSK ~ROGUE_FWT_TIMESTAMP_TIME_CLRMSK

/* Debug-info sub-fields */
/*
 * Bit 0: ROGUE_CR_EVENT_STATUS_MMU_PAGE_FAULT bit from ROGUE_CR_EVENT_STATUS
 * register
 */
#define ROGUE_FWT_DEBUG_INFO_MMU_PAGE_FAULT_SHIFT (0U)
#define ROGUE_FWT_DEBUG_INFO_MMU_PAGE_FAULT_SET \
	BIT(ROGUE_FWT_DEBUG_INFO_MMU_PAGE_FAULT_SHIFT)

/* Bit 1: ROGUE_CR_BIF_MMU_ENTRY_PENDING bit from ROGUE_CR_BIF_MMU_ENTRY register */
#define ROGUE_FWT_DEBUG_INFO_MMU_ENTRY_PENDING_SHIFT (1U)
#define ROGUE_FWT_DEBUG_INFO_MMU_ENTRY_PENDING_SET \
	BIT(ROGUE_FWT_DEBUG_INFO_MMU_ENTRY_PENDING_SHIFT)

/* Bit 2: ROGUE_CR_SLAVE_EVENT register is non-zero */
#define ROGUE_FWT_DEBUG_INFO_SLAVE_EVENTS_SHIFT (2U)
#define ROGUE_FWT_DEBUG_INFO_SLAVE_EVENTS_SET \
	BIT(ROGUE_FWT_DEBUG_INFO_SLAVE_EVENTS_SHIFT)

/* Bit 3-15: Unused bits */

#define ROGUE_FWT_DEBUG_INFO_STR_MAXLEN 64
#define ROGUE_FWT_DEBUG_INFO_STR_PREPEND " (debug info: "
#define ROGUE_FWT_DEBUG_INFO_STR_APPEND ")"

/*
 ******************************************************************************
 * HWR Data
 ******************************************************************************
 */
enum rogue_hwrtype {
	ROGUE_HWRTYPE_UNKNOWNFAILURE = 0,
	ROGUE_HWRTYPE_OVERRUN = 1,
	ROGUE_HWRTYPE_POLLFAILURE = 2,
	ROGUE_HWRTYPE_BIF0FAULT = 3,
	ROGUE_HWRTYPE_BIF1FAULT = 4,
	ROGUE_HWRTYPE_TEXASBIF0FAULT = 5,
	ROGUE_HWRTYPE_MMUFAULT = 6,
	ROGUE_HWRTYPE_MMUMETAFAULT = 7,
	ROGUE_HWRTYPE_MIPSTLBFAULT = 8,
	ROGUE_HWRTYPE_ECCFAULT = 9,
	ROGUE_HWRTYPE_MMURISCVFAULT = 10,
};

#define ROGUE_FWIF_HWRTYPE_BIF_BANK_GET(hwr_type) \
	(((hwr_type) == ROGUE_HWRTYPE_BIF0FAULT) ? 0 : 1)

#define ROGUE_FWIF_HWRTYPE_PAGE_FAULT_GET(hwr_type)       \
	((((hwr_type) == ROGUE_HWRTYPE_BIF0FAULT) ||      \
	  ((hwr_type) == ROGUE_HWRTYPE_BIF1FAULT) ||      \
	  ((hwr_type) == ROGUE_HWRTYPE_TEXASBIF0FAULT) || \
	  ((hwr_type) == ROGUE_HWRTYPE_MMUFAULT) ||       \
	  ((hwr_type) == ROGUE_HWRTYPE_MMUMETAFAULT) ||   \
	  ((hwr_type) == ROGUE_HWRTYPE_MIPSTLBFAULT) ||   \
	  ((hwr_type) == ROGUE_HWRTYPE_MMURISCVFAULT))    \
		 ? true                                   \
		 : false)

struct rogue_bifinfo {
	aligned_u64 bif_req_status;
	aligned_u64 bif_mmu_status;
	aligned_u64 pc_address; /* phys address of the page catalogue */
	aligned_u64 reserved;
};

struct rogue_eccinfo {
	u32 fault_gpu;
};

struct rogue_mmuinfo {
	aligned_u64 mmu_status[2];
	aligned_u64 pc_address; /* phys address of the page catalogue */
	aligned_u64 reserved;
};

struct rogue_pollinfo {
	u32 thread_num;
	u32 cr_poll_addr;
	u32 cr_poll_mask;
	u32 cr_poll_last_value;
	aligned_u64 reserved;
} __aligned(8);

struct rogue_tlbinfo {
	u32 bad_addr;
	u32 entry_lo;
};

struct rogue_hwrinfo {
	union {
		struct rogue_bifinfo bif_info;
		struct rogue_mmuinfo mmu_info;
		struct rogue_pollinfo poll_info;
		struct rogue_tlbinfo tlb_info;
		struct rogue_eccinfo ecc_info;
	} hwr_data;

	aligned_u64 cr_timer;
	aligned_u64 os_timer;
	u32 frame_num;
	u32 pid;
	u32 active_hwrt_data;
	u32 hwr_number;
	u32 event_status;
	u32 hwr_recovery_flags;
	enum rogue_hwrtype hwr_type;
	u32 dm;
	u32 core_id;
	aligned_u64 cr_time_of_kick;
	aligned_u64 cr_time_hw_reset_start;
	aligned_u64 cr_time_hw_reset_finish;
	aligned_u64 cr_time_freelist_ready;
	aligned_u64 reserved[2];
} __aligned(8);

/* Number of first HWR logs recorded (never overwritten by newer logs) */
#define ROGUE_FWIF_HWINFO_MAX_FIRST 8U
/* Number of latest HWR logs (older logs are overwritten by newer logs) */
#define ROGUE_FWIF_HWINFO_MAX_LAST 8U
/* Total number of HWR logs stored in a buffer */
#define ROGUE_FWIF_HWINFO_MAX \
	(ROGUE_FWIF_HWINFO_MAX_FIRST + ROGUE_FWIF_HWINFO_MAX_LAST)
/* Index of the last log in the HWR log buffer */
#define ROGUE_FWIF_HWINFO_LAST_INDEX (ROGUE_FWIF_HWINFO_MAX - 1U)

struct rogue_fwif_hwrinfobuf {
	struct rogue_hwrinfo hwr_info[ROGUE_FWIF_HWINFO_MAX];
	u32 hwr_counter;
	u32 write_index;
	u32 dd_req_count;
	u32 hwr_info_buf_flags; /* Compatibility and other flags */
	u32 hwr_dm_locked_up_count[PVR_FWIF_DM_MAX];
	u32 hwr_dm_overran_count[PVR_FWIF_DM_MAX];
	u32 hwr_dm_recovered_count[PVR_FWIF_DM_MAX];
	u32 hwr_dm_false_detect_count[PVR_FWIF_DM_MAX];
} __aligned(8);

#define ROGUE_FWIF_CTXSWITCH_PROFILE_FAST_EN (1)
#define ROGUE_FWIF_CTXSWITCH_PROFILE_MEDIUM_EN (2)
#define ROGUE_FWIF_CTXSWITCH_PROFILE_SLOW_EN (3)
#define ROGUE_FWIF_CTXSWITCH_PROFILE_NODELAY_EN (4)

#define ROGUE_FWIF_CDM_ARBITRATION_TASK_DEMAND_EN (1)
#define ROGUE_FWIF_CDM_ARBITRATION_ROUND_ROBIN_EN (2)

#define ROGUE_FWIF_ISP_SCHEDMODE_VER1_IPP (1)
#define ROGUE_FWIF_ISP_SCHEDMODE_VER2_ISP (2)
/*
 ******************************************************************************
 * ROGUE firmware Init Config Data
 ******************************************************************************
 */

/* Flag definitions affecting the firmware globally */
#define ROGUE_FWIF_INICFG_CTXSWITCH_MODE_RAND BIT(0)
#define ROGUE_FWIF_INICFG_CTXSWITCH_SRESET_EN BIT(1)
#define ROGUE_FWIF_INICFG_HWPERF_EN BIT(2)
#define ROGUE_FWIF_INICFG_DM_KILL_MODE_RAND_EN BIT(3)
#define ROGUE_FWIF_INICFG_POW_RASCALDUST BIT(4)
/* Bit 5 is reserved. */
#define ROGUE_FWIF_INICFG_FBCDC_V3_1_EN BIT(6)
#define ROGUE_FWIF_INICFG_CHECK_MLIST_EN BIT(7)
#define ROGUE_FWIF_INICFG_DISABLE_CLKGATING_EN BIT(8)
/* Bit 9 is reserved. */
/* Bit 10 is reserved. */
/* Bit 11 is reserved. */
#define ROGUE_FWIF_INICFG_REGCONFIG_EN BIT(12)
#define ROGUE_FWIF_INICFG_ASSERT_ON_OUTOFMEMORY BIT(13)
#define ROGUE_FWIF_INICFG_HWP_DISABLE_FILTER BIT(14)
/* Bit 15 is reserved. */
#define ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_SHIFT (16)
#define ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_FAST \
	(ROGUE_FWIF_CTXSWITCH_PROFILE_FAST_EN    \
	 << ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_SHIFT)
#define ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_MEDIUM \
	(ROGUE_FWIF_CTXSWITCH_PROFILE_MEDIUM_EN    \
	 << ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_SHIFT)
#define ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_SLOW \
	(ROGUE_FWIF_CTXSWITCH_PROFILE_SLOW_EN    \
	 << ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_SHIFT)
#define ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_NODELAY \
	(ROGUE_FWIF_CTXSWITCH_PROFILE_NODELAY_EN    \
	 << ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_SHIFT)
#define ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_MASK \
	(7 << ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_SHIFT)
#define ROGUE_FWIF_INICFG_DISABLE_DM_OVERLAP BIT(19)
#define ROGUE_FWIF_INICFG_ASSERT_ON_HWR_TRIGGER BIT(20)
#define ROGUE_FWIF_INICFG_FABRIC_COHERENCY_ENABLED BIT(21)
#define ROGUE_FWIF_INICFG_VALIDATE_IRQ BIT(22)
#define ROGUE_FWIF_INICFG_DISABLE_PDP_EN BIT(23)
#define ROGUE_FWIF_INICFG_SPU_POWER_STATE_MASK_CHANGE_EN BIT(24)
#define ROGUE_FWIF_INICFG_WORKEST BIT(25)
#define ROGUE_FWIF_INICFG_PDVFS BIT(26)
#define ROGUE_FWIF_INICFG_CDM_ARBITRATION_SHIFT (27)
#define ROGUE_FWIF_INICFG_CDM_ARBITRATION_TASK_DEMAND \
	(ROGUE_FWIF_CDM_ARBITRATION_TASK_DEMAND_EN    \
	 << ROGUE_FWIF_INICFG_CDM_ARBITRATION_SHIFT)
#define ROGUE_FWIF_INICFG_CDM_ARBITRATION_ROUND_ROBIN \
	(ROGUE_FWIF_CDM_ARBITRATION_ROUND_ROBIN_EN    \
	 << ROGUE_FWIF_INICFG_CDM_ARBITRATION_SHIFT)
#define ROGUE_FWIF_INICFG_CDM_ARBITRATION_MASK \
	(3 << ROGUE_FWIF_INICFG_CDM_ARBITRATION_SHIFT)
#define ROGUE_FWIF_INICFG_ISPSCHEDMODE_SHIFT (29)
#define ROGUE_FWIF_INICFG_ISPSCHEDMODE_NONE (0)
#define ROGUE_FWIF_INICFG_ISPSCHEDMODE_VER1_IPP \
	(ROGUE_FWIF_ISP_SCHEDMODE_VER1_IPP      \
	 << ROGUE_FWIF_INICFG_ISPSCHEDMODE_SHIFT)
#define ROGUE_FWIF_INICFG_ISPSCHEDMODE_VER2_ISP \
	(ROGUE_FWIF_ISP_SCHEDMODE_VER2_ISP      \
	 << ROGUE_FWIF_INICFG_ISPSCHEDMODE_SHIFT)
#define ROGUE_FWIF_INICFG_ISPSCHEDMODE_MASK        \
	(ROGUE_FWIF_INICFG_ISPSCHEDMODE_VER1_IPP | \
	 ROGUE_FWIF_INICFG_ISPSCHEDMODE_VER2_ISP)
#define ROGUE_FWIF_INICFG_VALIDATE_SOCUSC_TIMER BIT(31)

#define ROGUE_FWIF_INICFG_ALL (0xFFFFFFFFU)

/* Extended Flag definitions affecting the firmware globally */
#define ROGUE_FWIF_INICFG_EXT_TFBC_CONTROL_SHIFT (0)
/* [7]   YUV10 override
 * [6:4] Quality
 * [3]   Quality enable
 * [2:1] Compression scheme
 * [0]   Lossy group
 */
#define ROGUE_FWIF_INICFG_EXT_TFBC_CONTROL_MASK (0xFF)
#define ROGUE_FWIF_INICFG_EXT_ALL (ROGUE_FWIF_INICFG_EXT_TFBC_CONTROL_MASK)

/* Flag definitions affecting only workloads submitted by a particular OS */
#define ROGUE_FWIF_INICFG_OS_CTXSWITCH_TDM_EN BIT(0)
#define ROGUE_FWIF_INICFG_OS_CTXSWITCH_GEOM_EN BIT(1)
#define ROGUE_FWIF_INICFG_OS_CTXSWITCH_FRAG_EN BIT(2)
#define ROGUE_FWIF_INICFG_OS_CTXSWITCH_CDM_EN BIT(3)

#define ROGUE_FWIF_INICFG_OS_LOW_PRIO_CS_TDM BIT(4)
#define ROGUE_FWIF_INICFG_OS_LOW_PRIO_CS_GEOM BIT(5)
#define ROGUE_FWIF_INICFG_OS_LOW_PRIO_CS_FRAG BIT(6)
#define ROGUE_FWIF_INICFG_OS_LOW_PRIO_CS_CDM BIT(7)

#define ROGUE_FWIF_INICFG_OS_ALL (0xFF)

#define ROGUE_FWIF_INICFG_OS_CTXSWITCH_DM_ALL     \
	(ROGUE_FWIF_INICFG_OS_CTXSWITCH_TDM_EN |  \
	 ROGUE_FWIF_INICFG_OS_CTXSWITCH_GEOM_EN | \
	 ROGUE_FWIF_INICFG_OS_CTXSWITCH_FRAG_EN |   \
	 ROGUE_FWIF_INICFG_OS_CTXSWITCH_CDM_EN)

#define ROGUE_FWIF_INICFG_OS_CTXSWITCH_CLRMSK \
	~(ROGUE_FWIF_INICFG_OS_CTXSWITCH_DM_ALL)

#define ROGUE_FWIF_FILTCFG_TRUNCATE_HALF BIT(3)
#define ROGUE_FWIF_FILTCFG_TRUNCATE_INT BIT(2)
#define ROGUE_FWIF_FILTCFG_NEW_FILTER_MODE BIT(1)

enum rogue_activepm_conf {
	ROGUE_ACTIVEPM_FORCE_OFF = 0,
	ROGUE_ACTIVEPM_FORCE_ON = 1,
	ROGUE_ACTIVEPM_DEFAULT = 2
};

enum rogue_rd_power_island_conf {
	ROGUE_RD_POWER_ISLAND_FORCE_OFF = 0,
	ROGUE_RD_POWER_ISLAND_FORCE_ON = 1,
	ROGUE_RD_POWER_ISLAND_DEFAULT = 2
};

struct rogue_fw_register_list {
	/* Register number */
	u16 reg_num;
	/* Indirect register number (or 0 if not used) */
	u16 indirect_reg_num;
	/* Start value for indirect register */
	u16 indirect_start_val;
	/* End value for indirect register */
	u16 indirect_end_val;
};

struct rogue_fwif_dllist_node {
	u32 p;
	u32 n;
};

/*
 * This number is used to represent an invalid page catalogue physical address
 */
#define ROGUE_FWIF_INVALID_PC_PHYADDR 0xFFFFFFFFFFFFFFFFLLU

/* This number is used to represent unallocated page catalog base register */
#define ROGUE_FW_BIF_INVALID_PCSET 0xFFFFFFFFU

/* Firmware memory context. */
struct rogue_fwif_fwmemcontext {
	/* device physical address of context's page catalogue */
	aligned_u64 pc_dev_paddr;
	/*
	 * associated page catalog base register (ROGUE_FW_BIF_INVALID_PCSET ==
	 * unallocated)
	 */
	u32 page_cat_base_reg_set;
	/* breakpoint address */
	u32 breakpoint_addr;
	/* breakpoint handler address */
	u32 bp_handler_addr;
	/* DM and enable control for BP */
	u32 breakpoint_ctl;
	/* Compatibility and other flags */
	u32 fw_mem_ctx_flags;
	u32 padding;
} __aligned(8);

/*
 * FW context state flags
 */
#define ROGUE_FWIF_CONTEXT_FLAGS_NEED_RESUME (0x00000001U)
#define ROGUE_FWIF_CONTEXT_FLAGS_MC_NEED_RESUME_MASKFULL (0x000000FFU)
#define ROGUE_FWIF_CONTEXT_FLAGS_TDM_HEADER_STALE (0x00000100U)
#define ROGUE_FWIF_CONTEXT_FLAGS_LAST_KICK_SECURE (0x00000200U)

#define ROGUE_NUM_GEOM_CORES_MAX 4

/*
 * FW-accessible TA state which must be written out to memory on context store
 */
struct rogue_fwif_geom_ctx_state_per_geom {
	/* To store in mid-TA */
	aligned_u64 geom_reg_vdm_call_stack_pointer;
	/* Initial value (in case is 'lost' due to a lock-up */
	aligned_u64 geom_reg_vdm_call_stack_pointer_init;
	u32 geom_reg_vbs_so_prim[4];
	u16 geom_current_idx;
	u16 padding[3];
} __aligned(8);

struct rogue_fwif_geom_ctx_state {
	/* FW-accessible TA state which must be written out to memory on context store */
	struct rogue_fwif_geom_ctx_state_per_geom geom_core[ROGUE_NUM_GEOM_CORES_MAX];
} __aligned(8);

/*
 * FW-accessible ISP state which must be written out to memory on context store
 */
struct rogue_fwif_frag_ctx_state {
	u32 frag_reg_pm_deallocated_mask_status;
	u32 frag_reg_dm_pds_mtilefree_status;
	/* Compatibility and other flags */
	u32 ctx_state_flags;
	/*
	 * frag_reg_isp_store should be the last element of the structure as this
	 * is an array whose size is determined at runtime after detecting the
	 * ROGUE core
	 */
	u32 frag_reg_isp_store[];
} __aligned(8);

#define ROGUE_FWIF_CTX_USING_BUFFER_A (0)
#define ROGUE_FWIF_CTX_USING_BUFFER_B (1U)

struct rogue_fwif_compute_ctx_state {
	u32 ctx_state_flags; /* Target buffer and other flags */
};

struct rogue_fwif_fwcommoncontext {
	/* CCB details for this firmware context */
	u32 ccbctl_fw_addr; /* CCB control */
	u32 ccb_fw_addr; /* CCB base */
	struct rogue_fwif_dma_addr ccb_meta_dma_addr;

	/* Context suspend state */
	/* geom/frag context suspend state, read/written by FW */
	u32 context_state_addr __aligned(8);

	/* Flags e.g. for context switching */
	u32 fw_com_ctx_flags;
	u32 priority;
	u32 priority_seq_num;

	/* Framework state */
	/* Register updates for Framework */
	u32 rf_cmd_addr __aligned(8);

	/* Statistic updates waiting to be passed back to the host... */
	/* True when some stats are pending */
	bool stats_pending __aligned(4);
	/* Number of stores on this context since last update */
	s32 stats_num_stores;
	/* Number of OOMs on this context since last update */
	s32 stats_num_out_of_memory;
	/* Number of PRs on this context since last update */
	s32 stats_num_partial_renders;
	/* Data Master type */
	u32 dm;
	/* Device Virtual Address of the signal the context is waiting on */
	aligned_u64 wait_signal_address;
	/* List entry for the wait-signal list */
	struct rogue_fwif_dllist_node wait_signal_node __aligned(8);
	/* List entry for the buffer stalled list */
	struct rogue_fwif_dllist_node buf_stalled_node __aligned(8);
	/* Address of the circular buffer queue pointers */
	aligned_u64 cbuf_queue_ctrl_addr;

	aligned_u64 robustness_address;
	/* Max HWR deadline limit in ms */
	u32 max_deadline_ms;
	/* Following HWR circular buffer read-offset needs resetting */
	bool read_offset_needs_reset;

	/* List entry for the waiting list */
	struct rogue_fwif_dllist_node waiting_node __aligned(8);
	/* List entry for the run list */
	struct rogue_fwif_dllist_node run_node __aligned(8);
	/* UFO that last failed (or NULL) */
	struct rogue_fwif_ufo last_failed_ufo;

	/* Memory context */
	u32 fw_mem_context_fw_addr;

	/* References to the host side originators */
	/* the Server Common Context */
	u32 server_common_context_id;
	/* associated process ID */
	u32 pid;

	/* True when Geom DM OOM is not allowed */
	bool geom_oom_disabled __aligned(4);
} __aligned(8);

/* Firmware render context. */
struct rogue_fwif_fwrendercontext {
	/* Geometry firmware context. */
	struct rogue_fwif_fwcommoncontext geom_context;
	/* Fragment firmware context. */
	struct rogue_fwif_fwcommoncontext frag_context;

	struct rogue_fwif_static_rendercontext_state static_render_context_state;

	/* Number of commands submitted to the WorkEst FW CCB */
	u32 work_est_ccb_submitted;

	/* Compatibility and other flags */
	u32 fw_render_ctx_flags;
} __aligned(8);

/* Firmware compute context. */
struct rogue_fwif_fwcomputecontext {
	/* Firmware context for the CDM */
	struct rogue_fwif_fwcommoncontext cdm_context;

	struct rogue_fwif_static_computecontext_state
		static_compute_context_state;

	/* Number of commands submitted to the WorkEst FW CCB */
	u32 work_est_ccb_submitted;

	/* Compatibility and other flags */
	u32 compute_ctx_flags;

	u32 wgp_state;
	u32 wgp_checksum;
	u32 core_mask_a;
	u32 core_mask_b;
} __aligned(8);

/* Firmware TDM context. */
struct rogue_fwif_fwtdmcontext {
	/* Firmware context for the TDM */
	struct rogue_fwif_fwcommoncontext tdm_context;

	/* Number of commands submitted to the WorkEst FW CCB */
	u32 work_est_ccb_submitted;
} __aligned(8);

/* Firmware TQ3D context. */
struct rogue_fwif_fwtransfercontext {
	/* Firmware context for TQ3D. */
	struct rogue_fwif_fwcommoncontext tq_context;
} __aligned(8);

/*
 ******************************************************************************
 * Defines for CMD_TYPE corruption detection and forward compatibility check
 ******************************************************************************
 */

/*
 * CMD_TYPE 32bit contains:
 * 31:16	Reserved for magic value to detect corruption (16 bits)
 * 15		Reserved for ROGUE_CCB_TYPE_TASK (1 bit)
 * 14:0		Bits available for CMD_TYPEs (15 bits)
 */

/* Magic value to detect corruption */
#define ROGUE_CMD_MAGIC_DWORD (0x2ABC)
#define ROGUE_CMD_MAGIC_DWORD_MASK (0xFFFF0000U)
#define ROGUE_CMD_MAGIC_DWORD_SHIFT (16U)
#define ROGUE_CMD_MAGIC_DWORD_SHIFTED \
	(ROGUE_CMD_MAGIC_DWORD << ROGUE_CMD_MAGIC_DWORD_SHIFT)

/* Kernel CCB control for ROGUE */
struct rogue_fwif_ccb_ctl {
	/* write offset into array of commands (MUST be aligned to 16 bytes!) */
	u32 write_offset;
	/* Padding to ensure read and write offsets are in separate cache lines. */
	u8 padding[128 - sizeof(u32)];
	/* read offset into array of commands */
	u32 read_offset;
	/* Offset wrapping mask (Total capacity of the CCB - 1) */
	u32 wrap_mask;
	/* size of each command in bytes */
	u32 cmd_size;
	u32 padding2;
} __aligned(8);

/* Kernel CCB command structure for ROGUE */

#define ROGUE_FWIF_MMUCACHEDATA_FLAGS_PT (0x1U) /* MMU_CTRL_INVAL_PT_EN */
#define ROGUE_FWIF_MMUCACHEDATA_FLAGS_PD (0x2U) /* MMU_CTRL_INVAL_PD_EN */
#define ROGUE_FWIF_MMUCACHEDATA_FLAGS_PC (0x4U) /* MMU_CTRL_INVAL_PC_EN */

/*
 * can't use PM_TLB0 bit from BIFPM_CTRL reg because it collides with PT
 * bit from BIF_CTRL reg
 */
#define ROGUE_FWIF_MMUCACHEDATA_FLAGS_PMTLB (0x10)
/* BIF_CTRL_INVAL_TLB1_EN */
#define ROGUE_FWIF_MMUCACHEDATA_FLAGS_TLB \
	(ROGUE_FWIF_MMUCACHEDATA_FLAGS_PMTLB | 0x8)
/* MMU_CTRL_INVAL_ALL_CONTEXTS_EN */
#define ROGUE_FWIF_MMUCACHEDATA_FLAGS_CTX_ALL (0x800)

/* indicates FW should interrupt the host */
#define ROGUE_FWIF_MMUCACHEDATA_FLAGS_INTERRUPT (0x4000000U)

struct rogue_fwif_mmucachedata {
	u32 cache_flags;
	u32 mmu_cache_sync_fw_addr;
	u32 mmu_cache_sync_update_value;
};

#define ROGUE_FWIF_BPDATA_FLAGS_ENABLE BIT(0)
#define ROGUE_FWIF_BPDATA_FLAGS_WRITE BIT(1)
#define ROGUE_FWIF_BPDATA_FLAGS_CTL BIT(2)
#define ROGUE_FWIF_BPDATA_FLAGS_REGS BIT(3)

struct rogue_fwif_bpdata {
	/* Memory context */
	u32 fw_mem_context_fw_addr;
	/* Breakpoint address */
	u32 bp_addr;
	/* Breakpoint handler */
	u32 bp_handler_addr;
	/* Breakpoint control */
	u32 bp_dm;
	u32 bp_data_flags;
	/* Number of temporary registers to overallocate */
	u32 temp_regs;
	/* Number of shared registers to overallocate */
	u32 shared_regs;
	/* DM associated with the breakpoint */
	u32 dm;
};

#define ROGUE_FWIF_KCCB_CMD_KICK_DATA_MAX_NUM_CLEANUP_CTLS \
	(ROGUE_FWIF_PRBUFFER_MAXSUPPORTED + 1U) /* +1 is RTDATASET cleanup */

struct rogue_fwif_kccb_cmd_kick_data {
	/* address of the firmware context */
	u32 context_fw_addr;
	/* Client CCB woff update */
	u32 client_woff_update;
	/* Client CCB wrap mask update after CCCB growth */
	u32 client_wrap_mask_update;
	/* number of CleanupCtl pointers attached */
	u32 num_cleanup_ctl;
	/* CleanupCtl structures associated with command */
	u32 cleanup_ctl_fw_addr
		[ROGUE_FWIF_KCCB_CMD_KICK_DATA_MAX_NUM_CLEANUP_CTLS];
	/*
	 * offset to the CmdHeader which houses the workload estimation kick
	 * data.
	 */
	u32 work_est_cmd_header_offset;
};

struct rogue_fwif_kccb_cmd_combined_geom_frag_kick_data {
	struct rogue_fwif_kccb_cmd_kick_data geom_cmd_kick_data;
	struct rogue_fwif_kccb_cmd_kick_data frag_cmd_kick_data;
};

struct rogue_fwif_kccb_cmd_force_update_data {
	/* address of the firmware context */
	u32 context_fw_addr;
	/* Client CCB fence offset */
	u32 ccb_fence_offset;
};

enum rogue_fwif_cleanup_type {
	/* FW common context cleanup */
	ROGUE_FWIF_CLEANUP_FWCOMMONCONTEXT,
	/* FW HW RT data cleanup */
	ROGUE_FWIF_CLEANUP_HWRTDATA,
	/* FW freelist cleanup */
	ROGUE_FWIF_CLEANUP_FREELIST,
	/* FW ZS Buffer cleanup */
	ROGUE_FWIF_CLEANUP_ZSBUFFER,
};

struct rogue_fwif_cleanup_request {
	/* Cleanup type */
	enum rogue_fwif_cleanup_type cleanup_type;
	union {
		/* FW common context to cleanup */
		u32 context_fw_addr;
		/* HW RT to cleanup */
		u32 hwrt_data_fw_addr;
		/* Freelist to cleanup */
		u32 freelist_fw_addr;
		/* ZS Buffer to cleanup */
		u32 zs_buffer_fw_addr;
	} cleanup_data;
};

enum rogue_fwif_power_type {
	ROGUE_FWIF_POW_OFF_REQ = 1,
	ROGUE_FWIF_POW_FORCED_IDLE_REQ,
	ROGUE_FWIF_POW_NUM_UNITS_CHANGE,
	ROGUE_FWIF_POW_APM_LATENCY_CHANGE
};

enum rogue_fwif_power_force_idle_type {
	ROGUE_FWIF_POWER_FORCE_IDLE = 1,
	ROGUE_FWIF_POWER_CANCEL_FORCED_IDLE,
	ROGUE_FWIF_POWER_HOST_TIMEOUT,
};

struct rogue_fwif_power_request {
	/* Type of power request */
	enum rogue_fwif_power_type pow_type;
	union {
		/* Number of active Dusts */
		u32 num_of_dusts;
		/* If the operation is mandatory */
		bool forced __aligned(4);
		/*
		 * Type of Request. Consolidating Force Idle, Cancel Forced
		 * Idle, Host Timeout
		 */
		enum rogue_fwif_power_force_idle_type pow_request_type;
	} power_req_data;
};

struct rogue_fwif_slcflushinvaldata {
	/* Context to fence on (only useful when bDMContext == TRUE) */
	u32 context_fw_addr;
	/* Invalidate the cache as well as flushing */
	bool inval __aligned(4);
	/* The data to flush/invalidate belongs to a specific DM context */
	bool dm_context __aligned(4);
	/* Optional address of range (only useful when bDMContext == FALSE) */
	aligned_u64 address;
	/* Optional size of range (only useful when bDMContext == FALSE) */
	aligned_u64 size;
};

enum rogue_fwif_hwperf_update_config {
	ROGUE_FWIF_HWPERF_CTRL_TOGGLE = 0,
	ROGUE_FWIF_HWPERF_CTRL_SET = 1,
	ROGUE_FWIF_HWPERF_CTRL_EMIT_FEATURES_EV = 2
};

struct rogue_fwif_hwperf_ctrl {
	enum rogue_fwif_hwperf_update_config opcode; /* Control operation code */
	aligned_u64 mask; /* Mask of events to toggle */
};

struct rogue_fwif_hwperf_config_enable_blks {
	/* Number of ROGUE_HWPERF_CONFIG_MUX_CNTBLK in the array */
	u32 num_blocks;
	/* Address of the ROGUE_HWPERF_CONFIG_MUX_CNTBLK array */
	u32 block_configs_fw_addr;
};

struct rogue_fwif_hwperf_config_da_blks {
	/* Number of ROGUE_HWPERF_CONFIG_CNTBLK in the array */
	u32 num_blocks;
	/* Address of the ROGUE_HWPERF_CONFIG_CNTBLK array */
	u32 block_configs_fw_addr;
};

struct rogue_fwif_coreclkspeedchange_data {
	u32 new_clock_speed; /* New clock speed */
};

#define ROGUE_FWIF_HWPERF_CTRL_BLKS_MAX 16

struct rogue_fwif_hwperf_ctrl_blks {
	bool enable;
	/* Number of block IDs in the array */
	u32 num_blocks;
	/* Array of ROGUE_HWPERF_CNTBLK_ID values */
	u16 block_ids[ROGUE_FWIF_HWPERF_CTRL_BLKS_MAX];
};

struct rogue_fwif_hwperf_select_custom_cntrs {
	u16 custom_block;
	u16 num_counters;
	u32 custom_counter_ids_fw_addr;
};

struct rogue_fwif_zsbuffer_backing_data {
	u32 zs_buffer_fw_addr; /* ZS-Buffer FW address */

	bool done __aligned(4); /* action backing/unbacking succeeded */
};

struct rogue_fwif_freelist_gs_data {
	/* Freelist FW address */
	u32 freelist_fw_addr;
	/* Amount of the Freelist change */
	u32 delta_pages;
	/* New amount of pages on the freelist (including ready pages) */
	u32 new_pages;
	/* Number of ready pages to be held in reserve until OOM */
	u32 ready_pages;
};

#define MAX_FREELISTS_SIZE 3
#define MAX_HW_GEOM_FRAG_CONTEXTS_SIZE 3

#define ROGUE_FWIF_MAX_FREELISTS_TO_RECONSTRUCT \
	(MAX_HW_GEOM_FRAG_CONTEXTS_SIZE * MAX_FREELISTS_SIZE * 2U)
#define ROGUE_FWIF_FREELISTS_RECONSTRUCTION_FAILED_FLAG 0x80000000U

struct rogue_fwif_freelists_reconstruction_data {
	u32 freelist_count;
	u32 freelist_ids[ROGUE_FWIF_MAX_FREELISTS_TO_RECONSTRUCT];
};

struct rogue_fwif_write_offset_update_data {
	/*
	 * Context to that may need to be resumed following write offset update
	 */
	u32 context_fw_addr;
} __aligned(8);

/*
 ******************************************************************************
 * Proactive DVFS Structures
 ******************************************************************************
 */
#define NUM_OPP_VALUES 16

struct pdvfs_opp {
	u32 volt; /* V  */
	u32 freq; /* Hz */
} __aligned(8);

struct rogue_fwif_pdvfs_opp {
	struct pdvfs_opp opp_values[NUM_OPP_VALUES];
	u32 min_opp_point;
	u32 max_opp_point;
} __aligned(8);

struct rogue_fwif_pdvfs_max_freq_data {
	u32 max_opp_point;
} __aligned(8);

struct rogue_fwif_pdvfs_min_freq_data {
	u32 min_opp_point;
} __aligned(8);

/*
 ******************************************************************************
 * Register configuration structures
 ******************************************************************************
 */

#define ROGUE_FWIF_REG_CFG_MAX_SIZE 512

enum rogue_fwif_regdata_cmd_type {
	ROGUE_FWIF_REGCFG_CMD_ADD = 101,
	ROGUE_FWIF_REGCFG_CMD_CLEAR = 102,
	ROGUE_FWIF_REGCFG_CMD_ENABLE = 103,
	ROGUE_FWIF_REGCFG_CMD_DISABLE = 104
};

enum rogue_fwif_reg_cfg_type {
	/* Sidekick power event */
	ROGUE_FWIF_REG_CFG_TYPE_PWR_ON = 0,
	/* Rascal / dust power event */
	ROGUE_FWIF_REG_CFG_TYPE_DUST_CHANGE,
	/* Geometry kick */
	ROGUE_FWIF_REG_CFG_TYPE_GEOM,
	/* Fragment kick */
	ROGUE_FWIF_REG_CFG_TYPE_FRAG,
	/* Compute kick */
	ROGUE_FWIF_REG_CFG_TYPE_CDM,
	/* TLA kick */
	ROGUE_FWIF_REG_CFG_TYPE_TLA,
	/* TDM kick */
	ROGUE_FWIF_REG_CFG_TYPE_TDM,
	/* Applies to all types. Keep as last element */
	ROGUE_FWIF_REG_CFG_TYPE_ALL
};

struct rogue_fwif_reg_cfg_rec {
	u64 sddr;
	u64 mask;
	u64 value;
};

struct rogue_fwif_regconfig_data {
	enum rogue_fwif_regdata_cmd_type cmd_type;
	enum rogue_fwif_reg_cfg_type reg_config_type;
	struct rogue_fwif_reg_cfg_rec reg_config __aligned(8);
};

struct rogue_fwif_reg_cfg {
	/*
	 * PDump WRW command write granularity is 32 bits.
	 * Add padding to ensure array size is 32 bit granular.
	 */
	u8 num_regs_type[ALIGN((u32)ROGUE_FWIF_REG_CFG_TYPE_ALL,
			       sizeof(u32))] __aligned(8);
	struct rogue_fwif_reg_cfg_rec
		reg_configs[ROGUE_FWIF_REG_CFG_MAX_SIZE] __aligned(8);
} __aligned(8);

enum rogue_fwif_os_state_change {
	ROGUE_FWIF_OS_ONLINE = 1,
	ROGUE_FWIF_OS_OFFLINE
};

struct rogue_fwif_os_state_change_data {
	u32 osid;
	enum rogue_fwif_os_state_change new_os_state;
} __aligned(8);

enum rogue_fwif_counter_dump_request {
	ROGUE_FWIF_PWR_COUNTER_DUMP_START = 1,
	ROGUE_FWIF_PWR_COUNTER_DUMP_STOP,
	ROGUE_FWIF_PWR_COUNTER_DUMP_SAMPLE,
};

struct rogue_fwif_counter_dump_data {
	enum rogue_fwif_counter_dump_request counter_dump_request;
} __aligned(8);

enum rogue_fwif_kccb_cmd_type {
	/* Common commands */
	ROGUE_FWIF_KCCB_CMD_KICK = 101U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	ROGUE_FWIF_KCCB_CMD_MMUCACHE = 102U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	ROGUE_FWIF_KCCB_CMD_BP = 103U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* SLC flush and invalidation request */
	ROGUE_FWIF_KCCB_CMD_SLCFLUSHINVAL = 105U |
					    ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/*
	 * Requests cleanup of a FW resource (type specified in the command
	 * data)
	 */
	ROGUE_FWIF_KCCB_CMD_CLEANUP = 106U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Power request */
	ROGUE_FWIF_KCCB_CMD_POW = 107U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Backing for on-demand ZS-Buffer done */
	ROGUE_FWIF_KCCB_CMD_ZSBUFFER_BACKING_UPDATE =
		108U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Unbacking for on-demand ZS-Buffer done */
	ROGUE_FWIF_KCCB_CMD_ZSBUFFER_UNBACKING_UPDATE =
		109U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Freelist Grow done */
	ROGUE_FWIF_KCCB_CMD_FREELIST_GROW_UPDATE =
		110U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Freelists Reconstruction done */
	ROGUE_FWIF_KCCB_CMD_FREELISTS_RECONSTRUCTION_UPDATE =
		112U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/*
	 * Informs the firmware that the host has added more data to a CDM2
	 * Circular Buffer
	 */
	ROGUE_FWIF_KCCB_CMD_NOTIFY_WRITE_OFFSET_UPDATE =
		114U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Health check request */
	ROGUE_FWIF_KCCB_CMD_HEALTH_CHECK = 115U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Forcing signalling of all unmet UFOs for a given CCB offset */
	ROGUE_FWIF_KCCB_CMD_FORCE_UPDATE = 116U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,

	/* There is a geometry and a fragment command in this single kick */
	ROGUE_FWIF_KCCB_CMD_COMBINED_GEOM_FRAG_KICK = 117U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Informs the FW that a Guest OS has come online / offline. */
	ROGUE_FWIF_KCCB_CMD_OS_ONLINE_STATE_CONFIGURE	= 118U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,

	/* Commands only permitted to the native or host OS */
	ROGUE_FWIF_KCCB_CMD_REGCONFIG = 200U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,

	/* Configure HWPerf events (to be generated) and HWPerf buffer address (if required) */
	ROGUE_FWIF_KCCB_CMD_HWPERF_UPDATE_CONFIG = 201U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,

	/* Enable or disable multiple HWPerf blocks (reusing existing configuration) */
	ROGUE_FWIF_KCCB_CMD_HWPERF_CTRL_BLKS = 203U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Core clock speed change event */
	ROGUE_FWIF_KCCB_CMD_CORECLKSPEEDCHANGE = 204U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,

	/*
	 * Ask the firmware to update its cached ui32LogType value from the (shared)
	 * tracebuf control structure
	 */
	ROGUE_FWIF_KCCB_CMD_LOGTYPE_UPDATE = 206U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Set a maximum frequency/OPP point */
	ROGUE_FWIF_KCCB_CMD_PDVFS_LIMIT_MAX_FREQ = 207U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/*
	 * Changes the relative scheduling priority for a particular OSid. It can
	 * only be serviced for the Host DDK
	 */
	ROGUE_FWIF_KCCB_CMD_OSID_PRIORITY_CHANGE = 208U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Set or clear firmware state flags */
	ROGUE_FWIF_KCCB_CMD_STATEFLAGS_CTRL = 209U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,

	/* Set a minimum frequency/OPP point */
	ROGUE_FWIF_KCCB_CMD_PDVFS_LIMIT_MIN_FREQ = 212U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Configure Periodic Hardware Reset behaviour */
	ROGUE_FWIF_KCCB_CMD_PHR_CFG = 213U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,

	/* Configure Safety Firmware Watchdog */
	ROGUE_FWIF_KCCB_CMD_WDG_CFG = 215U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Controls counter dumping in the FW */
	ROGUE_FWIF_KCCB_CMD_COUNTER_DUMP = 216U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Configure, clear and enable multiple HWPerf blocks */
	ROGUE_FWIF_KCCB_CMD_HWPERF_CONFIG_ENABLE_BLKS = 217U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Configure the custom counters for HWPerf */
	ROGUE_FWIF_KCCB_CMD_HWPERF_SELECT_CUSTOM_CNTRS = 218U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,

	/* Configure directly addressable counters for HWPerf */
	ROGUE_FWIF_KCCB_CMD_HWPERF_CONFIG_BLKS = 220U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
};

#define ROGUE_FWIF_LAST_ALLOWED_GUEST_KCCB_CMD \
	(ROGUE_FWIF_KCCB_CMD_REGCONFIG - 1)

/* Kernel CCB command packet */
struct rogue_fwif_kccb_cmd {
	/* Command type */
	enum rogue_fwif_kccb_cmd_type cmd_type;
	/* Compatibility and other flags */
	u32 kccb_flags;

	/*
	 * NOTE: Make sure that uCmdData is the last member of this struct
	 * This is to calculate actual command size for device mem copy.
	 * (Refer ROGUEGetCmdMemCopySize())
	 */
	union {
		/* Data for Kick command */
		struct rogue_fwif_kccb_cmd_kick_data cmd_kick_data;
		/* Data for combined geom/frag Kick command */
		struct rogue_fwif_kccb_cmd_combined_geom_frag_kick_data
			combined_geom_frag_cmd_kick_data;
		/* Data for MMU cache command */
		struct rogue_fwif_mmucachedata mmu_cache_data;
		/* Data for Breakpoint Commands */
		struct rogue_fwif_bpdata bp_data;
		/* Data for SLC Flush/Inval commands */
		struct rogue_fwif_slcflushinvaldata slc_flush_inval_data;
		/* Data for cleanup commands */
		struct rogue_fwif_cleanup_request cleanup_data;
		/* Data for power request commands */
		struct rogue_fwif_power_request pow_data;
		/* Data for HWPerf control command */
		struct rogue_fwif_hwperf_ctrl hw_perf_ctrl;
		/*
		 * Data for HWPerf configure, clear and enable performance
		 * counter block command
		 */
		struct rogue_fwif_hwperf_config_enable_blks
			hw_perf_cfg_enable_blks;
		/*
		 * Data for HWPerf enable or disable performance counter block
		 * commands
		 */
		struct rogue_fwif_hwperf_ctrl_blks hw_perf_ctrl_blks;
		/* Data for HWPerf configure the custom counters to read */
		struct rogue_fwif_hwperf_select_custom_cntrs
			hw_perf_select_cstm_cntrs;
		/* Data for HWPerf configure Directly Addressable blocks */
		struct rogue_fwif_hwperf_config_da_blks hw_perf_cfg_da_blks;
		/* Data for core clock speed change */
		struct rogue_fwif_coreclkspeedchange_data
			core_clk_speed_change_data;
		/* Feedback for Z/S Buffer backing/unbacking */
		struct rogue_fwif_zsbuffer_backing_data zs_buffer_backing_data;
		/* Feedback for Freelist grow/shrink */
		struct rogue_fwif_freelist_gs_data free_list_gs_data;
		/* Feedback for Freelists reconstruction*/
		struct rogue_fwif_freelists_reconstruction_data
			free_lists_reconstruction_data;
		/* Data for custom register configuration */
		struct rogue_fwif_regconfig_data reg_config_data;
		/* Data for informing the FW about the write offset update */
		struct rogue_fwif_write_offset_update_data
			write_offset_update_data;
		/* Data for setting the max frequency/OPP */
		struct rogue_fwif_pdvfs_max_freq_data pdvfs_max_freq_data;
		/* Data for setting the min frequency/OPP */
		struct rogue_fwif_pdvfs_min_freq_data pdvfs_min_freq_data;
		/* Data for updating the Guest Online states */
		struct rogue_fwif_os_state_change_data cmd_os_online_state_data;
		/* Dev address for TBI buffer allocated on demand */
		u32 tbi_buffer_fw_addr;
		/* Data for dumping of register ranges */
		struct rogue_fwif_counter_dump_data counter_dump_config_data;
		/* Data for signalling all unmet fences for a given CCB */
		struct rogue_fwif_kccb_cmd_force_update_data force_update_data;
	} cmd_data __aligned(8);
} __aligned(8);

PVR_FW_STRUCT_SIZE_ASSERT(struct rogue_fwif_kccb_cmd);

/*
 ******************************************************************************
 * Firmware CCB command structure for ROGUE
 ******************************************************************************
 */

struct rogue_fwif_fwccb_cmd_zsbuffer_backing_data {
	u32 zs_buffer_id;
};

struct rogue_fwif_fwccb_cmd_freelist_gs_data {
	u32 freelist_id;
};

struct rogue_fwif_fwccb_cmd_freelists_reconstruction_data {
	u32 freelist_count;
	u32 hwr_counter;
	u32 freelist_ids[ROGUE_FWIF_MAX_FREELISTS_TO_RECONSTRUCT];
};

/* 1 if a page fault happened */
#define ROGUE_FWIF_FWCCB_CMD_CONTEXT_RESET_FLAG_PF BIT(0)
/* 1 if applicable to all contexts */
#define ROGUE_FWIF_FWCCB_CMD_CONTEXT_RESET_FLAG_ALL_CTXS BIT(1)

struct rogue_fwif_fwccb_cmd_context_reset_data {
	/* Context affected by the reset */
	u32 server_common_context_id;
	/* Reason for reset */
	enum rogue_context_reset_reason reset_reason;
	/* Data Master affected by the reset */
	u32 dm;
	/* Job ref running at the time of reset */
	u32 reset_job_ref;
	/* ROGUE_FWIF_FWCCB_CMD_CONTEXT_RESET_FLAG bitfield */
	u32 flags;
	/* At what page catalog address */
	aligned_u64 pc_address;
	/* Page fault address (only when applicable) */
	aligned_u64 fault_address;
};

struct rogue_fwif_fwccb_cmd_fw_pagefault_data {
	/* Page fault address */
	u64 fw_fault_addr;
};

enum rogue_fwif_fwccb_cmd_type {
	/* Requests ZSBuffer to be backed with physical pages */
	ROGUE_FWIF_FWCCB_CMD_ZSBUFFER_BACKING = 101U |
						ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Requests ZSBuffer to be unbacked */
	ROGUE_FWIF_FWCCB_CMD_ZSBUFFER_UNBACKING = 102U |
						  ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Requests an on-demand freelist grow/shrink */
	ROGUE_FWIF_FWCCB_CMD_FREELIST_GROW = 103U |
					     ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Requests freelists reconstruction */
	ROGUE_FWIF_FWCCB_CMD_FREELISTS_RECONSTRUCTION =
		104U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Notifies host of a HWR event on a context */
	ROGUE_FWIF_FWCCB_CMD_CONTEXT_RESET_NOTIFICATION =
		105U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Requests an on-demand debug dump */
	ROGUE_FWIF_FWCCB_CMD_DEBUG_DUMP = 106U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	/* Requests an on-demand update on process stats */
	ROGUE_FWIF_FWCCB_CMD_UPDATE_STATS = 107U |
					    ROGUE_CMD_MAGIC_DWORD_SHIFTED,

	ROGUE_FWIF_FWCCB_CMD_CORE_CLK_RATE_CHANGE =
		108U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
	ROGUE_FWIF_FWCCB_CMD_REQUEST_GPU_RESTART =
		109U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,

	/* Notifies host of a FW pagefault */
	ROGUE_FWIF_FWCCB_CMD_CONTEXT_FW_PF_NOTIFICATION =
		112U | ROGUE_CMD_MAGIC_DWORD_SHIFTED,
};

enum rogue_fwif_fwccb_cmd_update_stats_type {
	/*
	 * PVRSRVStatsUpdateRenderContextStats should increase the value of the
	 * ui32TotalNumPartialRenders stat
	 */
	ROGUE_FWIF_FWCCB_CMD_UPDATE_NUM_PARTIAL_RENDERS = 1,
	/*
	 * PVRSRVStatsUpdateRenderContextStats should increase the value of the
	 * ui32TotalNumOutOfMemory stat
	 */
	ROGUE_FWIF_FWCCB_CMD_UPDATE_NUM_OUT_OF_MEMORY,
	/*
	 * PVRSRVStatsUpdateRenderContextStats should increase the value of the
	 * ui32NumGeomStores stat
	 */
	ROGUE_FWIF_FWCCB_CMD_UPDATE_NUM_GEOM_STORES,
	/*
	 * PVRSRVStatsUpdateRenderContextStats should increase the value of the
	 * ui32NumFragStores stat
	 */
	ROGUE_FWIF_FWCCB_CMD_UPDATE_NUM_FRAG_STORES,
	/*
	 * PVRSRVStatsUpdateRenderContextStats should increase the value of the
	 * ui32NumCDMStores stat
	 */
	ROGUE_FWIF_FWCCB_CMD_UPDATE_NUM_CDM_STORES,
	/*
	 * PVRSRVStatsUpdateRenderContextStats should increase the value of the
	 * ui32NumTDMStores stat
	 */
	ROGUE_FWIF_FWCCB_CMD_UPDATE_NUM_TDM_STORES
};

struct rogue_fwif_fwccb_cmd_update_stats_data {
	/* Element to update */
	enum rogue_fwif_fwccb_cmd_update_stats_type element_to_update;
	/* The pid of the process whose stats are being updated */
	u32 pid_owner;
	/* Adjustment to be made to the statistic */
	s32 adjustment_value;
};

struct rogue_fwif_fwccb_cmd_core_clk_rate_change_data {
	u32 core_clk_rate;
} __aligned(8);

struct rogue_fwif_fwccb_cmd {
	/* Command type */
	enum rogue_fwif_fwccb_cmd_type cmd_type;
	/* Compatibility and other flags */
	u32 fwccb_flags;

	union {
		/* Data for Z/S-Buffer on-demand (un)backing*/
		struct rogue_fwif_fwccb_cmd_zsbuffer_backing_data
			cmd_zs_buffer_backing;
		/* Data for on-demand freelist grow/shrink */
		struct rogue_fwif_fwccb_cmd_freelist_gs_data cmd_free_list_gs;
		/* Data for freelists reconstruction */
		struct rogue_fwif_fwccb_cmd_freelists_reconstruction_data
			cmd_freelists_reconstruction;
		/* Data for context reset notification */
		struct rogue_fwif_fwccb_cmd_context_reset_data
			cmd_context_reset_notification;
		/* Data for updating process stats */
		struct rogue_fwif_fwccb_cmd_update_stats_data
			cmd_update_stats_data;
		struct rogue_fwif_fwccb_cmd_core_clk_rate_change_data
			cmd_core_clk_rate_change;
		struct rogue_fwif_fwccb_cmd_fw_pagefault_data cmd_fw_pagefault;
	} cmd_data __aligned(8);
} __aligned(8);

PVR_FW_STRUCT_SIZE_ASSERT(struct rogue_fwif_fwccb_cmd);

/*
 ******************************************************************************
 * Workload estimation Firmware CCB command structure for ROGUE
 ******************************************************************************
 */
struct rogue_fwif_workest_fwccb_cmd {
	/* Index for return data array */
	u16 return_data_index;
	/* The cycles the workload took on the hardware */
	u32 cycles_taken;
};

/*
 ******************************************************************************
 * Client CCB commands for ROGUE
 ******************************************************************************
 */

/*
 * Required memory alignment for 64-bit variables accessible by Meta
 * (The gcc meta aligns 64-bit variables to 64-bit; therefore, memory shared
 * between the host and meta that contains 64-bit variables has to maintain
 * this alignment)
 */
#define ROGUE_FWIF_FWALLOC_ALIGN sizeof(u64)

#define ROGUE_CCB_TYPE_TASK BIT(15)
#define ROGUE_CCB_FWALLOC_ALIGN(size)                \
	(((size) + (ROGUE_FWIF_FWALLOC_ALIGN - 1)) & \
	 ~(ROGUE_FWIF_FWALLOC_ALIGN - 1))

#define ROGUE_FWIF_CCB_CMD_TYPE_GEOM \
	(201U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK)
#define ROGUE_FWIF_CCB_CMD_TYPE_TQ_3D \
	(202U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK)
#define ROGUE_FWIF_CCB_CMD_TYPE_FRAG \
	(203U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK)
#define ROGUE_FWIF_CCB_CMD_TYPE_FRAG_PR \
	(204U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK)
#define ROGUE_FWIF_CCB_CMD_TYPE_CDM \
	(205U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK)
#define ROGUE_FWIF_CCB_CMD_TYPE_TQ_TDM \
	(206U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK)
#define ROGUE_FWIF_CCB_CMD_TYPE_FBSC_INVALIDATE \
	(207U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK)
#define ROGUE_FWIF_CCB_CMD_TYPE_TQ_2D \
	(208U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK)
#define ROGUE_FWIF_CCB_CMD_TYPE_PRE_TIMESTAMP \
	(209U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK)
#define ROGUE_FWIF_CCB_CMD_TYPE_NULL \
	(210U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK)
#define ROGUE_FWIF_CCB_CMD_TYPE_ABORT \
	(211U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK)

/* Leave a gap between CCB specific commands and generic commands */
#define ROGUE_FWIF_CCB_CMD_TYPE_FENCE (212U | ROGUE_CMD_MAGIC_DWORD_SHIFTED)
#define ROGUE_FWIF_CCB_CMD_TYPE_UPDATE (213U | ROGUE_CMD_MAGIC_DWORD_SHIFTED)
#define ROGUE_FWIF_CCB_CMD_TYPE_RMW_UPDATE \
	(214U | ROGUE_CMD_MAGIC_DWORD_SHIFTED)
#define ROGUE_FWIF_CCB_CMD_TYPE_FENCE_PR (215U | ROGUE_CMD_MAGIC_DWORD_SHIFTED)
#define ROGUE_FWIF_CCB_CMD_TYPE_PRIORITY (216U | ROGUE_CMD_MAGIC_DWORD_SHIFTED)
/*
 * Pre and Post timestamp commands are supposed to sandwich the DM cmd. The
 * padding code with the CCB wrap upsets the FW if we don't have the task type
 * bit cleared for POST_TIMESTAMPs. That's why we have 2 different cmd types.
 */
#define ROGUE_FWIF_CCB_CMD_TYPE_POST_TIMESTAMP \
	(217U | ROGUE_CMD_MAGIC_DWORD_SHIFTED)
#define ROGUE_FWIF_CCB_CMD_TYPE_UNFENCED_UPDATE \
	(218U | ROGUE_CMD_MAGIC_DWORD_SHIFTED)
#define ROGUE_FWIF_CCB_CMD_TYPE_UNFENCED_RMW_UPDATE \
	(219U | ROGUE_CMD_MAGIC_DWORD_SHIFTED)

#define ROGUE_FWIF_CCB_CMD_TYPE_PADDING (221U | ROGUE_CMD_MAGIC_DWORD_SHIFTED)

struct rogue_fwif_workest_kick_data {
	/* Index for the KM Workload estimation return data array */
	u16 return_data_index __aligned(8);
	/* Predicted time taken to do the work in cycles */
	u32 cycles_prediction __aligned(8);
	/* Deadline for the workload */
	aligned_u64 deadline;
};

struct rogue_fwif_ccb_cmd_header {
	u32 cmd_type;
	u32 cmd_size;
	/*
	 * external job reference - provided by client and used in debug for
	 * tracking submitted work
	 */
	u32 ext_job_ref;
	/*
	 * internal job reference - generated by services and used in debug for
	 * tracking submitted work
	 */
	u32 int_job_ref;
	/* Workload Estimation - Workload Estimation Data */
	struct rogue_fwif_workest_kick_data work_est_kick_data __aligned(8);
};

/*
 ******************************************************************************
 * Client CCB commands which are only required by the kernel
 ******************************************************************************
 */
struct rogue_fwif_cmd_priority {
	s32 priority;
};

/*
 ******************************************************************************
 * Signature and Checksums Buffer
 ******************************************************************************
 */
struct rogue_fwif_sigbuf_ctl {
	/* Ptr to Signature Buffer memory */
	u32 buffer_fw_addr;
	/* Amount of space left for storing regs in the buffer */
	u32 left_size_in_regs;
} __aligned(8);

struct rogue_fwif_counter_dump_ctl {
	/* Ptr to counter dump buffer */
	u32 buffer_fw_addr;
	/* Amount of space for storing in the buffer */
	u32 size_in_dwords;
} __aligned(8);

struct rogue_fwif_firmware_gcov_ctl {
	/* Ptr to firmware gcov buffer */
	u32 buffer_fw_addr;
	/* Amount of space for storing in the buffer */
	u32 size;
} __aligned(8);

/*
 *****************************************************************************
 * ROGUE Compatibility checks
 *****************************************************************************
 */

/*
 * WARNING: Whenever the layout of ROGUE_FWIF_COMPCHECKS_BVNC changes, the
 * following define should be increased by 1 to indicate to the compatibility
 * logic that layout has changed.
 */
#define ROGUE_FWIF_COMPCHECKS_LAYOUT_VERSION 3

struct rogue_fwif_compchecks_bvnc {
	/* WARNING: This field must be defined as first one in this structure */
	u32 layout_version;
	aligned_u64 bvnc;
} __aligned(8);

struct rogue_fwif_init_options {
	u8 os_count_support;
	u8 padding[7];
} __aligned(8);

#define ROGUE_FWIF_COMPCHECKS_BVNC_DECLARE_AND_INIT(name) \
	struct rogue_fwif_compchecks_bvnc(name) = {       \
		ROGUE_FWIF_COMPCHECKS_LAYOUT_VERSION,     \
		0,                                        \
	}

static inline void rogue_fwif_compchecks_bvnc_init(struct rogue_fwif_compchecks_bvnc *compchecks)
{
	compchecks->layout_version = ROGUE_FWIF_COMPCHECKS_LAYOUT_VERSION;
	compchecks->bvnc = 0;
}

struct rogue_fwif_compchecks {
	/* hardware BVNC (from the ROGUE registers) */
	struct rogue_fwif_compchecks_bvnc hw_bvnc;
	/* firmware BVNC */
	struct rogue_fwif_compchecks_bvnc fw_bvnc;
	/* identifier of the FW processor version */
	u32 fw_processor_version;
	/* software DDK version */
	u32 ddk_version;
	/* software DDK build no. */
	u32 ddk_build;
	/* build options bit-field */
	u32 build_options;
	/* initialisation options bit-field */
	struct rogue_fwif_init_options init_options;
	/* Information is valid */
	bool updated __aligned(4);
	u32 padding;
} __aligned(8);

/*
 ******************************************************************************
 * Updated configuration post FW data init.
 ******************************************************************************
 */
struct rogue_fwif_runtime_cfg {
	/* APM latency in ms before signalling IDLE to the host */
	u32 active_pm_latency_ms;
	/* Compatibility and other flags */
	u32 runtime_cfg_flags;
	/*
	 * If set, APM latency does not reset to system default each GPU power
	 * transition
	 */
	bool active_pm_latency_persistant __aligned(4);
	/* Core clock speed, currently only used to calculate timer ticks */
	u32 core_clock_speed;
	/* Last number of dusts change requested by the host */
	u32 default_dusts_num_init;
	/* Periodic Hardware Reset configuration values */
	u32 phr_mode;
	/* New number of milliseconds C/S is allowed to last */
	u32 hcs_deadline_ms;
	/* The watchdog period in microseconds */
	u32 wdg_period_us;
	/* Array of priorities per OS */
	u32 osid_priority[ROGUE_FW_MAX_NUM_OS];
	/* On-demand allocated HWPerf buffer address, to be passed to the FW */
	u32 hwperf_buf_fw_addr;

	bool padding __aligned(4);
};

/*
 *****************************************************************************
 * Control data for ROGUE
 *****************************************************************************
 */

#define ROGUE_FWIF_HWR_DEBUG_DUMP_ALL (99999U)

enum rogue_fwif_tpu_dm {
	ROGUE_FWIF_TPU_DM_PDM = 0,
	ROGUE_FWIF_TPU_DM_VDM = 1,
	ROGUE_FWIF_TPU_DM_CDM = 2,
	ROGUE_FWIF_TPU_DM_TDM = 3,
	ROGUE_FWIF_TPU_DM_LAST
};

enum rogue_fwif_gpio_val_mode {
	/* No GPIO validation */
	ROGUE_FWIF_GPIO_VAL_OFF = 0,
	/*
	 * Simple test case that initiates by sending data via the GPIO and then
	 * sends back any data received over the GPIO
	 */
	ROGUE_FWIF_GPIO_VAL_GENERAL = 1,
	/*
	 * More complex test case that writes and reads data across the entire
	 * GPIO AP address range.
	 */
	ROGUE_FWIF_GPIO_VAL_AP = 2,
	/* Validates the GPIO Testbench. */
	ROGUE_FWIF_GPIO_VAL_TESTBENCH = 5,
	/* Send and then receive each byte in the range 0-255. */
	ROGUE_FWIF_GPIO_VAL_LOOPBACK = 6,
	/* Send and then receive each power-of-2 byte in the range 0-255. */
	ROGUE_FWIF_GPIO_VAL_LOOPBACK_LITE = 7,
	ROGUE_FWIF_GPIO_VAL_LAST
};

enum fw_perf_conf {
	FW_PERF_CONF_NONE = 0,
	FW_PERF_CONF_ICACHE = 1,
	FW_PERF_CONF_DCACHE = 2,
	FW_PERF_CONF_JTLB_INSTR = 5,
	FW_PERF_CONF_INSTRUCTIONS = 6
};

enum fw_boot_stage {
	FW_BOOT_STAGE_TLB_INIT_FAILURE = -2,
	FW_BOOT_STAGE_NOT_AVAILABLE = -1,
	FW_BOOT_NOT_STARTED = 0,
	FW_BOOT_BLDR_STARTED = 1,
	FW_BOOT_CACHE_DONE,
	FW_BOOT_TLB_DONE,
	FW_BOOT_MAIN_STARTED,
	FW_BOOT_ALIGNCHECKS_DONE,
	FW_BOOT_INIT_DONE,
};

/*
 * Kernel CCB return slot responses. Usage of bit-fields instead of bare
 * integers allows FW to possibly pack-in several responses for each single kCCB
 * command.
 */
/* Command executed (return status from FW) */
#define ROGUE_FWIF_KCCB_RTN_SLOT_CMD_EXECUTED BIT(0)
/* A cleanup was requested but resource busy */
#define ROGUE_FWIF_KCCB_RTN_SLOT_CLEANUP_BUSY BIT(1)
/* Poll failed in FW for a HW operation to complete */
#define ROGUE_FWIF_KCCB_RTN_SLOT_POLL_FAILURE BIT(2)
/* Reset value of a kCCB return slot (set by host) */
#define ROGUE_FWIF_KCCB_RTN_SLOT_NO_RESPONSE 0x0U

struct rogue_fwif_connection_ctl {
	/* Fw-Os connection states */
	enum rogue_fwif_connection_fw_state connection_fw_state;
	enum rogue_fwif_connection_os_state connection_os_state;
	u32 alive_fw_token;
	u32 alive_os_token;
} __aligned(8);

struct rogue_fwif_osinit {
	/* Kernel CCB */
	u32 kernel_ccbctl_fw_addr;
	u32 kernel_ccb_fw_addr;
	u32 kernel_ccb_rtn_slots_fw_addr;

	/* Firmware CCB */
	u32 firmware_ccbctl_fw_addr;
	u32 firmware_ccb_fw_addr;

	/* Workload Estimation Firmware CCB */
	u32 work_est_firmware_ccbctl_fw_addr;
	u32 work_est_firmware_ccb_fw_addr;

	u32 rogue_fwif_hwr_info_buf_ctl_fw_addr;

	u32 hwr_debug_dump_limit;

	u32 fw_os_data_fw_addr;

	/* Compatibility checks to be populated by the Firmware */
	struct rogue_fwif_compchecks rogue_comp_checks;
} __aligned(8);

/* BVNC Features */
struct rogue_hwperf_bvnc_block {
	/* Counter block ID, see ROGUE_HWPERF_CNTBLK_ID */
	u16 block_id;

	/* Number of counters in this block type */
	u16 num_counters;

	/* Number of blocks of this type */
	u16 num_blocks;

	u16 reserved;
};

#define ROGUE_HWPERF_MAX_BVNC_LEN (24)

#define ROGUE_HWPERF_MAX_BVNC_BLOCK_LEN (16U)

/* BVNC Features */
struct rogue_hwperf_bvnc {
	/* BVNC string */
	char bvnc_string[ROGUE_HWPERF_MAX_BVNC_LEN];
	/* See ROGUE_HWPERF_FEATURE_FLAGS */
	u32 bvnc_km_feature_flags;
	/* Number of blocks described in aBvncBlocks */
	u16 num_bvnc_blocks;
	/* Number of GPU cores present */
	u16 bvnc_gpu_cores;
	/* Supported Performance Blocks for BVNC */
	struct rogue_hwperf_bvnc_block
		bvnc_blocks[ROGUE_HWPERF_MAX_BVNC_BLOCK_LEN];
};

PVR_FW_STRUCT_SIZE_ASSERT(struct rogue_hwperf_bvnc);

struct rogue_fwif_sysinit {
	/* Fault read address */
	aligned_u64 fault_phys_addr;

	/* PDS execution base */
	aligned_u64 pds_exec_base;
	/* UCS execution base */
	aligned_u64 usc_exec_base;
	/* FBCDC bindless texture state table base */
	aligned_u64 fbcdc_state_table_base;
	aligned_u64 fbcdc_large_state_table_base;
	/* Texture state base */
	aligned_u64 texture_heap_base;

	/* Event filter for Firmware events */
	u64 hw_perf_filter;

	aligned_u64 slc3_fence_dev_addr;

	u32 tpu_trilinear_frac_mask[ROGUE_FWIF_TPU_DM_LAST] __aligned(8);

	/* Signature and Checksum Buffers for DMs */
	struct rogue_fwif_sigbuf_ctl sigbuf_ctl[PVR_FWIF_DM_MAX];

	struct rogue_fwif_pdvfs_opp pdvfs_opp_info;

	struct rogue_fwif_dma_addr coremem_data_store;

	struct rogue_fwif_counter_dump_ctl counter_dump_ctl;

	u32 filter_flags;

	u32 runtime_cfg_fw_addr;

	u32 trace_buf_ctl_fw_addr;
	u32 fw_sys_data_fw_addr;

	u32 gpu_util_fw_cb_ctl_fw_addr;
	u32 reg_cfg_fw_addr;
	u32 hwperf_ctl_fw_addr;

	u32 align_checks;

	/* Core clock speed at FW boot time */
	u32 initial_core_clock_speed;

	/* APM latency in ms before signalling IDLE to the host */
	u32 active_pm_latency_ms;

	/* Flag to be set by the Firmware after successful start */
	bool firmware_started __aligned(4);

	/* Host/FW Trace synchronisation Partition Marker */
	u32 marker_val;

	/* Firmware initialization complete time */
	u32 firmware_started_timestamp;

	u32 jones_disable_mask;

	/* Firmware performance counter config */
	enum fw_perf_conf firmware_perf;

	/*
	 * FW Pointer to memory containing core clock rate in Hz.
	 * Firmware (PDVFS) updates the memory when running on non primary FW
	 * thread to communicate to host driver.
	 */
	u32 core_clock_rate_fw_addr;

	enum rogue_fwif_gpio_val_mode gpio_validation_mode;

	/* Used in HWPerf for decoding BVNC Features */
	struct rogue_hwperf_bvnc bvnc_km_feature_flags;

	/* Value to write into ROGUE_CR_TFBC_COMPRESSION_CONTROL */
	u32 tfbc_compression_control;
} __aligned(8);

/*
 *****************************************************************************
 * Timer correlation shared data and defines
 *****************************************************************************
 */

struct rogue_fwif_time_corr {
	aligned_u64 os_timestamp;
	aligned_u64 os_mono_timestamp;
	aligned_u64 cr_timestamp;

	/*
	 * Utility variable used to convert CR timer deltas to OS timer deltas
	 * (nS), where the deltas are relative to the timestamps above:
	 * deltaOS = (deltaCR * K) >> decimal_shift, see full explanation below
	 */
	aligned_u64 cr_delta_to_os_delta_kns;

	u32 core_clock_speed;
	u32 reserved;
} __aligned(8);

/*
 * The following macros are used to help converting FW timestamps to the Host
 * time domain. On the FW the ROGUE_CR_TIMER counter is used to keep track of
 * time; it increments by 1 every 256 GPU clock ticks, so the general
 * formula to perform the conversion is:
 *
 * [ GPU clock speed in Hz, if (scale == 10^9) then deltaOS is in nS,
 *   otherwise if (scale == 10^6) then deltaOS is in uS ]
 *
 *             deltaCR * 256                                   256 * scale
 *  deltaOS = --------------- * scale = deltaCR * K    [ K = --------------- ]
 *             GPUclockspeed                                  GPUclockspeed
 *
 * The actual K is multiplied by 2^20 (and deltaCR * K is divided by 2^20)
 * to get some better accuracy and to avoid returning 0 in the integer
 * division 256000000/GPUfreq if GPUfreq is greater than 256MHz.
 * This is the same as keeping K as a decimal number.
 *
 * The maximum deltaOS is slightly more than 5hrs for all GPU frequencies
 * (deltaCR * K is more or less a constant), and it's relative to the base
 * OS timestamp sampled as a part of the timer correlation data.
 * This base is refreshed on GPU power-on, DVFS transition and periodic
 * frequency calibration (executed every few seconds if the FW is doing
 * some work), so as long as the GPU is doing something and one of these
 * events is triggered then deltaCR * K will not overflow and deltaOS will be
 * correct.
 */

#define ROGUE_FWIF_CRDELTA_TO_OSDELTA_ACCURACY_SHIFT (20)

#define ROGUE_FWIF_GET_DELTA_OSTIME_NS(delta_cr, k) \
	(((delta_cr) * (k)) >> ROGUE_FWIF_CRDELTA_TO_OSDELTA_ACCURACY_SHIFT)

/*
 ******************************************************************************
 * GPU Utilisation
 ******************************************************************************
 */

/* See rogue_common.h for a list of GPU states */
#define ROGUE_FWIF_GPU_UTIL_TIME_MASK \
	(0xFFFFFFFFFFFFFFFFull & ~ROGUE_FWIF_GPU_UTIL_STATE_MASK)

#define ROGUE_FWIF_GPU_UTIL_GET_TIME(word) \
	((word)(&ROGUE_FWIF_GPU_UTIL_TIME_MASK))
#define ROGUE_FWIF_GPU_UTIL_GET_STATE(word) \
	((word)(&ROGUE_FWIF_GPU_UTIL_STATE_MASK))

/*
 * The OS timestamps computed by the FW are approximations of the real time,
 * which means they could be slightly behind or ahead the real timer on the
 * Host. In some cases we can perform subtractions between FW approximated
 * timestamps and real OS timestamps, so we need a form of protection against
 * negative results if for instance the FW one is a bit ahead of time.
 */
#define ROGUE_FWIF_GPU_UTIL_GET_PERIOD(newtime, oldtime) \
	(((newtime) > (oldtime)) ? ((newtime) - (oldtime)) : 0U)

#define ROGUE_FWIF_GPU_UTIL_MAKE_WORD(time, state) \
	(ROGUE_FWIF_GPU_UTIL_GET_TIME(time) |      \
	 ROGUE_FWIF_GPU_UTIL_GET_STATE(state))

/*
 * The timer correlation array must be big enough to ensure old entries won't be
 * overwritten before all the HWPerf events linked to those entries are
 * processed by the MISR. The update frequency of this array depends on how fast
 * the system can change state (basically how small the APM latency is) and
 * perform DVFS transitions.
 *
 * The minimum size is 2 (not 1) to avoid race conditions between the FW reading
 * an entry while the Host is updating it. With 2 entries in the worst case the
 * FW will read old data, which is still quite ok if the Host is updating the
 * timer correlation at that time.
 */
#define ROGUE_FWIF_TIME_CORR_ARRAY_SIZE 256U
#define ROGUE_FWIF_TIME_CORR_CURR_INDEX(seqcount) \
	((seqcount) % ROGUE_FWIF_TIME_CORR_ARRAY_SIZE)

/* Make sure the timer correlation array size is a power of 2 */
static_assert((ROGUE_FWIF_TIME_CORR_ARRAY_SIZE &
	       (ROGUE_FWIF_TIME_CORR_ARRAY_SIZE - 1U)) == 0U,
	      "ROGUE_FWIF_TIME_CORR_ARRAY_SIZE must be a power of two");

struct rogue_fwif_gpu_util_fwcb {
	struct rogue_fwif_time_corr time_corr[ROGUE_FWIF_TIME_CORR_ARRAY_SIZE];
	u32 time_corr_seq_count;

	/* Compatibility and other flags */
	u32 gpu_util_flags;

	/* Last GPU state + OS time of the last state update */
	aligned_u64 last_word;

	/* Counters for the amount of time the GPU was active/idle/blocked */
	aligned_u64 stats_counters[PVR_FWIF_GPU_UTIL_STATE_NUM];
} __aligned(8);

struct rogue_fwif_rta_ctl {
	/* Render number */
	u32 render_target_index;
	/* index in RTA */
	u32 current_render_target;
	/* total active RTs */
	u32 active_render_targets;
	/* total active RTs from the first TA kick, for OOM */
	u32 cumul_active_render_targets;
	/* Array of valid RT indices */
	u32 valid_render_targets_fw_addr;
	/* Array of number of occurred partial renders per render target */
	u32 rta_num_partial_renders_fw_addr;
	/* Number of render targets in the array */
	u32 max_rts;
	/* Compatibility and other flags */
	u32 rta_ctl_flags;
} __aligned(8);

struct rogue_fwif_freelist {
	aligned_u64 freelist_dev_addr;
	aligned_u64 current_dev_addr;
	u32 current_stack_top;
	u32 max_pages;
	u32 grow_pages;
	/* HW pages */
	u32 current_pages;
	u32 allocated_page_count;
	u32 allocated_mmu_page_count;
	u32 freelist_id;

	bool grow_pending __aligned(4);
	/* Pages that should be used only when OOM is reached */
	u32 ready_pages;
	/* Compatibility and other flags */
	u32 freelist_flags;
	/* PM Global PB on which Freelist is loaded */
	u32 pm_global_pb;
	u32 padding;
} __aligned(8);

/*
 ******************************************************************************
 * HWRTData
 ******************************************************************************
 */

/* HWRTData flags */
/* Deprecated flags 1:0 */
#define HWRTDATA_HAS_LAST_GEOM BIT(2)
#define HWRTDATA_PARTIAL_RENDERED BIT(3)
#define HWRTDATA_DISABLE_TILE_REORDERING BIT(4)
#define HWRTDATA_NEED_BRN65101_BLIT BIT(5)
#define HWRTDATA_FIRST_BRN65101_STRIP BIT(6)
#define HWRTDATA_NEED_BRN67182_2ND_RENDER BIT(7)

enum rogue_fwif_rtdata_state {
	ROGUE_FWIF_RTDATA_STATE_NONE = 0,
	ROGUE_FWIF_RTDATA_STATE_KICK_GEOM,
	ROGUE_FWIF_RTDATA_STATE_KICK_GEOM_FIRST,
	ROGUE_FWIF_RTDATA_STATE_GEOM_FINISHED,
	ROGUE_FWIF_RTDATA_STATE_KICK_FRAG,
	ROGUE_FWIF_RTDATA_STATE_FRAG_FINISHED,
	ROGUE_FWIF_RTDATA_STATE_FRAG_CONTEXT_STORED,
	ROGUE_FWIF_RTDATA_STATE_GEOM_OUTOFMEM,
	ROGUE_FWIF_RTDATA_STATE_PARTIALRENDERFINISHED,
	/*
	 * In case of HWR, we can't set the RTDATA state to NONE, as this will
	 * cause any TA to become a first TA. To ensure all related TA's are
	 * skipped, we use the HWR state
	 */
	ROGUE_FWIF_RTDATA_STATE_HWR,
	ROGUE_FWIF_RTDATA_STATE_UNKNOWN = 0x7FFFFFFFU
};

struct rogue_fwif_hwrtdata_common {
	bool geom_caches_need_zeroing __aligned(4);

	u32 screen_pixel_max;
	aligned_u64 multi_sample_ctl;
	u64 flipped_multi_sample_ctl;
	u32 tpc_stride;
	u32 tpc_size;
	u32 te_screen;
	u32 mtile_stride;
	u32 teaa;
	u32 te_mtile1;
	u32 te_mtile2;
	u32 isp_merge_lower_x;
	u32 isp_merge_lower_y;
	u32 isp_merge_upper_x;
	u32 isp_merge_upper_y;
	u32 isp_merge_scale_x;
	u32 isp_merge_scale_y;
	u32 rgn_header_size;
	u32 isp_mtile_size;
	u32 padding;
} __aligned(8);

struct rogue_fwif_hwrtdata {
	/* MList Data Store */
	aligned_u64 pm_mlist_dev_addr;

	aligned_u64 vce_cat_base[4];
	aligned_u64 vce_last_cat_base[4];
	aligned_u64 te_cat_base[4];
	aligned_u64 te_last_cat_base[4];
	aligned_u64 alist_cat_base;
	aligned_u64 alist_last_cat_base;

	aligned_u64 pm_alist_stack_pointer;
	u32 pm_mlist_stack_pointer;

	u32 hwrt_data_common_fw_addr;

	u32 hwrt_data_flags;
	enum rogue_fwif_rtdata_state state;

	u32 freelists_fw_addr[MAX_FREELISTS_SIZE] __aligned(8);
	u32 freelist_hwr_snapshot[MAX_FREELISTS_SIZE];

	aligned_u64 vheap_table_dev_addr;

	struct rogue_fwif_rta_ctl rta_ctl;

	aligned_u64 tail_ptrs_dev_addr;
	aligned_u64 macrotile_array_dev_addr;
	aligned_u64 rgn_header_dev_addr;
	aligned_u64 rtc_dev_addr;

	u32 owner_geom_not_used_by_host __aligned(8);

	bool geom_caches_need_zeroing __aligned(4);

	struct rogue_fwif_cleanup_ctl cleanup_state __aligned(64);
} __aligned(8);

/*
 ******************************************************************************
 * Sync checkpoints
 ******************************************************************************
 */

#define PVR_SYNC_CHECKPOINT_UNDEF 0x000
#define PVR_SYNC_CHECKPOINT_ACTIVE 0xac1     /* Checkpoint has not signaled. */
#define PVR_SYNC_CHECKPOINT_SIGNALED 0x519   /* Checkpoint has signaled. */
#define PVR_SYNC_CHECKPOINT_ERRORED 0xeff    /* Checkpoint has been errored. */

#include "pvr_rogue_fwif_check.h"

#endif /* PVR_ROGUE_FWIF_H */