Contributors: 8
Author Tokens Token Proportion Commits Commit Proportion
Matthew Brost 163 55.44% 1 4.35%
Matt Roper 65 22.11% 2 8.70%
Thomas Hellstrom 22 7.48% 3 13.04%
Anshuman Gupta 21 7.14% 4 17.39%
Rodrigo Vivi 14 4.76% 8 34.78%
Matthew Auld 6 2.04% 3 13.04%
Himal Prasad Ghimiray 2 0.68% 1 4.35%
Badal Nilawar 1 0.34% 1 4.35%
Total 294 23


/* SPDX-License-Identifier: MIT */
/*
 * Copyright © 2022 Intel Corporation
 */

#ifndef _XE_PM_H_
#define _XE_PM_H_

#include <linux/cleanup.h>
#include <linux/pm_runtime.h>

#define DEFAULT_VRAM_THRESHOLD 300 /* in MB */

struct xe_device;

int xe_pm_suspend(struct xe_device *xe);
int xe_pm_resume(struct xe_device *xe);

int xe_pm_init_early(struct xe_device *xe);
int xe_pm_init(struct xe_device *xe);
void xe_pm_fini(struct xe_device *xe);
bool xe_pm_runtime_suspended(struct xe_device *xe);
int xe_pm_runtime_suspend(struct xe_device *xe);
int xe_pm_runtime_resume(struct xe_device *xe);
void xe_pm_runtime_get(struct xe_device *xe);
int xe_pm_runtime_get_ioctl(struct xe_device *xe);
void xe_pm_runtime_put(struct xe_device *xe);
bool xe_pm_runtime_get_if_active(struct xe_device *xe);
bool xe_pm_runtime_get_if_in_use(struct xe_device *xe);
void xe_pm_runtime_get_noresume(struct xe_device *xe);
bool xe_pm_runtime_resume_and_get(struct xe_device *xe);
void xe_pm_assert_unbounded_bridge(struct xe_device *xe);
int xe_pm_set_vram_threshold(struct xe_device *xe, u32 threshold);
void xe_pm_d3cold_allowed_toggle(struct xe_device *xe);
bool xe_rpm_reclaim_safe(const struct xe_device *xe);
struct task_struct *xe_pm_read_callback_task(struct xe_device *xe);
int xe_pm_block_on_suspend(struct xe_device *xe);
void xe_pm_might_block_on_suspend(void);
int xe_pm_module_init(void);

static inline void __xe_pm_runtime_noop(struct xe_device *xe) {}

DEFINE_GUARD(xe_pm_runtime, struct xe_device *,
	     xe_pm_runtime_get(_T), xe_pm_runtime_put(_T))
DEFINE_GUARD(xe_pm_runtime_noresume, struct xe_device *,
	     xe_pm_runtime_get_noresume(_T), xe_pm_runtime_put(_T))
DEFINE_GUARD_COND(xe_pm_runtime, _ioctl, xe_pm_runtime_get_ioctl(_T), _RET >= 0)

/*
 * Used when a function needs to release runtime PM in all possible cases
 * and error paths, but the wakeref was already acquired by a different
 * function (i.e., get() has already happened so only a put() is needed).
 */
DEFINE_GUARD(xe_pm_runtime_release_only, struct xe_device *,
	     __xe_pm_runtime_noop(_T), xe_pm_runtime_put(_T));

#endif