Contributors: 16
Author |
Tokens |
Token Proportion |
Commits |
Commit Proportion |
Chris Wilson |
190 |
42.51% |
16 |
34.78% |
Imre Deak |
104 |
23.27% |
4 |
8.70% |
Michał Winiarski |
37 |
8.28% |
2 |
4.35% |
Michal Wajdeczko |
37 |
8.28% |
4 |
8.70% |
Jani Nikula |
17 |
3.80% |
6 |
13.04% |
Robert Beckett |
15 |
3.36% |
1 |
2.17% |
Zhenyu Wang |
11 |
2.46% |
1 |
2.17% |
Janusz Krzysztofik |
10 |
2.24% |
3 |
6.52% |
Tvrtko A. Ursulin |
7 |
1.57% |
2 |
4.35% |
Dave Airlie |
4 |
0.89% |
1 |
2.17% |
José Roberto de Souza |
4 |
0.89% |
1 |
2.17% |
David Weinehall |
4 |
0.89% |
1 |
2.17% |
Thomas Zimmermann |
3 |
0.67% |
1 |
2.17% |
Ben Widawsky |
2 |
0.45% |
1 |
2.17% |
Eric Anholt |
1 |
0.22% |
1 |
2.17% |
Mika Kuoppala |
1 |
0.22% |
1 |
2.17% |
Total |
447 |
|
46 |
|
// SPDX-License-Identifier: MIT
/*
* Copyright © 2019 Intel Corporation
*/
#include <linux/device.h>
#include <drm/drm_drv.h>
#include "i915_drv.h"
#include "i915_utils.h"
#define FDO_BUG_MSG "Please file a bug on drm/i915; see " FDO_BUG_URL " for details."
void
__i915_printk(struct drm_i915_private *dev_priv, const char *level,
const char *fmt, ...)
{
static bool shown_bug_once;
struct device *kdev = dev_priv->drm.dev;
bool is_error = level[1] <= KERN_ERR[1];
bool is_debug = level[1] == KERN_DEBUG[1];
struct va_format vaf;
va_list args;
if (is_debug && !drm_debug_enabled(DRM_UT_DRIVER))
return;
va_start(args, fmt);
vaf.fmt = fmt;
vaf.va = &args;
if (is_error)
dev_printk(level, kdev, "%pV", &vaf);
else
dev_printk(level, kdev, "[" DRM_NAME ":%ps] %pV",
__builtin_return_address(0), &vaf);
va_end(args);
if (is_error && !shown_bug_once) {
/*
* Ask the user to file a bug report for the error, except
* if they may have caused the bug by fiddling with unsafe
* module parameters.
*/
if (!test_taint(TAINT_USER))
dev_notice(kdev, "%s", FDO_BUG_MSG);
shown_bug_once = true;
}
}
void add_taint_for_CI(struct drm_i915_private *i915, unsigned int taint)
{
__i915_printk(i915, KERN_NOTICE, "CI tainted:%#x by %pS\n",
taint, (void *)_RET_IP_);
/* Failures that occur during fault injection testing are expected */
if (!i915_error_injected())
__add_taint_for_CI(taint);
}
#if IS_ENABLED(CONFIG_DRM_I915_DEBUG)
static unsigned int i915_probe_fail_count;
int __i915_inject_probe_error(struct drm_i915_private *i915, int err,
const char *func, int line)
{
if (i915_probe_fail_count >= i915_modparams.inject_probe_failure)
return 0;
if (++i915_probe_fail_count < i915_modparams.inject_probe_failure)
return 0;
__i915_printk(i915, KERN_INFO,
"Injecting failure %d at checkpoint %u [%s:%d]\n",
err, i915_modparams.inject_probe_failure, func, line);
i915_modparams.inject_probe_failure = 0;
return err;
}
bool i915_error_injected(void)
{
return i915_probe_fail_count && !i915_modparams.inject_probe_failure;
}
#endif
void cancel_timer(struct timer_list *t)
{
if (!timer_active(t))
return;
del_timer(t);
WRITE_ONCE(t->expires, 0);
}
void set_timer_ms(struct timer_list *t, unsigned long timeout)
{
if (!timeout) {
cancel_timer(t);
return;
}
timeout = msecs_to_jiffies(timeout);
/*
* Paranoia to make sure the compiler computes the timeout before
* loading 'jiffies' as jiffies is volatile and may be updated in
* the background by a timer tick. All to reduce the complexity
* of the addition and reduce the risk of losing a jiffie.
*/
barrier();
/* Keep t->expires = 0 reserved to indicate a canceled timer. */
mod_timer(t, jiffies + timeout ?: 1);
}
bool i915_vtd_active(struct drm_i915_private *i915)
{
if (device_iommu_mapped(i915->drm.dev))
return true;
/* Running as a guest, we assume the host is enforcing VT'd */
return i915_run_as_guest();
}