Contributors: 35
Author Tokens Token Proportion Commits Commit Proportion
Rusty Russell 95 19.92% 7 11.67%
Steven Rostedt 54 11.32% 3 5.00%
Linus Torvalds 45 9.43% 5 8.33%
Sergey Senozhatsky 37 7.76% 1 1.67%
Arjan van de Ven 28 5.87% 1 1.67%
James Bottomley 26 5.45% 2 3.33%
Linus Torvalds (pre-git) 22 4.61% 5 8.33%
Ingo Molnar 17 3.56% 2 3.33%
Masami Hiramatsu 16 3.35% 1 1.67%
Paul Mackerras 14 2.94% 1 1.67%
Andrew Morton 13 2.73% 1 1.67%
Christophe Leroy 12 2.52% 5 8.33%
Daniel Borkmann 12 2.52% 1 1.67%
Alexei Starovoitov 12 2.52% 1 1.67%
David Daney 11 2.31% 1 1.67%
Santosh Sivaraj 10 2.10% 1 1.67%
Borislav Petkov 7 1.47% 1 1.67%
Frédéric Weisbecker 6 1.26% 2 3.33%
Mathieu Desnoyers 5 1.05% 1 1.67%
Kefeng Wang 5 1.05% 2 3.33%
Uwe Kleine-König 4 0.84% 1 1.67%
Thomas Gleixner 3 0.63% 2 3.33%
Dan Aloni 3 0.63% 1 1.67%
Anton Blanchard 3 0.63% 1 1.67%
Prasanna S. Panchamukhi 3 0.63% 1 1.67%
Christoph Hellwig 3 0.63% 1 1.67%
Kyle McMartin 2 0.42% 1 1.67%
Nathan Chancellor 2 0.42% 1 1.67%
Zhou Chengming 1 0.21% 1 1.67%
Dmitri Vorobiev 1 0.21% 1 1.67%
Paul Gortmaker 1 0.21% 1 1.67%
Andi Kleen 1 0.21% 1 1.67%
Thomas Meyer 1 0.21% 1 1.67%
Marcin Nowakowski 1 0.21% 1 1.67%
Jiri Olsa 1 0.21% 1 1.67%
Total 477 60


// SPDX-License-Identifier: GPL-2.0-or-later
/* Rewritten by Rusty Russell, on the backs of many others...
   Copyright (C) 2001 Rusty Russell, 2002 Rusty Russell IBM.

*/
#include <linux/elf.h>
#include <linux/ftrace.h>
#include <linux/memory.h>
#include <linux/extable.h>
#include <linux/module.h>
#include <linux/mutex.h>
#include <linux/init.h>
#include <linux/kprobes.h>
#include <linux/filter.h>

#include <asm/sections.h>
#include <linux/uaccess.h>

/*
 * mutex protecting text section modification (dynamic code patching).
 * some users need to sleep (allocating memory...) while they hold this lock.
 *
 * Note: Also protects SMP-alternatives modification on x86.
 *
 * NOT exported to modules - patching kernel text is a really delicate matter.
 */
DEFINE_MUTEX(text_mutex);

extern struct exception_table_entry __start___ex_table[];
extern struct exception_table_entry __stop___ex_table[];

/* Cleared by build time tools if the table is already sorted. */
u32 __initdata __visible main_extable_sort_needed = 1;

/* Sort the kernel's built-in exception table */
void __init sort_main_extable(void)
{
	if (main_extable_sort_needed &&
	    &__stop___ex_table > &__start___ex_table) {
		pr_notice("Sorting __ex_table...\n");
		sort_extable(__start___ex_table, __stop___ex_table);
	}
}

/* Given an address, look for it in the kernel exception table */
const
struct exception_table_entry *search_kernel_exception_table(unsigned long addr)
{
	return search_extable(__start___ex_table,
			      __stop___ex_table - __start___ex_table, addr);
}

/* Given an address, look for it in the exception tables. */
const struct exception_table_entry *search_exception_tables(unsigned long addr)
{
	const struct exception_table_entry *e;

	e = search_kernel_exception_table(addr);
	if (!e)
		e = search_module_extables(addr);
	if (!e)
		e = search_bpf_extables(addr);
	return e;
}

int notrace core_kernel_text(unsigned long addr)
{
	if (is_kernel_text(addr))
		return 1;

	if (system_state < SYSTEM_FREEING_INITMEM &&
	    is_kernel_inittext(addr))
		return 1;
	return 0;
}

int __kernel_text_address(unsigned long addr)
{
	if (kernel_text_address(addr))
		return 1;
	/*
	 * There might be init symbols in saved stacktraces.
	 * Give those symbols a chance to be printed in
	 * backtraces (such as lockdep traces).
	 *
	 * Since we are after the module-symbols check, there's
	 * no danger of address overlap:
	 */
	if (is_kernel_inittext(addr))
		return 1;
	return 0;
}

int kernel_text_address(unsigned long addr)
{
	bool no_rcu;
	int ret = 1;

	if (core_kernel_text(addr))
		return 1;

	/*
	 * If a stack dump happens while RCU is not watching, then
	 * RCU needs to be notified that it requires to start
	 * watching again. This can happen either by tracing that
	 * triggers a stack trace, or a WARN() that happens during
	 * coming back from idle, or cpu on or offlining.
	 *
	 * is_module_text_address() as well as the kprobe slots,
	 * is_bpf_text_address() and is_bpf_image_address require
	 * RCU to be watching.
	 */
	no_rcu = !rcu_is_watching();

	/* Treat this like an NMI as it can happen anywhere */
	if (no_rcu)
		ct_nmi_enter();

	if (is_module_text_address(addr))
		goto out;
	if (is_ftrace_trampoline(addr))
		goto out;
	if (is_kprobe_optinsn_slot(addr) || is_kprobe_insn_slot(addr))
		goto out;
	if (is_bpf_text_address(addr))
		goto out;
	ret = 0;
out:
	if (no_rcu)
		ct_nmi_exit();

	return ret;
}

/*
 * On some architectures (PPC64, IA64, PARISC) function pointers
 * are actually only tokens to some data that then holds the
 * real function address. As a result, to find if a function
 * pointer is part of the kernel text, we need to do some
 * special dereferencing first.
 */
#ifdef CONFIG_HAVE_FUNCTION_DESCRIPTORS
void *dereference_function_descriptor(void *ptr)
{
	func_desc_t *desc = ptr;
	void *p;

	if (!get_kernel_nofault(p, (void *)&desc->addr))
		ptr = p;
	return ptr;
}
EXPORT_SYMBOL_GPL(dereference_function_descriptor);

void *dereference_kernel_function_descriptor(void *ptr)
{
	if (ptr < (void *)__start_opd || ptr >= (void *)__end_opd)
		return ptr;

	return dereference_function_descriptor(ptr);
}
#endif

int func_ptr_is_kernel_text(void *ptr)
{
	unsigned long addr;
	addr = (unsigned long) dereference_function_descriptor(ptr);
	if (core_kernel_text(addr))
		return 1;
	return is_module_text_address(addr);
}