Contributors: 17
	  
        
          | Author | 
          Tokens | 
          Token Proportion | 
          Commits | 
          Commit Proportion | 
        
	  
	  
        
        
          | David Howells | 
          150 | 
          25.86% | 
          1 | 
          4.35% | 
        
        
          | Al Viro | 
          133 | 
          22.93% | 
          2 | 
          8.70% | 
        
        
          | Kees Cook | 
          115 | 
          19.83% | 
          4 | 
          17.39% | 
        
        
          | Dave P Martin | 
          47 | 
          8.10% | 
          1 | 
          4.35% | 
        
        
          | Roman Zippel | 
          36 | 
          6.21% | 
          1 | 
          4.35% | 
        
        
          | Jeremy Fitzhardinge | 
          25 | 
          4.31% | 
          1 | 
          4.35% | 
        
        
          | Keun-o Park | 
          17 | 
          2.93% | 
          1 | 
          4.35% | 
        
        
          | Mark Rutland | 
          16 | 
          2.76% | 
          3 | 
          13.04% | 
        
        
          | Kyle Huey | 
          13 | 
          2.24% | 
          1 | 
          4.35% | 
        
        
          | Andrew Lutomirski | 
          10 | 
          1.72% | 
          1 | 
          4.35% | 
        
        
          | Peter Zijlstra | 
          5 | 
          0.86% | 
          1 | 
          4.35% | 
        
        
          | Roland McGrath | 
          3 | 
          0.52% | 
          1 | 
          4.35% | 
        
        
          | Steven Rostedt | 
          3 | 
          0.52% | 
          1 | 
          4.35% | 
        
        
          | Thomas Gleixner | 
          3 | 
          0.52% | 
          1 | 
          4.35% | 
        
        
          | OGAWA Hirofumi | 
          2 | 
          0.34% | 
          1 | 
          4.35% | 
        
        
          | Greg Kroah-Hartman | 
          1 | 
          0.17% | 
          1 | 
          4.35% | 
        
        
          | Alexander (Sasha) Levin | 
          1 | 
          0.17% | 
          1 | 
          4.35% | 
        
	  
	  
        
          | Total | 
          580 | 
           | 
          23 | 
           | 
	    
	  
    
 
/* SPDX-License-Identifier: GPL-2.0 */
/* thread_info.h: common low-level thread information accessors
 *
 * Copyright (C) 2002  David Howells (dhowells@redhat.com)
 * - Incorporating suggestions made by Linus Torvalds
 */
#ifndef _LINUX_THREAD_INFO_H
#define _LINUX_THREAD_INFO_H
#include <linux/types.h>
#include <linux/bug.h>
#include <linux/restart_block.h>
#ifdef CONFIG_THREAD_INFO_IN_TASK
/*
 * For CONFIG_THREAD_INFO_IN_TASK kernels we need <asm/current.h> for the
 * definition of current, but for !CONFIG_THREAD_INFO_IN_TASK kernels,
 * including <asm/current.h> can cause a circular dependency on some platforms.
 */
#include <asm/current.h>
#define current_thread_info() ((struct thread_info *)current)
#endif
#include <linux/bitops.h>
/*
 * For per-arch arch_within_stack_frames() implementations, defined in
 * asm/thread_info.h.
 */
enum {
	BAD_STACK = -1,
	NOT_STACK = 0,
	GOOD_FRAME,
	GOOD_STACK,
};
#include <asm/thread_info.h>
#ifdef __KERNEL__
#ifndef THREAD_ALIGN
#define THREAD_ALIGN	THREAD_SIZE
#endif
#define THREADINFO_GFP		(GFP_KERNEL_ACCOUNT | __GFP_ZERO)
/*
 * flag set/clear/test wrappers
 * - pass TIF_xxxx constants to these functions
 */
static inline void set_ti_thread_flag(struct thread_info *ti, int flag)
{
	set_bit(flag, (unsigned long *)&ti->flags);
}
static inline void clear_ti_thread_flag(struct thread_info *ti, int flag)
{
	clear_bit(flag, (unsigned long *)&ti->flags);
}
static inline void update_ti_thread_flag(struct thread_info *ti, int flag,
					 bool value)
{
	if (value)
		set_ti_thread_flag(ti, flag);
	else
		clear_ti_thread_flag(ti, flag);
}
static inline int test_and_set_ti_thread_flag(struct thread_info *ti, int flag)
{
	return test_and_set_bit(flag, (unsigned long *)&ti->flags);
}
static inline int test_and_clear_ti_thread_flag(struct thread_info *ti, int flag)
{
	return test_and_clear_bit(flag, (unsigned long *)&ti->flags);
}
static inline int test_ti_thread_flag(struct thread_info *ti, int flag)
{
	return test_bit(flag, (unsigned long *)&ti->flags);
}
#define set_thread_flag(flag) \
	set_ti_thread_flag(current_thread_info(), flag)
#define clear_thread_flag(flag) \
	clear_ti_thread_flag(current_thread_info(), flag)
#define update_thread_flag(flag, value) \
	update_ti_thread_flag(current_thread_info(), flag, value)
#define test_and_set_thread_flag(flag) \
	test_and_set_ti_thread_flag(current_thread_info(), flag)
#define test_and_clear_thread_flag(flag) \
	test_and_clear_ti_thread_flag(current_thread_info(), flag)
#define test_thread_flag(flag) \
	test_ti_thread_flag(current_thread_info(), flag)
#define tif_need_resched() test_thread_flag(TIF_NEED_RESCHED)
#ifndef CONFIG_HAVE_ARCH_WITHIN_STACK_FRAMES
static inline int arch_within_stack_frames(const void * const stack,
					   const void * const stackend,
					   const void *obj, unsigned long len)
{
	return 0;
}
#endif
#ifdef CONFIG_HARDENED_USERCOPY
extern void __check_object_size(const void *ptr, unsigned long n,
					bool to_user);
static __always_inline void check_object_size(const void *ptr, unsigned long n,
					      bool to_user)
{
	if (!__builtin_constant_p(n))
		__check_object_size(ptr, n, to_user);
}
#else
static inline void check_object_size(const void *ptr, unsigned long n,
				     bool to_user)
{ }
#endif /* CONFIG_HARDENED_USERCOPY */
extern void __compiletime_error("copy source size is too small")
__bad_copy_from(void);
extern void __compiletime_error("copy destination size is too small")
__bad_copy_to(void);
static inline void copy_overflow(int size, unsigned long count)
{
	WARN(1, "Buffer overflow detected (%d < %lu)!\n", size, count);
}
static __always_inline bool
check_copy_size(const void *addr, size_t bytes, bool is_source)
{
	int sz = __compiletime_object_size(addr);
	if (unlikely(sz >= 0 && sz < bytes)) {
		if (!__builtin_constant_p(bytes))
			copy_overflow(sz, bytes);
		else if (is_source)
			__bad_copy_from();
		else
			__bad_copy_to();
		return false;
	}
	check_object_size(addr, bytes, is_source);
	return true;
}
#ifndef arch_setup_new_exec
static inline void arch_setup_new_exec(void) { }
#endif
#endif	/* __KERNEL__ */
#endif /* _LINUX_THREAD_INFO_H */