Contributors: 25
	  
        
          | Author | 
          Tokens | 
          Token Proportion | 
          Commits | 
          Commit Proportion | 
        
	  
	  
        
        
          | Frédéric Weisbecker | 
          139 | 
          24.26% | 
          12 | 
          19.05% | 
        
        
          | Zhen Lei | 
          48 | 
          8.38% | 
          3 | 
          4.76% | 
        
        
          | Glauber de Oliveira Costa | 
          46 | 
          8.03% | 
          1 | 
          1.59% | 
        
        
          | Linus Torvalds (pre-git) | 
          43 | 
          7.50% | 
          11 | 
          17.46% | 
        
        
          | Martin Schwidefsky | 
          39 | 
          6.81% | 
          2 | 
          3.17% | 
        
        
          | Andrew Morton | 
          39 | 
          6.81% | 
          3 | 
          4.76% | 
        
        
          | Keika Kobayashi | 
          35 | 
          6.11% | 
          1 | 
          1.59% | 
        
        
          | Yinghai Lu | 
          30 | 
          5.24% | 
          3 | 
          4.76% | 
        
        
          | Bitao Hu | 
          25 | 
          4.36% | 
          1 | 
          1.59% | 
        
        
          | Roland McGrath | 
          20 | 
          3.49% | 
          1 | 
          1.59% | 
        
        
          | Kamezawa Hiroyuki | 
          20 | 
          3.49% | 
          2 | 
          3.17% | 
        
        
          | Thomas Gleixner | 
          20 | 
          3.49% | 
          7 | 
          11.11% | 
        
        
          | Josh Don | 
          16 | 
          2.79% | 
          2 | 
          3.17% | 
        
        
          | Ingo Molnar | 
          14 | 
          2.44% | 
          2 | 
          3.17% | 
        
        
          | Michael Neuling | 
          9 | 
          1.57% | 
          1 | 
          1.59% | 
        
        
          | Alexey Dobriyan | 
          7 | 
          1.22% | 
          2 | 
          3.17% | 
        
        
          | Linus Torvalds | 
          7 | 
          1.22% | 
          1 | 
          1.59% | 
        
        
          | Alan Mayer | 
          4 | 
          0.70% | 
          1 | 
          1.59% | 
        
        
          | Eric Dumazet | 
          4 | 
          0.70% | 
          1 | 
          1.59% | 
        
        
          | Michal Hocko | 
          2 | 
          0.35% | 
          1 | 
          1.59% | 
        
        
          | Christoph Lameter | 
          2 | 
          0.35% | 
          1 | 
          1.59% | 
        
        
          | Greg Kroah-Hartman | 
          1 | 
          0.17% | 
          1 | 
          1.59% | 
        
        
          | Paul Mackerras | 
          1 | 
          0.17% | 
          1 | 
          1.59% | 
        
        
          | Rusty Russell | 
          1 | 
          0.17% | 
          1 | 
          1.59% | 
        
        
          | Ryota Ozaki | 
          1 | 
          0.17% | 
          1 | 
          1.59% | 
        
	  
	  
        
          | Total | 
          573 | 
           | 
          63 | 
           | 
	    
	  
    
 
/* SPDX-License-Identifier: GPL-2.0 */
#ifndef _LINUX_KERNEL_STAT_H
#define _LINUX_KERNEL_STAT_H
#include <linux/smp.h>
#include <linux/threads.h>
#include <linux/percpu.h>
#include <linux/interrupt.h>
#include <linux/sched.h>
#include <linux/vtime.h>
#include <asm/irq.h>
/*
 * 'kernel_stat.h' contains the definitions needed for doing
 * some kernel statistics (CPU usage, context switches ...),
 * used by rstatd/perfmeter
 */
enum cpu_usage_stat {
	CPUTIME_USER,
	CPUTIME_NICE,
	CPUTIME_SYSTEM,
	CPUTIME_SOFTIRQ,
	CPUTIME_IRQ,
	CPUTIME_IDLE,
	CPUTIME_IOWAIT,
	CPUTIME_STEAL,
	CPUTIME_GUEST,
	CPUTIME_GUEST_NICE,
#ifdef CONFIG_SCHED_CORE
	CPUTIME_FORCEIDLE,
#endif
	NR_STATS,
};
struct kernel_cpustat {
	u64 cpustat[NR_STATS];
};
struct kernel_stat {
	unsigned long irqs_sum;
	unsigned int softirqs[NR_SOFTIRQS];
};
DECLARE_PER_CPU(struct kernel_stat, kstat);
DECLARE_PER_CPU(struct kernel_cpustat, kernel_cpustat);
/* Must have preemption disabled for this to be meaningful. */
#define kstat_this_cpu this_cpu_ptr(&kstat)
#define kcpustat_this_cpu this_cpu_ptr(&kernel_cpustat)
#define kstat_cpu(cpu) per_cpu(kstat, cpu)
#define kcpustat_cpu(cpu) per_cpu(kernel_cpustat, cpu)
extern unsigned long long nr_context_switches_cpu(int cpu);
extern unsigned long long nr_context_switches(void);
extern unsigned int kstat_irqs_cpu(unsigned int irq, int cpu);
extern void kstat_incr_irq_this_cpu(unsigned int irq);
static inline void kstat_incr_softirqs_this_cpu(unsigned int irq)
{
	__this_cpu_inc(kstat.softirqs[irq]);
}
static inline unsigned int kstat_softirqs_cpu(unsigned int irq, int cpu)
{
       return kstat_cpu(cpu).softirqs[irq];
}
static inline unsigned int kstat_cpu_softirqs_sum(int cpu)
{
	int i;
	unsigned int sum = 0;
	for (i = 0; i < NR_SOFTIRQS; i++)
		sum += kstat_softirqs_cpu(i, cpu);
	return sum;
}
#ifdef CONFIG_GENERIC_IRQ_STAT_SNAPSHOT
extern void kstat_snapshot_irqs(void);
extern unsigned int kstat_get_irq_since_snapshot(unsigned int irq);
#else
static inline void kstat_snapshot_irqs(void) { }
static inline unsigned int kstat_get_irq_since_snapshot(unsigned int irq) { return 0; }
#endif
/*
 * Number of interrupts per specific IRQ source, since bootup
 */
extern unsigned int kstat_irqs_usr(unsigned int irq);
/*
 * Number of interrupts per cpu, since bootup
 */
static inline unsigned long kstat_cpu_irqs_sum(unsigned int cpu)
{
	return kstat_cpu(cpu).irqs_sum;
}
#ifdef CONFIG_VIRT_CPU_ACCOUNTING_GEN
extern u64 kcpustat_field(struct kernel_cpustat *kcpustat,
			  enum cpu_usage_stat usage, int cpu);
extern void kcpustat_cpu_fetch(struct kernel_cpustat *dst, int cpu);
#else
static inline u64 kcpustat_field(struct kernel_cpustat *kcpustat,
				 enum cpu_usage_stat usage, int cpu)
{
	return kcpustat->cpustat[usage];
}
static inline void kcpustat_cpu_fetch(struct kernel_cpustat *dst, int cpu)
{
	*dst = kcpustat_cpu(cpu);
}
#endif
extern void account_user_time(struct task_struct *, u64);
extern void account_guest_time(struct task_struct *, u64);
extern void account_system_time(struct task_struct *, int, u64);
extern void account_system_index_time(struct task_struct *, u64,
				      enum cpu_usage_stat);
extern void account_steal_time(u64);
extern void account_idle_time(u64);
extern u64 get_idle_time(struct kernel_cpustat *kcs, int cpu);
#ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE
static inline void account_process_tick(struct task_struct *tsk, int user)
{
	vtime_flush(tsk);
}
#else
extern void account_process_tick(struct task_struct *, int user);
#endif
extern void account_idle_ticks(unsigned long ticks);
#ifdef CONFIG_SCHED_CORE
extern void __account_forceidle_time(struct task_struct *tsk, u64 delta);
#endif
#endif /* _LINUX_KERNEL_STAT_H */