Contributors: 9
	  
        
          | Author | 
          Tokens | 
          Token Proportion | 
          Commits | 
          Commit Proportion | 
        
	  
	  
        
        
          | Peter Zijlstra | 
          208 | 
          71.48% | 
          9 | 
          42.86% | 
        
        
          | Frédéric Weisbecker | 
          33 | 
          11.34% | 
          4 | 
          19.05% | 
        
        
          | Sebastian Andrzej Siewior | 
          15 | 
          5.15% | 
          2 | 
          9.52% | 
        
        
          | Huang Ying | 
          14 | 
          4.81% | 
          1 | 
          4.76% | 
        
        
          | Ingo Molnar | 
          9 | 
          3.09% | 
          1 | 
          4.76% | 
        
        
          | Ilya Leoshkevich | 
          6 | 
          2.06% | 
          1 | 
          4.76% | 
        
        
          | Steven Rostedt | 
          4 | 
          1.37% | 
          1 | 
          4.76% | 
        
        
          | Greg Kroah-Hartman | 
          1 | 
          0.34% | 
          1 | 
          4.76% | 
        
        
          | James Hogan | 
          1 | 
          0.34% | 
          1 | 
          4.76% | 
        
	  
	  
        
          | Total | 
          291 | 
           | 
          21 | 
           | 
	    
	  
    
 
/* SPDX-License-Identifier: GPL-2.0 */
#ifndef _LINUX_IRQ_WORK_H
#define _LINUX_IRQ_WORK_H
#include <linux/smp_types.h>
#include <linux/rcuwait.h>
/*
 * An entry can be in one of four states:
 *
 * free	     NULL, 0 -> {claimed}       : free to be used
 * claimed   NULL, 3 -> {pending}       : claimed to be enqueued
 * pending   next, 3 -> {busy}          : queued, pending callback
 * busy      NULL, 2 -> {free, claimed} : callback in progress, can be claimed
 */
struct irq_work {
	struct __call_single_node node;
	void (*func)(struct irq_work *);
	struct rcuwait irqwait;
};
#define __IRQ_WORK_INIT(_func, _flags) (struct irq_work){	\
	.node = { .u_flags = (_flags), },			\
	.func = (_func),					\
	.irqwait = __RCUWAIT_INITIALIZER(irqwait),		\
}
#define IRQ_WORK_INIT(_func) __IRQ_WORK_INIT(_func, 0)
#define IRQ_WORK_INIT_LAZY(_func) __IRQ_WORK_INIT(_func, IRQ_WORK_LAZY)
#define IRQ_WORK_INIT_HARD(_func) __IRQ_WORK_INIT(_func, IRQ_WORK_HARD_IRQ)
#define DEFINE_IRQ_WORK(name, _f)				\
	struct irq_work name = IRQ_WORK_INIT(_f)
static inline
void init_irq_work(struct irq_work *work, void (*func)(struct irq_work *))
{
	*work = IRQ_WORK_INIT(func);
}
static inline bool irq_work_is_pending(struct irq_work *work)
{
	return atomic_read(&work->node.a_flags) & IRQ_WORK_PENDING;
}
static inline bool irq_work_is_busy(struct irq_work *work)
{
	return atomic_read(&work->node.a_flags) & IRQ_WORK_BUSY;
}
static inline bool irq_work_is_hard(struct irq_work *work)
{
	return atomic_read(&work->node.a_flags) & IRQ_WORK_HARD_IRQ;
}
bool irq_work_queue(struct irq_work *work);
bool irq_work_queue_on(struct irq_work *work, int cpu);
void irq_work_tick(void);
void irq_work_sync(struct irq_work *work);
#ifdef CONFIG_IRQ_WORK
#include <asm/irq_work.h>
void irq_work_run(void);
bool irq_work_needs_cpu(void);
void irq_work_single(void *arg);
void arch_irq_work_raise(void);
#else
static inline bool irq_work_needs_cpu(void) { return false; }
static inline void irq_work_run(void) { }
static inline void irq_work_single(void *arg) { }
#endif
#endif /* _LINUX_IRQ_WORK_H */