Contributors: 7
	  
        
          | Author | 
          Tokens | 
          Token Proportion | 
          Commits | 
          Commit Proportion | 
        
	  
	  
        
        
          | Davidlohr Bueso A | 
          160 | 
          90.40% | 
          5 | 
          45.45% | 
        
        
          | Peter Zijlstra | 
          11 | 
          6.21% | 
          1 | 
          9.09% | 
        
        
          | Ingo Molnar | 
          2 | 
          1.13% | 
          1 | 
          9.09% | 
        
        
          | Joel A Fernandes | 
          1 | 
          0.56% | 
          1 | 
          9.09% | 
        
        
          | Eric W. Biedermann | 
          1 | 
          0.56% | 
          1 | 
          9.09% | 
        
        
          | Paolo Bonzini | 
          1 | 
          0.56% | 
          1 | 
          9.09% | 
        
        
          | Greg Kroah-Hartman | 
          1 | 
          0.56% | 
          1 | 
          9.09% | 
        
	  
	  
        
          | Total | 
          177 | 
           | 
          11 | 
           | 
	    
	  
    
 
/* SPDX-License-Identifier: GPL-2.0 */
#ifndef _LINUX_RCUWAIT_H_
#define _LINUX_RCUWAIT_H_
#include <linux/rcupdate.h>
#include <linux/sched/signal.h>
/*
 * rcuwait provides a way of blocking and waking up a single
 * task in an rcu-safe manner.
 *
 * The only time @task is non-nil is when a user is blocked (or
 * checking if it needs to) on a condition, and reset as soon as we
 * know that the condition has succeeded and are awoken.
 */
struct rcuwait {
	struct task_struct __rcu *task;
};
#define __RCUWAIT_INITIALIZER(name)		\
	{ .task = NULL, }
static inline void rcuwait_init(struct rcuwait *w)
{
	w->task = NULL;
}
/*
 * Note: this provides no serialization and, just as with waitqueues,
 * requires care to estimate as to whether or not the wait is active.
 */
static inline int rcuwait_active(struct rcuwait *w)
{
	return !!rcu_access_pointer(w->task);
}
extern int rcuwait_wake_up(struct rcuwait *w);
/*
 * The caller is responsible for locking around rcuwait_wait_event(),
 * and [prepare_to/finish]_rcuwait() such that writes to @task are
 * properly serialized.
 */
static inline void prepare_to_rcuwait(struct rcuwait *w)
{
	rcu_assign_pointer(w->task, current);
}
extern void finish_rcuwait(struct rcuwait *w);
#define ___rcuwait_wait_event(w, condition, state, ret, cmd)		\
({									\
	long __ret = ret;						\
	prepare_to_rcuwait(w);						\
	for (;;) {							\
		/*							\
		 * Implicit barrier (A) pairs with (B) in		\
		 * rcuwait_wake_up().					\
		 */							\
		set_current_state(state);				\
		if (condition)						\
			break;						\
									\
		if (signal_pending_state(state, current)) {		\
			__ret = -EINTR;					\
			break;						\
		}							\
									\
		cmd;							\
	}								\
	finish_rcuwait(w);						\
	__ret;								\
})
#define rcuwait_wait_event(w, condition, state)				\
	___rcuwait_wait_event(w, condition, state, 0, schedule())
#define __rcuwait_wait_event_timeout(w, condition, state, timeout)	\
	___rcuwait_wait_event(w, ___wait_cond_timeout(condition),	\
			      state, timeout,				\
			      __ret = schedule_timeout(__ret))
#define rcuwait_wait_event_timeout(w, condition, state, timeout)	\
({									\
	long __ret = timeout;						\
	if (!___wait_cond_timeout(condition))				\
		__ret = __rcuwait_wait_event_timeout(w, condition,	\
						     state, timeout);	\
	__ret;								\
})
#endif /* _LINUX_RCUWAIT_H_ */