Contributors: 7
	  
        
          | Author | 
          Tokens | 
          Token Proportion | 
          Commits | 
          Commit Proportion | 
        
	  
	  
        
        
          | Adrian Hunter | 
          89 | 
          35.32% | 
          2 | 
          20.00% | 
        
        
          | Peter Zijlstra | 
          88 | 
          34.92% | 
          2 | 
          20.00% | 
        
        
          | Jeremy Fitzhardinge | 
          47 | 
          18.65% | 
          2 | 
          20.00% | 
        
        
          | Roman Zippel | 
          13 | 
          5.16% | 
          1 | 
          10.00% | 
        
        
          | Vincenzo Frascino | 
          7 | 
          2.78% | 
          1 | 
          10.00% | 
        
        
          | Haozhong Zhang | 
          7 | 
          2.78% | 
          1 | 
          10.00% | 
        
        
          | Greg Kroah-Hartman | 
          1 | 
          0.40% | 
          1 | 
          10.00% | 
        
	  
	  
        
          | Total | 
          252 | 
           | 
          10 | 
           | 
	    
	  
    
 
/* SPDX-License-Identifier: GPL-2.0 */
#ifndef __VDSO_MATH64_H
#define __VDSO_MATH64_H
static __always_inline u32
__iter_div_u64_rem(u64 dividend, u32 divisor, u64 *remainder)
{
	u32 ret = 0;
	while (dividend >= divisor) {
		/* The following asm() prevents the compiler from
		   optimising this loop into a modulo operation.  */
		asm("" : "+rm"(dividend));
		dividend -= divisor;
		ret++;
	}
	*remainder = dividend;
	return ret;
}
#if defined(CONFIG_ARCH_SUPPORTS_INT128) && defined(__SIZEOF_INT128__)
#ifndef mul_u64_u32_add_u64_shr
static __always_inline u64 mul_u64_u32_add_u64_shr(u64 a, u32 mul, u64 b, unsigned int shift)
{
	return (u64)((((unsigned __int128)a * mul) + b) >> shift);
}
#endif /* mul_u64_u32_add_u64_shr */
#else
#ifndef mul_u64_u32_add_u64_shr
#ifndef mul_u32_u32
static inline u64 mul_u32_u32(u32 a, u32 b)
{
	return (u64)a * b;
}
#define mul_u32_u32 mul_u32_u32
#endif
static __always_inline u64 mul_u64_u32_add_u64_shr(u64 a, u32 mul, u64 b, unsigned int shift)
{
	u32 ah = a >> 32, al = a;
	bool ovf;
	u64 ret;
	ovf = __builtin_add_overflow(mul_u32_u32(al, mul), b, &ret);
	ret >>= shift;
	if (ovf && shift)
		ret += 1ULL << (64 - shift);
	if (ah)
		ret += mul_u32_u32(ah, mul) << (32 - shift);
	return ret;
}
#endif /* mul_u64_u32_add_u64_shr */
#endif
#endif /* __VDSO_MATH64_H */