Contributors: 14
Author |
Tokens |
Token Proportion |
Commits |
Commit Proportion |
Lorenzo Stoakes |
35 |
25.93% |
2 |
13.33% |
Vivek Kasireddy |
33 |
24.44% |
1 |
6.67% |
David Herrmann |
29 |
21.48% |
1 |
6.67% |
Mike Kravetz |
17 |
12.59% |
1 |
6.67% |
Hugh Dickins |
6 |
4.44% |
1 |
6.67% |
Linus Torvalds (pre-git) |
2 |
1.48% |
1 |
6.67% |
Luca Vizzarro |
2 |
1.48% |
1 |
6.67% |
Marc-André Lureau |
2 |
1.48% |
1 |
6.67% |
John Hubbard |
2 |
1.48% |
1 |
6.67% |
Christoph Hellwig |
2 |
1.48% |
1 |
6.67% |
Linus Torvalds |
2 |
1.48% |
1 |
6.67% |
Greg Kroah-Hartman |
1 |
0.74% |
1 |
6.67% |
Matt Mackall |
1 |
0.74% |
1 |
6.67% |
Matthew Wilcox |
1 |
0.74% |
1 |
6.67% |
Total |
135 |
|
15 |
|
/* SPDX-License-Identifier: GPL-2.0 */
#ifndef __LINUX_MEMFD_H
#define __LINUX_MEMFD_H
#include <linux/file.h>
#ifdef CONFIG_MEMFD_CREATE
extern long memfd_fcntl(struct file *file, unsigned int cmd, unsigned int arg);
struct folio *memfd_alloc_folio(struct file *memfd, pgoff_t idx);
/*
* Check for any existing seals on mmap, return an error if access is denied due
* to sealing, or 0 otherwise.
*
* We also update VMA flags if appropriate by manipulating the VMA flags pointed
* to by vm_flags_ptr.
*/
int memfd_check_seals_mmap(struct file *file, unsigned long *vm_flags_ptr);
#else
static inline long memfd_fcntl(struct file *f, unsigned int c, unsigned int a)
{
return -EINVAL;
}
static inline struct folio *memfd_alloc_folio(struct file *memfd, pgoff_t idx)
{
return ERR_PTR(-EINVAL);
}
static inline int memfd_check_seals_mmap(struct file *file,
unsigned long *vm_flags_ptr)
{
return 0;
}
#endif
#endif /* __LINUX_MEMFD_H */