Contributors: 5
Author |
Tokens |
Token Proportion |
Commits |
Commit Proportion |
Mark Salter |
166 |
91.71% |
2 |
33.33% |
Tom Lendacky |
5 |
2.76% |
1 |
16.67% |
Rik Van Riel |
5 |
2.76% |
1 |
16.67% |
Juergen Gross |
4 |
2.21% |
1 |
16.67% |
Greg Kroah-Hartman |
1 |
0.55% |
1 |
16.67% |
Total |
181 |
|
6 |
|
/* SPDX-License-Identifier: GPL-2.0 */
#ifndef _ASM_EARLY_IOREMAP_H_
#define _ASM_EARLY_IOREMAP_H_
#include <linux/types.h>
/*
* early_ioremap() and early_iounmap() are for temporary early boot-time
* mappings, before the real ioremap() is functional.
*/
extern void __iomem *early_ioremap(resource_size_t phys_addr,
unsigned long size);
extern void *early_memremap(resource_size_t phys_addr,
unsigned long size);
extern void *early_memremap_ro(resource_size_t phys_addr,
unsigned long size);
extern void *early_memremap_prot(resource_size_t phys_addr,
unsigned long size, unsigned long prot_val);
extern void early_iounmap(void __iomem *addr, unsigned long size);
extern void early_memunmap(void *addr, unsigned long size);
#if defined(CONFIG_GENERIC_EARLY_IOREMAP) && defined(CONFIG_MMU)
/* Arch-specific initialization */
extern void early_ioremap_init(void);
/* Generic initialization called by architecture code */
extern void early_ioremap_setup(void);
/*
* Called as last step in paging_init() so library can act
* accordingly for subsequent map/unmap requests.
*/
extern void early_ioremap_reset(void);
/*
* Early copy from unmapped memory to kernel mapped memory.
*/
extern void copy_from_early_mem(void *dest, phys_addr_t src,
unsigned long size);
#else
static inline void early_ioremap_init(void) { }
static inline void early_ioremap_setup(void) { }
static inline void early_ioremap_reset(void) { }
#endif
#endif /* _ASM_EARLY_IOREMAP_H_ */