Contributors: 8
Author Tokens Token Proportion Commits Commit Proportion
Chris Zankel 104 51.23% 4 25.00%
Max Filippov 57 28.08% 5 31.25%
Baoquan He 28 13.79% 1 6.25%
Adrian Bunk 5 2.46% 2 12.50%
Dan J Williams 4 1.97% 1 6.25%
Paul Gortmaker 3 1.48% 1 6.25%
Mike Rapoport 1 0.49% 1 6.25%
Christoph Hellwig 1 0.49% 1 6.25%
Total 203 16


/*
 * include/asm-xtensa/io.h
 *
 * This file is subject to the terms and conditions of the GNU General Public
 * License.  See the file "COPYING" in the main directory of this archive
 * for more details.
 *
 * Copyright (C) 2001 - 2005 Tensilica Inc.
 */

#ifndef _XTENSA_IO_H
#define _XTENSA_IO_H

#include <asm/byteorder.h>
#include <asm/page.h>
#include <asm/vectors.h>
#include <linux/bug.h>
#include <linux/kernel.h>
#include <linux/pgtable.h>

#include <linux/types.h>

#define IOADDR(x)		(XCHAL_KIO_BYPASS_VADDR + (x))
#define IO_SPACE_LIMIT ~0
#define PCI_IOBASE		((void __iomem *)XCHAL_KIO_BYPASS_VADDR)

#ifdef CONFIG_MMU
/*
 * I/O memory mapping functions.
 */
void __iomem *ioremap_prot(phys_addr_t phys_addr, size_t size,
			   unsigned long prot);
#define ioremap_prot ioremap_prot
#define iounmap iounmap

static inline void __iomem *ioremap(unsigned long offset, unsigned long size)
{
	if (offset >= XCHAL_KIO_PADDR
	    && offset - XCHAL_KIO_PADDR < XCHAL_KIO_SIZE)
		return (void*)(offset-XCHAL_KIO_PADDR+XCHAL_KIO_BYPASS_VADDR);
	else
		return ioremap_prot(offset, size,
			pgprot_val(pgprot_noncached(PAGE_KERNEL)));
}
#define ioremap ioremap

static inline void __iomem *ioremap_cache(unsigned long offset,
		unsigned long size)
{
	if (offset >= XCHAL_KIO_PADDR
	    && offset - XCHAL_KIO_PADDR < XCHAL_KIO_SIZE)
		return (void*)(offset-XCHAL_KIO_PADDR+XCHAL_KIO_CACHED_VADDR);
	else
		return ioremap_prot(offset, size, pgprot_val(PAGE_KERNEL));

}
#define ioremap_cache ioremap_cache
#endif /* CONFIG_MMU */

#include <asm-generic/io.h>

#endif	/* _XTENSA_IO_H */