2024-09-09 08:52:07 +00:00
|
|
|
/*
|
|
|
|
* Copyright 2010 Tilera Corporation. All Rights Reserved.
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU General Public License
|
|
|
|
* as published by the Free Software Foundation, version 2.
|
|
|
|
*
|
|
|
|
* This program is distributed in the hope that it will be useful, but
|
|
|
|
* WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
|
|
|
|
* NON INFRINGEMENT. See the GNU General Public License for
|
|
|
|
* more details.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef _ASM_TILE_IO_H
|
|
|
|
#define _ASM_TILE_IO_H
|
|
|
|
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <linux/bug.h>
|
|
|
|
#include <asm/page.h>
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
/* Maximum PCI I/O space address supported. */
|
|
|
|
#define IO_SPACE_LIMIT 0xffffffff
|
2024-09-09 08:52:07 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Convert a physical pointer to a virtual kernel pointer for /dev/mem
|
|
|
|
* access.
|
|
|
|
*/
|
|
|
|
#define xlate_dev_mem_ptr(p) __va(p)
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Convert a virtual cached pointer to an uncached pointer.
|
|
|
|
*/
|
|
|
|
#define xlate_dev_kmem_ptr(p) p
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Change "struct page" to physical address.
|
|
|
|
*/
|
|
|
|
#define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT)
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Some places try to pass in an loff_t for PHYSADDR (?!), so we cast it to
|
|
|
|
* long before casting it to a pointer to avoid compiler warnings.
|
|
|
|
*/
|
|
|
|
#if CHIP_HAS_MMIO()
|
|
|
|
extern void __iomem *ioremap(resource_size_t offset, unsigned long size);
|
|
|
|
extern void __iomem *ioremap_prot(resource_size_t offset, unsigned long size,
|
|
|
|
pgprot_t pgprot);
|
|
|
|
extern void iounmap(volatile void __iomem *addr);
|
|
|
|
#else
|
|
|
|
#define ioremap(physaddr, size) ((void __iomem *)(unsigned long)(physaddr))
|
|
|
|
#define iounmap(addr) ((void)0)
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#define ioremap_nocache(physaddr, size) ioremap(physaddr, size)
|
|
|
|
#define ioremap_wc(physaddr, size) ioremap(physaddr, size)
|
|
|
|
#define ioremap_writethrough(physaddr, size) ioremap(physaddr, size)
|
|
|
|
#define ioremap_fullcache(physaddr, size) ioremap(physaddr, size)
|
|
|
|
|
|
|
|
#define mmiowb()
|
|
|
|
|
|
|
|
/* Conversion between virtual and physical mappings. */
|
|
|
|
#define mm_ptov(addr) ((void *)phys_to_virt(addr))
|
|
|
|
#define mm_vtop(addr) ((unsigned long)virt_to_phys(addr))
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
#if CHIP_HAS_MMIO()
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We use inline assembly to guarantee that the compiler does not
|
|
|
|
* split an access into multiple byte-sized accesses as it might
|
|
|
|
* sometimes do if a register data structure is marked "packed".
|
|
|
|
* Obviously on tile we can't tolerate such an access being
|
|
|
|
* actually unaligned, but we want to avoid the case where the
|
|
|
|
* compiler conservatively would generate multiple accesses even
|
|
|
|
* for an aligned read or write.
|
|
|
|
*/
|
|
|
|
|
|
|
|
static inline u8 __raw_readb(const volatile void __iomem *addr)
|
|
|
|
{
|
|
|
|
return *(const volatile u8 __force *)addr;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u16 __raw_readw(const volatile void __iomem *addr)
|
|
|
|
{
|
|
|
|
u16 ret;
|
|
|
|
asm volatile("ld2u %0, %1" : "=r" (ret) : "r" (addr));
|
|
|
|
barrier();
|
|
|
|
return le16_to_cpu(ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 __raw_readl(const volatile void __iomem *addr)
|
|
|
|
{
|
|
|
|
u32 ret;
|
|
|
|
/* Sign-extend to conform to u32 ABI sign-extension convention. */
|
|
|
|
asm volatile("ld4s %0, %1" : "=r" (ret) : "r" (addr));
|
|
|
|
barrier();
|
|
|
|
return le32_to_cpu(ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u64 __raw_readq(const volatile void __iomem *addr)
|
|
|
|
{
|
|
|
|
u64 ret;
|
|
|
|
asm volatile("ld %0, %1" : "=r" (ret) : "r" (addr));
|
|
|
|
barrier();
|
|
|
|
return le64_to_cpu(ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void __raw_writeb(u8 val, volatile void __iomem *addr)
|
|
|
|
{
|
|
|
|
*(volatile u8 __force *)addr = val;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void __raw_writew(u16 val, volatile void __iomem *addr)
|
|
|
|
{
|
|
|
|
asm volatile("st2 %0, %1" :: "r" (addr), "r" (cpu_to_le16(val)));
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void __raw_writel(u32 val, volatile void __iomem *addr)
|
|
|
|
{
|
|
|
|
asm volatile("st4 %0, %1" :: "r" (addr), "r" (cpu_to_le32(val)));
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void __raw_writeq(u64 val, volatile void __iomem *addr)
|
|
|
|
{
|
|
|
|
asm volatile("st %0, %1" :: "r" (addr), "r" (cpu_to_le64(val)));
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The on-chip I/O hardware on tilegx is configured with VA=PA for the
|
|
|
|
* kernel's PA range. The low-level APIs and field names use "va" and
|
|
|
|
* "void *" nomenclature, to be consistent with the general notion
|
|
|
|
* that the addresses in question are virtualizable, but in the kernel
|
|
|
|
* context we are actually manipulating PA values. (In other contexts,
|
|
|
|
* e.g. access from user space, we do in fact use real virtual addresses
|
|
|
|
* in the va fields.) To allow readers of the code to understand what's
|
|
|
|
* happening, we direct their attention to this comment by using the
|
|
|
|
* following two functions that just duplicate __va() and __pa().
|
|
|
|
*/
|
|
|
|
typedef unsigned long tile_io_addr_t;
|
|
|
|
static inline tile_io_addr_t va_to_tile_io_addr(void *va)
|
|
|
|
{
|
|
|
|
BUILD_BUG_ON(sizeof(phys_addr_t) != sizeof(tile_io_addr_t));
|
|
|
|
return __pa(va);
|
|
|
|
}
|
|
|
|
static inline void *tile_io_addr_to_va(tile_io_addr_t tile_io_addr)
|
|
|
|
{
|
|
|
|
return __va(tile_io_addr);
|
|
|
|
}
|
|
|
|
|
|
|
|
#else /* CHIP_HAS_MMIO() */
|
|
|
|
|
2024-09-09 08:52:07 +00:00
|
|
|
#ifdef CONFIG_PCI
|
|
|
|
|
|
|
|
extern u8 _tile_readb(unsigned long addr);
|
|
|
|
extern u16 _tile_readw(unsigned long addr);
|
|
|
|
extern u32 _tile_readl(unsigned long addr);
|
|
|
|
extern u64 _tile_readq(unsigned long addr);
|
|
|
|
extern void _tile_writeb(u8 val, unsigned long addr);
|
|
|
|
extern void _tile_writew(u16 val, unsigned long addr);
|
|
|
|
extern void _tile_writel(u32 val, unsigned long addr);
|
|
|
|
extern void _tile_writeq(u64 val, unsigned long addr);
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
#define __raw_readb(addr) _tile_readb((unsigned long)addr)
|
|
|
|
#define __raw_readw(addr) _tile_readw((unsigned long)addr)
|
|
|
|
#define __raw_readl(addr) _tile_readl((unsigned long)addr)
|
|
|
|
#define __raw_readq(addr) _tile_readq((unsigned long)addr)
|
|
|
|
#define __raw_writeb(val, addr) _tile_writeb(val, (unsigned long)addr)
|
|
|
|
#define __raw_writew(val, addr) _tile_writew(val, (unsigned long)addr)
|
|
|
|
#define __raw_writel(val, addr) _tile_writel(val, (unsigned long)addr)
|
|
|
|
#define __raw_writeq(val, addr) _tile_writeq(val, (unsigned long)addr)
|
|
|
|
|
|
|
|
#else /* CONFIG_PCI */
|
2024-09-09 08:52:07 +00:00
|
|
|
|
|
|
|
/*
|
2024-09-09 08:57:42 +00:00
|
|
|
* The tilepro architecture does not support IOMEM unless PCI is enabled.
|
2024-09-09 08:52:07 +00:00
|
|
|
* Unfortunately we can't yet simply not declare these methods,
|
|
|
|
* since some generic code that compiles into the kernel, but
|
|
|
|
* we never run, uses them unconditionally.
|
|
|
|
*/
|
|
|
|
|
|
|
|
static inline int iomem_panic(void)
|
|
|
|
{
|
|
|
|
panic("readb/writeb and friends do not exist on tile without PCI");
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
static inline u8 readb(unsigned long addr)
|
2024-09-09 08:52:07 +00:00
|
|
|
{
|
|
|
|
return iomem_panic();
|
|
|
|
}
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
static inline u16 _readw(unsigned long addr)
|
2024-09-09 08:52:07 +00:00
|
|
|
{
|
|
|
|
return iomem_panic();
|
|
|
|
}
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
static inline u32 readl(unsigned long addr)
|
2024-09-09 08:52:07 +00:00
|
|
|
{
|
|
|
|
return iomem_panic();
|
|
|
|
}
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
static inline u64 readq(unsigned long addr)
|
2024-09-09 08:52:07 +00:00
|
|
|
{
|
|
|
|
return iomem_panic();
|
|
|
|
}
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
static inline void writeb(u8 val, unsigned long addr)
|
2024-09-09 08:52:07 +00:00
|
|
|
{
|
|
|
|
iomem_panic();
|
|
|
|
}
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
static inline void writew(u16 val, unsigned long addr)
|
2024-09-09 08:52:07 +00:00
|
|
|
{
|
|
|
|
iomem_panic();
|
|
|
|
}
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
static inline void writel(u32 val, unsigned long addr)
|
2024-09-09 08:52:07 +00:00
|
|
|
{
|
|
|
|
iomem_panic();
|
|
|
|
}
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
static inline void writeq(u64 val, unsigned long addr)
|
2024-09-09 08:52:07 +00:00
|
|
|
{
|
|
|
|
iomem_panic();
|
|
|
|
}
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
#endif /* CONFIG_PCI */
|
2024-09-09 08:52:07 +00:00
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
#endif /* CHIP_HAS_MMIO() */
|
|
|
|
|
|
|
|
#define readb __raw_readb
|
|
|
|
#define readw __raw_readw
|
|
|
|
#define readl __raw_readl
|
|
|
|
#define readq __raw_readq
|
|
|
|
#define writeb __raw_writeb
|
|
|
|
#define writew __raw_writew
|
|
|
|
#define writel __raw_writel
|
|
|
|
#define writeq __raw_writeq
|
2024-09-09 08:52:07 +00:00
|
|
|
|
|
|
|
#define readb_relaxed readb
|
|
|
|
#define readw_relaxed readw
|
|
|
|
#define readl_relaxed readl
|
|
|
|
#define readq_relaxed readq
|
|
|
|
|
|
|
|
#define ioread8 readb
|
|
|
|
#define ioread16 readw
|
|
|
|
#define ioread32 readl
|
|
|
|
#define ioread64 readq
|
|
|
|
#define iowrite8 writeb
|
|
|
|
#define iowrite16 writew
|
|
|
|
#define iowrite32 writel
|
|
|
|
#define iowrite64 writeq
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
#if CHIP_HAS_MMIO() || defined(CONFIG_PCI)
|
|
|
|
|
|
|
|
static inline void memset_io(volatile void *dst, int val, size_t len)
|
2024-09-09 08:52:07 +00:00
|
|
|
{
|
2024-09-09 08:57:42 +00:00
|
|
|
size_t x;
|
2024-09-09 08:52:07 +00:00
|
|
|
BUG_ON((unsigned long)dst & 0x3);
|
|
|
|
val = (val & 0xff) * 0x01010101;
|
|
|
|
for (x = 0; x < len; x += 4)
|
|
|
|
writel(val, dst + x);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void memcpy_fromio(void *dst, const volatile void __iomem *src,
|
|
|
|
size_t len)
|
|
|
|
{
|
2024-09-09 08:57:42 +00:00
|
|
|
size_t x;
|
2024-09-09 08:52:07 +00:00
|
|
|
BUG_ON((unsigned long)src & 0x3);
|
|
|
|
for (x = 0; x < len; x += 4)
|
|
|
|
*(u32 *)(dst + x) = readl(src + x);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void memcpy_toio(volatile void __iomem *dst, const void *src,
|
|
|
|
size_t len)
|
|
|
|
{
|
2024-09-09 08:57:42 +00:00
|
|
|
size_t x;
|
2024-09-09 08:52:07 +00:00
|
|
|
BUG_ON((unsigned long)dst & 0x3);
|
|
|
|
for (x = 0; x < len; x += 4)
|
|
|
|
writel(*(u32 *)(src + x), dst + x);
|
|
|
|
}
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
#if CHIP_HAS_MMIO() && defined(CONFIG_TILE_PCI_IO)
|
|
|
|
|
|
|
|
static inline u8 inb(unsigned long addr)
|
|
|
|
{
|
|
|
|
return readb((volatile void __iomem *) addr);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u16 inw(unsigned long addr)
|
|
|
|
{
|
|
|
|
return readw((volatile void __iomem *) addr);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 inl(unsigned long addr)
|
|
|
|
{
|
|
|
|
return readl((volatile void __iomem *) addr);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void outb(u8 b, unsigned long addr)
|
|
|
|
{
|
|
|
|
writeb(b, (volatile void __iomem *) addr);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void outw(u16 b, unsigned long addr)
|
|
|
|
{
|
|
|
|
writew(b, (volatile void __iomem *) addr);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void outl(u32 b, unsigned long addr)
|
|
|
|
{
|
|
|
|
writel(b, (volatile void __iomem *) addr);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void insb(unsigned long addr, void *buffer, int count)
|
|
|
|
{
|
|
|
|
if (count) {
|
|
|
|
u8 *buf = buffer;
|
|
|
|
do {
|
|
|
|
u8 x = inb(addr);
|
|
|
|
*buf++ = x;
|
|
|
|
} while (--count);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void insw(unsigned long addr, void *buffer, int count)
|
|
|
|
{
|
|
|
|
if (count) {
|
|
|
|
u16 *buf = buffer;
|
|
|
|
do {
|
|
|
|
u16 x = inw(addr);
|
|
|
|
*buf++ = x;
|
|
|
|
} while (--count);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void insl(unsigned long addr, void *buffer, int count)
|
|
|
|
{
|
|
|
|
if (count) {
|
|
|
|
u32 *buf = buffer;
|
|
|
|
do {
|
|
|
|
u32 x = inl(addr);
|
|
|
|
*buf++ = x;
|
|
|
|
} while (--count);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void outsb(unsigned long addr, const void *buffer, int count)
|
|
|
|
{
|
|
|
|
if (count) {
|
|
|
|
const u8 *buf = buffer;
|
|
|
|
do {
|
|
|
|
outb(*buf++, addr);
|
|
|
|
} while (--count);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void outsw(unsigned long addr, const void *buffer, int count)
|
|
|
|
{
|
|
|
|
if (count) {
|
|
|
|
const u16 *buf = buffer;
|
|
|
|
do {
|
|
|
|
outw(*buf++, addr);
|
|
|
|
} while (--count);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void outsl(unsigned long addr, const void *buffer, int count)
|
|
|
|
{
|
|
|
|
if (count) {
|
|
|
|
const u32 *buf = buffer;
|
|
|
|
do {
|
|
|
|
outl(*buf++, addr);
|
|
|
|
} while (--count);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
extern void __iomem *ioport_map(unsigned long port, unsigned int len);
|
|
|
|
extern void ioport_unmap(void __iomem *addr);
|
|
|
|
|
|
|
|
#else
|
|
|
|
|
2024-09-09 08:52:07 +00:00
|
|
|
/*
|
2024-09-09 08:57:42 +00:00
|
|
|
* The TilePro architecture does not support IOPORT, even with PCI.
|
2024-09-09 08:52:07 +00:00
|
|
|
* Unfortunately we can't yet simply not declare these methods,
|
|
|
|
* since some generic code that compiles into the kernel, but
|
|
|
|
* we never run, uses them unconditionally.
|
|
|
|
*/
|
|
|
|
|
|
|
|
static inline long ioport_panic(void)
|
|
|
|
{
|
2024-09-09 08:57:42 +00:00
|
|
|
#ifdef __tilegx__
|
|
|
|
panic("PCI IO space support is disabled. Configure the kernel with"
|
|
|
|
" CONFIG_TILE_PCI_IO to enable it");
|
|
|
|
#else
|
2024-09-09 08:52:07 +00:00
|
|
|
panic("inb/outb and friends do not exist on tile");
|
2024-09-09 08:57:42 +00:00
|
|
|
#endif
|
2024-09-09 08:52:07 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void __iomem *ioport_map(unsigned long port, unsigned int len)
|
|
|
|
{
|
|
|
|
pr_info("ioport_map: mapping IO resources is unsupported on tile.\n");
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void ioport_unmap(void __iomem *addr)
|
|
|
|
{
|
|
|
|
ioport_panic();
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u8 inb(unsigned long addr)
|
|
|
|
{
|
|
|
|
return ioport_panic();
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u16 inw(unsigned long addr)
|
|
|
|
{
|
|
|
|
return ioport_panic();
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 inl(unsigned long addr)
|
|
|
|
{
|
|
|
|
return ioport_panic();
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void outb(u8 b, unsigned long addr)
|
|
|
|
{
|
|
|
|
ioport_panic();
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void outw(u16 b, unsigned long addr)
|
|
|
|
{
|
|
|
|
ioport_panic();
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void outl(u32 b, unsigned long addr)
|
|
|
|
{
|
|
|
|
ioport_panic();
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void insb(unsigned long addr, void *buffer, int count)
|
|
|
|
{
|
|
|
|
ioport_panic();
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void insw(unsigned long addr, void *buffer, int count)
|
|
|
|
{
|
|
|
|
ioport_panic();
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void insl(unsigned long addr, void *buffer, int count)
|
|
|
|
{
|
|
|
|
ioport_panic();
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void outsb(unsigned long addr, const void *buffer, int count)
|
|
|
|
{
|
|
|
|
ioport_panic();
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void outsw(unsigned long addr, const void *buffer, int count)
|
|
|
|
{
|
|
|
|
ioport_panic();
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void outsl(unsigned long addr, const void *buffer, int count)
|
|
|
|
{
|
|
|
|
ioport_panic();
|
|
|
|
}
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
#endif /* CHIP_HAS_MMIO() && defined(CONFIG_TILE_PCI_IO) */
|
|
|
|
|
|
|
|
#define inb_p(addr) inb(addr)
|
|
|
|
#define inw_p(addr) inw(addr)
|
|
|
|
#define inl_p(addr) inl(addr)
|
|
|
|
#define outb_p(x, addr) outb((x), (addr))
|
|
|
|
#define outw_p(x, addr) outw((x), (addr))
|
|
|
|
#define outl_p(x, addr) outl((x), (addr))
|
|
|
|
|
2024-09-09 08:52:07 +00:00
|
|
|
#define ioread16be(addr) be16_to_cpu(ioread16(addr))
|
|
|
|
#define ioread32be(addr) be32_to_cpu(ioread32(addr))
|
|
|
|
#define iowrite16be(v, addr) iowrite16(be16_to_cpu(v), (addr))
|
|
|
|
#define iowrite32be(v, addr) iowrite32(be32_to_cpu(v), (addr))
|
|
|
|
|
|
|
|
#define ioread8_rep(p, dst, count) \
|
|
|
|
insb((unsigned long) (p), (dst), (count))
|
|
|
|
#define ioread16_rep(p, dst, count) \
|
|
|
|
insw((unsigned long) (p), (dst), (count))
|
|
|
|
#define ioread32_rep(p, dst, count) \
|
|
|
|
insl((unsigned long) (p), (dst), (count))
|
|
|
|
|
|
|
|
#define iowrite8_rep(p, src, count) \
|
|
|
|
outsb((unsigned long) (p), (src), (count))
|
|
|
|
#define iowrite16_rep(p, src, count) \
|
|
|
|
outsw((unsigned long) (p), (src), (count))
|
|
|
|
#define iowrite32_rep(p, src, count) \
|
|
|
|
outsl((unsigned long) (p), (src), (count))
|
|
|
|
|
|
|
|
#define virt_to_bus virt_to_phys
|
|
|
|
#define bus_to_virt phys_to_virt
|
|
|
|
|
|
|
|
#endif /* _ASM_TILE_IO_H */
|