2024-09-09 08:52:07 +00:00
|
|
|
/*
|
|
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
|
|
* for more details.
|
|
|
|
*
|
|
|
|
* Copyright (C) 2006, 07 Ralf Baechle <ralf@linux-mips.org>
|
|
|
|
* Copyright (C) 2007 Lemote, Inc. & Institute of Computing Technology
|
|
|
|
* Author: Fuxin Zhang, zhangfx@lemote.com
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
#ifndef __ASM_MACH_LOONGSON_DMA_COHERENCE_H
|
|
|
|
#define __ASM_MACH_LOONGSON_DMA_COHERENCE_H
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
#ifdef CONFIG_SWIOTLB
|
|
|
|
#include <linux/swiotlb.h>
|
|
|
|
#endif
|
|
|
|
|
2024-09-09 08:52:07 +00:00
|
|
|
struct device;
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
extern dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr);
|
|
|
|
extern phys_addr_t dma_to_phys(struct device *dev, dma_addr_t daddr);
|
2024-09-09 08:52:07 +00:00
|
|
|
static inline dma_addr_t plat_map_dma_mem(struct device *dev, void *addr,
|
|
|
|
size_t size)
|
|
|
|
{
|
2024-09-09 08:57:42 +00:00
|
|
|
#ifdef CONFIG_CPU_LOONGSON3
|
|
|
|
return virt_to_phys(addr);
|
|
|
|
#else
|
2024-09-09 08:52:07 +00:00
|
|
|
return virt_to_phys(addr) | 0x80000000;
|
2024-09-09 08:57:42 +00:00
|
|
|
#endif
|
2024-09-09 08:52:07 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline dma_addr_t plat_map_dma_mem_page(struct device *dev,
|
|
|
|
struct page *page)
|
|
|
|
{
|
2024-09-09 08:57:42 +00:00
|
|
|
#ifdef CONFIG_CPU_LOONGSON3
|
|
|
|
return page_to_phys(page);
|
|
|
|
#else
|
2024-09-09 08:52:07 +00:00
|
|
|
return page_to_phys(page) | 0x80000000;
|
2024-09-09 08:57:42 +00:00
|
|
|
#endif
|
2024-09-09 08:52:07 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline unsigned long plat_dma_addr_to_phys(struct device *dev,
|
|
|
|
dma_addr_t dma_addr)
|
|
|
|
{
|
2024-09-09 08:57:42 +00:00
|
|
|
#if defined(CONFIG_CPU_LOONGSON3) && defined(CONFIG_64BIT)
|
|
|
|
return dma_addr;
|
|
|
|
#elif defined(CONFIG_CPU_LOONGSON2F) && defined(CONFIG_64BIT)
|
2024-09-09 08:52:07 +00:00
|
|
|
return (dma_addr > 0x8fffffff) ? dma_addr : (dma_addr & 0x0fffffff);
|
|
|
|
#else
|
|
|
|
return dma_addr & 0x7fffffff;
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void plat_unmap_dma_mem(struct device *dev, dma_addr_t dma_addr,
|
|
|
|
size_t size, enum dma_data_direction direction)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int plat_dma_supported(struct device *dev, u64 mask)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* we fall back to GFP_DMA when the mask isn't all 1s,
|
|
|
|
* so we can't guarantee allocations that must be
|
|
|
|
* within a tighter range than GFP_DMA..
|
|
|
|
*/
|
|
|
|
if (mask < DMA_BIT_MASK(24))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int plat_device_is_coherent(struct device *dev)
|
|
|
|
{
|
2024-09-09 08:57:42 +00:00
|
|
|
#ifdef CONFIG_DMA_NONCOHERENT
|
2024-09-09 08:52:07 +00:00
|
|
|
return 0;
|
2024-09-09 08:57:42 +00:00
|
|
|
#else
|
|
|
|
return 1;
|
|
|
|
#endif /* CONFIG_DMA_NONCOHERENT */
|
2024-09-09 08:52:07 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#endif /* __ASM_MACH_LOONGSON_DMA_COHERENCE_H */
|