static inline dma_addr_t dma_map_single(struct device *dev, void *cpu_addr,
size_t size, enum dma_data_direction dir)
{
+ BUG_ON(!valid_dma_direction(dir));
+
if (!arch_is_coherent())
dma_cache_maint(cpu_addr, size, dir);
static inline dma_addr_t dma_map_page(struct device *dev, struct page *page,
unsigned long offset, size_t size, enum dma_data_direction dir)
{
+ BUG_ON(!valid_dma_direction(dir));
+
if (!arch_is_coherent())
dma_cache_maint(page_address(page) + offset, size, dir);
dma_addr_t handle, unsigned long offset, size_t size,
enum dma_data_direction dir)
{
+ BUG_ON(!valid_dma_direction(dir));
+
if (!dmabounce_sync_for_cpu(dev, handle, offset, size, dir))
return;
dma_addr_t handle, unsigned long offset, size_t size,
enum dma_data_direction dir)
{
+ BUG_ON(!valid_dma_direction(dir));
+
if (!dmabounce_sync_for_device(dev, handle, offset, size, dir))
return;