m68knommu: fix DMA support for ColdFire
authorGreg Ungerer <gerg@uclinux.org>
Fri, 17 Apr 2009 13:11:38 +0000 (23:11 +1000)
committerGreg Ungerer <gerg@uclinux.org>
Wed, 22 Apr 2009 04:45:08 +0000 (14:45 +1000)
ColdFire CPU family members support DMA (all those with the FEC ethernet
core use it, the rest have dedicated DMA engines). The code support is
just missing a handful of routines for it to be usable by drivers.
Add the missing dma_ functions.

Signed-off-by: Greg Ungerer <gerg@uclinux.org>
arch/m68knommu/Kconfig
arch/m68knommu/kernel/dma.c

index 4beb59dfc6ec58328961a63c3c5c6a29fcc1bbb7..534376299a99e2b47ebde39e17e619e723bf7461 100644 (file)
@@ -16,6 +16,7 @@ config MMU
 
 config NO_DMA
        bool
+       depends on !COLDFIRE
        default y
 
 config FPU
index 93612580663851055d058e89321c791d6dfe4765..aaf38bbbb6cdff29e9630bd35f757cf1ec85cf0f 100644 (file)
@@ -7,10 +7,9 @@
 
 #include <linux/types.h>
 #include <linux/mm.h>
-#include <linux/string.h>
 #include <linux/device.h>
 #include <linux/dma-mapping.h>
-#include <asm/io.h>
+#include <asm/cacheflush.h>
 
 void *dma_alloc_coherent(struct device *dev, size_t size,
                           dma_addr_t *dma_handle, gfp_t gfp)
@@ -36,7 +35,39 @@ void dma_free_coherent(struct device *dev, size_t size,
        free_pages((unsigned long)vaddr, get_order(size));
 }
 
-void dma_sync_single_for_cpu(struct device *dev, dma_addr_t handle, size_t size, enum dma_data_direction dir)
+void dma_sync_single_for_device(struct device *dev, dma_addr_t handle,
+                               size_t size, enum dma_data_direction dir)
 {
+       switch (dir) {
+       case DMA_TO_DEVICE:
+               flush_dcache_range(handle, size);
+               break;
+       case DMA_FROM_DEVICE:
+               /* Should be clear already */
+               break;
+       default:
+               if (printk_ratelimit())
+                       printk("dma_sync_single_for_device: unsupported dir %u\n", dir);
+               break;
+       }
+}
+
+EXPORT_SYMBOL(dma_sync_single_for_device);
+dma_addr_t dma_map_single(struct device *dev, void *addr, size_t size,
+                         enum dma_data_direction dir)
+{
+       dma_addr_t handle = virt_to_phys(addr);
+       flush_dcache_range(handle, size);
+       return handle;
 }
+EXPORT_SYMBOL(dma_map_single);
 
+dma_addr_t dma_map_page(struct device *dev, struct page *page,
+                       unsigned long offset, size_t size,
+                       enum dma_data_direction dir)
+{
+       dma_addr_t handle = page_to_phys(page) + offset;
+       dma_sync_single_for_device(dev, handle, size, dir);
+       return handle;
+}
+EXPORT_SYMBOL(dma_map_page);