dma40: fix DMA API usage for LCLA
authorRabin Vincent <rabin.vincent@stericsson.com>
Tue, 25 Jan 2011 10:18:14 +0000 (11:18 +0100)
committerDan Williams <dan.j.williams@intel.com>
Mon, 31 Jan 2011 06:27:17 +0000 (22:27 -0800)
Map the buffer once and use dma_sync*() appropriately instead of mapping the
buffer over and over without unmapping it.

Acked-by: Per Forlin <per.forlin@stericsson.com>
Acked-by: Jonas Aaberg <jonas.aberg@stericsson.com>
Signed-off-by: Rabin Vincent <rabin.vincent@stericsson.com>
Signed-off-by: Linus Walleij <linus.walleij@stericsson.com>
Signed-off-by: Dan Williams <dan.j.williams@intel.com>
drivers/dma/ste_dma40.c

index d32a9ac86084a4203e522e2bb0a7d4b44245f4ef..f08e5c49c5d2b072ecd350b5def8df963fb3a34e 100644 (file)
@@ -128,6 +128,7 @@ struct d40_desc {
  */
 struct d40_lcla_pool {
        void            *base;
+       dma_addr_t      dma_addr;
        void            *base_unaligned;
        int              pages;
        spinlock_t       lock;
@@ -504,25 +505,25 @@ static void d40_desc_load(struct d40_chan *d40c, struct d40_desc *d40d)
 
                d40d->lli_current++;
                for (; d40d->lli_current < d40d->lli_len; d40d->lli_current++) {
-                       struct d40_log_lli *lcla;
+                       unsigned int lcla_offset = d40c->phy_chan->num * 1024 +
+                                                  8 * curr_lcla * 2;
+                       struct d40_lcla_pool *pool = &d40c->base->lcla_pool;
+                       struct d40_log_lli *lcla = pool->base + lcla_offset;
 
                        if (d40d->lli_current + 1 < d40d->lli_len)
                                next_lcla = d40_lcla_alloc_one(d40c, d40d);
                        else
                                next_lcla = -EINVAL;
 
-                       lcla = d40c->base->lcla_pool.base +
-                               d40c->phy_chan->num * 1024 +
-                               8 * curr_lcla * 2;
-
                        d40_log_lli_lcla_write(lcla,
                                               &d40d->lli_log.dst[d40d->lli_current],
                                               &d40d->lli_log.src[d40d->lli_current],
                                               next_lcla);
 
-                       (void) dma_map_single(d40c->base->dev, lcla,
-                                             2 * sizeof(struct d40_log_lli),
-                                             DMA_TO_DEVICE);
+                       dma_sync_single_range_for_device(d40c->base->dev,
+                                               pool->dma_addr, lcla_offset,
+                                               2 * sizeof(struct d40_log_lli),
+                                               DMA_TO_DEVICE);
 
                        curr_lcla = next_lcla;
 
@@ -2771,6 +2772,7 @@ static void __init d40_hw_init(struct d40_base *base)
 
 static int __init d40_lcla_allocate(struct d40_base *base)
 {
+       struct d40_lcla_pool *pool = &base->lcla_pool;
        unsigned long *page_list;
        int i, j;
        int ret = 0;
@@ -2835,6 +2837,15 @@ static int __init d40_lcla_allocate(struct d40_base *base)
                                                 LCLA_ALIGNMENT);
        }
 
+       pool->dma_addr = dma_map_single(base->dev, pool->base,
+                                       SZ_1K * base->num_phy_chans,
+                                       DMA_TO_DEVICE);
+       if (dma_mapping_error(base->dev, pool->dma_addr)) {
+               pool->dma_addr = 0;
+               ret = -ENOMEM;
+               goto failure;
+       }
+
        writel(virt_to_phys(base->lcla_pool.base),
               base->virtbase + D40_DREG_LCLA);
 failure:
@@ -2929,6 +2940,12 @@ failure:
                        kmem_cache_destroy(base->desc_slab);
                if (base->virtbase)
                        iounmap(base->virtbase);
+
+               if (base->lcla_pool.dma_addr)
+                       dma_unmap_single(base->dev, base->lcla_pool.dma_addr,
+                                        SZ_1K * base->num_phy_chans,
+                                        DMA_TO_DEVICE);
+
                if (!base->lcla_pool.base_unaligned && base->lcla_pool.base)
                        free_pages((unsigned long)base->lcla_pool.base,
                                   base->lcla_pool.pages);