dma40: combine desc init functions
[platform/adaptation/renesas_rcar/renesas_kernel.git] / drivers / dma / ste_dma40.c
index b585686..5259a98 100644 (file)
@@ -357,14 +357,12 @@ static int d40_pool_lli_alloc(struct d40_chan *d40c, struct d40_desc *d40d,
        }
 
        if (is_log) {
-               d40d->lli_log.src = PTR_ALIGN((struct d40_log_lli *) base,
-                                             align);
+               d40d->lli_log.src = PTR_ALIGN(base, align);
                d40d->lli_log.dst = d40d->lli_log.src + lli_len;
 
                d40d->lli_pool.dma_addr = 0;
        } else {
-               d40d->lli_phy.src = PTR_ALIGN((struct d40_phy_lli *)base,
-                                             align);
+               d40d->lli_phy.src = PTR_ALIGN(base, align);
                d40d->lli_phy.dst = d40d->lli_phy.src + lli_len;
 
                d40d->lli_pool.dma_addr = dma_map_single(d40c->base->dev,
@@ -1619,6 +1617,35 @@ static u32 stedma40_residue(struct dma_chan *chan)
        return bytes_left;
 }
 
+static struct d40_desc *
+d40_prep_desc(struct d40_chan *chan, struct scatterlist *sg,
+             unsigned int sg_len, unsigned long dma_flags)
+{
+       struct stedma40_chan_cfg *cfg = &chan->dma_cfg;
+       struct d40_desc *desc;
+
+       desc = d40_desc_get(chan);
+       if (!desc)
+               return NULL;
+
+       desc->lli_len = d40_sg_2_dmalen(sg, sg_len, cfg->src_info.data_width,
+                                       cfg->dst_info.data_width);
+       if (desc->lli_len < 0) {
+               chan_err(chan, "Unaligned size\n");
+               d40_desc_free(chan, desc);
+
+               return NULL;
+       }
+
+       desc->lli_current = 0;
+       desc->txd.flags = dma_flags;
+       desc->txd.tx_submit = d40_tx_submit;
+
+       dma_async_tx_descriptor_init(&desc->txd, &chan->chan);
+
+       return desc;
+}
+
 struct dma_async_tx_descriptor *stedma40_memcpy_sg(struct dma_chan *chan,
                                                   struct scatterlist *sgl_dst,
                                                   struct scatterlist *sgl_src,
@@ -1637,21 +1664,10 @@ struct dma_async_tx_descriptor *stedma40_memcpy_sg(struct dma_chan *chan,
        }
 
        spin_lock_irqsave(&d40c->lock, flags);
-       d40d = d40_desc_get(d40c);
-
-       if (d40d == NULL)
-               goto err;
 
-       d40d->lli_len = d40_sg_2_dmalen(sgl_dst, sgl_len,
-                                       d40c->dma_cfg.src_info.data_width,
-                                       d40c->dma_cfg.dst_info.data_width);
-       if (d40d->lli_len < 0) {
-               chan_err(d40c, "Unaligned size\n");
+       d40d = d40_prep_desc(d40c, sgl_dst, sgl_len, dma_flags);
+       if (!d40d)
                goto err;
-       }
-
-       d40d->lli_current = 0;
-       d40d->txd.flags = dma_flags;
 
        if (chan_is_logical(d40c)) {
 
@@ -1710,10 +1726,6 @@ struct dma_async_tx_descriptor *stedma40_memcpy_sg(struct dma_chan *chan,
                                           d40d->lli_pool.size, DMA_TO_DEVICE);
        }
 
-       dma_async_tx_descriptor_init(&d40d->txd, chan);
-
-       d40d->txd.tx_submit = d40_tx_submit;
-
        spin_unlock_irqrestore(&d40c->lock, flags);
 
        return &d40d->txd;
@@ -1865,108 +1877,19 @@ static struct dma_async_tx_descriptor *d40_prep_memcpy(struct dma_chan *chan,
                                                       size_t size,
                                                       unsigned long dma_flags)
 {
-       struct d40_desc *d40d;
-       struct d40_chan *d40c = container_of(chan, struct d40_chan,
-                                            chan);
-       unsigned long flags;
-
-       if (d40c->phy_chan == NULL) {
-               chan_err(d40c, "Channel is not allocated.\n");
-               return ERR_PTR(-EINVAL);
-       }
+       struct scatterlist dst_sg;
+       struct scatterlist src_sg;
 
-       spin_lock_irqsave(&d40c->lock, flags);
-       d40d = d40_desc_get(d40c);
-
-       if (d40d == NULL) {
-               chan_err(d40c, "Descriptor is NULL\n");
-               goto err;
-       }
+       sg_init_table(&dst_sg, 1);
+       sg_init_table(&src_sg, 1);
 
-       d40d->txd.flags = dma_flags;
-       d40d->lli_len = d40_size_2_dmalen(size,
-                                         d40c->dma_cfg.src_info.data_width,
-                                         d40c->dma_cfg.dst_info.data_width);
-       if (d40d->lli_len < 0) {
-               chan_err(d40c, "Unaligned size\n");
-               goto err;
-       }
+       sg_dma_address(&dst_sg) = dst;
+       sg_dma_address(&src_sg) = src;
 
+       sg_dma_len(&dst_sg) = size;
+       sg_dma_len(&src_sg) = size;
 
-       dma_async_tx_descriptor_init(&d40d->txd, chan);
-
-       d40d->txd.tx_submit = d40_tx_submit;
-
-       if (chan_is_logical(d40c)) {
-
-               if (d40_pool_lli_alloc(d40c,d40d, d40d->lli_len, true) < 0) {
-                       chan_err(d40c, "Out of memory\n");
-                       goto err;
-               }
-               d40d->lli_current = 0;
-
-               if (d40_log_buf_to_lli(d40d->lli_log.src,
-                                      src,
-                                      size,
-                                      d40c->log_def.lcsp1,
-                                      d40c->dma_cfg.src_info.data_width,
-                                      d40c->dma_cfg.dst_info.data_width,
-                                      true) == NULL)
-                       goto err;
-
-               if (d40_log_buf_to_lli(d40d->lli_log.dst,
-                                      dst,
-                                      size,
-                                      d40c->log_def.lcsp3,
-                                      d40c->dma_cfg.dst_info.data_width,
-                                      d40c->dma_cfg.src_info.data_width,
-                                      true) == NULL)
-                       goto err;
-
-       } else {
-
-               if (d40_pool_lli_alloc(d40c, d40d, d40d->lli_len, false) < 0) {
-                       chan_err(d40c, "Out of memory\n");
-                       goto err;
-               }
-
-               if (d40_phy_buf_to_lli(d40d->lli_phy.src,
-                                      src,
-                                      size,
-                                      d40c->dma_cfg.src_info.psize,
-                                      0,
-                                      d40c->src_def_cfg,
-                                      true,
-                                      d40c->dma_cfg.src_info.data_width,
-                                      d40c->dma_cfg.dst_info.data_width,
-                                      false) == NULL)
-                       goto err;
-
-               if (d40_phy_buf_to_lli(d40d->lli_phy.dst,
-                                      dst,
-                                      size,
-                                      d40c->dma_cfg.dst_info.psize,
-                                      0,
-                                      d40c->dst_def_cfg,
-                                      true,
-                                      d40c->dma_cfg.dst_info.data_width,
-                                      d40c->dma_cfg.src_info.data_width,
-                                      false) == NULL)
-                       goto err;
-
-               dma_sync_single_for_device(d40c->base->dev,
-                                          d40d->lli_pool.dma_addr,
-                                          d40d->lli_pool.size, DMA_TO_DEVICE);
-       }
-
-       spin_unlock_irqrestore(&d40c->lock, flags);
-       return &d40d->txd;
-
-err:
-       if (d40d)
-               d40_desc_free(d40c, d40d);
-       spin_unlock_irqrestore(&d40c->lock, flags);
-       return NULL;
+       return stedma40_memcpy_sg(chan, &dst_sg, &src_sg, 1, dma_flags);
 }
 
 static struct dma_async_tx_descriptor *
@@ -1991,21 +1914,11 @@ static int d40_prep_slave_sg_log(struct d40_desc *d40d,
        dma_addr_t dev_addr = 0;
        int total_size;
 
-       d40d->lli_len = d40_sg_2_dmalen(sgl, sg_len,
-                                       d40c->dma_cfg.src_info.data_width,
-                                       d40c->dma_cfg.dst_info.data_width);
-       if (d40d->lli_len < 0) {
-               chan_err(d40c, "Unaligned size\n");
-               return -EINVAL;
-       }
-
        if (d40_pool_lli_alloc(d40c, d40d, d40d->lli_len, true) < 0) {
                chan_err(d40c, "Out of memory\n");
                return -ENOMEM;
        }
 
-       d40d->lli_current = 0;
-
        if (direction == DMA_FROM_DEVICE)
                if (d40c->runtime_addr)
                        dev_addr = d40c->runtime_addr;
@@ -2045,21 +1958,11 @@ static int d40_prep_slave_sg_phy(struct d40_desc *d40d,
        dma_addr_t dst_dev_addr;
        int res;
 
-       d40d->lli_len = d40_sg_2_dmalen(sgl, sgl_len,
-                                       d40c->dma_cfg.src_info.data_width,
-                                       d40c->dma_cfg.dst_info.data_width);
-       if (d40d->lli_len < 0) {
-               chan_err(d40c, "Unaligned size\n");
-               return -EINVAL;
-       }
-
        if (d40_pool_lli_alloc(d40c, d40d, d40d->lli_len, false) < 0) {
                chan_err(d40c, "Out of memory\n");
                return -ENOMEM;
        }
 
-       d40d->lli_current = 0;
-
        if (direction == DMA_FROM_DEVICE) {
                dst_dev_addr = 0;
                if (d40c->runtime_addr)
@@ -2122,8 +2025,8 @@ static struct dma_async_tx_descriptor *d40_prep_slave_sg(struct dma_chan *chan,
        }
 
        spin_lock_irqsave(&d40c->lock, flags);
-       d40d = d40_desc_get(d40c);
 
+       d40d = d40_prep_desc(d40c, sgl, sg_len, dma_flags);
        if (d40d == NULL)
                goto err;
 
@@ -2139,12 +2042,6 @@ static struct dma_async_tx_descriptor *d40_prep_slave_sg(struct dma_chan *chan,
                goto err;
        }
 
-       d40d->txd.flags = dma_flags;
-
-       dma_async_tx_descriptor_init(&d40d->txd, chan);
-
-       d40d->txd.tx_submit = d40_tx_submit;
-
        spin_unlock_irqrestore(&d40c->lock, flags);
        return &d40d->txd;