Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit bc0fa814 authored by Dan Williams's avatar Dan Williams
Browse files

Merge branches 'amba' and 'dma40' into dmaengine

Loading
Loading
Loading
Loading
+8 −0
Original line number Diff line number Diff line
@@ -13,6 +13,14 @@
#include <linux/workqueue.h>
#include <linux/interrupt.h>

/*
 * Maxium size for a single dma descriptor
 * Size is limited to 16 bits.
 * Size is in the units of addr-widths (1,2,4,8 bytes)
 * Larger transfers will be split up to multiple linked desc
 */
#define STEDMA40_MAX_SEG_SIZE 0xFFFF

/* dev types for memcpy */
#define STEDMA40_DEV_DST_MEMORY (-1)
#define	STEDMA40_DEV_SRC_MEMORY (-1)
+490 −550

File changed.

Preview size limit exceeded, changes collapsed.

+152 −39
Original line number Diff line number Diff line
/*
 * Copyright (C) ST-Ericsson SA 2007-2010
 * Copyright (C) Ericsson AB 2007-2008
 * Copyright (C) ST-Ericsson SA 2008-2010
 * Author: Per Forlin <per.forlin@stericsson.com> for ST-Ericsson
 * Author: Jonas Aaberg <jonas.aberg@stericsson.com> for ST-Ericsson
 * License terms: GNU General Public License (GPL) version 2
@@ -554,8 +555,66 @@ static struct d40_desc *d40_last_queued(struct d40_chan *d40c)
	return d;
}

/* Support functions for logical channels */
static int d40_psize_2_burst_size(bool is_log, int psize)
{
	if (is_log) {
		if (psize == STEDMA40_PSIZE_LOG_1)
			return 1;
	} else {
		if (psize == STEDMA40_PSIZE_PHY_1)
			return 1;
	}

	return 2 << psize;
}

/*
 * The dma only supports transmitting packages up to
 * STEDMA40_MAX_SEG_SIZE << data_width. Calculate the total number of
 * dma elements required to send the entire sg list
 */
static int d40_size_2_dmalen(int size, u32 data_width1, u32 data_width2)
{
	int dmalen;
	u32 max_w = max(data_width1, data_width2);
	u32 min_w = min(data_width1, data_width2);
	u32 seg_max = ALIGN(STEDMA40_MAX_SEG_SIZE << min_w, 1 << max_w);

	if (seg_max > STEDMA40_MAX_SEG_SIZE)
		seg_max -= (1 << max_w);

	if (!IS_ALIGNED(size, 1 << max_w))
		return -EINVAL;

	if (size <= seg_max)
		dmalen = 1;
	else {
		dmalen = size / seg_max;
		if (dmalen * seg_max < size)
			dmalen++;
	}
	return dmalen;
}

static int d40_sg_2_dmalen(struct scatterlist *sgl, int sg_len,
			   u32 data_width1, u32 data_width2)
{
	struct scatterlist *sg;
	int i;
	int len = 0;
	int ret;

	for_each_sg(sgl, sg, sg_len, i) {
		ret = d40_size_2_dmalen(sg_dma_len(sg),
					data_width1, data_width2);
		if (ret < 0)
			return ret;
		len += ret;
	}
	return len;
}

/* Support functions for logical channels */

static int d40_channel_execute_command(struct d40_chan *d40c,
				       enum d40_command command)
@@ -1241,6 +1300,21 @@ static int d40_validate_conf(struct d40_chan *d40c,
		res = -EINVAL;
	}

	if (d40_psize_2_burst_size(is_log, conf->src_info.psize) *
	    (1 << conf->src_info.data_width) !=
	    d40_psize_2_burst_size(is_log, conf->dst_info.psize) *
	    (1 << conf->dst_info.data_width)) {
		/*
		 * The DMAC hardware only supports
		 * src (burst x width) == dst (burst x width)
		 */

		dev_err(&d40c->chan.dev->device,
			"[%s] src (burst x width) != dst (burst x width)\n",
			__func__);
		res = -EINVAL;
	}

	return res;
}

@@ -1638,13 +1712,21 @@ struct dma_async_tx_descriptor *stedma40_memcpy_sg(struct dma_chan *chan,
	if (d40d == NULL)
		goto err;

	d40d->lli_len = sgl_len;
	d40d->lli_len = d40_sg_2_dmalen(sgl_dst, sgl_len,
					d40c->dma_cfg.src_info.data_width,
					d40c->dma_cfg.dst_info.data_width);
	if (d40d->lli_len < 0) {
		dev_err(&d40c->chan.dev->device,
			"[%s] Unaligned size\n", __func__);
		goto err;
	}

	d40d->lli_current = 0;
	d40d->txd.flags = dma_flags;

	if (d40c->log_num != D40_PHY_CHAN) {

		if (d40_pool_lli_alloc(d40d, sgl_len, true) < 0) {
		if (d40_pool_lli_alloc(d40d, d40d->lli_len, true) < 0) {
			dev_err(&d40c->chan.dev->device,
				"[%s] Out of memory\n", __func__);
			goto err;
@@ -1654,15 +1736,17 @@ struct dma_async_tx_descriptor *stedma40_memcpy_sg(struct dma_chan *chan,
					 sgl_len,
					 d40d->lli_log.src,
					 d40c->log_def.lcsp1,
					 d40c->dma_cfg.src_info.data_width);
					 d40c->dma_cfg.src_info.data_width,
					 d40c->dma_cfg.dst_info.data_width);

		(void) d40_log_sg_to_lli(sgl_dst,
					 sgl_len,
					 d40d->lli_log.dst,
					 d40c->log_def.lcsp3,
					 d40c->dma_cfg.dst_info.data_width);
					 d40c->dma_cfg.dst_info.data_width,
					 d40c->dma_cfg.src_info.data_width);
	} else {
		if (d40_pool_lli_alloc(d40d, sgl_len, false) < 0) {
		if (d40_pool_lli_alloc(d40d, d40d->lli_len, false) < 0) {
			dev_err(&d40c->chan.dev->device,
				"[%s] Out of memory\n", __func__);
			goto err;
@@ -1675,6 +1759,7 @@ struct dma_async_tx_descriptor *stedma40_memcpy_sg(struct dma_chan *chan,
					virt_to_phys(d40d->lli_phy.src),
					d40c->src_def_cfg,
					d40c->dma_cfg.src_info.data_width,
					d40c->dma_cfg.dst_info.data_width,
					d40c->dma_cfg.src_info.psize);

		if (res < 0)
@@ -1687,6 +1772,7 @@ struct dma_async_tx_descriptor *stedma40_memcpy_sg(struct dma_chan *chan,
					virt_to_phys(d40d->lli_phy.dst),
					d40c->dst_def_cfg,
					d40c->dma_cfg.dst_info.data_width,
					d40c->dma_cfg.src_info.data_width,
					d40c->dma_cfg.dst_info.psize);

		if (res < 0)
@@ -1826,7 +1912,6 @@ static struct dma_async_tx_descriptor *d40_prep_memcpy(struct dma_chan *chan,
	struct d40_chan *d40c = container_of(chan, struct d40_chan,
					     chan);
	unsigned long flags;
	int err = 0;

	if (d40c->phy_chan == NULL) {
		dev_err(&d40c->chan.dev->device,
@@ -1844,6 +1929,15 @@ static struct dma_async_tx_descriptor *d40_prep_memcpy(struct dma_chan *chan,
	}

	d40d->txd.flags = dma_flags;
	d40d->lli_len = d40_size_2_dmalen(size,
					  d40c->dma_cfg.src_info.data_width,
					  d40c->dma_cfg.dst_info.data_width);
	if (d40d->lli_len < 0) {
		dev_err(&d40c->chan.dev->device,
			"[%s] Unaligned size\n", __func__);
		goto err;
	}


	dma_async_tx_descriptor_init(&d40d->txd, chan);

@@ -1851,37 +1945,40 @@ static struct dma_async_tx_descriptor *d40_prep_memcpy(struct dma_chan *chan,

	if (d40c->log_num != D40_PHY_CHAN) {

		if (d40_pool_lli_alloc(d40d, 1, true) < 0) {
		if (d40_pool_lli_alloc(d40d, d40d->lli_len, true) < 0) {
			dev_err(&d40c->chan.dev->device,
				"[%s] Out of memory\n", __func__);
			goto err;
		}
		d40d->lli_len = 1;
		d40d->lli_current = 0;

		d40_log_fill_lli(d40d->lli_log.src,
		if (d40_log_buf_to_lli(d40d->lli_log.src,
				       src,
				       size,
				       d40c->log_def.lcsp1,
				       d40c->dma_cfg.src_info.data_width,
				 true);
				       d40c->dma_cfg.dst_info.data_width,
				       true) == NULL)
			goto err;

		d40_log_fill_lli(d40d->lli_log.dst,
		if (d40_log_buf_to_lli(d40d->lli_log.dst,
				       dst,
				       size,
				       d40c->log_def.lcsp3,
				       d40c->dma_cfg.dst_info.data_width,
				 true);
				       d40c->dma_cfg.src_info.data_width,
				       true) == NULL)
			goto err;

	} else {

		if (d40_pool_lli_alloc(d40d, 1, false) < 0) {
		if (d40_pool_lli_alloc(d40d, d40d->lli_len, false) < 0) {
			dev_err(&d40c->chan.dev->device,
				"[%s] Out of memory\n", __func__);
			goto err;
		}

		err = d40_phy_fill_lli(d40d->lli_phy.src,
		if (d40_phy_buf_to_lli(d40d->lli_phy.src,
				       src,
				       size,
				       d40c->dma_cfg.src_info.psize,
@@ -1889,11 +1986,11 @@ static struct dma_async_tx_descriptor *d40_prep_memcpy(struct dma_chan *chan,
				       d40c->src_def_cfg,
				       true,
				       d40c->dma_cfg.src_info.data_width,
				       false);
		if (err)
			goto err_fill_lli;
				       d40c->dma_cfg.dst_info.data_width,
				       false) == NULL)
			goto err;

		err = d40_phy_fill_lli(d40d->lli_phy.dst,
		if (d40_phy_buf_to_lli(d40d->lli_phy.dst,
				       dst,
				       size,
				       d40c->dma_cfg.dst_info.psize,
@@ -1901,10 +1998,9 @@ static struct dma_async_tx_descriptor *d40_prep_memcpy(struct dma_chan *chan,
				       d40c->dst_def_cfg,
				       true,
				       d40c->dma_cfg.dst_info.data_width,
				       false);

		if (err)
			goto err_fill_lli;
				       d40c->dma_cfg.src_info.data_width,
				       false) == NULL)
			goto err;

		(void) dma_map_single(d40c->base->dev, d40d->lli_phy.src,
				      d40d->lli_pool.size, DMA_TO_DEVICE);
@@ -1913,9 +2009,6 @@ static struct dma_async_tx_descriptor *d40_prep_memcpy(struct dma_chan *chan,
	spin_unlock_irqrestore(&d40c->lock, flags);
	return &d40d->txd;

err_fill_lli:
	dev_err(&d40c->chan.dev->device,
		"[%s] Failed filling in PHY LLI\n", __func__);
err:
	if (d40d)
		d40_desc_free(d40c, d40d);
@@ -1945,13 +2038,21 @@ static int d40_prep_slave_sg_log(struct d40_desc *d40d,
	dma_addr_t dev_addr = 0;
	int total_size;

	if (d40_pool_lli_alloc(d40d, sg_len, true) < 0) {
	d40d->lli_len = d40_sg_2_dmalen(sgl, sg_len,
					d40c->dma_cfg.src_info.data_width,
					d40c->dma_cfg.dst_info.data_width);
	if (d40d->lli_len < 0) {
		dev_err(&d40c->chan.dev->device,
			"[%s] Unaligned size\n", __func__);
		return -EINVAL;
	}

	if (d40_pool_lli_alloc(d40d, d40d->lli_len, true) < 0) {
		dev_err(&d40c->chan.dev->device,
			"[%s] Out of memory\n", __func__);
		return -ENOMEM;
	}

	d40d->lli_len = sg_len;
	d40d->lli_current = 0;

	if (direction == DMA_FROM_DEVICE)
@@ -1993,13 +2094,21 @@ static int d40_prep_slave_sg_phy(struct d40_desc *d40d,
	dma_addr_t dst_dev_addr;
	int res;

	if (d40_pool_lli_alloc(d40d, sgl_len, false) < 0) {
	d40d->lli_len = d40_sg_2_dmalen(sgl, sgl_len,
					d40c->dma_cfg.src_info.data_width,
					d40c->dma_cfg.dst_info.data_width);
	if (d40d->lli_len < 0) {
		dev_err(&d40c->chan.dev->device,
			"[%s] Unaligned size\n", __func__);
		return -EINVAL;
	}

	if (d40_pool_lli_alloc(d40d, d40d->lli_len, false) < 0) {
		dev_err(&d40c->chan.dev->device,
			"[%s] Out of memory\n", __func__);
		return -ENOMEM;
	}

	d40d->lli_len = sgl_len;
	d40d->lli_current = 0;

	if (direction == DMA_FROM_DEVICE) {
@@ -2024,6 +2133,7 @@ static int d40_prep_slave_sg_phy(struct d40_desc *d40d,
				virt_to_phys(d40d->lli_phy.src),
				d40c->src_def_cfg,
				d40c->dma_cfg.src_info.data_width,
				d40c->dma_cfg.dst_info.data_width,
				d40c->dma_cfg.src_info.psize);
	if (res < 0)
		return res;
@@ -2035,6 +2145,7 @@ static int d40_prep_slave_sg_phy(struct d40_desc *d40d,
				virt_to_phys(d40d->lli_phy.dst),
				d40c->dst_def_cfg,
				d40c->dma_cfg.dst_info.data_width,
				d40c->dma_cfg.src_info.data_width,
				d40c->dma_cfg.dst_info.psize);
	if (res < 0)
		return res;
@@ -2244,6 +2355,8 @@ static void d40_set_runtime_config(struct dma_chan *chan,
			psize = STEDMA40_PSIZE_PHY_8;
		else if (config_maxburst >= 4)
			psize = STEDMA40_PSIZE_PHY_4;
		else if (config_maxburst >= 2)
			psize = STEDMA40_PSIZE_PHY_2;
		else
			psize = STEDMA40_PSIZE_PHY_1;
	}
+172 −74
Original line number Diff line number Diff line
/*
 * Copyright (C) ST-Ericsson SA 2007-2010
 * Author: Per Friden <per.friden@stericsson.com> for ST-Ericsson
 * Author: Per Forlin <per.forlin@stericsson.com> for ST-Ericsson
 * Author: Jonas Aaberg <jonas.aberg@stericsson.com> for ST-Ericsson
 * License terms: GNU General Public License (GPL) version 2
 */
@@ -122,7 +122,7 @@ void d40_phy_cfg(struct stedma40_chan_cfg *cfg,
	*dst_cfg = dst;
}

int d40_phy_fill_lli(struct d40_phy_lli *lli,
static int d40_phy_fill_lli(struct d40_phy_lli *lli,
			    dma_addr_t data,
			    u32 data_size,
			    int psize,
@@ -139,13 +139,6 @@ int d40_phy_fill_lli(struct d40_phy_lli *lli,
	else
		num_elems = 2 << psize;

	/*
	 * Size is 16bit. data_width is 8, 16, 32 or 64 bit
	 * Block large than 64 KiB must be split.
	 */
	if (data_size > (0xffff << data_width))
		return -EINVAL;

	/* Must be aligned */
	if (!IS_ALIGNED(data, 0x1 << data_width))
		return -EINVAL;
@@ -187,55 +180,118 @@ int d40_phy_fill_lli(struct d40_phy_lli *lli,
	return 0;
}

static int d40_seg_size(int size, int data_width1, int data_width2)
{
	u32 max_w = max(data_width1, data_width2);
	u32 min_w = min(data_width1, data_width2);
	u32 seg_max = ALIGN(STEDMA40_MAX_SEG_SIZE << min_w, 1 << max_w);

	if (seg_max > STEDMA40_MAX_SEG_SIZE)
		seg_max -= (1 << max_w);

	if (size <= seg_max)
		return size;

	if (size <= 2 * seg_max)
		return ALIGN(size / 2, 1 << max_w);

	return seg_max;
}

struct d40_phy_lli *d40_phy_buf_to_lli(struct d40_phy_lli *lli,
				       dma_addr_t addr,
				       u32 size,
				       int psize,
				       dma_addr_t lli_phys,
				       u32 reg_cfg,
				       bool term_int,
				       u32 data_width1,
				       u32 data_width2,
				       bool is_device)
{
	int err;
	dma_addr_t next = lli_phys;
	int size_rest = size;
	int size_seg = 0;

	do {
		size_seg = d40_seg_size(size_rest, data_width1, data_width2);
		size_rest -= size_seg;

		if (term_int && size_rest == 0)
			next = 0;
		else
			next = ALIGN(next + sizeof(struct d40_phy_lli),
				     D40_LLI_ALIGN);

		err = d40_phy_fill_lli(lli,
				       addr,
				       size_seg,
				       psize,
				       next,
				       reg_cfg,
				       !next,
				       data_width1,
				       is_device);

		if (err)
			goto err;

		lli++;
		if (!is_device)
			addr += size_seg;
	} while (size_rest);

	return lli;

 err:
	return NULL;
}

int d40_phy_sg_to_lli(struct scatterlist *sg,
		      int sg_len,
		      dma_addr_t target,
		      struct d40_phy_lli *lli,
		      struct d40_phy_lli *lli_sg,
		      dma_addr_t lli_phys,
		      u32 reg_cfg,
		      u32 data_width,
		      u32 data_width1,
		      u32 data_width2,
		      int psize)
{
	int total_size = 0;
	int i;
	struct scatterlist *current_sg = sg;
	dma_addr_t next_lli_phys;
	dma_addr_t dst;
	int err = 0;
	struct d40_phy_lli *lli = lli_sg;
	dma_addr_t l_phys = lli_phys;

	for_each_sg(sg, current_sg, sg_len, i) {

		total_size += sg_dma_len(current_sg);

		/* If this scatter list entry is the last one, no next link */
		if (sg_len - 1 == i)
			next_lli_phys = 0;
		else
			next_lli_phys = ALIGN(lli_phys + (i + 1) *
					      sizeof(struct d40_phy_lli),
					      D40_LLI_ALIGN);

		if (target)
			dst = target;
		else
			dst = sg_phys(current_sg);

		err = d40_phy_fill_lli(&lli[i],
		l_phys = ALIGN(lli_phys + (lli - lli_sg) *
			       sizeof(struct d40_phy_lli), D40_LLI_ALIGN);

		lli = d40_phy_buf_to_lli(lli,
					 dst,
					 sg_dma_len(current_sg),
					 psize,
				       next_lli_phys,
					 l_phys,
					 reg_cfg,
				       !next_lli_phys,
				       data_width,
					 sg_len - 1 == i,
					 data_width1,
					 data_width2,
					 target == dst);
		if (err)
			goto err;
		if (lli == NULL)
			return -EINVAL;
	}

	return total_size;
err:
	return err;
}


@@ -315,7 +371,7 @@ void d40_log_lli_lcla_write(struct d40_log_lli *lcla,
	writel(lli_dst->lcsp13, &lcla[1].lcsp13);
}

void d40_log_fill_lli(struct d40_log_lli *lli,
static void d40_log_fill_lli(struct d40_log_lli *lli,
			     dma_addr_t data, u32 data_size,
			     u32 reg_cfg,
			     u32 data_width,
@@ -326,6 +382,9 @@ void d40_log_fill_lli(struct d40_log_lli *lli,
	/* The number of elements to transfer */
	lli->lcsp02 = ((data_size >> data_width) <<
		       D40_MEM_LCSP0_ECNT_POS) & D40_MEM_LCSP0_ECNT_MASK;

	BUG_ON((data_size >> data_width) > STEDMA40_MAX_SEG_SIZE);

	/* 16 LSBs address of the current element */
	lli->lcsp02 |= data & D40_MEM_LCSP0_SPTR_MASK;
	/* 16 MSBs address of the current element */
@@ -348,55 +407,94 @@ int d40_log_sg_to_dev(struct scatterlist *sg,
	int total_size = 0;
	struct scatterlist *current_sg = sg;
	int i;
	struct d40_log_lli *lli_src = lli->src;
	struct d40_log_lli *lli_dst = lli->dst;

	for_each_sg(sg, current_sg, sg_len, i) {
		total_size += sg_dma_len(current_sg);

		if (direction == DMA_TO_DEVICE) {
			d40_log_fill_lli(&lli->src[i],
			lli_src =
				d40_log_buf_to_lli(lli_src,
						   sg_phys(current_sg),
						   sg_dma_len(current_sg),
						   lcsp->lcsp1, src_data_width,
						   dst_data_width,
						   true);
			d40_log_fill_lli(&lli->dst[i],
			lli_dst =
				d40_log_buf_to_lli(lli_dst,
						   dev_addr,
						   sg_dma_len(current_sg),
						   lcsp->lcsp3, dst_data_width,
						   src_data_width,
						   false);
		} else {
			d40_log_fill_lli(&lli->dst[i],
			lli_dst =
				d40_log_buf_to_lli(lli_dst,
						   sg_phys(current_sg),
						   sg_dma_len(current_sg),
						   lcsp->lcsp3, dst_data_width,
						   src_data_width,
						   true);
			d40_log_fill_lli(&lli->src[i],
			lli_src =
				d40_log_buf_to_lli(lli_src,
						   dev_addr,
						   sg_dma_len(current_sg),
						   lcsp->lcsp1, src_data_width,
						   dst_data_width,
						   false);
		}
	}
	return total_size;
}

struct d40_log_lli *d40_log_buf_to_lli(struct d40_log_lli *lli_sg,
				       dma_addr_t addr,
				       int size,
				       u32 lcsp13, /* src or dst*/
				       u32 data_width1,
				       u32 data_width2,
				       bool addr_inc)
{
	struct d40_log_lli *lli = lli_sg;
	int size_rest = size;
	int size_seg = 0;

	do {
		size_seg = d40_seg_size(size_rest, data_width1, data_width2);
		size_rest -= size_seg;

		d40_log_fill_lli(lli,
				 addr,
				 size_seg,
				 lcsp13, data_width1,
				 addr_inc);
		if (addr_inc)
			addr += size_seg;
		lli++;
	} while (size_rest);

	return lli;
}

int d40_log_sg_to_lli(struct scatterlist *sg,
		      int sg_len,
		      struct d40_log_lli *lli_sg,
		      u32 lcsp13, /* src or dst*/
		      u32 data_width)
		      u32 data_width1, u32 data_width2)
{
	int total_size = 0;
	struct scatterlist *current_sg = sg;
	int i;
	struct d40_log_lli *lli = lli_sg;

	for_each_sg(sg, current_sg, sg_len, i) {
		total_size += sg_dma_len(current_sg);

		d40_log_fill_lli(&lli_sg[i],
		lli = d40_log_buf_to_lli(lli,
					 sg_phys(current_sg),
					 sg_dma_len(current_sg),
				 lcsp13, data_width,
				 true);
					 lcsp13,
					 data_width1, data_width2, true);
	}
	return total_size;
}
+19 −17
Original line number Diff line number Diff line
@@ -292,17 +292,19 @@ int d40_phy_sg_to_lli(struct scatterlist *sg,
		      struct d40_phy_lli *lli,
		      dma_addr_t lli_phys,
		      u32 reg_cfg,
		      u32 data_width,
		      u32 data_width1,
		      u32 data_width2,
		      int psize);

int d40_phy_fill_lli(struct d40_phy_lli *lli,
struct d40_phy_lli *d40_phy_buf_to_lli(struct d40_phy_lli *lli,
				       dma_addr_t data,
				       u32 data_size,
				       int psize,
				       dma_addr_t next_lli,
				       u32 reg_cfg,
				       bool term_int,
		     u32 data_width,
				       u32 data_width1,
				       u32 data_width2,
				       bool is_device);

void d40_phy_lli_write(void __iomem *virtbase,
@@ -312,11 +314,11 @@ void d40_phy_lli_write(void __iomem *virtbase,

/* Logical channels */

void d40_log_fill_lli(struct d40_log_lli *lli,
		      dma_addr_t data,
		      u32 data_size,
		      u32 reg_cfg,
		      u32 data_width,
struct d40_log_lli *d40_log_buf_to_lli(struct d40_log_lli *lli_sg,
				       dma_addr_t addr,
				       int size,
				       u32 lcsp13, /* src or dst*/
				       u32 data_width1, u32 data_width2,
				       bool addr_inc);

int d40_log_sg_to_dev(struct scatterlist *sg,
@@ -332,7 +334,7 @@ int d40_log_sg_to_lli(struct scatterlist *sg,
		      int sg_len,
		      struct d40_log_lli *lli_sg,
		      u32 lcsp13, /* src or dst*/
		      u32 data_width);
		      u32 data_width1, u32 data_width2);

void d40_log_lli_lcpa_write(struct d40_log_lli_full *lcpa,
			    struct d40_log_lli *lli_dst,
Loading