Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit d909b347 authored by Dan Williams's avatar Dan Williams
Browse files

async_tx: kill ASYNC_TX_ASSUME_COHERENT



Remove the unused ASYNC_TX_ASSUME_COHERENT flag.  Async_tx is
meant to hide the difference between asynchronous hardware and synchronous
software operations, this flag requires clients to understand cache
coherency consequences of the async path.

Signed-off-by: default avatarDan Williams <dan.j.williams@intel.com>
Reviewed-by: default avatarHaavard Skinnemoen <hskinnemoen@atmel.com>
parent e73ef9ac
Loading
Loading
Loading
Loading
+5 −10
Original line number Diff line number Diff line
@@ -35,7 +35,7 @@
 * @src: src page
 * @offset: offset in pages to start transaction
 * @len: length in bytes
 * @flags: ASYNC_TX_ASSUME_COHERENT, ASYNC_TX_ACK, ASYNC_TX_DEP_ACK,
 * @flags: ASYNC_TX_ACK, ASYNC_TX_DEP_ACK,
 * @depend_tx: memcpy depends on the result of this transaction
 * @cb_fn: function to call when the memcpy completes
 * @cb_param: parameter to pass to the callback routine
@@ -55,20 +55,15 @@ async_memcpy(struct page *dest, struct page *src, unsigned int dest_offset,

	if (tx) { /* run the memcpy asynchronously */
		dma_addr_t addr;
		enum dma_data_direction dir;

		pr_debug("%s: (async) len: %zu\n", __FUNCTION__, len);

		dir = (flags & ASYNC_TX_ASSUME_COHERENT) ?
			DMA_NONE : DMA_FROM_DEVICE;

		addr = dma_map_page(device->dev, dest, dest_offset, len, dir);
		addr = dma_map_page(device->dev, dest, dest_offset, len,
				    DMA_FROM_DEVICE);
		tx->tx_set_dest(addr, tx, 0);

		dir = (flags & ASYNC_TX_ASSUME_COHERENT) ?
			DMA_NONE : DMA_TO_DEVICE;

		addr = dma_map_page(device->dev, src, src_offset, len, dir);
		addr = dma_map_page(device->dev, src, src_offset, len,
				    DMA_TO_DEVICE);
		tx->tx_set_src(addr, tx, 0);

		async_tx_submit(chan, tx, flags, depend_tx, cb_fn, cb_param);
+3 −5
Original line number Diff line number Diff line
@@ -35,7 +35,7 @@
 * @val: fill value
 * @offset: offset in pages to start transaction
 * @len: length in bytes
 * @flags: ASYNC_TX_ASSUME_COHERENT, ASYNC_TX_ACK, ASYNC_TX_DEP_ACK
 * @flags: ASYNC_TX_ACK, ASYNC_TX_DEP_ACK
 * @depend_tx: memset depends on the result of this transaction
 * @cb_fn: function to call when the memcpy completes
 * @cb_param: parameter to pass to the callback routine
@@ -55,13 +55,11 @@ async_memset(struct page *dest, int val, unsigned int offset,

	if (tx) { /* run the memset asynchronously */
		dma_addr_t dma_addr;
		enum dma_data_direction dir;

		pr_debug("%s: (async) len: %zu\n", __FUNCTION__, len);
		dir = (flags & ASYNC_TX_ASSUME_COHERENT) ?
			DMA_NONE : DMA_FROM_DEVICE;

		dma_addr = dma_map_page(device->dev, dest, offset, len, dir);
		dma_addr = dma_map_page(device->dev, dest, offset, len,
					DMA_FROM_DEVICE);
		tx->tx_set_dest(dma_addr, tx, 0);

		async_tx_submit(chan, tx, flags, depend_tx, cb_fn, cb_param);
+6 −16
Original line number Diff line number Diff line
@@ -42,23 +42,17 @@ do_async_xor(struct dma_async_tx_descriptor *tx, struct dma_device *device,
	dma_async_tx_callback cb_fn, void *cb_param)
{
	dma_addr_t dma_addr;
	enum dma_data_direction dir;
	int i;

	pr_debug("%s: len: %zu\n", __FUNCTION__, len);

	dir = (flags & ASYNC_TX_ASSUME_COHERENT) ?
		DMA_NONE : DMA_FROM_DEVICE;

	dma_addr = dma_map_page(device->dev, dest, offset, len, dir);
	dma_addr = dma_map_page(device->dev, dest, offset, len,
				DMA_FROM_DEVICE);
	tx->tx_set_dest(dma_addr, tx, 0);

	dir = (flags & ASYNC_TX_ASSUME_COHERENT) ?
		DMA_NONE : DMA_TO_DEVICE;

	for (i = 0; i < src_cnt; i++) {
		dma_addr = dma_map_page(device->dev, src_list[i],
			offset, len, dir);
			offset, len, DMA_TO_DEVICE);
		tx->tx_set_src(dma_addr, tx, i);
	}

@@ -106,7 +100,7 @@ do_sync_xor(struct page *dest, struct page **src_list, unsigned int offset,
 * @src_cnt: number of source pages
 * @len: length in bytes
 * @flags: ASYNC_TX_XOR_ZERO_DST, ASYNC_TX_XOR_DROP_DEST,
 *	ASYNC_TX_ASSUME_COHERENT, ASYNC_TX_ACK, ASYNC_TX_DEP_ACK
 *	ASYNC_TX_ACK, ASYNC_TX_DEP_ACK
 * @depend_tx: xor depends on the result of this transaction.
 * @cb_fn: function to call when the xor completes
 * @cb_param: parameter to pass to the callback routine
@@ -246,7 +240,7 @@ static int page_is_zero(struct page *p, unsigned int offset, size_t len)
 * @src_cnt: number of source pages
 * @len: length in bytes
 * @result: 0 if sum == 0 else non-zero
 * @flags: ASYNC_TX_ASSUME_COHERENT, ASYNC_TX_ACK, ASYNC_TX_DEP_ACK
 * @flags: ASYNC_TX_ACK, ASYNC_TX_DEP_ACK
 * @depend_tx: xor depends on the result of this transaction.
 * @cb_fn: function to call when the xor completes
 * @cb_param: parameter to pass to the callback routine
@@ -270,16 +264,12 @@ async_xor_zero_sum(struct page *dest, struct page **src_list,

	if (tx) {
		dma_addr_t dma_addr;
		enum dma_data_direction dir;

		pr_debug("%s: (async) len: %zu\n", __FUNCTION__, len);

		dir = (flags & ASYNC_TX_ASSUME_COHERENT) ?
			DMA_NONE : DMA_TO_DEVICE;

		for (i = 0; i < src_cnt; i++) {
			dma_addr = dma_map_page(device->dev, src_list[i],
				offset, len, dir);
				offset, len, DMA_TO_DEVICE);
			tx->tx_set_src(dma_addr, tx, i);
		}

+0 −2
Original line number Diff line number Diff line
@@ -47,7 +47,6 @@ struct dma_chan_ref {
 * address is an implied source, whereas the asynchronous case it must be listed
 * as a source.  The destination address must be the first address in the source
 * array.
 * @ASYNC_TX_ASSUME_COHERENT: skip cache maintenance operations
 * @ASYNC_TX_ACK: immediately ack the descriptor, precludes setting up a
 * dependency chain
 * @ASYNC_TX_DEP_ACK: ack the dependency descriptor.  Useful for chaining.
@@ -55,7 +54,6 @@ struct dma_chan_ref {
enum async_tx_flags {
	ASYNC_TX_XOR_ZERO_DST	 = (1 << 0),
	ASYNC_TX_XOR_DROP_DST	 = (1 << 1),
	ASYNC_TX_ASSUME_COHERENT = (1 << 2),
	ASYNC_TX_ACK		 = (1 << 3),
	ASYNC_TX_DEP_ACK	 = (1 << 4),
};