Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit d1998ef3 authored by Ben Collins's avatar Ben Collins Committed by Linus Torvalds
Browse files

[PATCH] ib_verbs: Use explicit if-else statements to avoid errors with do-while macros



At least on PPC, the "op ? op : dma" construct causes a compile failure
because the dma_* is a do{}while(0) macro.

This turns all of them into proper if/else to avoid this problem.

Signed-off-by: default avatarBen Collins <bcollins@ubuntu.com>
Signed-off-by: default avatarLinus Torvalds <torvalds@osdl.org>
parent cc016448
Loading
Loading
Loading
Loading
+39 −31
Original line number Original line Diff line number Diff line
@@ -1456,9 +1456,9 @@ struct ib_mr *ib_get_dma_mr(struct ib_pd *pd, int mr_access_flags);
 */
 */
static inline int ib_dma_mapping_error(struct ib_device *dev, u64 dma_addr)
static inline int ib_dma_mapping_error(struct ib_device *dev, u64 dma_addr)
{
{
	return dev->dma_ops ?
	if (dev->dma_ops)
		dev->dma_ops->mapping_error(dev, dma_addr) :
		return dev->dma_ops->mapping_error(dev, dma_addr);
		dma_mapping_error(dma_addr);
	return dma_mapping_error(dma_addr);
}
}


/**
/**
@@ -1472,9 +1472,9 @@ static inline u64 ib_dma_map_single(struct ib_device *dev,
				    void *cpu_addr, size_t size,
				    void *cpu_addr, size_t size,
				    enum dma_data_direction direction)
				    enum dma_data_direction direction)
{
{
	return dev->dma_ops ?
	if (dev->dma_ops)
		dev->dma_ops->map_single(dev, cpu_addr, size, direction) :
		return dev->dma_ops->map_single(dev, cpu_addr, size, direction);
		dma_map_single(dev->dma_device, cpu_addr, size, direction);
	return dma_map_single(dev->dma_device, cpu_addr, size, direction);
}
}


/**
/**
@@ -1488,8 +1488,9 @@ static inline void ib_dma_unmap_single(struct ib_device *dev,
				       u64 addr, size_t size,
				       u64 addr, size_t size,
				       enum dma_data_direction direction)
				       enum dma_data_direction direction)
{
{
	dev->dma_ops ?
	if (dev->dma_ops)
		dev->dma_ops->unmap_single(dev, addr, size, direction) :
		dev->dma_ops->unmap_single(dev, addr, size, direction);
	else
		dma_unmap_single(dev->dma_device, addr, size, direction);
		dma_unmap_single(dev->dma_device, addr, size, direction);
}
}


@@ -1507,9 +1508,9 @@ static inline u64 ib_dma_map_page(struct ib_device *dev,
				  size_t size,
				  size_t size,
					 enum dma_data_direction direction)
					 enum dma_data_direction direction)
{
{
	return dev->dma_ops ?
	if (dev->dma_ops)
		dev->dma_ops->map_page(dev, page, offset, size, direction) :
		return dev->dma_ops->map_page(dev, page, offset, size, direction);
		dma_map_page(dev->dma_device, page, offset, size, direction);
	return dma_map_page(dev->dma_device, page, offset, size, direction);
}
}


/**
/**
@@ -1523,8 +1524,9 @@ static inline void ib_dma_unmap_page(struct ib_device *dev,
				     u64 addr, size_t size,
				     u64 addr, size_t size,
				     enum dma_data_direction direction)
				     enum dma_data_direction direction)
{
{
	dev->dma_ops ?
	if (dev->dma_ops)
		dev->dma_ops->unmap_page(dev, addr, size, direction) :
		dev->dma_ops->unmap_page(dev, addr, size, direction);
	else
		dma_unmap_page(dev->dma_device, addr, size, direction);
		dma_unmap_page(dev->dma_device, addr, size, direction);
}
}


@@ -1539,9 +1541,9 @@ static inline int ib_dma_map_sg(struct ib_device *dev,
				struct scatterlist *sg, int nents,
				struct scatterlist *sg, int nents,
				enum dma_data_direction direction)
				enum dma_data_direction direction)
{
{
	return dev->dma_ops ?
	if (dev->dma_ops)
		dev->dma_ops->map_sg(dev, sg, nents, direction) :
		return dev->dma_ops->map_sg(dev, sg, nents, direction);
		dma_map_sg(dev->dma_device, sg, nents, direction);
	return dma_map_sg(dev->dma_device, sg, nents, direction);
}
}


/**
/**
@@ -1555,8 +1557,9 @@ static inline void ib_dma_unmap_sg(struct ib_device *dev,
				   struct scatterlist *sg, int nents,
				   struct scatterlist *sg, int nents,
				   enum dma_data_direction direction)
				   enum dma_data_direction direction)
{
{
	dev->dma_ops ?
	if (dev->dma_ops)
		dev->dma_ops->unmap_sg(dev, sg, nents, direction) :
		dev->dma_ops->unmap_sg(dev, sg, nents, direction);
	else
		dma_unmap_sg(dev->dma_device, sg, nents, direction);
		dma_unmap_sg(dev->dma_device, sg, nents, direction);
}
}


@@ -1568,8 +1571,9 @@ static inline void ib_dma_unmap_sg(struct ib_device *dev,
static inline u64 ib_sg_dma_address(struct ib_device *dev,
static inline u64 ib_sg_dma_address(struct ib_device *dev,
				    struct scatterlist *sg)
				    struct scatterlist *sg)
{
{
	return dev->dma_ops ?
	if (dev->dma_ops)
		dev->dma_ops->dma_address(dev, sg) : sg_dma_address(sg);
		return dev->dma_ops->dma_address(dev, sg);
	return sg_dma_address(sg);
}
}


/**
/**
@@ -1580,8 +1584,9 @@ static inline u64 ib_sg_dma_address(struct ib_device *dev,
static inline unsigned int ib_sg_dma_len(struct ib_device *dev,
static inline unsigned int ib_sg_dma_len(struct ib_device *dev,
					 struct scatterlist *sg)
					 struct scatterlist *sg)
{
{
	return dev->dma_ops ?
	if (dev->dma_ops)
		dev->dma_ops->dma_len(dev, sg) : sg_dma_len(sg);
		return dev->dma_ops->dma_len(dev, sg);
	return sg_dma_len(sg);
}
}


/**
/**
@@ -1596,8 +1601,9 @@ static inline void ib_dma_sync_single_for_cpu(struct ib_device *dev,
					      size_t size,
					      size_t size,
					      enum dma_data_direction dir)
					      enum dma_data_direction dir)
{
{
	dev->dma_ops ?
	if (dev->dma_ops)
		dev->dma_ops->sync_single_for_cpu(dev, addr, size, dir) :
		dev->dma_ops->sync_single_for_cpu(dev, addr, size, dir);
	else
		dma_sync_single_for_cpu(dev->dma_device, addr, size, dir);
		dma_sync_single_for_cpu(dev->dma_device, addr, size, dir);
}
}


@@ -1613,8 +1619,9 @@ static inline void ib_dma_sync_single_for_device(struct ib_device *dev,
						 size_t size,
						 size_t size,
						 enum dma_data_direction dir)
						 enum dma_data_direction dir)
{
{
	dev->dma_ops ?
	if (dev->dma_ops)
		dev->dma_ops->sync_single_for_device(dev, addr, size, dir) :
		dev->dma_ops->sync_single_for_device(dev, addr, size, dir);
	else
		dma_sync_single_for_device(dev->dma_device, addr, size, dir);
		dma_sync_single_for_device(dev->dma_device, addr, size, dir);
}
}


@@ -1630,9 +1637,9 @@ static inline void *ib_dma_alloc_coherent(struct ib_device *dev,
					   u64 *dma_handle,
					   u64 *dma_handle,
					   gfp_t flag)
					   gfp_t flag)
{
{
	return dev->dma_ops ?
	if (dev->dma_ops)
		dev->dma_ops->alloc_coherent(dev, size, dma_handle, flag) :
		return dev->dma_ops->alloc_coherent(dev, size, dma_handle, flag);
		dma_alloc_coherent(dev->dma_device, size, dma_handle, flag);
	return dma_alloc_coherent(dev->dma_device, size, dma_handle, flag);
}
}


/**
/**
@@ -1646,8 +1653,9 @@ static inline void ib_dma_free_coherent(struct ib_device *dev,
					size_t size, void *cpu_addr,
					size_t size, void *cpu_addr,
					u64 dma_handle)
					u64 dma_handle)
{
{
	dev->dma_ops ?
	if (dev->dma_ops)
		dev->dma_ops->free_coherent(dev, size, cpu_addr, dma_handle) :
		dev->dma_ops->free_coherent(dev, size, cpu_addr, dma_handle);
	else
		dma_free_coherent(dev->dma_device, size, cpu_addr, dma_handle);
		dma_free_coherent(dev->dma_device, size, cpu_addr, dma_handle);
}
}