Loading arch/arm/common/dmabounce.c +12 −36 Original line number Diff line number Diff line Loading @@ -468,45 +468,23 @@ void dma_sync_single_range_for_device(struct device *dev, dma_addr_t dma_addr, } EXPORT_SYMBOL(dma_sync_single_range_for_device); void dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int nents, enum dma_data_direction dir) int dmabounce_sync_for_cpu(struct device *dev, dma_addr_t addr, unsigned long off, size_t sz, enum dma_data_direction dir) { struct scatterlist *s; int i; dev_dbg(dev, "%s(sg=%p,nents=%d,dir=%x)\n", __func__, sg, nents, dir); BUG_ON(dir == DMA_NONE); for_each_sg(sg, s, nents, i) { dma_addr_t dma_addr = s->dma_address; unsigned int length = s->length; sync_single(dev, dma_addr, length, dir); } dev_dbg(dev, "%s(dma=%#lx,off=%#lx,sz=%zx,dir=%x)\n", __func__, addr, off, sz, dir); return sync_single(dev, addr, off + sz, dir); } EXPORT_SYMBOL(dmabounce_sync_for_cpu); void dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int nents, enum dma_data_direction dir) int dmabounce_sync_for_device(struct device *dev, dma_addr_t addr, unsigned long off, size_t sz, enum dma_data_direction dir) { struct scatterlist *s; int i; dev_dbg(dev, "%s(sg=%p,nents=%d,dir=%x)\n", __func__, sg, nents, dir); BUG_ON(dir == DMA_NONE); for_each_sg(sg, s, nents, i) { dma_addr_t dma_addr = s->dma_address; unsigned int length = s->length; sync_single(dev, dma_addr, length, dir); } dev_dbg(dev, "%s(dma=%#lx,off=%#lx,sz=%zx,dir=%x)\n", __func__, addr, off, sz, dir); return sync_single(dev, addr, off + sz, dir); } EXPORT_SYMBOL(dmabounce_sync_for_device); static int dmabounce_init_pool(struct dmabounce_pool *pool, struct device *dev, const char *name, Loading Loading @@ -618,8 +596,6 @@ dmabounce_unregister_dev(struct device *dev) EXPORT_SYMBOL(dma_map_single); EXPORT_SYMBOL(dma_unmap_single); EXPORT_SYMBOL(dma_sync_sg_for_cpu); EXPORT_SYMBOL(dma_sync_sg_for_device); EXPORT_SYMBOL(dmabounce_register_dev); EXPORT_SYMBOL(dmabounce_unregister_dev); Loading arch/arm/include/asm/dma-mapping.h +11 −0 Original line number Diff line number Diff line Loading @@ -410,6 +410,17 @@ extern void dmabounce_unregister_dev(struct device *); * */ extern int dma_needs_bounce(struct device*, dma_addr_t, size_t); /* * Private functions */ int dmabounce_sync_for_cpu(struct device *, dma_addr_t, unsigned long, size_t, enum dma_data_direction); int dmabounce_sync_for_device(struct device *, dma_addr_t, unsigned long, size_t, enum dma_data_direction); #else #define dmabounce_sync_for_cpu(dev,dma,off,sz,dir) (1) #define dmabounce_sync_for_device(dev,dma,off,sz,dir) (1) #endif /* CONFIG_DMABOUNCE */ #endif /* __KERNEL__ */ Loading arch/arm/mm/dma-mapping.c +8 −2 Original line number Diff line number Diff line Loading @@ -571,7 +571,6 @@ void dma_unmap_sg(struct device *dev, struct scatterlist *sg, int nents, } EXPORT_SYMBOL(dma_unmap_sg); #ifndef CONFIG_DMABOUNCE /** * dma_sync_sg_for_cpu * @dev: valid struct device pointer, or NULL for ISA and EISA-like devices Loading @@ -586,6 +585,10 @@ void dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int i; for_each_sg(sg, s, nents, i) { if (!dmabounce_sync_for_cpu(dev, sg_dma_address(s), 0, sg_dma_len(s), dir)) continue; if (!arch_is_coherent()) dma_cache_maint(sg_virt(s), s->length, dir); } Loading @@ -606,9 +609,12 @@ void dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int i; for_each_sg(sg, s, nents, i) { if (!dmabounce_sync_for_device(dev, sg_dma_address(s), 0, sg_dma_len(s), dir)) continue; if (!arch_is_coherent()) dma_cache_maint(sg_virt(s), s->length, dir); } } EXPORT_SYMBOL(dma_sync_sg_for_device); #endif Loading
arch/arm/common/dmabounce.c +12 −36 Original line number Diff line number Diff line Loading @@ -468,45 +468,23 @@ void dma_sync_single_range_for_device(struct device *dev, dma_addr_t dma_addr, } EXPORT_SYMBOL(dma_sync_single_range_for_device); void dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int nents, enum dma_data_direction dir) int dmabounce_sync_for_cpu(struct device *dev, dma_addr_t addr, unsigned long off, size_t sz, enum dma_data_direction dir) { struct scatterlist *s; int i; dev_dbg(dev, "%s(sg=%p,nents=%d,dir=%x)\n", __func__, sg, nents, dir); BUG_ON(dir == DMA_NONE); for_each_sg(sg, s, nents, i) { dma_addr_t dma_addr = s->dma_address; unsigned int length = s->length; sync_single(dev, dma_addr, length, dir); } dev_dbg(dev, "%s(dma=%#lx,off=%#lx,sz=%zx,dir=%x)\n", __func__, addr, off, sz, dir); return sync_single(dev, addr, off + sz, dir); } EXPORT_SYMBOL(dmabounce_sync_for_cpu); void dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int nents, enum dma_data_direction dir) int dmabounce_sync_for_device(struct device *dev, dma_addr_t addr, unsigned long off, size_t sz, enum dma_data_direction dir) { struct scatterlist *s; int i; dev_dbg(dev, "%s(sg=%p,nents=%d,dir=%x)\n", __func__, sg, nents, dir); BUG_ON(dir == DMA_NONE); for_each_sg(sg, s, nents, i) { dma_addr_t dma_addr = s->dma_address; unsigned int length = s->length; sync_single(dev, dma_addr, length, dir); } dev_dbg(dev, "%s(dma=%#lx,off=%#lx,sz=%zx,dir=%x)\n", __func__, addr, off, sz, dir); return sync_single(dev, addr, off + sz, dir); } EXPORT_SYMBOL(dmabounce_sync_for_device); static int dmabounce_init_pool(struct dmabounce_pool *pool, struct device *dev, const char *name, Loading Loading @@ -618,8 +596,6 @@ dmabounce_unregister_dev(struct device *dev) EXPORT_SYMBOL(dma_map_single); EXPORT_SYMBOL(dma_unmap_single); EXPORT_SYMBOL(dma_sync_sg_for_cpu); EXPORT_SYMBOL(dma_sync_sg_for_device); EXPORT_SYMBOL(dmabounce_register_dev); EXPORT_SYMBOL(dmabounce_unregister_dev); Loading
arch/arm/include/asm/dma-mapping.h +11 −0 Original line number Diff line number Diff line Loading @@ -410,6 +410,17 @@ extern void dmabounce_unregister_dev(struct device *); * */ extern int dma_needs_bounce(struct device*, dma_addr_t, size_t); /* * Private functions */ int dmabounce_sync_for_cpu(struct device *, dma_addr_t, unsigned long, size_t, enum dma_data_direction); int dmabounce_sync_for_device(struct device *, dma_addr_t, unsigned long, size_t, enum dma_data_direction); #else #define dmabounce_sync_for_cpu(dev,dma,off,sz,dir) (1) #define dmabounce_sync_for_device(dev,dma,off,sz,dir) (1) #endif /* CONFIG_DMABOUNCE */ #endif /* __KERNEL__ */ Loading
arch/arm/mm/dma-mapping.c +8 −2 Original line number Diff line number Diff line Loading @@ -571,7 +571,6 @@ void dma_unmap_sg(struct device *dev, struct scatterlist *sg, int nents, } EXPORT_SYMBOL(dma_unmap_sg); #ifndef CONFIG_DMABOUNCE /** * dma_sync_sg_for_cpu * @dev: valid struct device pointer, or NULL for ISA and EISA-like devices Loading @@ -586,6 +585,10 @@ void dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int i; for_each_sg(sg, s, nents, i) { if (!dmabounce_sync_for_cpu(dev, sg_dma_address(s), 0, sg_dma_len(s), dir)) continue; if (!arch_is_coherent()) dma_cache_maint(sg_virt(s), s->length, dir); } Loading @@ -606,9 +609,12 @@ void dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int i; for_each_sg(sg, s, nents, i) { if (!dmabounce_sync_for_device(dev, sg_dma_address(s), 0, sg_dma_len(s), dir)) continue; if (!arch_is_coherent()) dma_cache_maint(sg_virt(s), s->length, dir); } } EXPORT_SYMBOL(dma_sync_sg_for_device); #endif