Loading arch/arm/include/asm/dma-iommu.h +1 −0 Original line number Diff line number Diff line Loading @@ -7,6 +7,7 @@ #include <linux/mm_types.h> #include <linux/scatterlist.h> #include <linux/dma-debug.h> #include <linux/dma-mapping-fast.h> #include <linux/kref.h> #define ARM_MAPPING_ERROR (~(dma_addr_t)0x0) Loading arch/arm/mm/dma-mapping.c +5 −5 Original line number Diff line number Diff line Loading @@ -375,10 +375,10 @@ __dma_alloc_remap(struct page *page, size_t size, gfp_t gfp, pgprot_t prot, prot, caller); } static void __dma_free_remap(void *cpu_addr, size_t size) static void __dma_free_remap(void *cpu_addr, size_t size, bool no_warn) { dma_common_free_remap(cpu_addr, size, VM_ARM_DMA_CONSISTENT | VM_USERMAP); VM_ARM_DMA_CONSISTENT | VM_USERMAP, no_warn); } #define DEFAULT_DMA_COHERENT_POOL_SIZE SZ_256K Loading Loading @@ -624,7 +624,7 @@ static void __free_from_contiguous(struct device *dev, struct page *page, { if (want_vaddr) { if (PageHighMem(page)) __dma_free_remap(cpu_addr, size); __dma_free_remap(cpu_addr, size, true); else __dma_remap(page, size, PAGE_KERNEL); } Loading Loading @@ -716,7 +716,7 @@ static void *remap_allocator_alloc(struct arm_dma_alloc_args *args, static void remap_allocator_free(struct arm_dma_free_args *args) { if (args->want_vaddr) __dma_free_remap(args->cpu_addr, args->size); __dma_free_remap(args->cpu_addr, args->size, false); __dma_free_buffer(args->page, args->size); } Loading Loading @@ -1648,7 +1648,7 @@ void __arm_iommu_free_attrs(struct device *dev, size_t size, void *cpu_addr, if ((attrs & DMA_ATTR_NO_KERNEL_MAPPING) == 0) { dma_common_free_remap(cpu_addr, size, VM_ARM_DMA_CONSISTENT | VM_USERMAP); VM_ARM_DMA_CONSISTENT | VM_USERMAP, true); } __iommu_remove_mapping(dev, handle, size); Loading Loading
arch/arm/include/asm/dma-iommu.h +1 −0 Original line number Diff line number Diff line Loading @@ -7,6 +7,7 @@ #include <linux/mm_types.h> #include <linux/scatterlist.h> #include <linux/dma-debug.h> #include <linux/dma-mapping-fast.h> #include <linux/kref.h> #define ARM_MAPPING_ERROR (~(dma_addr_t)0x0) Loading
arch/arm/mm/dma-mapping.c +5 −5 Original line number Diff line number Diff line Loading @@ -375,10 +375,10 @@ __dma_alloc_remap(struct page *page, size_t size, gfp_t gfp, pgprot_t prot, prot, caller); } static void __dma_free_remap(void *cpu_addr, size_t size) static void __dma_free_remap(void *cpu_addr, size_t size, bool no_warn) { dma_common_free_remap(cpu_addr, size, VM_ARM_DMA_CONSISTENT | VM_USERMAP); VM_ARM_DMA_CONSISTENT | VM_USERMAP, no_warn); } #define DEFAULT_DMA_COHERENT_POOL_SIZE SZ_256K Loading Loading @@ -624,7 +624,7 @@ static void __free_from_contiguous(struct device *dev, struct page *page, { if (want_vaddr) { if (PageHighMem(page)) __dma_free_remap(cpu_addr, size); __dma_free_remap(cpu_addr, size, true); else __dma_remap(page, size, PAGE_KERNEL); } Loading Loading @@ -716,7 +716,7 @@ static void *remap_allocator_alloc(struct arm_dma_alloc_args *args, static void remap_allocator_free(struct arm_dma_free_args *args) { if (args->want_vaddr) __dma_free_remap(args->cpu_addr, args->size); __dma_free_remap(args->cpu_addr, args->size, false); __dma_free_buffer(args->page, args->size); } Loading Loading @@ -1648,7 +1648,7 @@ void __arm_iommu_free_attrs(struct device *dev, size_t size, void *cpu_addr, if ((attrs & DMA_ATTR_NO_KERNEL_MAPPING) == 0) { dma_common_free_remap(cpu_addr, size, VM_ARM_DMA_CONSISTENT | VM_USERMAP); VM_ARM_DMA_CONSISTENT | VM_USERMAP, true); } __iommu_remove_mapping(dev, handle, size); Loading