Loading arch/microblaze/include/asm/dma-mapping.h +2 −2 Original line number Diff line number Diff line Loading @@ -18,11 +18,11 @@ /* * Available generic sets of operations */ extern const struct dma_map_ops dma_direct_ops; extern const struct dma_map_ops dma_nommu_ops; static inline const struct dma_map_ops *get_arch_dma_ops(struct bus_type *bus) { return &dma_direct_ops; return &dma_nommu_ops; } #endif /* _ASM_MICROBLAZE_DMA_MAPPING_H */ arch/microblaze/kernel/dma.c +24 −24 Original line number Diff line number Diff line Loading @@ -17,7 +17,7 @@ #define NOT_COHERENT_CACHE static void *dma_direct_alloc_coherent(struct device *dev, size_t size, static void *dma_nommu_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_handle, gfp_t flag, unsigned long attrs) { Loading @@ -42,7 +42,7 @@ static void *dma_direct_alloc_coherent(struct device *dev, size_t size, #endif } static void dma_direct_free_coherent(struct device *dev, size_t size, static void dma_nommu_free_coherent(struct device *dev, size_t size, void *vaddr, dma_addr_t dma_handle, unsigned long attrs) { Loading @@ -69,7 +69,7 @@ static inline void __dma_sync(unsigned long paddr, } } static int dma_direct_map_sg(struct device *dev, struct scatterlist *sgl, static int dma_nommu_map_sg(struct device *dev, struct scatterlist *sgl, int nents, enum dma_data_direction direction, unsigned long attrs) { Loading @@ -89,12 +89,12 @@ static int dma_direct_map_sg(struct device *dev, struct scatterlist *sgl, return nents; } static int dma_direct_dma_supported(struct device *dev, u64 mask) static int dma_nommu_dma_supported(struct device *dev, u64 mask) { return 1; } static inline dma_addr_t dma_direct_map_page(struct device *dev, static inline dma_addr_t dma_nommu_map_page(struct device *dev, struct page *page, unsigned long offset, size_t size, Loading @@ -106,7 +106,7 @@ static inline dma_addr_t dma_direct_map_page(struct device *dev, return page_to_phys(page) + offset; } static inline void dma_direct_unmap_page(struct device *dev, static inline void dma_nommu_unmap_page(struct device *dev, dma_addr_t dma_address, size_t size, enum dma_data_direction direction, Loading @@ -122,7 +122,7 @@ static inline void dma_direct_unmap_page(struct device *dev, } static inline void dma_direct_sync_single_for_cpu(struct device *dev, dma_nommu_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction) { Loading @@ -136,7 +136,7 @@ dma_direct_sync_single_for_cpu(struct device *dev, } static inline void dma_direct_sync_single_for_device(struct device *dev, dma_nommu_sync_single_for_device(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction) { Loading @@ -150,7 +150,7 @@ dma_direct_sync_single_for_device(struct device *dev, } static inline void dma_direct_sync_sg_for_cpu(struct device *dev, dma_nommu_sync_sg_for_cpu(struct device *dev, struct scatterlist *sgl, int nents, enum dma_data_direction direction) { Loading @@ -164,7 +164,7 @@ dma_direct_sync_sg_for_cpu(struct device *dev, } static inline void dma_direct_sync_sg_for_device(struct device *dev, dma_nommu_sync_sg_for_device(struct device *dev, struct scatterlist *sgl, int nents, enum dma_data_direction direction) { Loading @@ -178,7 +178,7 @@ dma_direct_sync_sg_for_device(struct device *dev, } static int dma_direct_mmap_coherent(struct device *dev, struct vm_area_struct *vma, int dma_nommu_mmap_coherent(struct device *dev, struct vm_area_struct *vma, void *cpu_addr, dma_addr_t handle, size_t size, unsigned long attrs) { Loading @@ -204,20 +204,20 @@ int dma_direct_mmap_coherent(struct device *dev, struct vm_area_struct *vma, #endif } const struct dma_map_ops dma_direct_ops = { .alloc = dma_direct_alloc_coherent, .free = dma_direct_free_coherent, .mmap = dma_direct_mmap_coherent, .map_sg = dma_direct_map_sg, .dma_supported = dma_direct_dma_supported, .map_page = dma_direct_map_page, .unmap_page = dma_direct_unmap_page, .sync_single_for_cpu = dma_direct_sync_single_for_cpu, .sync_single_for_device = dma_direct_sync_single_for_device, .sync_sg_for_cpu = dma_direct_sync_sg_for_cpu, .sync_sg_for_device = dma_direct_sync_sg_for_device, const struct dma_map_ops dma_nommu_ops = { .alloc = dma_nommu_alloc_coherent, .free = dma_nommu_free_coherent, .mmap = dma_nommu_mmap_coherent, .map_sg = dma_nommu_map_sg, .dma_supported = dma_nommu_dma_supported, .map_page = dma_nommu_map_page, .unmap_page = dma_nommu_unmap_page, .sync_single_for_cpu = dma_nommu_sync_single_for_cpu, .sync_single_for_device = dma_nommu_sync_single_for_device, .sync_sg_for_cpu = dma_nommu_sync_sg_for_cpu, .sync_sg_for_device = dma_nommu_sync_sg_for_device, }; EXPORT_SYMBOL(dma_direct_ops); EXPORT_SYMBOL(dma_nommu_ops); /* Number of entries preallocated for DMA-API debugging */ #define PREALLOC_DMA_DEBUG_ENTRIES (1 << 16) Loading Loading
arch/microblaze/include/asm/dma-mapping.h +2 −2 Original line number Diff line number Diff line Loading @@ -18,11 +18,11 @@ /* * Available generic sets of operations */ extern const struct dma_map_ops dma_direct_ops; extern const struct dma_map_ops dma_nommu_ops; static inline const struct dma_map_ops *get_arch_dma_ops(struct bus_type *bus) { return &dma_direct_ops; return &dma_nommu_ops; } #endif /* _ASM_MICROBLAZE_DMA_MAPPING_H */
arch/microblaze/kernel/dma.c +24 −24 Original line number Diff line number Diff line Loading @@ -17,7 +17,7 @@ #define NOT_COHERENT_CACHE static void *dma_direct_alloc_coherent(struct device *dev, size_t size, static void *dma_nommu_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_handle, gfp_t flag, unsigned long attrs) { Loading @@ -42,7 +42,7 @@ static void *dma_direct_alloc_coherent(struct device *dev, size_t size, #endif } static void dma_direct_free_coherent(struct device *dev, size_t size, static void dma_nommu_free_coherent(struct device *dev, size_t size, void *vaddr, dma_addr_t dma_handle, unsigned long attrs) { Loading @@ -69,7 +69,7 @@ static inline void __dma_sync(unsigned long paddr, } } static int dma_direct_map_sg(struct device *dev, struct scatterlist *sgl, static int dma_nommu_map_sg(struct device *dev, struct scatterlist *sgl, int nents, enum dma_data_direction direction, unsigned long attrs) { Loading @@ -89,12 +89,12 @@ static int dma_direct_map_sg(struct device *dev, struct scatterlist *sgl, return nents; } static int dma_direct_dma_supported(struct device *dev, u64 mask) static int dma_nommu_dma_supported(struct device *dev, u64 mask) { return 1; } static inline dma_addr_t dma_direct_map_page(struct device *dev, static inline dma_addr_t dma_nommu_map_page(struct device *dev, struct page *page, unsigned long offset, size_t size, Loading @@ -106,7 +106,7 @@ static inline dma_addr_t dma_direct_map_page(struct device *dev, return page_to_phys(page) + offset; } static inline void dma_direct_unmap_page(struct device *dev, static inline void dma_nommu_unmap_page(struct device *dev, dma_addr_t dma_address, size_t size, enum dma_data_direction direction, Loading @@ -122,7 +122,7 @@ static inline void dma_direct_unmap_page(struct device *dev, } static inline void dma_direct_sync_single_for_cpu(struct device *dev, dma_nommu_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction) { Loading @@ -136,7 +136,7 @@ dma_direct_sync_single_for_cpu(struct device *dev, } static inline void dma_direct_sync_single_for_device(struct device *dev, dma_nommu_sync_single_for_device(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction) { Loading @@ -150,7 +150,7 @@ dma_direct_sync_single_for_device(struct device *dev, } static inline void dma_direct_sync_sg_for_cpu(struct device *dev, dma_nommu_sync_sg_for_cpu(struct device *dev, struct scatterlist *sgl, int nents, enum dma_data_direction direction) { Loading @@ -164,7 +164,7 @@ dma_direct_sync_sg_for_cpu(struct device *dev, } static inline void dma_direct_sync_sg_for_device(struct device *dev, dma_nommu_sync_sg_for_device(struct device *dev, struct scatterlist *sgl, int nents, enum dma_data_direction direction) { Loading @@ -178,7 +178,7 @@ dma_direct_sync_sg_for_device(struct device *dev, } static int dma_direct_mmap_coherent(struct device *dev, struct vm_area_struct *vma, int dma_nommu_mmap_coherent(struct device *dev, struct vm_area_struct *vma, void *cpu_addr, dma_addr_t handle, size_t size, unsigned long attrs) { Loading @@ -204,20 +204,20 @@ int dma_direct_mmap_coherent(struct device *dev, struct vm_area_struct *vma, #endif } const struct dma_map_ops dma_direct_ops = { .alloc = dma_direct_alloc_coherent, .free = dma_direct_free_coherent, .mmap = dma_direct_mmap_coherent, .map_sg = dma_direct_map_sg, .dma_supported = dma_direct_dma_supported, .map_page = dma_direct_map_page, .unmap_page = dma_direct_unmap_page, .sync_single_for_cpu = dma_direct_sync_single_for_cpu, .sync_single_for_device = dma_direct_sync_single_for_device, .sync_sg_for_cpu = dma_direct_sync_sg_for_cpu, .sync_sg_for_device = dma_direct_sync_sg_for_device, const struct dma_map_ops dma_nommu_ops = { .alloc = dma_nommu_alloc_coherent, .free = dma_nommu_free_coherent, .mmap = dma_nommu_mmap_coherent, .map_sg = dma_nommu_map_sg, .dma_supported = dma_nommu_dma_supported, .map_page = dma_nommu_map_page, .unmap_page = dma_nommu_unmap_page, .sync_single_for_cpu = dma_nommu_sync_single_for_cpu, .sync_single_for_device = dma_nommu_sync_single_for_device, .sync_sg_for_cpu = dma_nommu_sync_sg_for_cpu, .sync_sg_for_device = dma_nommu_sync_sg_for_device, }; EXPORT_SYMBOL(dma_direct_ops); EXPORT_SYMBOL(dma_nommu_ops); /* Number of entries preallocated for DMA-API debugging */ #define PREALLOC_DMA_DEBUG_ENTRIES (1 << 16) Loading