Loading drivers/iommu/arm-smmu.c +0 −10 Original line number Diff line number Diff line Loading @@ -3021,14 +3021,6 @@ static size_t arm_smmu_unmap(struct iommu_domain *domain, unsigned long iova, return ret; } static void arm_smmu_iotlb_sync(struct iommu_domain *domain) { struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain); if (smmu_domain->tlb_ops) smmu_domain->tlb_ops->tlb_sync(smmu_domain); } #define MAX_MAP_SG_BATCH_SIZE (SZ_4M) static size_t arm_smmu_map_sg(struct iommu_domain *domain, unsigned long iova, struct scatterlist *sg, unsigned int nents, int prot) Loading Loading @@ -4000,8 +3992,6 @@ static struct iommu_ops arm_smmu_ops = { .map = arm_smmu_map, .unmap = arm_smmu_unmap, .map_sg = arm_smmu_map_sg, .flush_iotlb_all = arm_smmu_iotlb_sync, .iotlb_sync = arm_smmu_iotlb_sync, .iova_to_phys = arm_smmu_iova_to_phys, .iova_to_phys_hard = arm_smmu_iova_to_phys_hard, .add_device = arm_smmu_add_device, Loading Loading
drivers/iommu/arm-smmu.c +0 −10 Original line number Diff line number Diff line Loading @@ -3021,14 +3021,6 @@ static size_t arm_smmu_unmap(struct iommu_domain *domain, unsigned long iova, return ret; } static void arm_smmu_iotlb_sync(struct iommu_domain *domain) { struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain); if (smmu_domain->tlb_ops) smmu_domain->tlb_ops->tlb_sync(smmu_domain); } #define MAX_MAP_SG_BATCH_SIZE (SZ_4M) static size_t arm_smmu_map_sg(struct iommu_domain *domain, unsigned long iova, struct scatterlist *sg, unsigned int nents, int prot) Loading Loading @@ -4000,8 +3992,6 @@ static struct iommu_ops arm_smmu_ops = { .map = arm_smmu_map, .unmap = arm_smmu_unmap, .map_sg = arm_smmu_map_sg, .flush_iotlb_all = arm_smmu_iotlb_sync, .iotlb_sync = arm_smmu_iotlb_sync, .iova_to_phys = arm_smmu_iova_to_phys, .iova_to_phys_hard = arm_smmu_iova_to_phys_hard, .add_device = arm_smmu_add_device, Loading