Loading drivers/staging/android/ion/msm/ion_iommu_map.c +3 −3 Original line number Diff line number Diff line Loading @@ -208,8 +208,8 @@ static int ion_iommu_map_iommu(struct ion_iommu_meta *meta, * biggest entry. To take advantage of bigger mapping sizes both the * VA and PA addresses have to be aligned to the biggest size. */ if (sg_dma_len(table->sgl) > align) align = sg_dma_len(table->sgl); if (table->sgl->length > align) align = table->sgl->length; ret = msm_allocate_iova_address(domain_num, partition_num, data->mapped_size, align, Loading Loading @@ -394,7 +394,7 @@ int ion_map_iommu(struct ion_client *client, struct ion_handle *handle, return PTR_ERR(table); for_each_sg(table->sgl, sg, table->nents, i) size += sg_dma_len(sg); size += sg->length; if (!msm_use_iommu()) { unsigned long pa = sg_dma_address(table->sgl); Loading Loading
drivers/staging/android/ion/msm/ion_iommu_map.c +3 −3 Original line number Diff line number Diff line Loading @@ -208,8 +208,8 @@ static int ion_iommu_map_iommu(struct ion_iommu_meta *meta, * biggest entry. To take advantage of bigger mapping sizes both the * VA and PA addresses have to be aligned to the biggest size. */ if (sg_dma_len(table->sgl) > align) align = sg_dma_len(table->sgl); if (table->sgl->length > align) align = table->sgl->length; ret = msm_allocate_iova_address(domain_num, partition_num, data->mapped_size, align, Loading Loading @@ -394,7 +394,7 @@ int ion_map_iommu(struct ion_client *client, struct ion_handle *handle, return PTR_ERR(table); for_each_sg(table->sgl, sg, table->nents, i) size += sg_dma_len(sg); size += sg->length; if (!msm_use_iommu()) { unsigned long pa = sg_dma_address(table->sgl); Loading