Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit bfa00489 authored by Marek Szyprowski's avatar Marek Szyprowski Committed by Joerg Roedel
Browse files

iommu/exynos: Rename variables to reflect their purpose



This patch renames some variables to make the code easier to understand.
'domain' is replaced by 'iommu_domain' (more generic entity) and really
meaningless 'priv' by 'domain' to reflect its purpose.

Signed-off-by: default avatarMarek Szyprowski <m.szyprowski@samsung.com>
Tested-by: default avatarJavier Martinez Canillas <javier.martinez@collabora.co.uk>
Signed-off-by: default avatarJoerg Roedel <jroedel@suse.de>
parent 469acebe
Loading
Loading
Loading
Loading
+94 −93
Original line number Original line Diff line number Diff line
@@ -435,8 +435,8 @@ static void __sysmmu_enable_nocount(struct sysmmu_drvdata *data)
		clk_disable(data->clk_master);
		clk_disable(data->clk_master);
}
}


static int __sysmmu_enable(struct sysmmu_drvdata *data,
static int __sysmmu_enable(struct sysmmu_drvdata *data, phys_addr_t pgtable,
			phys_addr_t pgtable, struct iommu_domain *domain)
			   struct iommu_domain *iommu_domain)
{
{
	int ret = 0;
	int ret = 0;
	unsigned long flags;
	unsigned long flags;
@@ -444,7 +444,7 @@ static int __sysmmu_enable(struct sysmmu_drvdata *data,
	spin_lock_irqsave(&data->lock, flags);
	spin_lock_irqsave(&data->lock, flags);
	if (set_sysmmu_active(data)) {
	if (set_sysmmu_active(data)) {
		data->pgtable = pgtable;
		data->pgtable = pgtable;
		data->domain = domain;
		data->domain = iommu_domain;


		__sysmmu_enable_nocount(data);
		__sysmmu_enable_nocount(data);


@@ -609,91 +609,91 @@ static inline void pgtable_flush(void *vastart, void *vaend)


static struct iommu_domain *exynos_iommu_domain_alloc(unsigned type)
static struct iommu_domain *exynos_iommu_domain_alloc(unsigned type)
{
{
	struct exynos_iommu_domain *exynos_domain;
	struct exynos_iommu_domain *domain;
	int i;
	int i;


	if (type != IOMMU_DOMAIN_UNMANAGED)
	if (type != IOMMU_DOMAIN_UNMANAGED)
		return NULL;
		return NULL;


	exynos_domain = kzalloc(sizeof(*exynos_domain), GFP_KERNEL);
	domain = kzalloc(sizeof(*domain), GFP_KERNEL);
	if (!exynos_domain)
	if (!domain)
		return NULL;
		return NULL;


	exynos_domain->pgtable = (sysmmu_pte_t *)__get_free_pages(GFP_KERNEL, 2);
	domain->pgtable = (sysmmu_pte_t *)__get_free_pages(GFP_KERNEL, 2);
	if (!exynos_domain->pgtable)
	if (!domain->pgtable)
		goto err_pgtable;
		goto err_pgtable;


	exynos_domain->lv2entcnt = (short *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, 1);
	domain->lv2entcnt = (short *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, 1);
	if (!exynos_domain->lv2entcnt)
	if (!domain->lv2entcnt)
		goto err_counter;
		goto err_counter;


	/* Workaround for System MMU v3.3 to prevent caching 1MiB mapping */
	/* Workaround for System MMU v3.3 to prevent caching 1MiB mapping */
	for (i = 0; i < NUM_LV1ENTRIES; i += 8) {
	for (i = 0; i < NUM_LV1ENTRIES; i += 8) {
		exynos_domain->pgtable[i + 0] = ZERO_LV2LINK;
		domain->pgtable[i + 0] = ZERO_LV2LINK;
		exynos_domain->pgtable[i + 1] = ZERO_LV2LINK;
		domain->pgtable[i + 1] = ZERO_LV2LINK;
		exynos_domain->pgtable[i + 2] = ZERO_LV2LINK;
		domain->pgtable[i + 2] = ZERO_LV2LINK;
		exynos_domain->pgtable[i + 3] = ZERO_LV2LINK;
		domain->pgtable[i + 3] = ZERO_LV2LINK;
		exynos_domain->pgtable[i + 4] = ZERO_LV2LINK;
		domain->pgtable[i + 4] = ZERO_LV2LINK;
		exynos_domain->pgtable[i + 5] = ZERO_LV2LINK;
		domain->pgtable[i + 5] = ZERO_LV2LINK;
		exynos_domain->pgtable[i + 6] = ZERO_LV2LINK;
		domain->pgtable[i + 6] = ZERO_LV2LINK;
		exynos_domain->pgtable[i + 7] = ZERO_LV2LINK;
		domain->pgtable[i + 7] = ZERO_LV2LINK;
	}
	}


	pgtable_flush(exynos_domain->pgtable, exynos_domain->pgtable + NUM_LV1ENTRIES);
	pgtable_flush(domain->pgtable, domain->pgtable + NUM_LV1ENTRIES);


	spin_lock_init(&exynos_domain->lock);
	spin_lock_init(&domain->lock);
	spin_lock_init(&exynos_domain->pgtablelock);
	spin_lock_init(&domain->pgtablelock);
	INIT_LIST_HEAD(&exynos_domain->clients);
	INIT_LIST_HEAD(&domain->clients);


	exynos_domain->domain.geometry.aperture_start = 0;
	domain->domain.geometry.aperture_start = 0;
	exynos_domain->domain.geometry.aperture_end   = ~0UL;
	domain->domain.geometry.aperture_end   = ~0UL;
	exynos_domain->domain.geometry.force_aperture = true;
	domain->domain.geometry.force_aperture = true;


	return &exynos_domain->domain;
	return &domain->domain;


err_counter:
err_counter:
	free_pages((unsigned long)exynos_domain->pgtable, 2);
	free_pages((unsigned long)domain->pgtable, 2);
err_pgtable:
err_pgtable:
	kfree(exynos_domain);
	kfree(domain);
	return NULL;
	return NULL;
}
}


static void exynos_iommu_domain_free(struct iommu_domain *domain)
static void exynos_iommu_domain_free(struct iommu_domain *iommu_domain)
{
{
	struct exynos_iommu_domain *priv = to_exynos_domain(domain);
	struct exynos_iommu_domain *domain = to_exynos_domain(iommu_domain);
	struct sysmmu_drvdata *data, *next;
	struct sysmmu_drvdata *data, *next;
	unsigned long flags;
	unsigned long flags;
	int i;
	int i;


	WARN_ON(!list_empty(&priv->clients));
	WARN_ON(!list_empty(&domain->clients));


	spin_lock_irqsave(&priv->lock, flags);
	spin_lock_irqsave(&domain->lock, flags);


	list_for_each_entry_safe(data, next, &priv->clients, domain_node) {
	list_for_each_entry_safe(data, next, &domain->clients, domain_node) {
		if (__sysmmu_disable(data))
		if (__sysmmu_disable(data))
			data->master = NULL;
			data->master = NULL;
		list_del_init(&data->domain_node);
		list_del_init(&data->domain_node);
	}
	}


	spin_unlock_irqrestore(&priv->lock, flags);
	spin_unlock_irqrestore(&domain->lock, flags);


	for (i = 0; i < NUM_LV1ENTRIES; i++)
	for (i = 0; i < NUM_LV1ENTRIES; i++)
		if (lv1ent_page(priv->pgtable + i))
		if (lv1ent_page(domain->pgtable + i))
			kmem_cache_free(lv2table_kmem_cache,
			kmem_cache_free(lv2table_kmem_cache,
				phys_to_virt(lv2table_base(priv->pgtable + i)));
				phys_to_virt(lv2table_base(domain->pgtable + i)));


	free_pages((unsigned long)priv->pgtable, 2);
	free_pages((unsigned long)domain->pgtable, 2);
	free_pages((unsigned long)priv->lv2entcnt, 1);
	free_pages((unsigned long)domain->lv2entcnt, 1);
	kfree(priv);
	kfree(domain);
}
}


static int exynos_iommu_attach_device(struct iommu_domain *domain,
static int exynos_iommu_attach_device(struct iommu_domain *iommu_domain,
				   struct device *dev)
				   struct device *dev)
{
{
	struct exynos_iommu_owner *owner = dev->archdata.iommu;
	struct exynos_iommu_owner *owner = dev->archdata.iommu;
	struct exynos_iommu_domain *priv = to_exynos_domain(domain);
	struct exynos_iommu_domain *domain = to_exynos_domain(iommu_domain);
	struct sysmmu_drvdata *data;
	struct sysmmu_drvdata *data;
	phys_addr_t pagetable = virt_to_phys(priv->pgtable);
	phys_addr_t pagetable = virt_to_phys(domain->pgtable);
	unsigned long flags;
	unsigned long flags;
	int ret = -ENODEV;
	int ret = -ENODEV;


@@ -702,13 +702,13 @@ static int exynos_iommu_attach_device(struct iommu_domain *domain,


	data = dev_get_drvdata(owner->sysmmu);
	data = dev_get_drvdata(owner->sysmmu);
	if (data) {
	if (data) {
		ret = __sysmmu_enable(data, pagetable, domain);
		ret = __sysmmu_enable(data, pagetable, iommu_domain);
		if (ret >= 0) {
		if (ret >= 0) {
			data->master = dev;
			data->master = dev;


			spin_lock_irqsave(&priv->lock, flags);
			spin_lock_irqsave(&domain->lock, flags);
			list_add_tail(&data->domain_node, &priv->clients);
			list_add_tail(&data->domain_node, &domain->clients);
			spin_unlock_irqrestore(&priv->lock, flags);
			spin_unlock_irqrestore(&domain->lock, flags);
		}
		}
	}
	}


@@ -724,11 +724,11 @@ static int exynos_iommu_attach_device(struct iommu_domain *domain,
	return ret;
	return ret;
}
}


static void exynos_iommu_detach_device(struct iommu_domain *domain,
static void exynos_iommu_detach_device(struct iommu_domain *iommu_domain,
				    struct device *dev)
				    struct device *dev)
{
{
	struct exynos_iommu_domain *priv = to_exynos_domain(domain);
	struct exynos_iommu_domain *domain = to_exynos_domain(iommu_domain);
	phys_addr_t pagetable = virt_to_phys(priv->pgtable);
	phys_addr_t pagetable = virt_to_phys(domain->pgtable);
	struct sysmmu_drvdata *data;
	struct sysmmu_drvdata *data;
	unsigned long flags;
	unsigned long flags;
	bool found = false;
	bool found = false;
@@ -736,8 +736,8 @@ static void exynos_iommu_detach_device(struct iommu_domain *domain,
	if (!has_sysmmu(dev))
	if (!has_sysmmu(dev))
		return;
		return;


	spin_lock_irqsave(&priv->lock, flags);
	spin_lock_irqsave(&domain->lock, flags);
	list_for_each_entry(data, &priv->clients, domain_node) {
	list_for_each_entry(data, &domain->clients, domain_node) {
		if (data->master == dev) {
		if (data->master == dev) {
			if (__sysmmu_disable(data)) {
			if (__sysmmu_disable(data)) {
				data->master = NULL;
				data->master = NULL;
@@ -747,7 +747,7 @@ static void exynos_iommu_detach_device(struct iommu_domain *domain,
			break;
			break;
		}
		}
	}
	}
	spin_unlock_irqrestore(&priv->lock, flags);
	spin_unlock_irqrestore(&domain->lock, flags);


	if (found)
	if (found)
		dev_dbg(dev, "%s: Detached IOMMU with pgtable %pa\n",
		dev_dbg(dev, "%s: Detached IOMMU with pgtable %pa\n",
@@ -756,7 +756,7 @@ static void exynos_iommu_detach_device(struct iommu_domain *domain,
		dev_err(dev, "%s: No IOMMU is attached\n", __func__);
		dev_err(dev, "%s: No IOMMU is attached\n", __func__);
}
}


static sysmmu_pte_t *alloc_lv2entry(struct exynos_iommu_domain *priv,
static sysmmu_pte_t *alloc_lv2entry(struct exynos_iommu_domain *domain,
		sysmmu_pte_t *sent, sysmmu_iova_t iova, short *pgcounter)
		sysmmu_pte_t *sent, sysmmu_iova_t iova, short *pgcounter)
{
{
	if (lv1ent_section(sent)) {
	if (lv1ent_section(sent)) {
@@ -799,17 +799,17 @@ static sysmmu_pte_t *alloc_lv2entry(struct exynos_iommu_domain *priv,
		if (need_flush_flpd_cache) {
		if (need_flush_flpd_cache) {
			struct sysmmu_drvdata *data;
			struct sysmmu_drvdata *data;


			spin_lock(&priv->lock);
			spin_lock(&domain->lock);
			list_for_each_entry(data, &priv->clients, domain_node)
			list_for_each_entry(data, &domain->clients, domain_node)
				sysmmu_tlb_invalidate_flpdcache(data, iova);
				sysmmu_tlb_invalidate_flpdcache(data, iova);
			spin_unlock(&priv->lock);
			spin_unlock(&domain->lock);
		}
		}
	}
	}


	return page_entry(sent, iova);
	return page_entry(sent, iova);
}
}


static int lv1set_section(struct exynos_iommu_domain *priv,
static int lv1set_section(struct exynos_iommu_domain *domain,
			  sysmmu_pte_t *sent, sysmmu_iova_t iova,
			  sysmmu_pte_t *sent, sysmmu_iova_t iova,
			  phys_addr_t paddr, short *pgcnt)
			  phys_addr_t paddr, short *pgcnt)
{
{
@@ -834,17 +834,17 @@ static int lv1set_section(struct exynos_iommu_domain *priv,


	pgtable_flush(sent, sent + 1);
	pgtable_flush(sent, sent + 1);


	spin_lock(&priv->lock);
	spin_lock(&domain->lock);
	if (lv1ent_page_zero(sent)) {
	if (lv1ent_page_zero(sent)) {
		struct sysmmu_drvdata *data;
		struct sysmmu_drvdata *data;
		/*
		/*
		 * Flushing FLPD cache in System MMU v3.3 that may cache a FLPD
		 * Flushing FLPD cache in System MMU v3.3 that may cache a FLPD
		 * entry by speculative prefetch of SLPD which has no mapping.
		 * entry by speculative prefetch of SLPD which has no mapping.
		 */
		 */
		list_for_each_entry(data, &priv->clients, domain_node)
		list_for_each_entry(data, &domain->clients, domain_node)
			sysmmu_tlb_invalidate_flpdcache(data, iova);
			sysmmu_tlb_invalidate_flpdcache(data, iova);
	}
	}
	spin_unlock(&priv->lock);
	spin_unlock(&domain->lock);


	return 0;
	return 0;
}
}
@@ -904,74 +904,75 @@ static int lv2set_page(sysmmu_pte_t *pent, phys_addr_t paddr, size_t size,
 *   than or equal to 128KiB.
 *   than or equal to 128KiB.
 * - Start address of an I/O virtual region must be aligned by 128KiB.
 * - Start address of an I/O virtual region must be aligned by 128KiB.
 */
 */
static int exynos_iommu_map(struct iommu_domain *domain, unsigned long l_iova,
static int exynos_iommu_map(struct iommu_domain *iommu_domain,
			 phys_addr_t paddr, size_t size, int prot)
			    unsigned long l_iova, phys_addr_t paddr, size_t size,
			    int prot)
{
{
	struct exynos_iommu_domain *priv = to_exynos_domain(domain);
	struct exynos_iommu_domain *domain = to_exynos_domain(iommu_domain);
	sysmmu_pte_t *entry;
	sysmmu_pte_t *entry;
	sysmmu_iova_t iova = (sysmmu_iova_t)l_iova;
	sysmmu_iova_t iova = (sysmmu_iova_t)l_iova;
	unsigned long flags;
	unsigned long flags;
	int ret = -ENOMEM;
	int ret = -ENOMEM;


	BUG_ON(priv->pgtable == NULL);
	BUG_ON(domain->pgtable == NULL);


	spin_lock_irqsave(&priv->pgtablelock, flags);
	spin_lock_irqsave(&domain->pgtablelock, flags);


	entry = section_entry(priv->pgtable, iova);
	entry = section_entry(domain->pgtable, iova);


	if (size == SECT_SIZE) {
	if (size == SECT_SIZE) {
		ret = lv1set_section(priv, entry, iova, paddr,
		ret = lv1set_section(domain, entry, iova, paddr,
					&priv->lv2entcnt[lv1ent_offset(iova)]);
				     &domain->lv2entcnt[lv1ent_offset(iova)]);
	} else {
	} else {
		sysmmu_pte_t *pent;
		sysmmu_pte_t *pent;


		pent = alloc_lv2entry(priv, entry, iova,
		pent = alloc_lv2entry(domain, entry, iova,
					&priv->lv2entcnt[lv1ent_offset(iova)]);
				      &domain->lv2entcnt[lv1ent_offset(iova)]);


		if (IS_ERR(pent))
		if (IS_ERR(pent))
			ret = PTR_ERR(pent);
			ret = PTR_ERR(pent);
		else
		else
			ret = lv2set_page(pent, paddr, size,
			ret = lv2set_page(pent, paddr, size,
					&priv->lv2entcnt[lv1ent_offset(iova)]);
				       &domain->lv2entcnt[lv1ent_offset(iova)]);
	}
	}


	if (ret)
	if (ret)
		pr_err("%s: Failed(%d) to map %#zx bytes @ %#x\n",
		pr_err("%s: Failed(%d) to map %#zx bytes @ %#x\n",
			__func__, ret, size, iova);
			__func__, ret, size, iova);


	spin_unlock_irqrestore(&priv->pgtablelock, flags);
	spin_unlock_irqrestore(&domain->pgtablelock, flags);


	return ret;
	return ret;
}
}


static void exynos_iommu_tlb_invalidate_entry(struct exynos_iommu_domain *priv,
static void exynos_iommu_tlb_invalidate_entry(struct exynos_iommu_domain *domain,
					      sysmmu_iova_t iova, size_t size)
					      sysmmu_iova_t iova, size_t size)
{
{
	struct sysmmu_drvdata *data;
	struct sysmmu_drvdata *data;
	unsigned long flags;
	unsigned long flags;


	spin_lock_irqsave(&priv->lock, flags);
	spin_lock_irqsave(&domain->lock, flags);


	list_for_each_entry(data, &priv->clients, domain_node)
	list_for_each_entry(data, &domain->clients, domain_node)
		sysmmu_tlb_invalidate_entry(data, iova, size);
		sysmmu_tlb_invalidate_entry(data, iova, size);


	spin_unlock_irqrestore(&priv->lock, flags);
	spin_unlock_irqrestore(&domain->lock, flags);
}
}


static size_t exynos_iommu_unmap(struct iommu_domain *domain,
static size_t exynos_iommu_unmap(struct iommu_domain *iommu_domain,
				 unsigned long l_iova, size_t size)
				 unsigned long l_iova, size_t size)
{
{
	struct exynos_iommu_domain *priv = to_exynos_domain(domain);
	struct exynos_iommu_domain *domain = to_exynos_domain(iommu_domain);
	sysmmu_iova_t iova = (sysmmu_iova_t)l_iova;
	sysmmu_iova_t iova = (sysmmu_iova_t)l_iova;
	sysmmu_pte_t *ent;
	sysmmu_pte_t *ent;
	size_t err_pgsize;
	size_t err_pgsize;
	unsigned long flags;
	unsigned long flags;


	BUG_ON(priv->pgtable == NULL);
	BUG_ON(domain->pgtable == NULL);


	spin_lock_irqsave(&priv->pgtablelock, flags);
	spin_lock_irqsave(&domain->pgtablelock, flags);


	ent = section_entry(priv->pgtable, iova);
	ent = section_entry(domain->pgtable, iova);


	if (lv1ent_section(ent)) {
	if (lv1ent_section(ent)) {
		if (WARN_ON(size < SECT_SIZE)) {
		if (WARN_ON(size < SECT_SIZE)) {
@@ -1005,7 +1006,7 @@ static size_t exynos_iommu_unmap(struct iommu_domain *domain,
		*ent = 0;
		*ent = 0;
		size = SPAGE_SIZE;
		size = SPAGE_SIZE;
		pgtable_flush(ent, ent + 1);
		pgtable_flush(ent, ent + 1);
		priv->lv2entcnt[lv1ent_offset(iova)] += 1;
		domain->lv2entcnt[lv1ent_offset(iova)] += 1;
		goto done;
		goto done;
	}
	}


@@ -1019,15 +1020,15 @@ static size_t exynos_iommu_unmap(struct iommu_domain *domain,
	pgtable_flush(ent, ent + SPAGES_PER_LPAGE);
	pgtable_flush(ent, ent + SPAGES_PER_LPAGE);


	size = LPAGE_SIZE;
	size = LPAGE_SIZE;
	priv->lv2entcnt[lv1ent_offset(iova)] += SPAGES_PER_LPAGE;
	domain->lv2entcnt[lv1ent_offset(iova)] += SPAGES_PER_LPAGE;
done:
done:
	spin_unlock_irqrestore(&priv->pgtablelock, flags);
	spin_unlock_irqrestore(&domain->pgtablelock, flags);


	exynos_iommu_tlb_invalidate_entry(priv, iova, size);
	exynos_iommu_tlb_invalidate_entry(domain, iova, size);


	return size;
	return size;
err:
err:
	spin_unlock_irqrestore(&priv->pgtablelock, flags);
	spin_unlock_irqrestore(&domain->pgtablelock, flags);


	pr_err("%s: Failed: size(%#zx) @ %#x is smaller than page size %#zx\n",
	pr_err("%s: Failed: size(%#zx) @ %#x is smaller than page size %#zx\n",
		__func__, size, iova, err_pgsize);
		__func__, size, iova, err_pgsize);
@@ -1035,17 +1036,17 @@ static size_t exynos_iommu_unmap(struct iommu_domain *domain,
	return 0;
	return 0;
}
}


static phys_addr_t exynos_iommu_iova_to_phys(struct iommu_domain *domain,
static phys_addr_t exynos_iommu_iova_to_phys(struct iommu_domain *iommu_domain,
					  dma_addr_t iova)
					  dma_addr_t iova)
{
{
	struct exynos_iommu_domain *priv = to_exynos_domain(domain);
	struct exynos_iommu_domain *domain = to_exynos_domain(iommu_domain);
	sysmmu_pte_t *entry;
	sysmmu_pte_t *entry;
	unsigned long flags;
	unsigned long flags;
	phys_addr_t phys = 0;
	phys_addr_t phys = 0;


	spin_lock_irqsave(&priv->pgtablelock, flags);
	spin_lock_irqsave(&domain->pgtablelock, flags);


	entry = section_entry(priv->pgtable, iova);
	entry = section_entry(domain->pgtable, iova);


	if (lv1ent_section(entry)) {
	if (lv1ent_section(entry)) {
		phys = section_phys(entry) + section_offs(iova);
		phys = section_phys(entry) + section_offs(iova);
@@ -1058,7 +1059,7 @@ static phys_addr_t exynos_iommu_iova_to_phys(struct iommu_domain *domain,
			phys = spage_phys(entry) + spage_offs(iova);
			phys = spage_phys(entry) + spage_offs(iova);
	}
	}


	spin_unlock_irqrestore(&priv->pgtablelock, flags);
	spin_unlock_irqrestore(&domain->pgtablelock, flags);


	return phys;
	return phys;
}
}