Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 9cabe89b authored by Joerg Roedel's avatar Joerg Roedel
Browse files

amd-iommu: move aperture_range allocation code to seperate function



This patch prepares the dynamic increasement of dma_ops domain
apertures.

Signed-off-by: default avatarJoerg Roedel <joerg.roedel@amd.com>
parent 803b8cb4
Loading
Loading
Loading
Loading
+59 −36
Original line number Original line Diff line number Diff line
@@ -620,6 +620,59 @@ static int init_unity_mappings_for_device(struct dma_ops_domain *dma_dom,
 * called with domain->lock held
 * called with domain->lock held
 */
 */


/*
 * This function is used to add a new aperture range to an existing
 * aperture in case of dma_ops domain allocation or address allocation
 * failure.
 */
static int alloc_new_range(struct dma_ops_domain *dma_dom,
			   bool populate, gfp_t gfp)
{
	int index = dma_dom->aperture_size >> APERTURE_RANGE_SHIFT;

	if (index >= APERTURE_MAX_RANGES)
		return -ENOMEM;

	dma_dom->aperture[index] = kzalloc(sizeof(struct aperture_range), gfp);
	if (!dma_dom->aperture[index])
		return -ENOMEM;

	dma_dom->aperture[index]->bitmap = (void *)get_zeroed_page(gfp);
	if (!dma_dom->aperture[index]->bitmap)
		goto out_free;

	dma_dom->aperture[index]->offset = dma_dom->aperture_size;

	if (populate) {
		unsigned long address = dma_dom->aperture_size;
		int i, num_ptes = APERTURE_RANGE_PAGES / 512;
		u64 *pte, *pte_page;

		for (i = 0; i < num_ptes; ++i) {
			pte = alloc_pte(&dma_dom->domain, address,
					&pte_page, gfp);
			if (!pte)
				goto out_free;

			dma_dom->aperture[index]->pte_pages[i] = pte_page;

			address += APERTURE_RANGE_SIZE / 64;
		}
	}

	dma_dom->aperture_size += APERTURE_RANGE_SIZE;

	return 0;

out_free:
	free_page((unsigned long)dma_dom->aperture[index]->bitmap);

	kfree(dma_dom->aperture[index]);
	dma_dom->aperture[index] = NULL;

	return -ENOMEM;
}

static unsigned long dma_ops_area_alloc(struct device *dev,
static unsigned long dma_ops_area_alloc(struct device *dev,
					struct dma_ops_domain *dom,
					struct dma_ops_domain *dom,
					unsigned int pages,
					unsigned int pages,
@@ -832,9 +885,6 @@ static struct dma_ops_domain *dma_ops_domain_alloc(struct amd_iommu *iommu,
						   unsigned order)
						   unsigned order)
{
{
	struct dma_ops_domain *dma_dom;
	struct dma_ops_domain *dma_dom;
	unsigned i, num_pte_pages;
	u64 *l2_pde;
	u64 address;


	/*
	/*
	 * Currently the DMA aperture must be between 32 MB and 1GB in size
	 * Currently the DMA aperture must be between 32 MB and 1GB in size
@@ -846,11 +896,6 @@ static struct dma_ops_domain *dma_ops_domain_alloc(struct amd_iommu *iommu,
	if (!dma_dom)
	if (!dma_dom)
		return NULL;
		return NULL;


	dma_dom->aperture[0] = kzalloc(sizeof(struct aperture_range),
				       GFP_KERNEL);
	if (!dma_dom->aperture[0])
		goto free_dma_dom;

	spin_lock_init(&dma_dom->domain.lock);
	spin_lock_init(&dma_dom->domain.lock);


	dma_dom->domain.id = domain_id_alloc();
	dma_dom->domain.id = domain_id_alloc();
@@ -862,10 +907,13 @@ static struct dma_ops_domain *dma_ops_domain_alloc(struct amd_iommu *iommu,
	dma_dom->domain.priv = dma_dom;
	dma_dom->domain.priv = dma_dom;
	if (!dma_dom->domain.pt_root)
	if (!dma_dom->domain.pt_root)
		goto free_dma_dom;
		goto free_dma_dom;
	dma_dom->aperture_size = APERTURE_RANGE_SIZE;

	dma_dom->aperture[0]->bitmap = (void *)get_zeroed_page(GFP_KERNEL);
	dma_dom->need_flush = false;
	if (!dma_dom->aperture[0]->bitmap)
	dma_dom->target_dev = 0xffff;

	if (alloc_new_range(dma_dom, true, GFP_KERNEL))
		goto free_dma_dom;
		goto free_dma_dom;

	/*
	/*
	 * mark the first page as allocated so we never return 0 as
	 * mark the first page as allocated so we never return 0 as
	 * a valid dma-address. So we can use 0 as error value
	 * a valid dma-address. So we can use 0 as error value
@@ -873,9 +921,6 @@ static struct dma_ops_domain *dma_ops_domain_alloc(struct amd_iommu *iommu,
	dma_dom->aperture[0]->bitmap[0] = 1;
	dma_dom->aperture[0]->bitmap[0] = 1;
	dma_dom->next_address = 0;
	dma_dom->next_address = 0;


	dma_dom->need_flush = false;
	dma_dom->target_dev = 0xffff;

	/* Intialize the exclusion range if necessary */
	/* Intialize the exclusion range if necessary */
	if (iommu->exclusion_start &&
	if (iommu->exclusion_start &&
	    iommu->exclusion_start < dma_dom->aperture_size) {
	    iommu->exclusion_start < dma_dom->aperture_size) {
@@ -886,28 +931,6 @@ static struct dma_ops_domain *dma_ops_domain_alloc(struct amd_iommu *iommu,
		dma_ops_reserve_addresses(dma_dom, startpage, pages);
		dma_ops_reserve_addresses(dma_dom, startpage, pages);
	}
	}


	/*
	 * At the last step, build the page tables so we don't need to
	 * allocate page table pages in the dma_ops mapping/unmapping
	 * path for the first 128MB of dma address space.
	 */
	num_pte_pages = dma_dom->aperture_size / (PAGE_SIZE * 512);

	l2_pde = (u64 *)get_zeroed_page(GFP_KERNEL);
	if (l2_pde == NULL)
		goto free_dma_dom;

	dma_dom->domain.pt_root[0] = IOMMU_L2_PDE(virt_to_phys(l2_pde));

	for (i = 0; i < num_pte_pages; ++i) {
		u64 **pte_page = &dma_dom->aperture[0]->pte_pages[i];
		*pte_page = (u64 *)get_zeroed_page(GFP_KERNEL);
		if (!*pte_page)
			goto free_dma_dom;
		address = virt_to_phys(*pte_page);
		l2_pde[i] = IOMMU_L1_PDE(address);
	}

	return dma_dom;
	return dma_dom;


free_dma_dom:
free_dma_dom: