Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 71b0bfe4 authored by Li Zhong's avatar Li Zhong Committed by Benjamin Herrenschmidt
Browse files

powerpc: implement vmemmap_free()



vmemmap_free() does the opposite of vmemap_populate().
This patch also puts vmemmap_free() and vmemmap_list_free() into
 CONFIG_MEMMORY_HOTPLUG.

Signed-off-by: default avatarLi Zhong <zhong@linux.vnet.ibm.com>
Cc: Nathan Fontenot <nfont@linux.vnet.ibm.com>
Acked-by: default avatarNathan Fontenot <nfont@linux.vnet.ibm.com>
Signed-off-by: default avatarBenjamin Herrenschmidt <benh@kernel.crashing.org>
parent ed5694a8
Loading
Loading
Loading
Loading
+64 −21
Original line number Original line Diff line number Diff line
@@ -298,6 +298,37 @@ static __meminit void vmemmap_list_populate(unsigned long phys,
	vmemmap_list = vmem_back;
	vmemmap_list = vmem_back;
}
}


int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node)
{
	unsigned long page_size = 1 << mmu_psize_defs[mmu_vmemmap_psize].shift;

	/* Align to the page size of the linear mapping. */
	start = _ALIGN_DOWN(start, page_size);

	pr_debug("vmemmap_populate %lx..%lx, node %d\n", start, end, node);

	for (; start < end; start += page_size) {
		void *p;

		if (vmemmap_populated(start, page_size))
			continue;

		p = vmemmap_alloc_block(page_size, node);
		if (!p)
			return -ENOMEM;

		vmemmap_list_populate(__pa(p), start, node);

		pr_debug("      * %016lx..%016lx allocated at %p\n",
			 start, start + page_size, p);

		vmemmap_create_mapping(start, page_size, __pa(p));
	}

	return 0;
}

#ifdef CONFIG_MEMORY_HOTPLUG
static unsigned long vmemmap_list_free(unsigned long start)
static unsigned long vmemmap_list_free(unsigned long start)
{
{
	struct vmemmap_backing *vmem_back, *vmem_back_prev;
	struct vmemmap_backing *vmem_back, *vmem_back_prev;
@@ -330,40 +361,52 @@ static unsigned long vmemmap_list_free(unsigned long start)
	return vmem_back->phys;
	return vmem_back->phys;
}
}


int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node)
void __ref vmemmap_free(unsigned long start, unsigned long end)
{
{
	unsigned long page_size = 1 << mmu_psize_defs[mmu_vmemmap_psize].shift;
	unsigned long page_size = 1 << mmu_psize_defs[mmu_vmemmap_psize].shift;


	/* Align to the page size of the linear mapping. */
	start = _ALIGN_DOWN(start, page_size);
	start = _ALIGN_DOWN(start, page_size);


	pr_debug("vmemmap_populate %lx..%lx, node %d\n", start, end, node);
	pr_debug("vmemmap_free %lx...%lx\n", start, end);


	for (; start < end; start += page_size) {
	for (; start < end; start += page_size) {
		void *p;
		unsigned long addr;


		/*
		 * the section has already be marked as invalid, so
		 * vmemmap_populated() true means some other sections still
		 * in this page, so skip it.
		 */
		if (vmemmap_populated(start, page_size))
		if (vmemmap_populated(start, page_size))
			continue;
			continue;


		p = vmemmap_alloc_block(page_size, node);
		addr = vmemmap_list_free(start);
		if (!p)
		if (addr) {
			return -ENOMEM;
			struct page *page = pfn_to_page(addr >> PAGE_SHIFT);


		vmemmap_list_populate(__pa(p), start, node);
			if (PageReserved(page)) {

				/* allocated from bootmem */
		pr_debug("      * %016lx..%016lx allocated at %p\n",
				if (page_size < PAGE_SIZE) {
			 start, start + page_size, p);
					/*

					 * this shouldn't happen, but if it is
		vmemmap_create_mapping(start, page_size, __pa(p));
					 * the case, leave the memory there
					 */
					WARN_ON_ONCE(1);
				} else {
					unsigned int nr_pages =
						1 << get_order(page_size);
					while (nr_pages--)
						free_reserved_page(page++);
				}
				}
			} else
				free_pages((unsigned long)(__va(addr)),
							get_order(page_size));


	return 0;
			vmemmap_remove_mapping(start, page_size);
		}
		}

void vmemmap_free(unsigned long start, unsigned long end)
{
	}
	}

}
#endif
void register_page_bootmem_memmap(unsigned long section_nr,
void register_page_bootmem_memmap(unsigned long section_nr,
				  struct page *start_page, unsigned long size)
				  struct page *start_page, unsigned long size)
{
{