Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 8a9ca34c authored by Tejun Heo's avatar Tejun Heo Committed by H. Peter Anvin
Browse files

memblock, x86: Replace __get_free_all_memory_range() with for_each_free_mem_range()



__get_free_all_memory_range() walks memblock, calculates free memory
areas and fills in the specified range.  It can be easily replaced
with for_each_free_mem_range().

Convert free_low_memory_core_early() and
add_highpages_with_active_regions() to for_each_free_mem_range().
This leaves __get_free_all_memory_range() without any user.  Kill it
and related functions.

Signed-off-by: default avatarTejun Heo <tj@kernel.org>
Link: http://lkml.kernel.org/r/1310462166-31469-10-git-send-email-tj@kernel.org


Cc: Yinghai Lu <yinghai@kernel.org>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: "H. Peter Anvin" <hpa@zytor.com>
Signed-off-by: default avatarH. Peter Anvin <hpa@linux.intel.com>
parent 64a02daa
Loading
Loading
Loading
Loading
+0 −3
Original line number Original line Diff line number Diff line
@@ -5,9 +5,6 @@


void memblock_x86_reserve_range(u64 start, u64 end, char *name);
void memblock_x86_reserve_range(u64 start, u64 end, char *name);
void memblock_x86_free_range(u64 start, u64 end);
void memblock_x86_free_range(u64 start, u64 end);
struct range;
int __get_free_all_memory_range(struct range **range, int nodeid,
			 unsigned long start_pfn, unsigned long end_pfn);


u64 memblock_x86_hole_size(u64 start, u64 end);
u64 memblock_x86_hole_size(u64 start, u64 end);
u64 memblock_x86_free_memory_in_range(u64 addr, u64 limit);
u64 memblock_x86_free_memory_in_range(u64 addr, u64 limit);
+11 −17
Original line number Original line Diff line number Diff line
@@ -427,23 +427,17 @@ static void __init add_one_highpage_init(struct page *page)
void __init add_highpages_with_active_regions(int nid,
void __init add_highpages_with_active_regions(int nid,
			 unsigned long start_pfn, unsigned long end_pfn)
			 unsigned long start_pfn, unsigned long end_pfn)
{
{
	struct range *range;
	phys_addr_t start, end;
	int nr_range;
	u64 i;
	int i;


	for_each_free_mem_range(i, nid, &start, &end, NULL) {
	nr_range = __get_free_all_memory_range(&range, nid, start_pfn, end_pfn);
		unsigned long pfn = clamp_t(unsigned long, PFN_UP(start),

					    start_pfn, end_pfn);
	for (i = 0; i < nr_range; i++) {
		unsigned long e_pfn = clamp_t(unsigned long, PFN_DOWN(end),
		struct page *page;
					      start_pfn, end_pfn);
		int node_pfn;
		for ( ; pfn < e_pfn; pfn++)

			if (pfn_valid(pfn))
		for (node_pfn = range[i].start; node_pfn < range[i].end;
				add_one_highpage_init(pfn_to_page(pfn));
		     node_pfn++) {
			if (!pfn_valid(node_pfn))
				continue;
			page = pfn_to_page(node_pfn);
			add_one_highpage_init(page);
		}
	}
	}
}
}
#else
#else
+0 −59
Original line number Original line Diff line number Diff line
@@ -30,65 +30,6 @@ static __init struct range *find_range_array(int count)
	return range;
	return range;
}
}


static void __init memblock_x86_subtract_reserved(struct range *range, int az)
{
	u64 final_start, final_end;
	struct memblock_region *r;

	/* Take out region array itself at first*/
	memblock_free_reserved_regions();

	memblock_dbg("Subtract (%ld early reservations)\n", memblock.reserved.cnt);

	for_each_memblock(reserved, r) {
		memblock_dbg("  [%010llx-%010llx]\n", (u64)r->base, (u64)r->base + r->size - 1);
		final_start = PFN_DOWN(r->base);
		final_end = PFN_UP(r->base + r->size);
		if (final_start >= final_end)
			continue;
		subtract_range(range, az, final_start, final_end);
	}

	/* Put region array back ? */
	memblock_reserve_reserved_regions();
}

static int __init count_early_node_map(int nodeid)
{
	int i, cnt = 0;

	for_each_mem_pfn_range(i, nodeid, NULL, NULL, NULL)
		cnt++;
	return cnt;
}

int __init __get_free_all_memory_range(struct range **rangep, int nodeid,
			 unsigned long start_pfn, unsigned long end_pfn)
{
	int count;
	struct range *range;
	int nr_range;

	count = (memblock.reserved.cnt + count_early_node_map(nodeid)) * 2;

	range = find_range_array(count);
	nr_range = 0;

	/*
	 * Use early_node_map[] and memblock.reserved.region to get range array
	 * at first
	 */
	nr_range = add_from_early_node_map(range, count, nr_range, nodeid);
	subtract_range(range, count, 0, start_pfn);
	subtract_range(range, count, end_pfn, -1ULL);

	memblock_x86_subtract_reserved(range, count);
	nr_range = clean_sort_range(range, count);

	*rangep = range;
	return nr_range;
}

static u64 __init __memblock_x86_memory_in_range(u64 addr, u64 limit, bool get_free)
static u64 __init __memblock_x86_memory_in_range(u64 addr, u64 limit, bool get_free)
{
{
	int i, count;
	int i, count;
+16 −12
Original line number Original line Diff line number Diff line
@@ -108,21 +108,25 @@ static void __init __free_pages_memory(unsigned long start, unsigned long end)


unsigned long __init free_low_memory_core_early(int nodeid)
unsigned long __init free_low_memory_core_early(int nodeid)
{
{
	int i;
	u64 start, end;
	unsigned long count = 0;
	unsigned long count = 0;
	struct range *range = NULL;
	phys_addr_t start, end;
	int nr_range;
	u64 i;


	nr_range = __get_free_all_memory_range(&range, nodeid, 0, max_low_pfn);
	/* free reserved array temporarily so that it's treated as free area */

	memblock_free_reserved_regions();
	for (i = 0; i < nr_range; i++) {

		start = range[i].start;
	for_each_free_mem_range(i, MAX_NUMNODES, &start, &end, NULL) {
		end = range[i].end;
		unsigned long start_pfn = PFN_UP(start);
		count += end - start;
		unsigned long end_pfn = min_t(unsigned long,
		__free_pages_memory(start, end);
					      PFN_DOWN(end), max_low_pfn);
		if (start_pfn < end_pfn) {
			__free_pages_memory(start_pfn, end_pfn);
			count += end_pfn - start_pfn;
		}
	}
	}


	/* put region array back? */
	memblock_reserve_reserved_regions();
	return count;
	return count;
}
}