Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 0f0a1024 authored by qctecmdr Service's avatar qctecmdr Service Committed by Gerrit - the friendly Code Review server
Browse files

Merge "mm: retry more before OOM in the presence of slow shrinkers"

parents c6217dd5 bc20a502
Loading
Loading
Loading
Loading
+47 −2
Original line number Diff line number Diff line
@@ -3602,6 +3602,46 @@ __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order,
	return NULL;
}

#ifdef CONFIG_HAVE_LOW_MEMORY_KILLER
static inline bool
should_compact_lmk_retry(struct alloc_context *ac, int order, int alloc_flags)
{
	struct zone *zone;
	struct zoneref *z;

	/* Let costly order requests check for compaction progress */
	if (order > PAGE_ALLOC_COSTLY_ORDER)
		return false;

	/*
	 * For (0 < order < PAGE_ALLOC_COSTLY_ORDER) allow the shrinkers
	 * to run and free up memory. Do not let these allocations fail
	 * if shrinkers can free up memory. This is similar to
	 * should_compact_retry implementation for !CONFIG_COMPACTION.
	 */
	for_each_zone_zonelist_nodemask(zone, z, ac->zonelist,
				ac->high_zoneidx, ac->nodemask) {
		unsigned long available;

		available = zone_reclaimable_pages(zone);
		available +=
			zone_page_state_snapshot(zone, NR_FREE_PAGES);

		if (__zone_watermark_ok(zone, 0, min_wmark_pages(zone),
			ac_classzone_idx(ac), alloc_flags, available))
			return true;
	}

	return false;
}
#else
static inline bool
should_compact_lmk_retry(struct alloc_context *ac, int order, int alloc_flags)
{
	return false;
}
#endif

static inline bool
should_compact_retry(struct alloc_context *ac, int order, int alloc_flags,
		     enum compact_result compact_result,
@@ -3617,6 +3657,9 @@ should_compact_retry(struct alloc_context *ac, int order, int alloc_flags,
	if (!order)
		return false;

	if (should_compact_lmk_retry(ac, order, alloc_flags))
		return true;

	if (compaction_made_progress(compact_result))
		(*compaction_retries)++;

@@ -3940,7 +3983,8 @@ should_reclaim_retry(gfp_t gfp_mask, unsigned order,
	 * their order will become available due to high fragmentation so
	 * always increment the no progress counter for them
	 */
	if (did_some_progress && order <= PAGE_ALLOC_COSTLY_ORDER)
	if ((did_some_progress && order <= PAGE_ALLOC_COSTLY_ORDER) ||
			IS_ENABLED(CONFIG_HAVE_LOW_MEMORY_KILLER))
		*no_progress_loops = 0;
	else
		(*no_progress_loops)++;
@@ -4221,7 +4265,8 @@ __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order,
	 * implementation of the compaction depends on the sufficient amount
	 * of free memory (see __compaction_suitable)
	 */
	if (did_some_progress > 0 &&
	if ((did_some_progress > 0 ||
			IS_ENABLED(CONFIG_HAVE_LOW_MEMORY_KILLER)) &&
			should_compact_retry(ac, order, alloc_flags,
				compact_result, &compact_priority,
				&compaction_retries))