mm: vmscan: rework compaction-ready signaling in direct reclaim
Page reclaim for a higher-order page runs until compaction is ready, then aborts and signals this situation through the return value of shrink_zones(). This is an oddly specific signal to encode in the return value of shrink_zones(), though, and can be quite confusing. Introduce sc->compaction_ready and signal the compactability of the zones out-of-band to free up the return value of shrink_zones() for actual zone reclaimability. Signed-off-by: Johannes Weiner <hannes@cmpxchg.org> Acked-by: Vlastimil Babka <vbabka@suse.cz> Cc: Mel Gorman <mgorman@suse.de> Cc: Rik van Riel <riel@redhat.com> Cc: Michal Hocko <mhocko@suse.cz> Acked-by: Minchan Kim <minchan@kernel.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
committed by
Linus Torvalds
parent
8d07429319
commit
0b06496a33
70
mm/vmscan.c
70
mm/vmscan.c
@@ -65,6 +65,9 @@ struct scan_control {
|
|||||||
/* Number of pages freed so far during a call to shrink_zones() */
|
/* Number of pages freed so far during a call to shrink_zones() */
|
||||||
unsigned long nr_reclaimed;
|
unsigned long nr_reclaimed;
|
||||||
|
|
||||||
|
/* One of the zones is ready for compaction */
|
||||||
|
int compaction_ready;
|
||||||
|
|
||||||
/* How many pages shrink_list() should reclaim */
|
/* How many pages shrink_list() should reclaim */
|
||||||
unsigned long nr_to_reclaim;
|
unsigned long nr_to_reclaim;
|
||||||
|
|
||||||
@@ -2292,15 +2295,11 @@ static void shrink_zone(struct zone *zone, struct scan_control *sc)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/* Returns true if compaction should go ahead for a high-order request */
|
/* Returns true if compaction should go ahead for a high-order request */
|
||||||
static inline bool compaction_ready(struct zone *zone, struct scan_control *sc)
|
static inline bool compaction_ready(struct zone *zone, int order)
|
||||||
{
|
{
|
||||||
unsigned long balance_gap, watermark;
|
unsigned long balance_gap, watermark;
|
||||||
bool watermark_ok;
|
bool watermark_ok;
|
||||||
|
|
||||||
/* Do not consider compaction for orders reclaim is meant to satisfy */
|
|
||||||
if (sc->order <= PAGE_ALLOC_COSTLY_ORDER)
|
|
||||||
return false;
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Compaction takes time to run and there are potentially other
|
* Compaction takes time to run and there are potentially other
|
||||||
* callers using the pages just freed. Continue reclaiming until
|
* callers using the pages just freed. Continue reclaiming until
|
||||||
@@ -2309,18 +2308,18 @@ static inline bool compaction_ready(struct zone *zone, struct scan_control *sc)
|
|||||||
*/
|
*/
|
||||||
balance_gap = min(low_wmark_pages(zone), DIV_ROUND_UP(
|
balance_gap = min(low_wmark_pages(zone), DIV_ROUND_UP(
|
||||||
zone->managed_pages, KSWAPD_ZONE_BALANCE_GAP_RATIO));
|
zone->managed_pages, KSWAPD_ZONE_BALANCE_GAP_RATIO));
|
||||||
watermark = high_wmark_pages(zone) + balance_gap + (2UL << sc->order);
|
watermark = high_wmark_pages(zone) + balance_gap + (2UL << order);
|
||||||
watermark_ok = zone_watermark_ok_safe(zone, 0, watermark, 0, 0);
|
watermark_ok = zone_watermark_ok_safe(zone, 0, watermark, 0, 0);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* If compaction is deferred, reclaim up to a point where
|
* If compaction is deferred, reclaim up to a point where
|
||||||
* compaction will have a chance of success when re-enabled
|
* compaction will have a chance of success when re-enabled
|
||||||
*/
|
*/
|
||||||
if (compaction_deferred(zone, sc->order))
|
if (compaction_deferred(zone, order))
|
||||||
return watermark_ok;
|
return watermark_ok;
|
||||||
|
|
||||||
/* If compaction is not ready to start, keep reclaiming */
|
/* If compaction is not ready to start, keep reclaiming */
|
||||||
if (!compaction_suitable(zone, sc->order))
|
if (!compaction_suitable(zone, order))
|
||||||
return false;
|
return false;
|
||||||
|
|
||||||
return watermark_ok;
|
return watermark_ok;
|
||||||
@@ -2341,20 +2340,14 @@ static inline bool compaction_ready(struct zone *zone, struct scan_control *sc)
|
|||||||
*
|
*
|
||||||
* If a zone is deemed to be full of pinned pages then just give it a light
|
* If a zone is deemed to be full of pinned pages then just give it a light
|
||||||
* scan then give up on it.
|
* scan then give up on it.
|
||||||
*
|
|
||||||
* This function returns true if a zone is being reclaimed for a costly
|
|
||||||
* high-order allocation and compaction is ready to begin. This indicates to
|
|
||||||
* the caller that it should consider retrying the allocation instead of
|
|
||||||
* further reclaim.
|
|
||||||
*/
|
*/
|
||||||
static bool shrink_zones(struct zonelist *zonelist, struct scan_control *sc)
|
static void shrink_zones(struct zonelist *zonelist, struct scan_control *sc)
|
||||||
{
|
{
|
||||||
struct zoneref *z;
|
struct zoneref *z;
|
||||||
struct zone *zone;
|
struct zone *zone;
|
||||||
unsigned long nr_soft_reclaimed;
|
unsigned long nr_soft_reclaimed;
|
||||||
unsigned long nr_soft_scanned;
|
unsigned long nr_soft_scanned;
|
||||||
unsigned long lru_pages = 0;
|
unsigned long lru_pages = 0;
|
||||||
bool aborted_reclaim = false;
|
|
||||||
struct reclaim_state *reclaim_state = current->reclaim_state;
|
struct reclaim_state *reclaim_state = current->reclaim_state;
|
||||||
gfp_t orig_mask;
|
gfp_t orig_mask;
|
||||||
struct shrink_control shrink = {
|
struct shrink_control shrink = {
|
||||||
@@ -2391,22 +2384,24 @@ static bool shrink_zones(struct zonelist *zonelist, struct scan_control *sc)
|
|||||||
if (sc->priority != DEF_PRIORITY &&
|
if (sc->priority != DEF_PRIORITY &&
|
||||||
!zone_reclaimable(zone))
|
!zone_reclaimable(zone))
|
||||||
continue; /* Let kswapd poll it */
|
continue; /* Let kswapd poll it */
|
||||||
if (IS_ENABLED(CONFIG_COMPACTION)) {
|
|
||||||
/*
|
/*
|
||||||
* If we already have plenty of memory free for
|
* If we already have plenty of memory free for
|
||||||
* compaction in this zone, don't free any more.
|
* compaction in this zone, don't free any more.
|
||||||
* Even though compaction is invoked for any
|
* Even though compaction is invoked for any
|
||||||
* non-zero order, only frequent costly order
|
* non-zero order, only frequent costly order
|
||||||
* reclamation is disruptive enough to become a
|
* reclamation is disruptive enough to become a
|
||||||
* noticeable problem, like transparent huge
|
* noticeable problem, like transparent huge
|
||||||
* page allocations.
|
* page allocations.
|
||||||
*/
|
*/
|
||||||
if ((zonelist_zone_idx(z) <= requested_highidx)
|
if (IS_ENABLED(CONFIG_COMPACTION) &&
|
||||||
&& compaction_ready(zone, sc)) {
|
sc->order > PAGE_ALLOC_COSTLY_ORDER &&
|
||||||
aborted_reclaim = true;
|
zonelist_zone_idx(z) <= requested_highidx &&
|
||||||
continue;
|
compaction_ready(zone, sc->order)) {
|
||||||
}
|
sc->compaction_ready = true;
|
||||||
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* This steals pages from memory cgroups over softlimit
|
* This steals pages from memory cgroups over softlimit
|
||||||
* and returns the number of reclaimed pages and
|
* and returns the number of reclaimed pages and
|
||||||
@@ -2444,8 +2439,6 @@ static bool shrink_zones(struct zonelist *zonelist, struct scan_control *sc)
|
|||||||
* promoted it to __GFP_HIGHMEM.
|
* promoted it to __GFP_HIGHMEM.
|
||||||
*/
|
*/
|
||||||
sc->gfp_mask = orig_mask;
|
sc->gfp_mask = orig_mask;
|
||||||
|
|
||||||
return aborted_reclaim;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* All zones in zonelist are unreclaimable? */
|
/* All zones in zonelist are unreclaimable? */
|
||||||
@@ -2489,7 +2482,6 @@ static unsigned long do_try_to_free_pages(struct zonelist *zonelist,
|
|||||||
{
|
{
|
||||||
unsigned long total_scanned = 0;
|
unsigned long total_scanned = 0;
|
||||||
unsigned long writeback_threshold;
|
unsigned long writeback_threshold;
|
||||||
bool aborted_reclaim;
|
|
||||||
|
|
||||||
delayacct_freepages_start();
|
delayacct_freepages_start();
|
||||||
|
|
||||||
@@ -2500,11 +2492,14 @@ static unsigned long do_try_to_free_pages(struct zonelist *zonelist,
|
|||||||
vmpressure_prio(sc->gfp_mask, sc->target_mem_cgroup,
|
vmpressure_prio(sc->gfp_mask, sc->target_mem_cgroup,
|
||||||
sc->priority);
|
sc->priority);
|
||||||
sc->nr_scanned = 0;
|
sc->nr_scanned = 0;
|
||||||
aborted_reclaim = shrink_zones(zonelist, sc);
|
shrink_zones(zonelist, sc);
|
||||||
|
|
||||||
total_scanned += sc->nr_scanned;
|
total_scanned += sc->nr_scanned;
|
||||||
if (sc->nr_reclaimed >= sc->nr_to_reclaim)
|
if (sc->nr_reclaimed >= sc->nr_to_reclaim)
|
||||||
goto out;
|
break;
|
||||||
|
|
||||||
|
if (sc->compaction_ready)
|
||||||
|
break;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* If we're getting trouble reclaiming, start doing
|
* If we're getting trouble reclaiming, start doing
|
||||||
@@ -2526,16 +2521,15 @@ static unsigned long do_try_to_free_pages(struct zonelist *zonelist,
|
|||||||
WB_REASON_TRY_TO_FREE_PAGES);
|
WB_REASON_TRY_TO_FREE_PAGES);
|
||||||
sc->may_writepage = 1;
|
sc->may_writepage = 1;
|
||||||
}
|
}
|
||||||
} while (--sc->priority >= 0 && !aborted_reclaim);
|
} while (--sc->priority >= 0);
|
||||||
|
|
||||||
out:
|
|
||||||
delayacct_freepages_end();
|
delayacct_freepages_end();
|
||||||
|
|
||||||
if (sc->nr_reclaimed)
|
if (sc->nr_reclaimed)
|
||||||
return sc->nr_reclaimed;
|
return sc->nr_reclaimed;
|
||||||
|
|
||||||
/* Aborted reclaim to try compaction? don't OOM, then */
|
/* Aborted reclaim to try compaction? don't OOM, then */
|
||||||
if (aborted_reclaim)
|
if (sc->compaction_ready)
|
||||||
return 1;
|
return 1;
|
||||||
|
|
||||||
/* top priority shrink_zones still had more to do? don't OOM, then */
|
/* top priority shrink_zones still had more to do? don't OOM, then */
|
||||||
|
Reference in New Issue
Block a user