diff options
author | Ingo Molnar <mingo@kernel.org> | 2018-10-04 08:23:03 +0200 |
---|---|---|
committer | Ingo Molnar <mingo@kernel.org> | 2018-10-04 08:23:03 +0200 |
commit | c0554d2d3db438623b4f2f9abc3d766b2b15d2fb (patch) | |
tree | d00dc324772b95373f4bdc566b3437e5bf637157 /mm/vmscan.c | |
parent | bef459026b161fbc39d20dcba698ed0cfffbac38 (diff) | |
parent | 1d2ba7fee28b3a47cca8e8f4f94a22d30b2b3a6f (diff) |
Merge branch 'linus' into x86/core, to pick up fixes
Signed-off-by: Ingo Molnar <mingo@kernel.org>
Diffstat (limited to 'mm/vmscan.c')
-rw-r--r-- | mm/vmscan.c | 11 |
1 files changed, 11 insertions, 0 deletions
diff --git a/mm/vmscan.c b/mm/vmscan.c index 7e7d25504651..c7ce2c161225 100644 --- a/mm/vmscan.c +++ b/mm/vmscan.c @@ -476,6 +476,17 @@ static unsigned long do_shrink_slab(struct shrink_control *shrinkctl, delta = freeable >> priority; delta *= 4; do_div(delta, shrinker->seeks); + + /* + * Make sure we apply some minimal pressure on default priority + * even on small cgroups. Stale objects are not only consuming memory + * by themselves, but can also hold a reference to a dying cgroup, + * preventing it from being reclaimed. A dying cgroup with all + * corresponding structures like per-cpu stats and kmem caches + * can be really big, so it may lead to a significant waste of memory. + */ + delta = max_t(unsigned long long, delta, min(freeable, batch_size)); + total_scan += delta; if (total_scan < 0) { pr_err("shrink_slab: %pF negative objects to delete nr=%ld\n", |