forked from luck/tmp_suning_uos_patched
slab: remove synchronous rcu_barrier() call in memcg cache release path
With kmem cgroup support enabled, kmem_caches can be created and destroyed frequently and a great number of near empty kmem_caches can accumulate if there are a lot of transient cgroups and the system is not under memory pressure. When memory reclaim starts under such conditions, it can lead to consecutive deactivation and destruction of many kmem_caches, easily hundreds of thousands on moderately large systems, exposing scalability issues in the current slab management code. This is one of the patches to address the issue. SLAB_DESTORY_BY_RCU caches need to flush all RCU operations before destruction because slab pages are freed through RCU and they need to be able to dereference the associated kmem_cache. Currently, it's done synchronously with rcu_barrier(). As rcu_barrier() is expensive time-wise, slab implements a batching mechanism so that rcu_barrier() can be done for multiple caches at the same time. Unfortunately, the rcu_barrier() is in synchronous path which is called while holding cgroup_mutex and the batching is too limited to be actually helpful. This patch updates the cache release path so that the batching is asynchronous and global. All SLAB_DESTORY_BY_RCU caches are queued globally and a work item consumes the list. The work item calls rcu_barrier() only once for all caches that are currently queued. * release_caches() is removed and shutdown_cache() now either directly release the cache or schedules a RCU callback to do that. This makes the cache inaccessible once shutdown_cache() is called and makes it impossible for shutdown_memcg_caches() to do memcg-specific cleanups afterwards. Move memcg-specific part into a helper, unlink_memcg_cache(), and make shutdown_cache() call it directly. Link: http://lkml.kernel.org/r/20170117235411.9408-4-tj@kernel.org Signed-off-by: Tejun Heo <tj@kernel.org> Reported-by: Jay Vana <jsvana@fb.com> Acked-by: Vladimir Davydov <vdavydov@tarantool.org> Cc: Christoph Lameter <cl@linux.com> Cc: Pekka Enberg <penberg@kernel.org> Cc: David Rientjes <rientjes@google.com> Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
bf5eb3de38
commit
657dc2f972
110
mm/slab_common.c
110
mm/slab_common.c
|
@ -30,6 +30,11 @@ LIST_HEAD(slab_caches);
|
||||||
DEFINE_MUTEX(slab_mutex);
|
DEFINE_MUTEX(slab_mutex);
|
||||||
struct kmem_cache *kmem_cache;
|
struct kmem_cache *kmem_cache;
|
||||||
|
|
||||||
|
static LIST_HEAD(slab_caches_to_rcu_destroy);
|
||||||
|
static void slab_caches_to_rcu_destroy_workfn(struct work_struct *work);
|
||||||
|
static DECLARE_WORK(slab_caches_to_rcu_destroy_work,
|
||||||
|
slab_caches_to_rcu_destroy_workfn);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Set of flags that will prevent slab merging
|
* Set of flags that will prevent slab merging
|
||||||
*/
|
*/
|
||||||
|
@ -215,6 +220,11 @@ int memcg_update_all_caches(int num_memcgs)
|
||||||
mutex_unlock(&slab_mutex);
|
mutex_unlock(&slab_mutex);
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void unlink_memcg_cache(struct kmem_cache *s)
|
||||||
|
{
|
||||||
|
list_del(&s->memcg_params.list);
|
||||||
|
}
|
||||||
#else
|
#else
|
||||||
static inline int init_memcg_params(struct kmem_cache *s,
|
static inline int init_memcg_params(struct kmem_cache *s,
|
||||||
struct mem_cgroup *memcg, struct kmem_cache *root_cache)
|
struct mem_cgroup *memcg, struct kmem_cache *root_cache)
|
||||||
|
@ -225,6 +235,10 @@ static inline int init_memcg_params(struct kmem_cache *s,
|
||||||
static inline void destroy_memcg_params(struct kmem_cache *s)
|
static inline void destroy_memcg_params(struct kmem_cache *s)
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static inline void unlink_memcg_cache(struct kmem_cache *s)
|
||||||
|
{
|
||||||
|
}
|
||||||
#endif /* CONFIG_MEMCG && !CONFIG_SLOB */
|
#endif /* CONFIG_MEMCG && !CONFIG_SLOB */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -461,27 +475,30 @@ kmem_cache_create(const char *name, size_t size, size_t align,
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(kmem_cache_create);
|
EXPORT_SYMBOL(kmem_cache_create);
|
||||||
|
|
||||||
static int shutdown_cache(struct kmem_cache *s,
|
static void slab_caches_to_rcu_destroy_workfn(struct work_struct *work)
|
||||||
struct list_head *release, bool *need_rcu_barrier)
|
|
||||||
{
|
|
||||||
if (__kmem_cache_shutdown(s) != 0)
|
|
||||||
return -EBUSY;
|
|
||||||
|
|
||||||
if (s->flags & SLAB_DESTROY_BY_RCU)
|
|
||||||
*need_rcu_barrier = true;
|
|
||||||
|
|
||||||
list_move(&s->list, release);
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void release_caches(struct list_head *release, bool need_rcu_barrier)
|
|
||||||
{
|
{
|
||||||
|
LIST_HEAD(to_destroy);
|
||||||
struct kmem_cache *s, *s2;
|
struct kmem_cache *s, *s2;
|
||||||
|
|
||||||
if (need_rcu_barrier)
|
/*
|
||||||
rcu_barrier();
|
* On destruction, SLAB_DESTROY_BY_RCU kmem_caches are put on the
|
||||||
|
* @slab_caches_to_rcu_destroy list. The slab pages are freed
|
||||||
|
* through RCU and and the associated kmem_cache are dereferenced
|
||||||
|
* while freeing the pages, so the kmem_caches should be freed only
|
||||||
|
* after the pending RCU operations are finished. As rcu_barrier()
|
||||||
|
* is a pretty slow operation, we batch all pending destructions
|
||||||
|
* asynchronously.
|
||||||
|
*/
|
||||||
|
mutex_lock(&slab_mutex);
|
||||||
|
list_splice_init(&slab_caches_to_rcu_destroy, &to_destroy);
|
||||||
|
mutex_unlock(&slab_mutex);
|
||||||
|
|
||||||
list_for_each_entry_safe(s, s2, release, list) {
|
if (list_empty(&to_destroy))
|
||||||
|
return;
|
||||||
|
|
||||||
|
rcu_barrier();
|
||||||
|
|
||||||
|
list_for_each_entry_safe(s, s2, &to_destroy, list) {
|
||||||
#ifdef SLAB_SUPPORTS_SYSFS
|
#ifdef SLAB_SUPPORTS_SYSFS
|
||||||
sysfs_slab_release(s);
|
sysfs_slab_release(s);
|
||||||
#else
|
#else
|
||||||
|
@ -490,6 +507,29 @@ static void release_caches(struct list_head *release, bool need_rcu_barrier)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static int shutdown_cache(struct kmem_cache *s)
|
||||||
|
{
|
||||||
|
if (__kmem_cache_shutdown(s) != 0)
|
||||||
|
return -EBUSY;
|
||||||
|
|
||||||
|
list_del(&s->list);
|
||||||
|
if (!is_root_cache(s))
|
||||||
|
unlink_memcg_cache(s);
|
||||||
|
|
||||||
|
if (s->flags & SLAB_DESTROY_BY_RCU) {
|
||||||
|
list_add_tail(&s->list, &slab_caches_to_rcu_destroy);
|
||||||
|
schedule_work(&slab_caches_to_rcu_destroy_work);
|
||||||
|
} else {
|
||||||
|
#ifdef SLAB_SUPPORTS_SYSFS
|
||||||
|
sysfs_slab_release(s);
|
||||||
|
#else
|
||||||
|
slab_kmem_cache_release(s);
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
#if defined(CONFIG_MEMCG) && !defined(CONFIG_SLOB)
|
#if defined(CONFIG_MEMCG) && !defined(CONFIG_SLOB)
|
||||||
/*
|
/*
|
||||||
* memcg_create_kmem_cache - Create a cache for a memory cgroup.
|
* memcg_create_kmem_cache - Create a cache for a memory cgroup.
|
||||||
|
@ -602,22 +642,8 @@ void memcg_deactivate_kmem_caches(struct mem_cgroup *memcg)
|
||||||
put_online_cpus();
|
put_online_cpus();
|
||||||
}
|
}
|
||||||
|
|
||||||
static int __shutdown_memcg_cache(struct kmem_cache *s,
|
|
||||||
struct list_head *release, bool *need_rcu_barrier)
|
|
||||||
{
|
|
||||||
BUG_ON(is_root_cache(s));
|
|
||||||
|
|
||||||
if (shutdown_cache(s, release, need_rcu_barrier))
|
|
||||||
return -EBUSY;
|
|
||||||
|
|
||||||
list_del(&s->memcg_params.list);
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
void memcg_destroy_kmem_caches(struct mem_cgroup *memcg)
|
void memcg_destroy_kmem_caches(struct mem_cgroup *memcg)
|
||||||
{
|
{
|
||||||
LIST_HEAD(release);
|
|
||||||
bool need_rcu_barrier = false;
|
|
||||||
struct kmem_cache *s, *s2;
|
struct kmem_cache *s, *s2;
|
||||||
|
|
||||||
get_online_cpus();
|
get_online_cpus();
|
||||||
|
@ -631,18 +657,15 @@ void memcg_destroy_kmem_caches(struct mem_cgroup *memcg)
|
||||||
* The cgroup is about to be freed and therefore has no charges
|
* The cgroup is about to be freed and therefore has no charges
|
||||||
* left. Hence, all its caches must be empty by now.
|
* left. Hence, all its caches must be empty by now.
|
||||||
*/
|
*/
|
||||||
BUG_ON(__shutdown_memcg_cache(s, &release, &need_rcu_barrier));
|
BUG_ON(shutdown_cache(s));
|
||||||
}
|
}
|
||||||
mutex_unlock(&slab_mutex);
|
mutex_unlock(&slab_mutex);
|
||||||
|
|
||||||
put_online_mems();
|
put_online_mems();
|
||||||
put_online_cpus();
|
put_online_cpus();
|
||||||
|
|
||||||
release_caches(&release, need_rcu_barrier);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static int shutdown_memcg_caches(struct kmem_cache *s,
|
static int shutdown_memcg_caches(struct kmem_cache *s)
|
||||||
struct list_head *release, bool *need_rcu_barrier)
|
|
||||||
{
|
{
|
||||||
struct memcg_cache_array *arr;
|
struct memcg_cache_array *arr;
|
||||||
struct kmem_cache *c, *c2;
|
struct kmem_cache *c, *c2;
|
||||||
|
@ -661,7 +684,7 @@ static int shutdown_memcg_caches(struct kmem_cache *s,
|
||||||
c = arr->entries[i];
|
c = arr->entries[i];
|
||||||
if (!c)
|
if (!c)
|
||||||
continue;
|
continue;
|
||||||
if (__shutdown_memcg_cache(c, release, need_rcu_barrier))
|
if (shutdown_cache(c))
|
||||||
/*
|
/*
|
||||||
* The cache still has objects. Move it to a temporary
|
* The cache still has objects. Move it to a temporary
|
||||||
* list so as not to try to destroy it for a second
|
* list so as not to try to destroy it for a second
|
||||||
|
@ -684,7 +707,7 @@ static int shutdown_memcg_caches(struct kmem_cache *s,
|
||||||
*/
|
*/
|
||||||
list_for_each_entry_safe(c, c2, &s->memcg_params.list,
|
list_for_each_entry_safe(c, c2, &s->memcg_params.list,
|
||||||
memcg_params.list)
|
memcg_params.list)
|
||||||
__shutdown_memcg_cache(c, release, need_rcu_barrier);
|
shutdown_cache(c);
|
||||||
|
|
||||||
list_splice(&busy, &s->memcg_params.list);
|
list_splice(&busy, &s->memcg_params.list);
|
||||||
|
|
||||||
|
@ -697,8 +720,7 @@ static int shutdown_memcg_caches(struct kmem_cache *s,
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
#else
|
#else
|
||||||
static inline int shutdown_memcg_caches(struct kmem_cache *s,
|
static inline int shutdown_memcg_caches(struct kmem_cache *s)
|
||||||
struct list_head *release, bool *need_rcu_barrier)
|
|
||||||
{
|
{
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
@ -714,8 +736,6 @@ void slab_kmem_cache_release(struct kmem_cache *s)
|
||||||
|
|
||||||
void kmem_cache_destroy(struct kmem_cache *s)
|
void kmem_cache_destroy(struct kmem_cache *s)
|
||||||
{
|
{
|
||||||
LIST_HEAD(release);
|
|
||||||
bool need_rcu_barrier = false;
|
|
||||||
int err;
|
int err;
|
||||||
|
|
||||||
if (unlikely(!s))
|
if (unlikely(!s))
|
||||||
|
@ -731,9 +751,9 @@ void kmem_cache_destroy(struct kmem_cache *s)
|
||||||
if (s->refcount)
|
if (s->refcount)
|
||||||
goto out_unlock;
|
goto out_unlock;
|
||||||
|
|
||||||
err = shutdown_memcg_caches(s, &release, &need_rcu_barrier);
|
err = shutdown_memcg_caches(s);
|
||||||
if (!err)
|
if (!err)
|
||||||
err = shutdown_cache(s, &release, &need_rcu_barrier);
|
err = shutdown_cache(s);
|
||||||
|
|
||||||
if (err) {
|
if (err) {
|
||||||
pr_err("kmem_cache_destroy %s: Slab cache still has objects\n",
|
pr_err("kmem_cache_destroy %s: Slab cache still has objects\n",
|
||||||
|
@ -745,8 +765,6 @@ void kmem_cache_destroy(struct kmem_cache *s)
|
||||||
|
|
||||||
put_online_mems();
|
put_online_mems();
|
||||||
put_online_cpus();
|
put_online_cpus();
|
||||||
|
|
||||||
release_caches(&release, need_rcu_barrier);
|
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(kmem_cache_destroy);
|
EXPORT_SYMBOL(kmem_cache_destroy);
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue
Block a user