mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 02:30:34 +02:00 
			
		
		
		
	slab: introduce __kmemcg_cache_deactivate()
__kmem_cache_shrink() is called with %true @deactivate only for memcg caches. Remove @deactivate from __kmem_cache_shrink() and introduce __kmemcg_cache_deactivate() instead. Each memcg-supporting allocator should implement it and it should deactivate and drain the cache. This is to allow memcg cache deactivation behavior to further deviate from simple shrinking without messing up __kmem_cache_shrink(). This is pure reorganization and doesn't introduce any observable behavior changes. v2: Dropped unnecessary ifdef in mm/slab.h as suggested by Vladimir. Link: http://lkml.kernel.org/r/20170117235411.9408-8-tj@kernel.org Signed-off-by: Tejun Heo <tj@kernel.org> Acked-by: Vladimir Davydov <vdavydov.dev@gmail.com> Cc: Christoph Lameter <cl@linux.com> Cc: Pekka Enberg <penberg@kernel.org> Cc: David Rientjes <rientjes@google.com> Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
		
							parent
							
								
									510ded33e0
								
							
						
					
					
						commit
						c9fc586403
					
				
					 5 changed files with 36 additions and 23 deletions
				
			
		
							
								
								
									
										11
									
								
								mm/slab.c
									
									
									
									
									
								
							
							
						
						
									
										11
									
								
								mm/slab.c
									
									
									
									
									
								
							| 
						 | 
					@ -2315,7 +2315,7 @@ static int drain_freelist(struct kmem_cache *cache,
 | 
				
			||||||
	return nr_freed;
 | 
						return nr_freed;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int __kmem_cache_shrink(struct kmem_cache *cachep, bool deactivate)
 | 
					int __kmem_cache_shrink(struct kmem_cache *cachep)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	int ret = 0;
 | 
						int ret = 0;
 | 
				
			||||||
	int node;
 | 
						int node;
 | 
				
			||||||
| 
						 | 
					@ -2333,9 +2333,16 @@ int __kmem_cache_shrink(struct kmem_cache *cachep, bool deactivate)
 | 
				
			||||||
	return (ret ? 1 : 0);
 | 
						return (ret ? 1 : 0);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					#ifdef CONFIG_MEMCG
 | 
				
			||||||
 | 
					void __kmemcg_cache_deactivate(struct kmem_cache *cachep)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						__kmem_cache_shrink(cachep);
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					#endif
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int __kmem_cache_shutdown(struct kmem_cache *cachep)
 | 
					int __kmem_cache_shutdown(struct kmem_cache *cachep)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	return __kmem_cache_shrink(cachep, false);
 | 
						return __kmem_cache_shrink(cachep);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void __kmem_cache_release(struct kmem_cache *cachep)
 | 
					void __kmem_cache_release(struct kmem_cache *cachep)
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -167,7 +167,8 @@ static inline unsigned long kmem_cache_flags(unsigned long object_size,
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int __kmem_cache_shutdown(struct kmem_cache *);
 | 
					int __kmem_cache_shutdown(struct kmem_cache *);
 | 
				
			||||||
void __kmem_cache_release(struct kmem_cache *);
 | 
					void __kmem_cache_release(struct kmem_cache *);
 | 
				
			||||||
int __kmem_cache_shrink(struct kmem_cache *, bool);
 | 
					int __kmem_cache_shrink(struct kmem_cache *);
 | 
				
			||||||
 | 
					void __kmemcg_cache_deactivate(struct kmem_cache *s);
 | 
				
			||||||
void slab_kmem_cache_release(struct kmem_cache *);
 | 
					void slab_kmem_cache_release(struct kmem_cache *);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
struct seq_file;
 | 
					struct seq_file;
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -646,7 +646,7 @@ void memcg_deactivate_kmem_caches(struct mem_cgroup *memcg)
 | 
				
			||||||
		if (!c)
 | 
							if (!c)
 | 
				
			||||||
			continue;
 | 
								continue;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		__kmem_cache_shrink(c, true);
 | 
							__kmemcg_cache_deactivate(c);
 | 
				
			||||||
		arr->entries[idx] = NULL;
 | 
							arr->entries[idx] = NULL;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
	mutex_unlock(&slab_mutex);
 | 
						mutex_unlock(&slab_mutex);
 | 
				
			||||||
| 
						 | 
					@ -794,7 +794,7 @@ int kmem_cache_shrink(struct kmem_cache *cachep)
 | 
				
			||||||
	get_online_cpus();
 | 
						get_online_cpus();
 | 
				
			||||||
	get_online_mems();
 | 
						get_online_mems();
 | 
				
			||||||
	kasan_cache_shrink(cachep);
 | 
						kasan_cache_shrink(cachep);
 | 
				
			||||||
	ret = __kmem_cache_shrink(cachep, false);
 | 
						ret = __kmem_cache_shrink(cachep);
 | 
				
			||||||
	put_online_mems();
 | 
						put_online_mems();
 | 
				
			||||||
	put_online_cpus();
 | 
						put_online_cpus();
 | 
				
			||||||
	return ret;
 | 
						return ret;
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -634,7 +634,7 @@ void __kmem_cache_release(struct kmem_cache *c)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int __kmem_cache_shrink(struct kmem_cache *d, bool deactivate)
 | 
					int __kmem_cache_shrink(struct kmem_cache *d)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	return 0;
 | 
						return 0;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
							
								
								
									
										39
									
								
								mm/slub.c
									
									
									
									
									
								
							
							
						
						
									
										39
									
								
								mm/slub.c
									
									
									
									
									
								
							| 
						 | 
					@ -3894,7 +3894,7 @@ EXPORT_SYMBOL(kfree);
 | 
				
			||||||
 * being allocated from last increasing the chance that the last objects
 | 
					 * being allocated from last increasing the chance that the last objects
 | 
				
			||||||
 * are freed in them.
 | 
					 * are freed in them.
 | 
				
			||||||
 */
 | 
					 */
 | 
				
			||||||
int __kmem_cache_shrink(struct kmem_cache *s, bool deactivate)
 | 
					int __kmem_cache_shrink(struct kmem_cache *s)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	int node;
 | 
						int node;
 | 
				
			||||||
	int i;
 | 
						int i;
 | 
				
			||||||
| 
						 | 
					@ -3906,21 +3906,6 @@ int __kmem_cache_shrink(struct kmem_cache *s, bool deactivate)
 | 
				
			||||||
	unsigned long flags;
 | 
						unsigned long flags;
 | 
				
			||||||
	int ret = 0;
 | 
						int ret = 0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (deactivate) {
 | 
					 | 
				
			||||||
		/*
 | 
					 | 
				
			||||||
		 * Disable empty slabs caching. Used to avoid pinning offline
 | 
					 | 
				
			||||||
		 * memory cgroups by kmem pages that can be freed.
 | 
					 | 
				
			||||||
		 */
 | 
					 | 
				
			||||||
		s->cpu_partial = 0;
 | 
					 | 
				
			||||||
		s->min_partial = 0;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		/*
 | 
					 | 
				
			||||||
		 * s->cpu_partial is checked locklessly (see put_cpu_partial),
 | 
					 | 
				
			||||||
		 * so we have to make sure the change is visible.
 | 
					 | 
				
			||||||
		 */
 | 
					 | 
				
			||||||
		synchronize_sched();
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	flush_all(s);
 | 
						flush_all(s);
 | 
				
			||||||
	for_each_kmem_cache_node(s, node, n) {
 | 
						for_each_kmem_cache_node(s, node, n) {
 | 
				
			||||||
		INIT_LIST_HEAD(&discard);
 | 
							INIT_LIST_HEAD(&discard);
 | 
				
			||||||
| 
						 | 
					@ -3971,13 +3956,33 @@ int __kmem_cache_shrink(struct kmem_cache *s, bool deactivate)
 | 
				
			||||||
	return ret;
 | 
						return ret;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					#ifdef CONFIG_MEMCG
 | 
				
			||||||
 | 
					void __kmemcg_cache_deactivate(struct kmem_cache *s)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						/*
 | 
				
			||||||
 | 
						 * Disable empty slabs caching. Used to avoid pinning offline
 | 
				
			||||||
 | 
						 * memory cgroups by kmem pages that can be freed.
 | 
				
			||||||
 | 
						 */
 | 
				
			||||||
 | 
						s->cpu_partial = 0;
 | 
				
			||||||
 | 
						s->min_partial = 0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						/*
 | 
				
			||||||
 | 
						 * s->cpu_partial is checked locklessly (see put_cpu_partial), so
 | 
				
			||||||
 | 
						 * we have to make sure the change is visible.
 | 
				
			||||||
 | 
						 */
 | 
				
			||||||
 | 
						synchronize_sched();
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						__kmem_cache_shrink(s);
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					#endif
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static int slab_mem_going_offline_callback(void *arg)
 | 
					static int slab_mem_going_offline_callback(void *arg)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct kmem_cache *s;
 | 
						struct kmem_cache *s;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	mutex_lock(&slab_mutex);
 | 
						mutex_lock(&slab_mutex);
 | 
				
			||||||
	list_for_each_entry(s, &slab_caches, list)
 | 
						list_for_each_entry(s, &slab_caches, list)
 | 
				
			||||||
		__kmem_cache_shrink(s, false);
 | 
							__kmem_cache_shrink(s);
 | 
				
			||||||
	mutex_unlock(&slab_mutex);
 | 
						mutex_unlock(&slab_mutex);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	return 0;
 | 
						return 0;
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue