mirror of
				https://github.com/torvalds/linux.git
				synced 2025-10-31 16:48:26 +02:00 
			
		
		
		
	 2e9a548009
			
		
	
	
		2e9a548009
		
	
	
	
	
		
			
			Add a new open coded iterator for kmem_cache which can be called from a
BPF program like below.  It doesn't take any argument and traverses all
kmem_cache entries.
  struct kmem_cache *pos;
  bpf_for_each(kmem_cache, pos) {
      ...
  }
As it needs to grab slab_mutex, it should be called from sleepable BPF
programs only.
Also update the existing iterator code to use the open coded version
internally as suggested by Andrii.
Signed-off-by: Namhyung Kim <namhyung@kernel.org>
Link: https://lore.kernel.org/r/20241030222819.1800667-1-namhyung@kernel.org
Signed-off-by: Alexei Starovoitov <ast@kernel.org>
		
	
			
		
			
				
	
	
		
			238 lines
		
	
	
	
		
			5.6 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			238 lines
		
	
	
	
		
			5.6 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0-only
 | |
| /* Copyright (c) 2024 Google */
 | |
| #include <linux/bpf.h>
 | |
| #include <linux/btf_ids.h>
 | |
| #include <linux/slab.h>
 | |
| #include <linux/kernel.h>
 | |
| #include <linux/seq_file.h>
 | |
| 
 | |
| #include "../../mm/slab.h" /* kmem_cache, slab_caches and slab_mutex */
 | |
| 
 | |
| /* open-coded version */
 | |
| struct bpf_iter_kmem_cache {
 | |
| 	__u64 __opaque[1];
 | |
| } __attribute__((aligned(8)));
 | |
| 
 | |
| struct bpf_iter_kmem_cache_kern {
 | |
| 	struct kmem_cache *pos;
 | |
| } __attribute__((aligned(8)));
 | |
| 
 | |
| #define KMEM_CACHE_POS_START  ((void *)1L)
 | |
| 
 | |
| __bpf_kfunc_start_defs();
 | |
| 
 | |
| __bpf_kfunc int bpf_iter_kmem_cache_new(struct bpf_iter_kmem_cache *it)
 | |
| {
 | |
| 	struct bpf_iter_kmem_cache_kern *kit = (void *)it;
 | |
| 
 | |
| 	BUILD_BUG_ON(sizeof(*kit) > sizeof(*it));
 | |
| 	BUILD_BUG_ON(__alignof__(*kit) != __alignof__(*it));
 | |
| 
 | |
| 	kit->pos = KMEM_CACHE_POS_START;
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| __bpf_kfunc struct kmem_cache *bpf_iter_kmem_cache_next(struct bpf_iter_kmem_cache *it)
 | |
| {
 | |
| 	struct bpf_iter_kmem_cache_kern *kit = (void *)it;
 | |
| 	struct kmem_cache *prev = kit->pos;
 | |
| 	struct kmem_cache *next;
 | |
| 	bool destroy = false;
 | |
| 
 | |
| 	if (!prev)
 | |
| 		return NULL;
 | |
| 
 | |
| 	mutex_lock(&slab_mutex);
 | |
| 
 | |
| 	if (list_empty(&slab_caches)) {
 | |
| 		mutex_unlock(&slab_mutex);
 | |
| 		return NULL;
 | |
| 	}
 | |
| 
 | |
| 	if (prev == KMEM_CACHE_POS_START)
 | |
| 		next = list_first_entry(&slab_caches, struct kmem_cache, list);
 | |
| 	else if (list_last_entry(&slab_caches, struct kmem_cache, list) == prev)
 | |
| 		next = NULL;
 | |
| 	else
 | |
| 		next = list_next_entry(prev, list);
 | |
| 
 | |
| 	/* boot_caches have negative refcount, don't touch them */
 | |
| 	if (next && next->refcount > 0)
 | |
| 		next->refcount++;
 | |
| 
 | |
| 	/* Skip kmem_cache_destroy() for active entries */
 | |
| 	if (prev && prev != KMEM_CACHE_POS_START) {
 | |
| 		if (prev->refcount > 1)
 | |
| 			prev->refcount--;
 | |
| 		else if (prev->refcount == 1)
 | |
| 			destroy = true;
 | |
| 	}
 | |
| 
 | |
| 	mutex_unlock(&slab_mutex);
 | |
| 
 | |
| 	if (destroy)
 | |
| 		kmem_cache_destroy(prev);
 | |
| 
 | |
| 	kit->pos = next;
 | |
| 	return next;
 | |
| }
 | |
| 
 | |
| __bpf_kfunc void bpf_iter_kmem_cache_destroy(struct bpf_iter_kmem_cache *it)
 | |
| {
 | |
| 	struct bpf_iter_kmem_cache_kern *kit = (void *)it;
 | |
| 	struct kmem_cache *s = kit->pos;
 | |
| 	bool destroy = false;
 | |
| 
 | |
| 	if (s == NULL || s == KMEM_CACHE_POS_START)
 | |
| 		return;
 | |
| 
 | |
| 	mutex_lock(&slab_mutex);
 | |
| 
 | |
| 	/* Skip kmem_cache_destroy() for active entries */
 | |
| 	if (s->refcount > 1)
 | |
| 		s->refcount--;
 | |
| 	else if (s->refcount == 1)
 | |
| 		destroy = true;
 | |
| 
 | |
| 	mutex_unlock(&slab_mutex);
 | |
| 
 | |
| 	if (destroy)
 | |
| 		kmem_cache_destroy(s);
 | |
| }
 | |
| 
 | |
| __bpf_kfunc_end_defs();
 | |
| 
 | |
| struct bpf_iter__kmem_cache {
 | |
| 	__bpf_md_ptr(struct bpf_iter_meta *, meta);
 | |
| 	__bpf_md_ptr(struct kmem_cache *, s);
 | |
| };
 | |
| 
 | |
| union kmem_cache_iter_priv {
 | |
| 	struct bpf_iter_kmem_cache it;
 | |
| 	struct bpf_iter_kmem_cache_kern kit;
 | |
| };
 | |
| 
 | |
| static void *kmem_cache_iter_seq_start(struct seq_file *seq, loff_t *pos)
 | |
| {
 | |
| 	loff_t cnt = 0;
 | |
| 	bool found = false;
 | |
| 	struct kmem_cache *s;
 | |
| 	union kmem_cache_iter_priv *p = seq->private;
 | |
| 
 | |
| 	mutex_lock(&slab_mutex);
 | |
| 
 | |
| 	/* Find an entry at the given position in the slab_caches list instead
 | |
| 	 * of keeping a reference (of the last visited entry, if any) out of
 | |
| 	 * slab_mutex. It might miss something if one is deleted in the middle
 | |
| 	 * while it releases the lock.  But it should be rare and there's not
 | |
| 	 * much we can do about it.
 | |
| 	 */
 | |
| 	list_for_each_entry(s, &slab_caches, list) {
 | |
| 		if (cnt == *pos) {
 | |
| 			/* Make sure this entry remains in the list by getting
 | |
| 			 * a new reference count.  Note that boot_cache entries
 | |
| 			 * have a negative refcount, so don't touch them.
 | |
| 			 */
 | |
| 			if (s->refcount > 0)
 | |
| 				s->refcount++;
 | |
| 			found = true;
 | |
| 			break;
 | |
| 		}
 | |
| 		cnt++;
 | |
| 	}
 | |
| 	mutex_unlock(&slab_mutex);
 | |
| 
 | |
| 	if (!found)
 | |
| 		s = NULL;
 | |
| 
 | |
| 	p->kit.pos = s;
 | |
| 	return s;
 | |
| }
 | |
| 
 | |
| static void kmem_cache_iter_seq_stop(struct seq_file *seq, void *v)
 | |
| {
 | |
| 	struct bpf_iter_meta meta;
 | |
| 	struct bpf_iter__kmem_cache ctx = {
 | |
| 		.meta = &meta,
 | |
| 		.s = v,
 | |
| 	};
 | |
| 	union kmem_cache_iter_priv *p = seq->private;
 | |
| 	struct bpf_prog *prog;
 | |
| 
 | |
| 	meta.seq = seq;
 | |
| 	prog = bpf_iter_get_info(&meta, true);
 | |
| 	if (prog && !ctx.s)
 | |
| 		bpf_iter_run_prog(prog, &ctx);
 | |
| 
 | |
| 	bpf_iter_kmem_cache_destroy(&p->it);
 | |
| }
 | |
| 
 | |
| static void *kmem_cache_iter_seq_next(struct seq_file *seq, void *v, loff_t *pos)
 | |
| {
 | |
| 	union kmem_cache_iter_priv *p = seq->private;
 | |
| 
 | |
| 	++*pos;
 | |
| 
 | |
| 	return bpf_iter_kmem_cache_next(&p->it);
 | |
| }
 | |
| 
 | |
| static int kmem_cache_iter_seq_show(struct seq_file *seq, void *v)
 | |
| {
 | |
| 	struct bpf_iter_meta meta;
 | |
| 	struct bpf_iter__kmem_cache ctx = {
 | |
| 		.meta = &meta,
 | |
| 		.s = v,
 | |
| 	};
 | |
| 	struct bpf_prog *prog;
 | |
| 	int ret = 0;
 | |
| 
 | |
| 	meta.seq = seq;
 | |
| 	prog = bpf_iter_get_info(&meta, false);
 | |
| 	if (prog)
 | |
| 		ret = bpf_iter_run_prog(prog, &ctx);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static const struct seq_operations kmem_cache_iter_seq_ops = {
 | |
| 	.start  = kmem_cache_iter_seq_start,
 | |
| 	.next   = kmem_cache_iter_seq_next,
 | |
| 	.stop   = kmem_cache_iter_seq_stop,
 | |
| 	.show   = kmem_cache_iter_seq_show,
 | |
| };
 | |
| 
 | |
| BTF_ID_LIST_GLOBAL_SINGLE(bpf_kmem_cache_btf_id, struct, kmem_cache)
 | |
| 
 | |
| static const struct bpf_iter_seq_info kmem_cache_iter_seq_info = {
 | |
| 	.seq_ops		= &kmem_cache_iter_seq_ops,
 | |
| 	.seq_priv_size		= sizeof(union kmem_cache_iter_priv),
 | |
| };
 | |
| 
 | |
| static void bpf_iter_kmem_cache_show_fdinfo(const struct bpf_iter_aux_info *aux,
 | |
| 					    struct seq_file *seq)
 | |
| {
 | |
| 	seq_puts(seq, "kmem_cache iter\n");
 | |
| }
 | |
| 
 | |
| DEFINE_BPF_ITER_FUNC(kmem_cache, struct bpf_iter_meta *meta,
 | |
| 		     struct kmem_cache *s)
 | |
| 
 | |
| static struct bpf_iter_reg bpf_kmem_cache_reg_info = {
 | |
| 	.target			= "kmem_cache",
 | |
| 	.feature		= BPF_ITER_RESCHED,
 | |
| 	.show_fdinfo		= bpf_iter_kmem_cache_show_fdinfo,
 | |
| 	.ctx_arg_info_size	= 1,
 | |
| 	.ctx_arg_info		= {
 | |
| 		{ offsetof(struct bpf_iter__kmem_cache, s),
 | |
| 		  PTR_TO_BTF_ID_OR_NULL | PTR_TRUSTED },
 | |
| 	},
 | |
| 	.seq_info		= &kmem_cache_iter_seq_info,
 | |
| };
 | |
| 
 | |
| static int __init bpf_kmem_cache_iter_init(void)
 | |
| {
 | |
| 	bpf_kmem_cache_reg_info.ctx_arg_info[0].btf_id = bpf_kmem_cache_btf_id[0];
 | |
| 	return bpf_iter_reg_target(&bpf_kmem_cache_reg_info);
 | |
| }
 | |
| 
 | |
| late_initcall(bpf_kmem_cache_iter_init);
 |