mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 10:40:15 +02:00 
			
		
		
		
	ring-buffer: Add test if range of boot buffer is valid
Add a test against the ring buffer memory range to see if it has valid data. The ring_buffer_meta structure is given a new field called "first_buffer" which holds the address of the first sub-buffer. This is used to both determine if the other fields are valid as well as finding the offset between the old addresses of the sub-buffer from the previous boot to the new addresses of the current boot. Since the values for nr_subbufs and subbuf_size is to be the same, check if the values in the meta page match the values calculated. Take the range of the first_buffer and the total size of all the buffers and make sure the saved head_buffer and commit_buffer fall in the range. Iterate through all the sub-buffers to make sure that the values in the sub-buffer "commit" field (the field that holds the amount of data on the sub-buffer) is within the end of the sub-buffer. Also check the index array to make sure that all the indexes are within nr_subbufs. Link: https://lkml.kernel.org/r/20240612232026.013843655@goodmis.org Cc: Masami Hiramatsu <mhiramat@kernel.org> Cc: Mark Rutland <mark.rutland@arm.com> Cc: Mathieu Desnoyers <mathieu.desnoyers@efficios.com> Cc: Andrew Morton <akpm@linux-foundation.org> Cc: Vincent Donnefort <vdonnefort@google.com> Cc: Joel Fernandes <joel@joelfernandes.org> Cc: Daniel Bristot de Oliveira <bristot@redhat.com> Cc: Ingo Molnar <mingo@kernel.org> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Vineeth Pillai <vineeth@bitbyteword.org> Cc: Youssef Esmat <youssefesmat@google.com> Cc: Beau Belgrave <beaub@linux.microsoft.com> Cc: Alexander Graf <graf@amazon.com> Cc: Baoquan He <bhe@redhat.com> Cc: Borislav Petkov <bp@alien8.de> Cc: "Paul E. McKenney" <paulmck@kernel.org> Cc: David Howells <dhowells@redhat.com> Cc: Mike Rapoport <rppt@kernel.org> Cc: Dave Hansen <dave.hansen@linux.intel.com> Cc: Tony Luck <tony.luck@intel.com> Cc: Guenter Roeck <linux@roeck-us.net> Cc: Ross Zwisler <zwisler@google.com> Cc: Kees Cook <keescook@chromium.org> Signed-off-by: Steven Rostedt (Google) <rostedt@goodmis.org>
This commit is contained in:
		
							parent
							
								
									950032ffce
								
							
						
					
					
						commit
						c76883f18e
					
				
					 1 changed files with 135 additions and 8 deletions
				
			
		| 
						 | 
					@ -45,6 +45,7 @@
 | 
				
			||||||
static void update_pages_handler(struct work_struct *work);
 | 
					static void update_pages_handler(struct work_struct *work);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
struct ring_buffer_meta {
 | 
					struct ring_buffer_meta {
 | 
				
			||||||
 | 
						unsigned long	first_buffer;
 | 
				
			||||||
	unsigned long	head_buffer;
 | 
						unsigned long	head_buffer;
 | 
				
			||||||
	unsigned long	commit_buffer;
 | 
						unsigned long	commit_buffer;
 | 
				
			||||||
	__u32		subbuf_size;
 | 
						__u32		subbuf_size;
 | 
				
			||||||
| 
						 | 
					@ -1618,21 +1619,103 @@ static void *rb_range_buffer(struct ring_buffer_per_cpu *cpu_buffer, int idx)
 | 
				
			||||||
	return (void *)ptr;
 | 
						return (void *)ptr;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					/*
 | 
				
			||||||
 | 
					 * See if the existing memory contains valid ring buffer data.
 | 
				
			||||||
 | 
					 * As the previous kernel must be the same as this kernel, all
 | 
				
			||||||
 | 
					 * the calculations (size of buffers and number of buffers)
 | 
				
			||||||
 | 
					 * must be the same.
 | 
				
			||||||
 | 
					 */
 | 
				
			||||||
 | 
					static bool rb_meta_valid(struct ring_buffer_meta *meta, int cpu,
 | 
				
			||||||
 | 
								  struct trace_buffer *buffer, int nr_pages)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						int subbuf_size = PAGE_SIZE;
 | 
				
			||||||
 | 
						struct buffer_data_page *subbuf;
 | 
				
			||||||
 | 
						unsigned long buffers_start;
 | 
				
			||||||
 | 
						unsigned long buffers_end;
 | 
				
			||||||
 | 
						int i;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						/* The subbuffer's size and number of subbuffers must match */
 | 
				
			||||||
 | 
						if (meta->subbuf_size != subbuf_size ||
 | 
				
			||||||
 | 
						    meta->nr_subbufs != nr_pages + 1) {
 | 
				
			||||||
 | 
							pr_info("Ring buffer boot meta [%d] mismatch of subbuf_size/nr_pages\n", cpu);
 | 
				
			||||||
 | 
							return false;
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						buffers_start = meta->first_buffer;
 | 
				
			||||||
 | 
						buffers_end = meta->first_buffer + (subbuf_size * meta->nr_subbufs);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						/* Is the head and commit buffers within the range of buffers? */
 | 
				
			||||||
 | 
						if (meta->head_buffer < buffers_start ||
 | 
				
			||||||
 | 
						    meta->head_buffer >= buffers_end) {
 | 
				
			||||||
 | 
							pr_info("Ring buffer boot meta [%d] head buffer out of range\n", cpu);
 | 
				
			||||||
 | 
							return false;
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						if (meta->commit_buffer < buffers_start ||
 | 
				
			||||||
 | 
						    meta->commit_buffer >= buffers_end) {
 | 
				
			||||||
 | 
							pr_info("Ring buffer boot meta [%d] commit buffer out of range\n", cpu);
 | 
				
			||||||
 | 
							return false;
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						subbuf = rb_subbufs_from_meta(meta);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						/* Is the meta buffers and the subbufs themselves have correct data? */
 | 
				
			||||||
 | 
						for (i = 0; i < meta->nr_subbufs; i++) {
 | 
				
			||||||
 | 
							if (meta->buffers[i] < 0 ||
 | 
				
			||||||
 | 
							    meta->buffers[i] >= meta->nr_subbufs) {
 | 
				
			||||||
 | 
								pr_info("Ring buffer boot meta [%d] array out of range\n", cpu);
 | 
				
			||||||
 | 
								return false;
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							if ((unsigned)local_read(&subbuf->commit) > subbuf_size) {
 | 
				
			||||||
 | 
								pr_info("Ring buffer boot meta [%d] buffer invalid commit\n", cpu);
 | 
				
			||||||
 | 
								return false;
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							subbuf = (void *)subbuf + subbuf_size;
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						pr_info("Ring buffer meta is from previous boot!\n");
 | 
				
			||||||
 | 
						return true;
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static void rb_range_meta_init(struct trace_buffer *buffer, int nr_pages)
 | 
					static void rb_range_meta_init(struct trace_buffer *buffer, int nr_pages)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct ring_buffer_meta *meta;
 | 
						struct ring_buffer_meta *meta;
 | 
				
			||||||
 | 
						unsigned long delta;
 | 
				
			||||||
	void *subbuf;
 | 
						void *subbuf;
 | 
				
			||||||
	int cpu;
 | 
						int cpu;
 | 
				
			||||||
	int i;
 | 
						int i;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	for (cpu = 0; cpu < nr_cpu_ids; cpu++) {
 | 
						for (cpu = 0; cpu < nr_cpu_ids; cpu++) {
 | 
				
			||||||
 | 
							void *next_meta;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		meta = rb_range_meta(buffer, nr_pages, cpu);
 | 
							meta = rb_range_meta(buffer, nr_pages, cpu);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							if (rb_meta_valid(meta, cpu, buffer, nr_pages)) {
 | 
				
			||||||
 | 
								/* Make the mappings match the current address */
 | 
				
			||||||
 | 
								subbuf = rb_subbufs_from_meta(meta);
 | 
				
			||||||
 | 
								delta = (unsigned long)subbuf - meta->first_buffer;
 | 
				
			||||||
 | 
								meta->first_buffer += delta;
 | 
				
			||||||
 | 
								meta->head_buffer += delta;
 | 
				
			||||||
 | 
								meta->commit_buffer += delta;
 | 
				
			||||||
 | 
								continue;
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							if (cpu < nr_cpu_ids - 1)
 | 
				
			||||||
 | 
								next_meta = rb_range_meta(buffer, nr_pages, cpu + 1);
 | 
				
			||||||
 | 
							else
 | 
				
			||||||
 | 
								next_meta = (void *)buffer->range_addr_end;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							memset(meta, 0, next_meta - (void *)meta);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		meta->nr_subbufs = nr_pages + 1;
 | 
							meta->nr_subbufs = nr_pages + 1;
 | 
				
			||||||
		meta->subbuf_size = PAGE_SIZE;
 | 
							meta->subbuf_size = PAGE_SIZE;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		subbuf = rb_subbufs_from_meta(meta);
 | 
							subbuf = rb_subbufs_from_meta(meta);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							meta->first_buffer = (unsigned long)subbuf;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		/*
 | 
							/*
 | 
				
			||||||
		 * The buffers[] array holds the order of the sub-buffers
 | 
							 * The buffers[] array holds the order of the sub-buffers
 | 
				
			||||||
		 * that are after the meta data. The sub-buffers may
 | 
							 * that are after the meta data. The sub-buffers may
 | 
				
			||||||
| 
						 | 
					@ -1724,10 +1807,26 @@ int ring_buffer_meta_seq_init(struct file *file, struct trace_buffer *buffer, in
 | 
				
			||||||
	return 0;
 | 
						return 0;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					/* Map the buffer_pages to the previous head and commit pages */
 | 
				
			||||||
 | 
					static void rb_meta_buffer_update(struct ring_buffer_per_cpu *cpu_buffer,
 | 
				
			||||||
 | 
									  struct buffer_page *bpage)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						struct ring_buffer_meta *meta = cpu_buffer->ring_meta;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						if (meta->head_buffer == (unsigned long)bpage->page)
 | 
				
			||||||
 | 
							cpu_buffer->head_page = bpage;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						if (meta->commit_buffer == (unsigned long)bpage->page) {
 | 
				
			||||||
 | 
							cpu_buffer->commit_page = bpage;
 | 
				
			||||||
 | 
							cpu_buffer->tail_page = bpage;
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static int __rb_allocate_pages(struct ring_buffer_per_cpu *cpu_buffer,
 | 
					static int __rb_allocate_pages(struct ring_buffer_per_cpu *cpu_buffer,
 | 
				
			||||||
		long nr_pages, struct list_head *pages)
 | 
							long nr_pages, struct list_head *pages)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct trace_buffer *buffer = cpu_buffer->buffer;
 | 
						struct trace_buffer *buffer = cpu_buffer->buffer;
 | 
				
			||||||
 | 
						struct ring_buffer_meta *meta = NULL;
 | 
				
			||||||
	struct buffer_page *bpage, *tmp;
 | 
						struct buffer_page *bpage, *tmp;
 | 
				
			||||||
	bool user_thread = current->mm != NULL;
 | 
						bool user_thread = current->mm != NULL;
 | 
				
			||||||
	gfp_t mflags;
 | 
						gfp_t mflags;
 | 
				
			||||||
| 
						 | 
					@ -1762,6 +1861,10 @@ static int __rb_allocate_pages(struct ring_buffer_per_cpu *cpu_buffer,
 | 
				
			||||||
	 */
 | 
						 */
 | 
				
			||||||
	if (user_thread)
 | 
						if (user_thread)
 | 
				
			||||||
		set_current_oom_origin();
 | 
							set_current_oom_origin();
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						if (buffer->range_addr_start)
 | 
				
			||||||
 | 
							meta = rb_range_meta(buffer, nr_pages, cpu_buffer->cpu);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	for (i = 0; i < nr_pages; i++) {
 | 
						for (i = 0; i < nr_pages; i++) {
 | 
				
			||||||
		struct page *page;
 | 
							struct page *page;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -1778,11 +1881,14 @@ static int __rb_allocate_pages(struct ring_buffer_per_cpu *cpu_buffer,
 | 
				
			||||||
		 */
 | 
							 */
 | 
				
			||||||
		list_add_tail(&bpage->list, pages);
 | 
							list_add_tail(&bpage->list, pages);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		if (buffer->range_addr_start) {
 | 
							if (meta) {
 | 
				
			||||||
			/* A range was given. Use that for the buffer page */
 | 
								/* A range was given. Use that for the buffer page */
 | 
				
			||||||
			bpage->page = rb_range_buffer(cpu_buffer, i + 1);
 | 
								bpage->page = rb_range_buffer(cpu_buffer, i + 1);
 | 
				
			||||||
			if (!bpage->page)
 | 
								if (!bpage->page)
 | 
				
			||||||
				goto free_pages;
 | 
									goto free_pages;
 | 
				
			||||||
 | 
								/* If this is valid from a previous boot */
 | 
				
			||||||
 | 
								if (meta->head_buffer)
 | 
				
			||||||
 | 
									rb_meta_buffer_update(cpu_buffer, bpage);
 | 
				
			||||||
			bpage->range = 1;
 | 
								bpage->range = 1;
 | 
				
			||||||
			bpage->id = i + 1;
 | 
								bpage->id = i + 1;
 | 
				
			||||||
		} else {
 | 
							} else {
 | 
				
			||||||
| 
						 | 
					@ -1844,6 +1950,7 @@ static struct ring_buffer_per_cpu *
 | 
				
			||||||
rb_allocate_cpu_buffer(struct trace_buffer *buffer, long nr_pages, int cpu)
 | 
					rb_allocate_cpu_buffer(struct trace_buffer *buffer, long nr_pages, int cpu)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct ring_buffer_per_cpu *cpu_buffer;
 | 
						struct ring_buffer_per_cpu *cpu_buffer;
 | 
				
			||||||
 | 
						struct ring_buffer_meta *meta;
 | 
				
			||||||
	struct buffer_page *bpage;
 | 
						struct buffer_page *bpage;
 | 
				
			||||||
	struct page *page;
 | 
						struct page *page;
 | 
				
			||||||
	int ret;
 | 
						int ret;
 | 
				
			||||||
| 
						 | 
					@ -1884,6 +1991,8 @@ rb_allocate_cpu_buffer(struct trace_buffer *buffer, long nr_pages, int cpu)
 | 
				
			||||||
		bpage->page = rb_range_buffer(cpu_buffer, 0);
 | 
							bpage->page = rb_range_buffer(cpu_buffer, 0);
 | 
				
			||||||
		if (!bpage->page)
 | 
							if (!bpage->page)
 | 
				
			||||||
			goto fail_free_reader;
 | 
								goto fail_free_reader;
 | 
				
			||||||
 | 
							if (cpu_buffer->ring_meta->head_buffer)
 | 
				
			||||||
 | 
								rb_meta_buffer_update(cpu_buffer, bpage);
 | 
				
			||||||
		bpage->range = 1;
 | 
							bpage->range = 1;
 | 
				
			||||||
	} else {
 | 
						} else {
 | 
				
			||||||
		page = alloc_pages_node(cpu_to_node(cpu),
 | 
							page = alloc_pages_node(cpu_to_node(cpu),
 | 
				
			||||||
| 
						 | 
					@ -1902,14 +2011,32 @@ rb_allocate_cpu_buffer(struct trace_buffer *buffer, long nr_pages, int cpu)
 | 
				
			||||||
	if (ret < 0)
 | 
						if (ret < 0)
 | 
				
			||||||
		goto fail_free_reader;
 | 
							goto fail_free_reader;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						/* If the boot meta was valid then this has already been updated */
 | 
				
			||||||
 | 
						meta = cpu_buffer->ring_meta;
 | 
				
			||||||
 | 
						if (!meta || !meta->head_buffer ||
 | 
				
			||||||
 | 
						    !cpu_buffer->head_page || !cpu_buffer->commit_page || !cpu_buffer->tail_page) {
 | 
				
			||||||
 | 
							if (meta && meta->head_buffer &&
 | 
				
			||||||
 | 
							    (cpu_buffer->head_page || cpu_buffer->commit_page || cpu_buffer->tail_page)) {
 | 
				
			||||||
 | 
								pr_warn("Ring buffer meta buffers not all mapped\n");
 | 
				
			||||||
 | 
								if (!cpu_buffer->head_page)
 | 
				
			||||||
 | 
									pr_warn("   Missing head_page\n");
 | 
				
			||||||
 | 
								if (!cpu_buffer->commit_page)
 | 
				
			||||||
 | 
									pr_warn("   Missing commit_page\n");
 | 
				
			||||||
 | 
								if (!cpu_buffer->tail_page)
 | 
				
			||||||
 | 
									pr_warn("   Missing tail_page\n");
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		cpu_buffer->head_page
 | 
							cpu_buffer->head_page
 | 
				
			||||||
			= list_entry(cpu_buffer->pages, struct buffer_page, list);
 | 
								= list_entry(cpu_buffer->pages, struct buffer_page, list);
 | 
				
			||||||
		cpu_buffer->tail_page = cpu_buffer->commit_page = cpu_buffer->head_page;
 | 
							cpu_buffer->tail_page = cpu_buffer->commit_page = cpu_buffer->head_page;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		rb_head_page_activate(cpu_buffer);
 | 
							rb_head_page_activate(cpu_buffer);
 | 
				
			||||||
	if (cpu_buffer->ring_meta) {
 | 
					
 | 
				
			||||||
		struct ring_buffer_meta *meta = cpu_buffer->ring_meta;
 | 
							if (cpu_buffer->ring_meta)
 | 
				
			||||||
			meta->commit_buffer = meta->head_buffer;
 | 
								meta->commit_buffer = meta->head_buffer;
 | 
				
			||||||
 | 
						} else {
 | 
				
			||||||
 | 
							/* The valid meta buffer still needs to activate the head page */
 | 
				
			||||||
 | 
							rb_head_page_activate(cpu_buffer);
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	return cpu_buffer;
 | 
						return cpu_buffer;
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue