mirror of
				https://github.com/torvalds/linux.git
				synced 2025-10-31 08:38:45 +02:00 
			
		
		
		
	 4afb85f5e3
			
		
	
	
		4afb85f5e3
		
	
	
	
	
		
			
			When using vmalloc with VM_ALLOW_HUGE_VMAP flag, it will set the alignment to PMD_SIZE internally, if it deems huge mappings to be eligible. Therefore, setting the alignment in execmem_vmalloc is redundant. Apart from this, it also reduces the probability of allocation in case vmalloc fails to allocate hugepages - in the fallback case, vmalloc tries to use the original alignment and allocate basepages, which unfortunately will again be PMD_SIZE passed over from execmem_vmalloc, thus constraining the search for a free space in vmalloc region. Therefore, remove this constraint. Link: https://lkml.kernel.org/r/20250918093453.75676-1-dev.jain@arm.com Signed-off-by: Dev Jain <dev.jain@arm.com> Reviewed-by: Mike Rapoport (Microsoft) <rppt@kernel.org> Cc: Ryan Roberts <ryan.roberts@arm.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
		
			
				
	
	
		
			593 lines
		
	
	
	
		
			14 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			593 lines
		
	
	
	
		
			14 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| /*
 | |
|  * Copyright (C) 2002 Richard Henderson
 | |
|  * Copyright (C) 2001 Rusty Russell, 2002, 2010 Rusty Russell IBM.
 | |
|  * Copyright (C) 2023 Luis Chamberlain <mcgrof@kernel.org>
 | |
|  * Copyright (C) 2024 Mike Rapoport IBM.
 | |
|  */
 | |
| 
 | |
| #define pr_fmt(fmt) "execmem: " fmt
 | |
| 
 | |
| #include <linux/mm.h>
 | |
| #include <linux/mutex.h>
 | |
| #include <linux/vmalloc.h>
 | |
| #include <linux/execmem.h>
 | |
| #include <linux/maple_tree.h>
 | |
| #include <linux/set_memory.h>
 | |
| #include <linux/moduleloader.h>
 | |
| #include <linux/text-patching.h>
 | |
| 
 | |
| #include <asm/tlbflush.h>
 | |
| 
 | |
| #include "internal.h"
 | |
| 
 | |
| static struct execmem_info *execmem_info __ro_after_init;
 | |
| static struct execmem_info default_execmem_info __ro_after_init;
 | |
| 
 | |
| #ifdef CONFIG_MMU
 | |
| static void *execmem_vmalloc(struct execmem_range *range, size_t size,
 | |
| 			     pgprot_t pgprot, unsigned long vm_flags)
 | |
| {
 | |
| 	bool kasan = range->flags & EXECMEM_KASAN_SHADOW;
 | |
| 	gfp_t gfp_flags = GFP_KERNEL | __GFP_NOWARN;
 | |
| 	unsigned int align = range->alignment;
 | |
| 	unsigned long start = range->start;
 | |
| 	unsigned long end = range->end;
 | |
| 	void *p;
 | |
| 
 | |
| 	if (kasan)
 | |
| 		vm_flags |= VM_DEFER_KMEMLEAK;
 | |
| 
 | |
| 	p = __vmalloc_node_range(size, align, start, end, gfp_flags,
 | |
| 				 pgprot, vm_flags, NUMA_NO_NODE,
 | |
| 				 __builtin_return_address(0));
 | |
| 	if (!p && range->fallback_start) {
 | |
| 		start = range->fallback_start;
 | |
| 		end = range->fallback_end;
 | |
| 		p = __vmalloc_node_range(size, align, start, end, gfp_flags,
 | |
| 					 pgprot, vm_flags, NUMA_NO_NODE,
 | |
| 					 __builtin_return_address(0));
 | |
| 	}
 | |
| 
 | |
| 	if (!p) {
 | |
| 		pr_warn_ratelimited("unable to allocate memory\n");
 | |
| 		return NULL;
 | |
| 	}
 | |
| 
 | |
| 	if (kasan && (kasan_alloc_module_shadow(p, size, GFP_KERNEL) < 0)) {
 | |
| 		vfree(p);
 | |
| 		return NULL;
 | |
| 	}
 | |
| 
 | |
| 	return p;
 | |
| }
 | |
| 
 | |
| struct vm_struct *execmem_vmap(size_t size)
 | |
| {
 | |
| 	struct execmem_range *range = &execmem_info->ranges[EXECMEM_MODULE_DATA];
 | |
| 	struct vm_struct *area;
 | |
| 
 | |
| 	area = __get_vm_area_node(size, range->alignment, PAGE_SHIFT, VM_ALLOC,
 | |
| 				  range->start, range->end, NUMA_NO_NODE,
 | |
| 				  GFP_KERNEL, __builtin_return_address(0));
 | |
| 	if (!area && range->fallback_start)
 | |
| 		area = __get_vm_area_node(size, range->alignment, PAGE_SHIFT, VM_ALLOC,
 | |
| 					  range->fallback_start, range->fallback_end,
 | |
| 					  NUMA_NO_NODE, GFP_KERNEL, __builtin_return_address(0));
 | |
| 
 | |
| 	return area;
 | |
| }
 | |
| #else
 | |
| static void *execmem_vmalloc(struct execmem_range *range, size_t size,
 | |
| 			     pgprot_t pgprot, unsigned long vm_flags)
 | |
| {
 | |
| 	return vmalloc(size);
 | |
| }
 | |
| #endif /* CONFIG_MMU */
 | |
| 
 | |
| #ifdef CONFIG_ARCH_HAS_EXECMEM_ROX
 | |
| struct execmem_cache {
 | |
| 	struct mutex mutex;
 | |
| 	struct maple_tree busy_areas;
 | |
| 	struct maple_tree free_areas;
 | |
| 	unsigned int pending_free_cnt;	/* protected by mutex */
 | |
| };
 | |
| 
 | |
| /* delay to schedule asynchronous free if fast path free fails */
 | |
| #define FREE_DELAY	(msecs_to_jiffies(10))
 | |
| 
 | |
| /* mark entries in busy_areas that should be freed asynchronously */
 | |
| #define PENDING_FREE_MASK	(1 << (PAGE_SHIFT - 1))
 | |
| 
 | |
| static struct execmem_cache execmem_cache = {
 | |
| 	.mutex = __MUTEX_INITIALIZER(execmem_cache.mutex),
 | |
| 	.busy_areas = MTREE_INIT_EXT(busy_areas, MT_FLAGS_LOCK_EXTERN,
 | |
| 				     execmem_cache.mutex),
 | |
| 	.free_areas = MTREE_INIT_EXT(free_areas, MT_FLAGS_LOCK_EXTERN,
 | |
| 				     execmem_cache.mutex),
 | |
| };
 | |
| 
 | |
| static inline unsigned long mas_range_len(struct ma_state *mas)
 | |
| {
 | |
| 	return mas->last - mas->index + 1;
 | |
| }
 | |
| 
 | |
| static int execmem_set_direct_map_valid(struct vm_struct *vm, bool valid)
 | |
| {
 | |
| 	unsigned int nr = (1 << get_vm_area_page_order(vm));
 | |
| 	unsigned int updated = 0;
 | |
| 	int err = 0;
 | |
| 
 | |
| 	for (int i = 0; i < vm->nr_pages; i += nr) {
 | |
| 		err = set_direct_map_valid_noflush(vm->pages[i], nr, valid);
 | |
| 		if (err)
 | |
| 			goto err_restore;
 | |
| 		updated += nr;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| 
 | |
| err_restore:
 | |
| 	for (int i = 0; i < updated; i += nr)
 | |
| 		set_direct_map_valid_noflush(vm->pages[i], nr, !valid);
 | |
| 
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| static int execmem_force_rw(void *ptr, size_t size)
 | |
| {
 | |
| 	unsigned int nr = PAGE_ALIGN(size) >> PAGE_SHIFT;
 | |
| 	unsigned long addr = (unsigned long)ptr;
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = set_memory_nx(addr, nr);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	return set_memory_rw(addr, nr);
 | |
| }
 | |
| 
 | |
| int execmem_restore_rox(void *ptr, size_t size)
 | |
| {
 | |
| 	unsigned int nr = PAGE_ALIGN(size) >> PAGE_SHIFT;
 | |
| 	unsigned long addr = (unsigned long)ptr;
 | |
| 
 | |
| 	return set_memory_rox(addr, nr);
 | |
| }
 | |
| 
 | |
| static void execmem_cache_clean(struct work_struct *work)
 | |
| {
 | |
| 	struct maple_tree *free_areas = &execmem_cache.free_areas;
 | |
| 	struct mutex *mutex = &execmem_cache.mutex;
 | |
| 	MA_STATE(mas, free_areas, 0, ULONG_MAX);
 | |
| 	void *area;
 | |
| 
 | |
| 	mutex_lock(mutex);
 | |
| 	mas_for_each(&mas, area, ULONG_MAX) {
 | |
| 		size_t size = mas_range_len(&mas);
 | |
| 
 | |
| 		if (IS_ALIGNED(size, PMD_SIZE) &&
 | |
| 		    IS_ALIGNED(mas.index, PMD_SIZE)) {
 | |
| 			struct vm_struct *vm = find_vm_area(area);
 | |
| 
 | |
| 			execmem_set_direct_map_valid(vm, true);
 | |
| 			mas_store_gfp(&mas, NULL, GFP_KERNEL);
 | |
| 			vfree(area);
 | |
| 		}
 | |
| 	}
 | |
| 	mutex_unlock(mutex);
 | |
| }
 | |
| 
 | |
| static DECLARE_WORK(execmem_cache_clean_work, execmem_cache_clean);
 | |
| 
 | |
| static int execmem_cache_add_locked(void *ptr, size_t size, gfp_t gfp_mask)
 | |
| {
 | |
| 	struct maple_tree *free_areas = &execmem_cache.free_areas;
 | |
| 	unsigned long addr = (unsigned long)ptr;
 | |
| 	MA_STATE(mas, free_areas, addr - 1, addr + 1);
 | |
| 	unsigned long lower, upper;
 | |
| 	void *area = NULL;
 | |
| 
 | |
| 	lower = addr;
 | |
| 	upper = addr + size - 1;
 | |
| 
 | |
| 	area = mas_walk(&mas);
 | |
| 	if (area && mas.last == addr - 1)
 | |
| 		lower = mas.index;
 | |
| 
 | |
| 	area = mas_next(&mas, ULONG_MAX);
 | |
| 	if (area && mas.index == addr + size)
 | |
| 		upper = mas.last;
 | |
| 
 | |
| 	mas_set_range(&mas, lower, upper);
 | |
| 	return mas_store_gfp(&mas, (void *)lower, gfp_mask);
 | |
| }
 | |
| 
 | |
| static int execmem_cache_add(void *ptr, size_t size, gfp_t gfp_mask)
 | |
| {
 | |
| 	guard(mutex)(&execmem_cache.mutex);
 | |
| 
 | |
| 	return execmem_cache_add_locked(ptr, size, gfp_mask);
 | |
| }
 | |
| 
 | |
| static bool within_range(struct execmem_range *range, struct ma_state *mas,
 | |
| 			 size_t size)
 | |
| {
 | |
| 	unsigned long addr = mas->index;
 | |
| 
 | |
| 	if (addr >= range->start && addr + size < range->end)
 | |
| 		return true;
 | |
| 
 | |
| 	if (range->fallback_start &&
 | |
| 	    addr >= range->fallback_start && addr + size < range->fallback_end)
 | |
| 		return true;
 | |
| 
 | |
| 	return false;
 | |
| }
 | |
| 
 | |
| static void *__execmem_cache_alloc(struct execmem_range *range, size_t size)
 | |
| {
 | |
| 	struct maple_tree *free_areas = &execmem_cache.free_areas;
 | |
| 	struct maple_tree *busy_areas = &execmem_cache.busy_areas;
 | |
| 	MA_STATE(mas_free, free_areas, 0, ULONG_MAX);
 | |
| 	MA_STATE(mas_busy, busy_areas, 0, ULONG_MAX);
 | |
| 	struct mutex *mutex = &execmem_cache.mutex;
 | |
| 	unsigned long addr, last, area_size = 0;
 | |
| 	void *area, *ptr = NULL;
 | |
| 	int err;
 | |
| 
 | |
| 	mutex_lock(mutex);
 | |
| 	mas_for_each(&mas_free, area, ULONG_MAX) {
 | |
| 		area_size = mas_range_len(&mas_free);
 | |
| 
 | |
| 		if (area_size >= size && within_range(range, &mas_free, size))
 | |
| 			break;
 | |
| 	}
 | |
| 
 | |
| 	if (area_size < size)
 | |
| 		goto out_unlock;
 | |
| 
 | |
| 	addr = mas_free.index;
 | |
| 	last = mas_free.last;
 | |
| 
 | |
| 	/* insert allocated size to busy_areas at range [addr, addr + size) */
 | |
| 	mas_set_range(&mas_busy, addr, addr + size - 1);
 | |
| 	err = mas_store_gfp(&mas_busy, (void *)addr, GFP_KERNEL);
 | |
| 	if (err)
 | |
| 		goto out_unlock;
 | |
| 
 | |
| 	mas_store_gfp(&mas_free, NULL, GFP_KERNEL);
 | |
| 	if (area_size > size) {
 | |
| 		void *ptr = (void *)(addr + size);
 | |
| 
 | |
| 		/*
 | |
| 		 * re-insert remaining free size to free_areas at range
 | |
| 		 * [addr + size, last]
 | |
| 		 */
 | |
| 		mas_set_range(&mas_free, addr + size, last);
 | |
| 		err = mas_store_gfp(&mas_free, ptr, GFP_KERNEL);
 | |
| 		if (err) {
 | |
| 			mas_store_gfp(&mas_busy, NULL, GFP_KERNEL);
 | |
| 			goto out_unlock;
 | |
| 		}
 | |
| 	}
 | |
| 	ptr = (void *)addr;
 | |
| 
 | |
| out_unlock:
 | |
| 	mutex_unlock(mutex);
 | |
| 	return ptr;
 | |
| }
 | |
| 
 | |
| static int execmem_cache_populate(struct execmem_range *range, size_t size)
 | |
| {
 | |
| 	unsigned long vm_flags = VM_ALLOW_HUGE_VMAP;
 | |
| 	struct vm_struct *vm;
 | |
| 	size_t alloc_size;
 | |
| 	int err = -ENOMEM;
 | |
| 	void *p;
 | |
| 
 | |
| 	alloc_size = round_up(size, PMD_SIZE);
 | |
| 	p = execmem_vmalloc(range, alloc_size, PAGE_KERNEL, vm_flags);
 | |
| 	if (!p) {
 | |
| 		alloc_size = size;
 | |
| 		p = execmem_vmalloc(range, alloc_size, PAGE_KERNEL, vm_flags);
 | |
| 	}
 | |
| 
 | |
| 	if (!p)
 | |
| 		return err;
 | |
| 
 | |
| 	vm = find_vm_area(p);
 | |
| 	if (!vm)
 | |
| 		goto err_free_mem;
 | |
| 
 | |
| 	/* fill memory with instructions that will trap */
 | |
| 	execmem_fill_trapping_insns(p, alloc_size);
 | |
| 
 | |
| 	err = set_memory_rox((unsigned long)p, vm->nr_pages);
 | |
| 	if (err)
 | |
| 		goto err_free_mem;
 | |
| 
 | |
| 	err = execmem_cache_add(p, alloc_size, GFP_KERNEL);
 | |
| 	if (err)
 | |
| 		goto err_reset_direct_map;
 | |
| 
 | |
| 	return 0;
 | |
| 
 | |
| err_reset_direct_map:
 | |
| 	execmem_set_direct_map_valid(vm, true);
 | |
| err_free_mem:
 | |
| 	vfree(p);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| static void *execmem_cache_alloc(struct execmem_range *range, size_t size)
 | |
| {
 | |
| 	void *p;
 | |
| 	int err;
 | |
| 
 | |
| 	p = __execmem_cache_alloc(range, size);
 | |
| 	if (p)
 | |
| 		return p;
 | |
| 
 | |
| 	err = execmem_cache_populate(range, size);
 | |
| 	if (err)
 | |
| 		return NULL;
 | |
| 
 | |
| 	return __execmem_cache_alloc(range, size);
 | |
| }
 | |
| 
 | |
| static inline bool is_pending_free(void *ptr)
 | |
| {
 | |
| 	return ((unsigned long)ptr & PENDING_FREE_MASK);
 | |
| }
 | |
| 
 | |
| static inline void *pending_free_set(void *ptr)
 | |
| {
 | |
| 	return (void *)((unsigned long)ptr | PENDING_FREE_MASK);
 | |
| }
 | |
| 
 | |
| static inline void *pending_free_clear(void *ptr)
 | |
| {
 | |
| 	return (void *)((unsigned long)ptr & ~PENDING_FREE_MASK);
 | |
| }
 | |
| 
 | |
| static int __execmem_cache_free(struct ma_state *mas, void *ptr, gfp_t gfp_mask)
 | |
| {
 | |
| 	size_t size = mas_range_len(mas);
 | |
| 	int err;
 | |
| 
 | |
| 	err = execmem_force_rw(ptr, size);
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 
 | |
| 	execmem_fill_trapping_insns(ptr, size);
 | |
| 	execmem_restore_rox(ptr, size);
 | |
| 
 | |
| 	err = execmem_cache_add_locked(ptr, size, gfp_mask);
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 
 | |
| 	mas_store_gfp(mas, NULL, gfp_mask);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void execmem_cache_free_slow(struct work_struct *work);
 | |
| static DECLARE_DELAYED_WORK(execmem_cache_free_work, execmem_cache_free_slow);
 | |
| 
 | |
| static void execmem_cache_free_slow(struct work_struct *work)
 | |
| {
 | |
| 	struct maple_tree *busy_areas = &execmem_cache.busy_areas;
 | |
| 	MA_STATE(mas, busy_areas, 0, ULONG_MAX);
 | |
| 	void *area;
 | |
| 
 | |
| 	guard(mutex)(&execmem_cache.mutex);
 | |
| 
 | |
| 	if (!execmem_cache.pending_free_cnt)
 | |
| 		return;
 | |
| 
 | |
| 	mas_for_each(&mas, area, ULONG_MAX) {
 | |
| 		if (!is_pending_free(area))
 | |
| 			continue;
 | |
| 
 | |
| 		area = pending_free_clear(area);
 | |
| 		if (__execmem_cache_free(&mas, area, GFP_KERNEL))
 | |
| 			continue;
 | |
| 
 | |
| 		execmem_cache.pending_free_cnt--;
 | |
| 	}
 | |
| 
 | |
| 	if (execmem_cache.pending_free_cnt)
 | |
| 		schedule_delayed_work(&execmem_cache_free_work, FREE_DELAY);
 | |
| 	else
 | |
| 		schedule_work(&execmem_cache_clean_work);
 | |
| }
 | |
| 
 | |
| static bool execmem_cache_free(void *ptr)
 | |
| {
 | |
| 	struct maple_tree *busy_areas = &execmem_cache.busy_areas;
 | |
| 	unsigned long addr = (unsigned long)ptr;
 | |
| 	MA_STATE(mas, busy_areas, addr, addr);
 | |
| 	void *area;
 | |
| 	int err;
 | |
| 
 | |
| 	guard(mutex)(&execmem_cache.mutex);
 | |
| 
 | |
| 	area = mas_walk(&mas);
 | |
| 	if (!area)
 | |
| 		return false;
 | |
| 
 | |
| 	err = __execmem_cache_free(&mas, area, GFP_KERNEL | __GFP_NORETRY);
 | |
| 	if (err) {
 | |
| 		/*
 | |
| 		 * mas points to exact slot we've got the area from, nothing
 | |
| 		 * else can modify the tree because of the mutex, so there
 | |
| 		 * won't be any allocations in mas_store_gfp() and it will just
 | |
| 		 * change the pointer.
 | |
| 		 */
 | |
| 		area = pending_free_set(area);
 | |
| 		mas_store_gfp(&mas, area, GFP_KERNEL);
 | |
| 		execmem_cache.pending_free_cnt++;
 | |
| 		schedule_delayed_work(&execmem_cache_free_work, FREE_DELAY);
 | |
| 		return true;
 | |
| 	}
 | |
| 
 | |
| 	schedule_work(&execmem_cache_clean_work);
 | |
| 
 | |
| 	return true;
 | |
| }
 | |
| 
 | |
| #else /* CONFIG_ARCH_HAS_EXECMEM_ROX */
 | |
| /*
 | |
|  * when ROX cache is not used the permissions defined by architectures for
 | |
|  * execmem ranges that are updated before use (e.g. EXECMEM_MODULE_TEXT) must
 | |
|  * be writable anyway
 | |
|  */
 | |
| static inline int execmem_force_rw(void *ptr, size_t size)
 | |
| {
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void *execmem_cache_alloc(struct execmem_range *range, size_t size)
 | |
| {
 | |
| 	return NULL;
 | |
| }
 | |
| 
 | |
| static bool execmem_cache_free(void *ptr)
 | |
| {
 | |
| 	return false;
 | |
| }
 | |
| #endif /* CONFIG_ARCH_HAS_EXECMEM_ROX */
 | |
| 
 | |
| void *execmem_alloc(enum execmem_type type, size_t size)
 | |
| {
 | |
| 	struct execmem_range *range = &execmem_info->ranges[type];
 | |
| 	bool use_cache = range->flags & EXECMEM_ROX_CACHE;
 | |
| 	unsigned long vm_flags = VM_FLUSH_RESET_PERMS;
 | |
| 	pgprot_t pgprot = range->pgprot;
 | |
| 	void *p = NULL;
 | |
| 
 | |
| 	size = PAGE_ALIGN(size);
 | |
| 
 | |
| 	if (use_cache)
 | |
| 		p = execmem_cache_alloc(range, size);
 | |
| 	else
 | |
| 		p = execmem_vmalloc(range, size, pgprot, vm_flags);
 | |
| 
 | |
| 	return kasan_reset_tag(p);
 | |
| }
 | |
| 
 | |
| void *execmem_alloc_rw(enum execmem_type type, size_t size)
 | |
| {
 | |
| 	void *p __free(execmem) = execmem_alloc(type, size);
 | |
| 	int err;
 | |
| 
 | |
| 	if (!p)
 | |
| 		return NULL;
 | |
| 
 | |
| 	err = execmem_force_rw(p, size);
 | |
| 	if (err)
 | |
| 		return NULL;
 | |
| 
 | |
| 	return no_free_ptr(p);
 | |
| }
 | |
| 
 | |
| void execmem_free(void *ptr)
 | |
| {
 | |
| 	/*
 | |
| 	 * This memory may be RO, and freeing RO memory in an interrupt is not
 | |
| 	 * supported by vmalloc.
 | |
| 	 */
 | |
| 	WARN_ON(in_interrupt());
 | |
| 
 | |
| 	if (!execmem_cache_free(ptr))
 | |
| 		vfree(ptr);
 | |
| }
 | |
| 
 | |
| bool execmem_is_rox(enum execmem_type type)
 | |
| {
 | |
| 	return !!(execmem_info->ranges[type].flags & EXECMEM_ROX_CACHE);
 | |
| }
 | |
| 
 | |
| static bool execmem_validate(struct execmem_info *info)
 | |
| {
 | |
| 	struct execmem_range *r = &info->ranges[EXECMEM_DEFAULT];
 | |
| 
 | |
| 	if (!r->alignment || !r->start || !r->end || !pgprot_val(r->pgprot)) {
 | |
| 		pr_crit("Invalid parameters for execmem allocator, module loading will fail");
 | |
| 		return false;
 | |
| 	}
 | |
| 
 | |
| 	if (!IS_ENABLED(CONFIG_ARCH_HAS_EXECMEM_ROX)) {
 | |
| 		for (int i = EXECMEM_DEFAULT; i < EXECMEM_TYPE_MAX; i++) {
 | |
| 			r = &info->ranges[i];
 | |
| 
 | |
| 			if (r->flags & EXECMEM_ROX_CACHE) {
 | |
| 				pr_warn_once("ROX cache is not supported\n");
 | |
| 				r->flags &= ~EXECMEM_ROX_CACHE;
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return true;
 | |
| }
 | |
| 
 | |
| static void execmem_init_missing(struct execmem_info *info)
 | |
| {
 | |
| 	struct execmem_range *default_range = &info->ranges[EXECMEM_DEFAULT];
 | |
| 
 | |
| 	for (int i = EXECMEM_DEFAULT + 1; i < EXECMEM_TYPE_MAX; i++) {
 | |
| 		struct execmem_range *r = &info->ranges[i];
 | |
| 
 | |
| 		if (!r->start) {
 | |
| 			if (i == EXECMEM_MODULE_DATA)
 | |
| 				r->pgprot = PAGE_KERNEL;
 | |
| 			else
 | |
| 				r->pgprot = default_range->pgprot;
 | |
| 			r->alignment = default_range->alignment;
 | |
| 			r->start = default_range->start;
 | |
| 			r->end = default_range->end;
 | |
| 			r->flags = default_range->flags;
 | |
| 			r->fallback_start = default_range->fallback_start;
 | |
| 			r->fallback_end = default_range->fallback_end;
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| struct execmem_info * __weak execmem_arch_setup(void)
 | |
| {
 | |
| 	return NULL;
 | |
| }
 | |
| 
 | |
| static void __init __execmem_init(void)
 | |
| {
 | |
| 	struct execmem_info *info = execmem_arch_setup();
 | |
| 
 | |
| 	if (!info) {
 | |
| 		info = execmem_info = &default_execmem_info;
 | |
| 		info->ranges[EXECMEM_DEFAULT].start = VMALLOC_START;
 | |
| 		info->ranges[EXECMEM_DEFAULT].end = VMALLOC_END;
 | |
| 		info->ranges[EXECMEM_DEFAULT].pgprot = PAGE_KERNEL_EXEC;
 | |
| 		info->ranges[EXECMEM_DEFAULT].alignment = 1;
 | |
| 	}
 | |
| 
 | |
| 	if (!execmem_validate(info))
 | |
| 		return;
 | |
| 
 | |
| 	execmem_init_missing(info);
 | |
| 
 | |
| 	execmem_info = info;
 | |
| }
 | |
| 
 | |
| #ifdef CONFIG_ARCH_WANTS_EXECMEM_LATE
 | |
| static int __init execmem_late_init(void)
 | |
| {
 | |
| 	__execmem_init();
 | |
| 	return 0;
 | |
| }
 | |
| core_initcall(execmem_late_init);
 | |
| #else
 | |
| void __init execmem_init(void)
 | |
| {
 | |
| 	__execmem_init();
 | |
| }
 | |
| #endif
 |