forked from mirrors/linux
		
	 4ab5f8ec7d
			
		
	
	
		4ab5f8ec7d
		
	
	
	
	
		
			
			Patch series "mm, dma, arm64: Reduce ARCH_KMALLOC_MINALIGN to 8", v7. A series reducing the kmalloc() minimum alignment on arm64 to 8 (from 128). This patch (of 17): In preparation for supporting a kmalloc() minimum alignment smaller than the arch DMA alignment, decouple the two definitions. This requires that either the kmalloc() caches are aligned to a (run-time) cache-line size or the DMA API bounces unaligned kmalloc() allocations. Subsequent patches will implement both options. After this patch, ARCH_DMA_MINALIGN is expected to be used in static alignment annotations and defined by an architecture to be the maximum alignment for all supported configurations/SoCs in a single Image. Architectures opting in to a smaller ARCH_KMALLOC_MINALIGN will need to define its value in the arch headers. Since ARCH_DMA_MINALIGN is now always defined, adjust the #ifdef in dma_get_cache_alignment() so that there is no change for architectures not requiring a minimum DMA alignment. Link: https://lkml.kernel.org/r/20230612153201.554742-1-catalin.marinas@arm.com Link: https://lkml.kernel.org/r/20230612153201.554742-2-catalin.marinas@arm.com Signed-off-by: Catalin Marinas <catalin.marinas@arm.com> Tested-by: Isaac J. Manjarres <isaacmanjarres@google.com> Cc: Vlastimil Babka <vbabka@suse.cz> Cc: Christoph Hellwig <hch@lst.de> Cc: Robin Murphy <robin.murphy@arm.com> Cc: Alasdair Kergon <agk@redhat.com> Cc: Ard Biesheuvel <ardb@kernel.org> Cc: Arnd Bergmann <arnd@arndb.de> Cc: Daniel Vetter <daniel@ffwll.ch> Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org> Cc: Herbert Xu <herbert@gondor.apana.org.au> Cc: Joerg Roedel <joro@8bytes.org> Cc: Jonathan Cameron <jic23@kernel.org> Cc: Marc Zyngier <maz@kernel.org> Cc: Mark Brown <broonie@kernel.org> Cc: Mike Snitzer <snitzer@kernel.org> Cc: Rafael J. Wysocki <rafael@kernel.org> Cc: Saravana Kannan <saravanak@google.com> Cc: Will Deacon <will@kernel.org> Cc: Jerry Snitselaar <jsnitsel@redhat.com> Cc: Jonathan Cameron <Jonathan.Cameron@huawei.com> Cc: Lars-Peter Clausen <lars@metafoo.de> Cc: Logan Gunthorpe <logang@deltatee.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
		
			
				
	
	
		
			107 lines
		
	
	
	
		
			3 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			107 lines
		
	
	
	
		
			3 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /* SPDX-License-Identifier: GPL-2.0 */
 | |
| #ifndef __LINUX_CACHE_H
 | |
| #define __LINUX_CACHE_H
 | |
| 
 | |
| #include <uapi/linux/kernel.h>
 | |
| #include <asm/cache.h>
 | |
| 
 | |
| #ifndef L1_CACHE_ALIGN
 | |
| #define L1_CACHE_ALIGN(x) __ALIGN_KERNEL(x, L1_CACHE_BYTES)
 | |
| #endif
 | |
| 
 | |
| #ifndef SMP_CACHE_BYTES
 | |
| #define SMP_CACHE_BYTES L1_CACHE_BYTES
 | |
| #endif
 | |
| 
 | |
| /*
 | |
|  * __read_mostly is used to keep rarely changing variables out of frequently
 | |
|  * updated cachelines. Its use should be reserved for data that is used
 | |
|  * frequently in hot paths. Performance traces can help decide when to use
 | |
|  * this. You want __read_mostly data to be tightly packed, so that in the
 | |
|  * best case multiple frequently read variables for a hot path will be next
 | |
|  * to each other in order to reduce the number of cachelines needed to
 | |
|  * execute a critical path. We should be mindful and selective of its use.
 | |
|  * ie: if you're going to use it please supply a *good* justification in your
 | |
|  * commit log
 | |
|  */
 | |
| #ifndef __read_mostly
 | |
| #define __read_mostly
 | |
| #endif
 | |
| 
 | |
| /*
 | |
|  * __ro_after_init is used to mark things that are read-only after init (i.e.
 | |
|  * after mark_rodata_ro() has been called). These are effectively read-only,
 | |
|  * but may get written to during init, so can't live in .rodata (via "const").
 | |
|  */
 | |
| #ifndef __ro_after_init
 | |
| #define __ro_after_init __section(".data..ro_after_init")
 | |
| #endif
 | |
| 
 | |
| #ifndef ____cacheline_aligned
 | |
| #define ____cacheline_aligned __attribute__((__aligned__(SMP_CACHE_BYTES)))
 | |
| #endif
 | |
| 
 | |
| #ifndef ____cacheline_aligned_in_smp
 | |
| #ifdef CONFIG_SMP
 | |
| #define ____cacheline_aligned_in_smp ____cacheline_aligned
 | |
| #else
 | |
| #define ____cacheline_aligned_in_smp
 | |
| #endif /* CONFIG_SMP */
 | |
| #endif
 | |
| 
 | |
| #ifndef __cacheline_aligned
 | |
| #define __cacheline_aligned					\
 | |
|   __attribute__((__aligned__(SMP_CACHE_BYTES),			\
 | |
| 		 __section__(".data..cacheline_aligned")))
 | |
| #endif /* __cacheline_aligned */
 | |
| 
 | |
| #ifndef __cacheline_aligned_in_smp
 | |
| #ifdef CONFIG_SMP
 | |
| #define __cacheline_aligned_in_smp __cacheline_aligned
 | |
| #else
 | |
| #define __cacheline_aligned_in_smp
 | |
| #endif /* CONFIG_SMP */
 | |
| #endif
 | |
| 
 | |
| /*
 | |
|  * The maximum alignment needed for some critical structures
 | |
|  * These could be inter-node cacheline sizes/L3 cacheline
 | |
|  * size etc.  Define this in asm/cache.h for your arch
 | |
|  */
 | |
| #ifndef INTERNODE_CACHE_SHIFT
 | |
| #define INTERNODE_CACHE_SHIFT L1_CACHE_SHIFT
 | |
| #endif
 | |
| 
 | |
| #if !defined(____cacheline_internodealigned_in_smp)
 | |
| #if defined(CONFIG_SMP)
 | |
| #define ____cacheline_internodealigned_in_smp \
 | |
| 	__attribute__((__aligned__(1 << (INTERNODE_CACHE_SHIFT))))
 | |
| #else
 | |
| #define ____cacheline_internodealigned_in_smp
 | |
| #endif
 | |
| #endif
 | |
| 
 | |
| #ifndef CONFIG_ARCH_HAS_CACHE_LINE_SIZE
 | |
| #define cache_line_size()	L1_CACHE_BYTES
 | |
| #endif
 | |
| 
 | |
| /*
 | |
|  * Helper to add padding within a struct to ensure data fall into separate
 | |
|  * cachelines.
 | |
|  */
 | |
| #if defined(CONFIG_SMP)
 | |
| struct cacheline_padding {
 | |
| 	char x[0];
 | |
| } ____cacheline_internodealigned_in_smp;
 | |
| #define CACHELINE_PADDING(name)		struct cacheline_padding name
 | |
| #else
 | |
| #define CACHELINE_PADDING(name)
 | |
| #endif
 | |
| 
 | |
| #ifdef ARCH_DMA_MINALIGN
 | |
| #define ARCH_HAS_DMA_MINALIGN
 | |
| #else
 | |
| #define ARCH_DMA_MINALIGN __alignof__(unsigned long long)
 | |
| #endif
 | |
| 
 | |
| #endif /* __LINUX_CACHE_H */
 |