forked from mirrors/linux
		
	- move the USB special case that bounced DMA through a device
    bar into the USB code instead of handling it in the common
    DMA code (Laurentiu Tudor and Fredrik Noring)
  - don't dip into the global CMA pool for single page allocations
    (Nicolin Chen)
  - fix a crash when allocating memory for the atomic pool failed
    during boot (Florian Fainelli)
  - move support for MIPS-style uncached segments to the common
    code and use that for MIPS and nios2 (me)
  - make support for DMA_ATTR_NON_CONSISTENT and
    DMA_ATTR_NO_KERNEL_MAPPING generic (me)
  - convert nds32 to the generic remapping allocator (me)
 -----BEGIN PGP SIGNATURE-----
 
 iQI/BAABCgApFiEEgdbnc3r/njty3Iq9D55TZVIEUYMFAl0nPqgLHGhjaEBsc3Qu
 ZGUACgkQD55TZVIEUYNj2hAAxIv2O3wv6V5xhzWwOVo8e/xW1ZLlGAF0/z92u0do
 32Tm8jkdAGjZDnyxam7qisMSIjCNykpauQzVVxyUNBRSsn1V5t7KSaH3/OXCOVcr
 x2VWBirxGO2BbRseaCBjIcA/2qna+VIDGFcNXCtf6rM00YUK6qaJzkMwBKQAeYcM
 uJMJkaf8qaW4hygLJP8axXiGFdIJyFNLAlJ+ok6kYsJHHJNceOp0bo3CDa2mJBK9
 IhraK2zVkyE5EQkQM5cE/Kw1ppPelUKUkHwjgM4wpz2b18WbLu11nKP0hmUcvKRQ
 heY8xWiKxN0QTgS03ou7EVylyrSAE4dIKgzuA4VO32QCGsWypcAg4iU6s5TX6p9g
 tZEW2ckE6wbmRdQPyKoDpZg299/eQjRHc4MAA1yinT8tFMokw2tk8Fq1FWyltwL1
 8EiP5oNs2qUNvNgqUresl6/f6YOacFi1Q6IhgBVj6d6lyhMhlsHfW4w1XA1siv/I
 6l4qJbLohYab6hY7i+mBOd8iG/KrAlr4P6admnv2jDchswbb5t2j+ABE9xv++PFi
 u1HFqMlxqdWQaXGca2UeCUxUjkwO9N+kHpP+VRz+6D2b64dtCWSu8CN23sYXm2tO
 ubWIlrQQZPhhMkoFg7XqKSTacd+ut+SXN9Nxsyv548ETV0l1xbiLRHIbhyoIESD5
 RAI=
 =01Fr
 -----END PGP SIGNATURE-----
Merge tag 'dma-mapping-5.3' of git://git.infradead.org/users/hch/dma-mapping
Pull dma-mapping updates from Christoph Hellwig:
 - move the USB special case that bounced DMA through a device bar into
   the USB code instead of handling it in the common DMA code (Laurentiu
   Tudor and Fredrik Noring)
 - don't dip into the global CMA pool for single page allocations
   (Nicolin Chen)
 - fix a crash when allocating memory for the atomic pool failed during
   boot (Florian Fainelli)
 - move support for MIPS-style uncached segments to the common code and
   use that for MIPS and nios2 (me)
 - make support for DMA_ATTR_NON_CONSISTENT and
   DMA_ATTR_NO_KERNEL_MAPPING generic (me)
 - convert nds32 to the generic remapping allocator (me)
* tag 'dma-mapping-5.3' of git://git.infradead.org/users/hch/dma-mapping: (29 commits)
  dma-mapping: mark dma_alloc_need_uncached as __always_inline
  MIPS: only select ARCH_HAS_UNCACHED_SEGMENT for non-coherent platforms
  usb: host: Fix excessive alignment restriction for local memory allocations
  lib/genalloc.c: Add algorithm, align and zeroed family of DMA allocators
  nios2: use the generic uncached segment support in dma-direct
  nds32: use the generic remapping allocator for coherent DMA allocations
  arc: use the generic remapping allocator for coherent DMA allocations
  dma-direct: handle DMA_ATTR_NO_KERNEL_MAPPING in common code
  dma-direct: handle DMA_ATTR_NON_CONSISTENT in common code
  dma-mapping: add a dma_alloc_need_uncached helper
  openrisc: remove the partial DMA_ATTR_NON_CONSISTENT support
  arc: remove the partial DMA_ATTR_NON_CONSISTENT support
  arm-nommu: remove the partial DMA_ATTR_NON_CONSISTENT support
  ARM: dma-mapping: allow larger DMA mask than supported
  dma-mapping: truncate dma masks to what dma_addr_t can hold
  iommu/dma: Apply dma_{alloc,free}_contiguous functions
  dma-remap: Avoid de-referencing NULL atomic_pool
  MIPS: use the generic uncached segment support in dma-direct
  dma-direct: provide generic support for uncached kernel segments
  au1100fb: fix DMA API abuse
  ...
		
	
			
		
			
				
	
	
		
			222 lines
		
	
	
	
		
			7.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			222 lines
		
	
	
	
		
			7.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/* SPDX-License-Identifier: GPL-2.0-only */
 | 
						|
/*
 | 
						|
 * Basic general purpose allocator for managing special purpose
 | 
						|
 * memory, for example, memory that is not managed by the regular
 | 
						|
 * kmalloc/kfree interface.  Uses for this includes on-device special
 | 
						|
 * memory, uncached memory etc.
 | 
						|
 *
 | 
						|
 * It is safe to use the allocator in NMI handlers and other special
 | 
						|
 * unblockable contexts that could otherwise deadlock on locks.  This
 | 
						|
 * is implemented by using atomic operations and retries on any
 | 
						|
 * conflicts.  The disadvantage is that there may be livelocks in
 | 
						|
 * extreme cases.  For better scalability, one allocator can be used
 | 
						|
 * for each CPU.
 | 
						|
 *
 | 
						|
 * The lockless operation only works if there is enough memory
 | 
						|
 * available.  If new memory is added to the pool a lock has to be
 | 
						|
 * still taken.  So any user relying on locklessness has to ensure
 | 
						|
 * that sufficient memory is preallocated.
 | 
						|
 *
 | 
						|
 * The basic atomic operation of this allocator is cmpxchg on long.
 | 
						|
 * On architectures that don't have NMI-safe cmpxchg implementation,
 | 
						|
 * the allocator can NOT be used in NMI handler.  So code uses the
 | 
						|
 * allocator in NMI handler should depend on
 | 
						|
 * CONFIG_ARCH_HAVE_NMI_SAFE_CMPXCHG.
 | 
						|
 */
 | 
						|
 | 
						|
 | 
						|
#ifndef __GENALLOC_H__
 | 
						|
#define __GENALLOC_H__
 | 
						|
 | 
						|
#include <linux/types.h>
 | 
						|
#include <linux/spinlock_types.h>
 | 
						|
#include <linux/atomic.h>
 | 
						|
 | 
						|
struct device;
 | 
						|
struct device_node;
 | 
						|
struct gen_pool;
 | 
						|
 | 
						|
/**
 | 
						|
 * typedef genpool_algo_t: Allocation callback function type definition
 | 
						|
 * @map: Pointer to bitmap
 | 
						|
 * @size: The bitmap size in bits
 | 
						|
 * @start: The bitnumber to start searching at
 | 
						|
 * @nr: The number of zeroed bits we're looking for
 | 
						|
 * @data: optional additional data used by the callback
 | 
						|
 * @pool: the pool being allocated from
 | 
						|
 */
 | 
						|
typedef unsigned long (*genpool_algo_t)(unsigned long *map,
 | 
						|
			unsigned long size,
 | 
						|
			unsigned long start,
 | 
						|
			unsigned int nr,
 | 
						|
			void *data, struct gen_pool *pool,
 | 
						|
			unsigned long start_addr);
 | 
						|
 | 
						|
/*
 | 
						|
 *  General purpose special memory pool descriptor.
 | 
						|
 */
 | 
						|
struct gen_pool {
 | 
						|
	spinlock_t lock;
 | 
						|
	struct list_head chunks;	/* list of chunks in this pool */
 | 
						|
	int min_alloc_order;		/* minimum allocation order */
 | 
						|
 | 
						|
	genpool_algo_t algo;		/* allocation function */
 | 
						|
	void *data;
 | 
						|
 | 
						|
	const char *name;
 | 
						|
};
 | 
						|
 | 
						|
/*
 | 
						|
 *  General purpose special memory pool chunk descriptor.
 | 
						|
 */
 | 
						|
struct gen_pool_chunk {
 | 
						|
	struct list_head next_chunk;	/* next chunk in pool */
 | 
						|
	atomic_long_t avail;
 | 
						|
	phys_addr_t phys_addr;		/* physical starting address of memory chunk */
 | 
						|
	void *owner;			/* private data to retrieve at alloc time */
 | 
						|
	unsigned long start_addr;	/* start address of memory chunk */
 | 
						|
	unsigned long end_addr;		/* end address of memory chunk (inclusive) */
 | 
						|
	unsigned long bits[0];		/* bitmap for allocating memory chunk */
 | 
						|
};
 | 
						|
 | 
						|
/*
 | 
						|
 *  gen_pool data descriptor for gen_pool_first_fit_align.
 | 
						|
 */
 | 
						|
struct genpool_data_align {
 | 
						|
	int align;		/* alignment by bytes for starting address */
 | 
						|
};
 | 
						|
 | 
						|
/*
 | 
						|
 *  gen_pool data descriptor for gen_pool_fixed_alloc.
 | 
						|
 */
 | 
						|
struct genpool_data_fixed {
 | 
						|
	unsigned long offset;		/* The offset of the specific region */
 | 
						|
};
 | 
						|
 | 
						|
extern struct gen_pool *gen_pool_create(int, int);
 | 
						|
extern phys_addr_t gen_pool_virt_to_phys(struct gen_pool *pool, unsigned long);
 | 
						|
extern int gen_pool_add_owner(struct gen_pool *, unsigned long, phys_addr_t,
 | 
						|
			     size_t, int, void *);
 | 
						|
 | 
						|
static inline int gen_pool_add_virt(struct gen_pool *pool, unsigned long addr,
 | 
						|
		phys_addr_t phys, size_t size, int nid)
 | 
						|
{
 | 
						|
	return gen_pool_add_owner(pool, addr, phys, size, nid, NULL);
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
 * gen_pool_add - add a new chunk of special memory to the pool
 | 
						|
 * @pool: pool to add new memory chunk to
 | 
						|
 * @addr: starting address of memory chunk to add to pool
 | 
						|
 * @size: size in bytes of the memory chunk to add to pool
 | 
						|
 * @nid: node id of the node the chunk structure and bitmap should be
 | 
						|
 *       allocated on, or -1
 | 
						|
 *
 | 
						|
 * Add a new chunk of special memory to the specified pool.
 | 
						|
 *
 | 
						|
 * Returns 0 on success or a -ve errno on failure.
 | 
						|
 */
 | 
						|
static inline int gen_pool_add(struct gen_pool *pool, unsigned long addr,
 | 
						|
			       size_t size, int nid)
 | 
						|
{
 | 
						|
	return gen_pool_add_virt(pool, addr, -1, size, nid);
 | 
						|
}
 | 
						|
extern void gen_pool_destroy(struct gen_pool *);
 | 
						|
unsigned long gen_pool_alloc_algo_owner(struct gen_pool *pool, size_t size,
 | 
						|
		genpool_algo_t algo, void *data, void **owner);
 | 
						|
 | 
						|
static inline unsigned long gen_pool_alloc_owner(struct gen_pool *pool,
 | 
						|
		size_t size, void **owner)
 | 
						|
{
 | 
						|
	return gen_pool_alloc_algo_owner(pool, size, pool->algo, pool->data,
 | 
						|
			owner);
 | 
						|
}
 | 
						|
 | 
						|
static inline unsigned long gen_pool_alloc_algo(struct gen_pool *pool,
 | 
						|
		size_t size, genpool_algo_t algo, void *data)
 | 
						|
{
 | 
						|
	return gen_pool_alloc_algo_owner(pool, size, algo, data, NULL);
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
 * gen_pool_alloc - allocate special memory from the pool
 | 
						|
 * @pool: pool to allocate from
 | 
						|
 * @size: number of bytes to allocate from the pool
 | 
						|
 *
 | 
						|
 * Allocate the requested number of bytes from the specified pool.
 | 
						|
 * Uses the pool allocation function (with first-fit algorithm by default).
 | 
						|
 * Can not be used in NMI handler on architectures without
 | 
						|
 * NMI-safe cmpxchg implementation.
 | 
						|
 */
 | 
						|
static inline unsigned long gen_pool_alloc(struct gen_pool *pool, size_t size)
 | 
						|
{
 | 
						|
	return gen_pool_alloc_algo(pool, size, pool->algo, pool->data);
 | 
						|
}
 | 
						|
 | 
						|
extern void *gen_pool_dma_alloc(struct gen_pool *pool, size_t size,
 | 
						|
		dma_addr_t *dma);
 | 
						|
extern void *gen_pool_dma_alloc_algo(struct gen_pool *pool, size_t size,
 | 
						|
		dma_addr_t *dma, genpool_algo_t algo, void *data);
 | 
						|
extern void *gen_pool_dma_alloc_align(struct gen_pool *pool, size_t size,
 | 
						|
		dma_addr_t *dma, int align);
 | 
						|
extern void *gen_pool_dma_zalloc(struct gen_pool *pool, size_t size, dma_addr_t *dma);
 | 
						|
extern void *gen_pool_dma_zalloc_algo(struct gen_pool *pool, size_t size,
 | 
						|
		dma_addr_t *dma, genpool_algo_t algo, void *data);
 | 
						|
extern void *gen_pool_dma_zalloc_align(struct gen_pool *pool, size_t size,
 | 
						|
		dma_addr_t *dma, int align);
 | 
						|
extern void gen_pool_free_owner(struct gen_pool *pool, unsigned long addr,
 | 
						|
		size_t size, void **owner);
 | 
						|
static inline void gen_pool_free(struct gen_pool *pool, unsigned long addr,
 | 
						|
                size_t size)
 | 
						|
{
 | 
						|
	gen_pool_free_owner(pool, addr, size, NULL);
 | 
						|
}
 | 
						|
 | 
						|
extern void gen_pool_for_each_chunk(struct gen_pool *,
 | 
						|
	void (*)(struct gen_pool *, struct gen_pool_chunk *, void *), void *);
 | 
						|
extern size_t gen_pool_avail(struct gen_pool *);
 | 
						|
extern size_t gen_pool_size(struct gen_pool *);
 | 
						|
 | 
						|
extern void gen_pool_set_algo(struct gen_pool *pool, genpool_algo_t algo,
 | 
						|
		void *data);
 | 
						|
 | 
						|
extern unsigned long gen_pool_first_fit(unsigned long *map, unsigned long size,
 | 
						|
		unsigned long start, unsigned int nr, void *data,
 | 
						|
		struct gen_pool *pool, unsigned long start_addr);
 | 
						|
 | 
						|
extern unsigned long gen_pool_fixed_alloc(unsigned long *map,
 | 
						|
		unsigned long size, unsigned long start, unsigned int nr,
 | 
						|
		void *data, struct gen_pool *pool, unsigned long start_addr);
 | 
						|
 | 
						|
extern unsigned long gen_pool_first_fit_align(unsigned long *map,
 | 
						|
		unsigned long size, unsigned long start, unsigned int nr,
 | 
						|
		void *data, struct gen_pool *pool, unsigned long start_addr);
 | 
						|
 | 
						|
 | 
						|
extern unsigned long gen_pool_first_fit_order_align(unsigned long *map,
 | 
						|
		unsigned long size, unsigned long start, unsigned int nr,
 | 
						|
		void *data, struct gen_pool *pool, unsigned long start_addr);
 | 
						|
 | 
						|
extern unsigned long gen_pool_best_fit(unsigned long *map, unsigned long size,
 | 
						|
		unsigned long start, unsigned int nr, void *data,
 | 
						|
		struct gen_pool *pool, unsigned long start_addr);
 | 
						|
 | 
						|
 | 
						|
extern struct gen_pool *devm_gen_pool_create(struct device *dev,
 | 
						|
		int min_alloc_order, int nid, const char *name);
 | 
						|
extern struct gen_pool *gen_pool_get(struct device *dev, const char *name);
 | 
						|
 | 
						|
bool addr_in_gen_pool(struct gen_pool *pool, unsigned long start,
 | 
						|
			size_t size);
 | 
						|
 | 
						|
#ifdef CONFIG_OF
 | 
						|
extern struct gen_pool *of_gen_pool_get(struct device_node *np,
 | 
						|
	const char *propname, int index);
 | 
						|
#else
 | 
						|
static inline struct gen_pool *of_gen_pool_get(struct device_node *np,
 | 
						|
	const char *propname, int index)
 | 
						|
{
 | 
						|
	return NULL;
 | 
						|
}
 | 
						|
#endif
 | 
						|
#endif /* __GENALLOC_H__ */
 |