forked from mirrors/linux
		
	ARM: dma-mapping: remove custom consistent dma region
This patch changes dma-mapping subsystem to use generic vmalloc areas for all consistent dma allocations. This increases the total size limit of the consistent allocations and removes platform hacks and a lot of duplicated code. Atomic allocations are served from special pool preallocated on boot, because vmalloc areas cannot be reliably created in atomic context. Signed-off-by: Marek Szyprowski <m.szyprowski@samsung.com> Reviewed-by: Kyungmin Park <kyungmin.park@samsung.com> Reviewed-by: Minchan Kim <minchan@kernel.org>
This commit is contained in:
		
							parent
							
								
									5e6cafc83e
								
							
						
					
					
						commit
						e9da6e9905
					
				
					 6 changed files with 188 additions and 330 deletions
				
			
		|  | @ -526,7 +526,7 @@ bytes respectively. Such letter suffixes can also be entirely omitted. | |||
| 
 | ||||
| 	coherent_pool=nn[KMG]	[ARM,KNL] | ||||
| 			Sets the size of memory pool for coherent, atomic dma | ||||
| 			allocations if Contiguous Memory Allocator (CMA) is used. | ||||
| 			allocations, by default set to 256K. | ||||
| 
 | ||||
| 	code_bytes	[X86] How many bytes of object code to print | ||||
| 			in an oops report. | ||||
|  |  | |||
|  | @ -226,7 +226,7 @@ static inline int dma_mmap_writecombine(struct device *dev, struct vm_area_struc | |||
|  * DMA region above it's default value of 2MB. It must be called before the | ||||
|  * memory allocator is initialised, i.e. before any core_initcall. | ||||
|  */ | ||||
| extern void __init init_consistent_dma_size(unsigned long size); | ||||
| static inline void init_consistent_dma_size(unsigned long size) { } | ||||
| 
 | ||||
| /*
 | ||||
|  * For SA-1111, IXP425, and ADI systems  the dma-mapping functions are "magic" | ||||
|  |  | |||
|  | @ -22,6 +22,7 @@ | |||
| #include <linux/memblock.h> | ||||
| #include <linux/slab.h> | ||||
| #include <linux/iommu.h> | ||||
| #include <linux/io.h> | ||||
| #include <linux/vmalloc.h> | ||||
| 
 | ||||
| #include <asm/memory.h> | ||||
|  | @ -217,115 +218,70 @@ static void __dma_free_buffer(struct page *page, size_t size) | |||
| } | ||||
| 
 | ||||
| #ifdef CONFIG_MMU | ||||
| 
 | ||||
| #define CONSISTENT_OFFSET(x)	(((unsigned long)(x) - consistent_base) >> PAGE_SHIFT) | ||||
| #define CONSISTENT_PTE_INDEX(x) (((unsigned long)(x) - consistent_base) >> PMD_SHIFT) | ||||
| 
 | ||||
| /*
 | ||||
|  * These are the page tables (2MB each) covering uncached, DMA consistent allocations | ||||
|  */ | ||||
| static pte_t **consistent_pte; | ||||
| 
 | ||||
| #define DEFAULT_CONSISTENT_DMA_SIZE SZ_2M | ||||
| 
 | ||||
| static unsigned long consistent_base = CONSISTENT_END - DEFAULT_CONSISTENT_DMA_SIZE; | ||||
| 
 | ||||
| void __init init_consistent_dma_size(unsigned long size) | ||||
| { | ||||
| 	unsigned long base = CONSISTENT_END - ALIGN(size, SZ_2M); | ||||
| 
 | ||||
| 	BUG_ON(consistent_pte); /* Check we're called before DMA region init */ | ||||
| 	BUG_ON(base < VMALLOC_END); | ||||
| 
 | ||||
| 	/* Grow region to accommodate specified size  */ | ||||
| 	if (base < consistent_base) | ||||
| 		consistent_base = base; | ||||
| } | ||||
| 
 | ||||
| #include "vmregion.h" | ||||
| 
 | ||||
| static struct arm_vmregion_head consistent_head = { | ||||
| 	.vm_lock	= __SPIN_LOCK_UNLOCKED(&consistent_head.vm_lock), | ||||
| 	.vm_list	= LIST_HEAD_INIT(consistent_head.vm_list), | ||||
| 	.vm_end		= CONSISTENT_END, | ||||
| }; | ||||
| 
 | ||||
| #ifdef CONFIG_HUGETLB_PAGE | ||||
| #error ARM Coherent DMA allocator does not (yet) support huge TLB | ||||
| #endif | ||||
| 
 | ||||
| /*
 | ||||
|  * Initialise the consistent memory allocation. | ||||
|  */ | ||||
| static int __init consistent_init(void) | ||||
| { | ||||
| 	int ret = 0; | ||||
| 	pgd_t *pgd; | ||||
| 	pud_t *pud; | ||||
| 	pmd_t *pmd; | ||||
| 	pte_t *pte; | ||||
| 	int i = 0; | ||||
| 	unsigned long base = consistent_base; | ||||
| 	unsigned long num_ptes = (CONSISTENT_END - base) >> PMD_SHIFT; | ||||
| 
 | ||||
| 	if (IS_ENABLED(CONFIG_CMA) && !IS_ENABLED(CONFIG_ARM_DMA_USE_IOMMU)) | ||||
| 		return 0; | ||||
| 
 | ||||
| 	consistent_pte = kmalloc(num_ptes * sizeof(pte_t), GFP_KERNEL); | ||||
| 	if (!consistent_pte) { | ||||
| 		pr_err("%s: no memory\n", __func__); | ||||
| 		return -ENOMEM; | ||||
| 	} | ||||
| 
 | ||||
| 	pr_debug("DMA memory: 0x%08lx - 0x%08lx:\n", base, CONSISTENT_END); | ||||
| 	consistent_head.vm_start = base; | ||||
| 
 | ||||
| 	do { | ||||
| 		pgd = pgd_offset(&init_mm, base); | ||||
| 
 | ||||
| 		pud = pud_alloc(&init_mm, pgd, base); | ||||
| 		if (!pud) { | ||||
| 			pr_err("%s: no pud tables\n", __func__); | ||||
| 			ret = -ENOMEM; | ||||
| 			break; | ||||
| 		} | ||||
| 
 | ||||
| 		pmd = pmd_alloc(&init_mm, pud, base); | ||||
| 		if (!pmd) { | ||||
| 			pr_err("%s: no pmd tables\n", __func__); | ||||
| 			ret = -ENOMEM; | ||||
| 			break; | ||||
| 		} | ||||
| 		WARN_ON(!pmd_none(*pmd)); | ||||
| 
 | ||||
| 		pte = pte_alloc_kernel(pmd, base); | ||||
| 		if (!pte) { | ||||
| 			pr_err("%s: no pte tables\n", __func__); | ||||
| 			ret = -ENOMEM; | ||||
| 			break; | ||||
| 		} | ||||
| 
 | ||||
| 		consistent_pte[i++] = pte; | ||||
| 		base += PMD_SIZE; | ||||
| 	} while (base < CONSISTENT_END); | ||||
| 
 | ||||
| 	return ret; | ||||
| } | ||||
| core_initcall(consistent_init); | ||||
| 
 | ||||
| static void *__alloc_from_contiguous(struct device *dev, size_t size, | ||||
| 				     pgprot_t prot, struct page **ret_page); | ||||
| 
 | ||||
| static struct arm_vmregion_head coherent_head = { | ||||
| 	.vm_lock	= __SPIN_LOCK_UNLOCKED(&coherent_head.vm_lock), | ||||
| 	.vm_list	= LIST_HEAD_INIT(coherent_head.vm_list), | ||||
| static void *__alloc_remap_buffer(struct device *dev, size_t size, gfp_t gfp, | ||||
| 				 pgprot_t prot, struct page **ret_page, | ||||
| 				 const void *caller); | ||||
| 
 | ||||
| static void * | ||||
| __dma_alloc_remap(struct page *page, size_t size, gfp_t gfp, pgprot_t prot, | ||||
| 	const void *caller) | ||||
| { | ||||
| 	struct vm_struct *area; | ||||
| 	unsigned long addr; | ||||
| 
 | ||||
| 	/*
 | ||||
| 	 * DMA allocation can be mapped to user space, so lets | ||||
| 	 * set VM_USERMAP flags too. | ||||
| 	 */ | ||||
| 	area = get_vm_area_caller(size, VM_ARM_DMA_CONSISTENT | VM_USERMAP, | ||||
| 				  caller); | ||||
| 	if (!area) | ||||
| 		return NULL; | ||||
| 	addr = (unsigned long)area->addr; | ||||
| 	area->phys_addr = __pfn_to_phys(page_to_pfn(page)); | ||||
| 
 | ||||
| 	if (ioremap_page_range(addr, addr + size, area->phys_addr, prot)) { | ||||
| 		vunmap((void *)addr); | ||||
| 		return NULL; | ||||
| 	} | ||||
| 	return (void *)addr; | ||||
| } | ||||
| 
 | ||||
| static void __dma_free_remap(void *cpu_addr, size_t size) | ||||
| { | ||||
| 	unsigned int flags = VM_ARM_DMA_CONSISTENT | VM_USERMAP; | ||||
| 	struct vm_struct *area = find_vm_area(cpu_addr); | ||||
| 	if (!area || (area->flags & flags) != flags) { | ||||
| 		WARN(1, "trying to free invalid coherent area: %p\n", cpu_addr); | ||||
| 		return; | ||||
| 	} | ||||
| 	unmap_kernel_range((unsigned long)cpu_addr, size); | ||||
| 	vunmap(cpu_addr); | ||||
| } | ||||
| 
 | ||||
| struct dma_pool { | ||||
| 	size_t size; | ||||
| 	spinlock_t lock; | ||||
| 	unsigned long *bitmap; | ||||
| 	unsigned long nr_pages; | ||||
| 	void *vaddr; | ||||
| 	struct page *page; | ||||
| }; | ||||
| 
 | ||||
| static size_t coherent_pool_size = DEFAULT_CONSISTENT_DMA_SIZE / 8; | ||||
| static struct dma_pool atomic_pool = { | ||||
| 	.size = SZ_256K, | ||||
| }; | ||||
| 
 | ||||
| static int __init early_coherent_pool(char *p) | ||||
| { | ||||
| 	coherent_pool_size = memparse(p, &p); | ||||
| 	atomic_pool.size = memparse(p, &p); | ||||
| 	return 0; | ||||
| } | ||||
| early_param("coherent_pool", early_coherent_pool); | ||||
|  | @ -333,32 +289,45 @@ early_param("coherent_pool", early_coherent_pool); | |||
| /*
 | ||||
|  * Initialise the coherent pool for atomic allocations. | ||||
|  */ | ||||
| static int __init coherent_init(void) | ||||
| static int __init atomic_pool_init(void) | ||||
| { | ||||
| 	struct dma_pool *pool = &atomic_pool; | ||||
| 	pgprot_t prot = pgprot_dmacoherent(pgprot_kernel); | ||||
| 	size_t size = coherent_pool_size; | ||||
| 	unsigned long nr_pages = pool->size >> PAGE_SHIFT; | ||||
| 	unsigned long *bitmap; | ||||
| 	struct page *page; | ||||
| 	void *ptr; | ||||
| 	int bitmap_size = BITS_TO_LONGS(nr_pages) * sizeof(long); | ||||
| 
 | ||||
| 	if (!IS_ENABLED(CONFIG_CMA)) | ||||
| 		return 0; | ||||
| 	bitmap = kzalloc(bitmap_size, GFP_KERNEL); | ||||
| 	if (!bitmap) | ||||
| 		goto no_bitmap; | ||||
| 
 | ||||
| 	ptr = __alloc_from_contiguous(NULL, size, prot, &page); | ||||
| 	if (IS_ENABLED(CONFIG_CMA)) | ||||
| 		ptr = __alloc_from_contiguous(NULL, pool->size, prot, &page); | ||||
| 	else | ||||
| 		ptr = __alloc_remap_buffer(NULL, pool->size, GFP_KERNEL, prot, | ||||
| 					   &page, NULL); | ||||
| 	if (ptr) { | ||||
| 		coherent_head.vm_start = (unsigned long) ptr; | ||||
| 		coherent_head.vm_end = (unsigned long) ptr + size; | ||||
| 		printk(KERN_INFO "DMA: preallocated %u KiB pool for atomic coherent allocations\n", | ||||
| 		       (unsigned)size / 1024); | ||||
| 		spin_lock_init(&pool->lock); | ||||
| 		pool->vaddr = ptr; | ||||
| 		pool->page = page; | ||||
| 		pool->bitmap = bitmap; | ||||
| 		pool->nr_pages = nr_pages; | ||||
| 		pr_info("DMA: preallocated %u KiB pool for atomic coherent allocations\n", | ||||
| 		       (unsigned)pool->size / 1024); | ||||
| 		return 0; | ||||
| 	} | ||||
| 	printk(KERN_ERR "DMA: failed to allocate %u KiB pool for atomic coherent allocation\n", | ||||
| 	       (unsigned)size / 1024); | ||||
| 	kfree(bitmap); | ||||
| no_bitmap: | ||||
| 	pr_err("DMA: failed to allocate %u KiB pool for atomic coherent allocation\n", | ||||
| 	       (unsigned)pool->size / 1024); | ||||
| 	return -ENOMEM; | ||||
| } | ||||
| /*
 | ||||
|  * CMA is activated by core_initcall, so we must be called after it. | ||||
|  */ | ||||
| postcore_initcall(coherent_init); | ||||
| postcore_initcall(atomic_pool_init); | ||||
| 
 | ||||
| struct dma_contig_early_reserve { | ||||
| 	phys_addr_t base; | ||||
|  | @ -406,112 +375,6 @@ void __init dma_contiguous_remap(void) | |||
| 	} | ||||
| } | ||||
| 
 | ||||
| static void * | ||||
| __dma_alloc_remap(struct page *page, size_t size, gfp_t gfp, pgprot_t prot, | ||||
| 	const void *caller) | ||||
| { | ||||
| 	struct arm_vmregion *c; | ||||
| 	size_t align; | ||||
| 	int bit; | ||||
| 
 | ||||
| 	if (!consistent_pte) { | ||||
| 		pr_err("%s: not initialised\n", __func__); | ||||
| 		dump_stack(); | ||||
| 		return NULL; | ||||
| 	} | ||||
| 
 | ||||
| 	/*
 | ||||
| 	 * Align the virtual region allocation - maximum alignment is | ||||
| 	 * a section size, minimum is a page size.  This helps reduce | ||||
| 	 * fragmentation of the DMA space, and also prevents allocations | ||||
| 	 * smaller than a section from crossing a section boundary. | ||||
| 	 */ | ||||
| 	bit = fls(size - 1); | ||||
| 	if (bit > SECTION_SHIFT) | ||||
| 		bit = SECTION_SHIFT; | ||||
| 	align = 1 << bit; | ||||
| 
 | ||||
| 	/*
 | ||||
| 	 * Allocate a virtual address in the consistent mapping region. | ||||
| 	 */ | ||||
| 	c = arm_vmregion_alloc(&consistent_head, align, size, | ||||
| 			    gfp & ~(__GFP_DMA | __GFP_HIGHMEM), caller); | ||||
| 	if (c) { | ||||
| 		pte_t *pte; | ||||
| 		int idx = CONSISTENT_PTE_INDEX(c->vm_start); | ||||
| 		u32 off = CONSISTENT_OFFSET(c->vm_start) & (PTRS_PER_PTE-1); | ||||
| 
 | ||||
| 		pte = consistent_pte[idx] + off; | ||||
| 		c->priv = page; | ||||
| 
 | ||||
| 		do { | ||||
| 			BUG_ON(!pte_none(*pte)); | ||||
| 
 | ||||
| 			set_pte_ext(pte, mk_pte(page, prot), 0); | ||||
| 			page++; | ||||
| 			pte++; | ||||
| 			off++; | ||||
| 			if (off >= PTRS_PER_PTE) { | ||||
| 				off = 0; | ||||
| 				pte = consistent_pte[++idx]; | ||||
| 			} | ||||
| 		} while (size -= PAGE_SIZE); | ||||
| 
 | ||||
| 		dsb(); | ||||
| 
 | ||||
| 		return (void *)c->vm_start; | ||||
| 	} | ||||
| 	return NULL; | ||||
| } | ||||
| 
 | ||||
| static void __dma_free_remap(void *cpu_addr, size_t size) | ||||
| { | ||||
| 	struct arm_vmregion *c; | ||||
| 	unsigned long addr; | ||||
| 	pte_t *ptep; | ||||
| 	int idx; | ||||
| 	u32 off; | ||||
| 
 | ||||
| 	c = arm_vmregion_find_remove(&consistent_head, (unsigned long)cpu_addr); | ||||
| 	if (!c) { | ||||
| 		pr_err("%s: trying to free invalid coherent area: %p\n", | ||||
| 		       __func__, cpu_addr); | ||||
| 		dump_stack(); | ||||
| 		return; | ||||
| 	} | ||||
| 
 | ||||
| 	if ((c->vm_end - c->vm_start) != size) { | ||||
| 		pr_err("%s: freeing wrong coherent size (%ld != %d)\n", | ||||
| 		       __func__, c->vm_end - c->vm_start, size); | ||||
| 		dump_stack(); | ||||
| 		size = c->vm_end - c->vm_start; | ||||
| 	} | ||||
| 
 | ||||
| 	idx = CONSISTENT_PTE_INDEX(c->vm_start); | ||||
| 	off = CONSISTENT_OFFSET(c->vm_start) & (PTRS_PER_PTE-1); | ||||
| 	ptep = consistent_pte[idx] + off; | ||||
| 	addr = c->vm_start; | ||||
| 	do { | ||||
| 		pte_t pte = ptep_get_and_clear(&init_mm, addr, ptep); | ||||
| 
 | ||||
| 		ptep++; | ||||
| 		addr += PAGE_SIZE; | ||||
| 		off++; | ||||
| 		if (off >= PTRS_PER_PTE) { | ||||
| 			off = 0; | ||||
| 			ptep = consistent_pte[++idx]; | ||||
| 		} | ||||
| 
 | ||||
| 		if (pte_none(pte) || !pte_present(pte)) | ||||
| 			pr_crit("%s: bad page in kernel page table\n", | ||||
| 				__func__); | ||||
| 	} while (size -= PAGE_SIZE); | ||||
| 
 | ||||
| 	flush_tlb_kernel_range(c->vm_start, c->vm_end); | ||||
| 
 | ||||
| 	arm_vmregion_free(&consistent_head, c); | ||||
| } | ||||
| 
 | ||||
| static int __dma_update_pte(pte_t *pte, pgtable_t token, unsigned long addr, | ||||
| 			    void *data) | ||||
| { | ||||
|  | @ -552,16 +415,17 @@ static void *__alloc_remap_buffer(struct device *dev, size_t size, gfp_t gfp, | |||
| 	return ptr; | ||||
| } | ||||
| 
 | ||||
| static void *__alloc_from_pool(struct device *dev, size_t size, | ||||
| 			       struct page **ret_page, const void *caller) | ||||
| static void *__alloc_from_pool(size_t size, struct page **ret_page) | ||||
| { | ||||
| 	struct arm_vmregion *c; | ||||
| 	struct dma_pool *pool = &atomic_pool; | ||||
| 	unsigned int count = PAGE_ALIGN(size) >> PAGE_SHIFT; | ||||
| 	unsigned int pageno; | ||||
| 	unsigned long flags; | ||||
| 	void *ptr = NULL; | ||||
| 	size_t align; | ||||
| 
 | ||||
| 	if (!coherent_head.vm_start) { | ||||
| 		printk(KERN_ERR "%s: coherent pool not initialised!\n", | ||||
| 		       __func__); | ||||
| 		dump_stack(); | ||||
| 	if (!pool->vaddr) { | ||||
| 		WARN(1, "coherent pool not initialised!\n"); | ||||
| 		return NULL; | ||||
| 	} | ||||
| 
 | ||||
|  | @ -571,35 +435,41 @@ static void *__alloc_from_pool(struct device *dev, size_t size, | |||
| 	 * size. This helps reduce fragmentation of the DMA space. | ||||
| 	 */ | ||||
| 	align = PAGE_SIZE << get_order(size); | ||||
| 	c = arm_vmregion_alloc(&coherent_head, align, size, 0, caller); | ||||
| 	if (c) { | ||||
| 		void *ptr = (void *)c->vm_start; | ||||
| 		struct page *page = virt_to_page(ptr); | ||||
| 		*ret_page = page; | ||||
| 		return ptr; | ||||
| 
 | ||||
| 	spin_lock_irqsave(&pool->lock, flags); | ||||
| 	pageno = bitmap_find_next_zero_area(pool->bitmap, pool->nr_pages, | ||||
| 					    0, count, (1 << align) - 1); | ||||
| 	if (pageno < pool->nr_pages) { | ||||
| 		bitmap_set(pool->bitmap, pageno, count); | ||||
| 		ptr = pool->vaddr + PAGE_SIZE * pageno; | ||||
| 		*ret_page = pool->page + pageno; | ||||
| 	} | ||||
| 	return NULL; | ||||
| 	spin_unlock_irqrestore(&pool->lock, flags); | ||||
| 
 | ||||
| 	return ptr; | ||||
| } | ||||
| 
 | ||||
| static int __free_from_pool(void *cpu_addr, size_t size) | ||||
| static int __free_from_pool(void *start, size_t size) | ||||
| { | ||||
| 	unsigned long start = (unsigned long)cpu_addr; | ||||
| 	unsigned long end = start + size; | ||||
| 	struct arm_vmregion *c; | ||||
| 	struct dma_pool *pool = &atomic_pool; | ||||
| 	unsigned long pageno, count; | ||||
| 	unsigned long flags; | ||||
| 
 | ||||
| 	if (start < coherent_head.vm_start || end > coherent_head.vm_end) | ||||
| 	if (start < pool->vaddr || start > pool->vaddr + pool->size) | ||||
| 		return 0; | ||||
| 
 | ||||
| 	c = arm_vmregion_find_remove(&coherent_head, (unsigned long)start); | ||||
| 
 | ||||
| 	if ((c->vm_end - c->vm_start) != size) { | ||||
| 		printk(KERN_ERR "%s: freeing wrong coherent size (%ld != %d)\n", | ||||
| 		       __func__, c->vm_end - c->vm_start, size); | ||||
| 		dump_stack(); | ||||
| 		size = c->vm_end - c->vm_start; | ||||
| 	if (start + size > pool->vaddr + pool->size) { | ||||
| 		WARN(1, "freeing wrong coherent size from pool\n"); | ||||
| 		return 0; | ||||
| 	} | ||||
| 
 | ||||
| 	arm_vmregion_free(&coherent_head, c); | ||||
| 	pageno = (start - pool->vaddr) >> PAGE_SHIFT; | ||||
| 	count = size >> PAGE_SHIFT; | ||||
| 
 | ||||
| 	spin_lock_irqsave(&pool->lock, flags); | ||||
| 	bitmap_clear(pool->bitmap, pageno, count); | ||||
| 	spin_unlock_irqrestore(&pool->lock, flags); | ||||
| 
 | ||||
| 	return 1; | ||||
| } | ||||
| 
 | ||||
|  | @ -644,7 +514,7 @@ static inline pgprot_t __get_dma_pgprot(struct dma_attrs *attrs, pgprot_t prot) | |||
| 
 | ||||
| #define __get_dma_pgprot(attrs, prot)	__pgprot(0) | ||||
| #define __alloc_remap_buffer(dev, size, gfp, prot, ret, c)	NULL | ||||
| #define __alloc_from_pool(dev, size, ret_page, c)		NULL | ||||
| #define __alloc_from_pool(size, ret_page)			NULL | ||||
| #define __alloc_from_contiguous(dev, size, prot, ret)		NULL | ||||
| #define __free_from_pool(cpu_addr, size)			0 | ||||
| #define __free_from_contiguous(dev, page, size)			do { } while (0) | ||||
|  | @ -702,10 +572,10 @@ static void *__dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, | |||
| 
 | ||||
| 	if (arch_is_coherent() || nommu()) | ||||
| 		addr = __alloc_simple_buffer(dev, size, gfp, &page); | ||||
| 	else if (gfp & GFP_ATOMIC) | ||||
| 		addr = __alloc_from_pool(size, &page); | ||||
| 	else if (!IS_ENABLED(CONFIG_CMA)) | ||||
| 		addr = __alloc_remap_buffer(dev, size, gfp, prot, &page, caller); | ||||
| 	else if (gfp & GFP_ATOMIC) | ||||
| 		addr = __alloc_from_pool(dev, size, &page, caller); | ||||
| 	else | ||||
| 		addr = __alloc_from_contiguous(dev, size, prot, &page); | ||||
| 
 | ||||
|  | @ -998,9 +868,6 @@ static int arm_dma_set_mask(struct device *dev, u64 dma_mask) | |||
| 
 | ||||
| static int __init dma_debug_do_init(void) | ||||
| { | ||||
| #ifdef CONFIG_MMU | ||||
| 	arm_vmregion_create_proc("dma-mappings", &consistent_head); | ||||
| #endif | ||||
| 	dma_debug_init(PREALLOC_DMA_DEBUG_ENTRIES); | ||||
| 	return 0; | ||||
| } | ||||
|  | @ -1117,61 +984,32 @@ static int __iommu_free_buffer(struct device *dev, struct page **pages, size_t s | |||
|  * Create a CPU mapping for a specified pages | ||||
|  */ | ||||
| static void * | ||||
| __iommu_alloc_remap(struct page **pages, size_t size, gfp_t gfp, pgprot_t prot) | ||||
| __iommu_alloc_remap(struct page **pages, size_t size, gfp_t gfp, pgprot_t prot, | ||||
| 		    const void *caller) | ||||
| { | ||||
| 	struct arm_vmregion *c; | ||||
| 	size_t align; | ||||
| 	size_t count = size >> PAGE_SHIFT; | ||||
| 	int bit; | ||||
| 	unsigned int i, nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; | ||||
| 	struct vm_struct *area; | ||||
| 	unsigned long p; | ||||
| 
 | ||||
| 	if (!consistent_pte[0]) { | ||||
| 		pr_err("%s: not initialised\n", __func__); | ||||
| 		dump_stack(); | ||||
| 	area = get_vm_area_caller(size, VM_ARM_DMA_CONSISTENT | VM_USERMAP, | ||||
| 				  caller); | ||||
| 	if (!area) | ||||
| 		return NULL; | ||||
| 
 | ||||
| 	area->pages = pages; | ||||
| 	area->nr_pages = nr_pages; | ||||
| 	p = (unsigned long)area->addr; | ||||
| 
 | ||||
| 	for (i = 0; i < nr_pages; i++) { | ||||
| 		phys_addr_t phys = __pfn_to_phys(page_to_pfn(pages[i])); | ||||
| 		if (ioremap_page_range(p, p + PAGE_SIZE, phys, prot)) | ||||
| 			goto err; | ||||
| 		p += PAGE_SIZE; | ||||
| 	} | ||||
| 
 | ||||
| 	/*
 | ||||
| 	 * Align the virtual region allocation - maximum alignment is | ||||
| 	 * a section size, minimum is a page size.  This helps reduce | ||||
| 	 * fragmentation of the DMA space, and also prevents allocations | ||||
| 	 * smaller than a section from crossing a section boundary. | ||||
| 	 */ | ||||
| 	bit = fls(size - 1); | ||||
| 	if (bit > SECTION_SHIFT) | ||||
| 		bit = SECTION_SHIFT; | ||||
| 	align = 1 << bit; | ||||
| 
 | ||||
| 	/*
 | ||||
| 	 * Allocate a virtual address in the consistent mapping region. | ||||
| 	 */ | ||||
| 	c = arm_vmregion_alloc(&consistent_head, align, size, | ||||
| 			    gfp & ~(__GFP_DMA | __GFP_HIGHMEM), NULL); | ||||
| 	if (c) { | ||||
| 		pte_t *pte; | ||||
| 		int idx = CONSISTENT_PTE_INDEX(c->vm_start); | ||||
| 		int i = 0; | ||||
| 		u32 off = CONSISTENT_OFFSET(c->vm_start) & (PTRS_PER_PTE-1); | ||||
| 
 | ||||
| 		pte = consistent_pte[idx] + off; | ||||
| 		c->priv = pages; | ||||
| 
 | ||||
| 		do { | ||||
| 			BUG_ON(!pte_none(*pte)); | ||||
| 
 | ||||
| 			set_pte_ext(pte, mk_pte(pages[i], prot), 0); | ||||
| 			pte++; | ||||
| 			off++; | ||||
| 			i++; | ||||
| 			if (off >= PTRS_PER_PTE) { | ||||
| 				off = 0; | ||||
| 				pte = consistent_pte[++idx]; | ||||
| 			} | ||||
| 		} while (i < count); | ||||
| 
 | ||||
| 		dsb(); | ||||
| 
 | ||||
| 		return (void *)c->vm_start; | ||||
| 	} | ||||
| 	return area->addr; | ||||
| err: | ||||
| 	unmap_kernel_range((unsigned long)area->addr, size); | ||||
| 	vunmap(area->addr); | ||||
| 	return NULL; | ||||
| } | ||||
| 
 | ||||
|  | @ -1230,6 +1068,16 @@ static int __iommu_remove_mapping(struct device *dev, dma_addr_t iova, size_t si | |||
| 	return 0; | ||||
| } | ||||
| 
 | ||||
| static struct page **__iommu_get_pages(void *cpu_addr) | ||||
| { | ||||
| 	struct vm_struct *area; | ||||
| 
 | ||||
| 	area = find_vm_area(cpu_addr); | ||||
| 	if (area && (area->flags & VM_ARM_DMA_CONSISTENT)) | ||||
| 		return area->pages; | ||||
| 	return NULL; | ||||
| } | ||||
| 
 | ||||
| static void *arm_iommu_alloc_attrs(struct device *dev, size_t size, | ||||
| 	    dma_addr_t *handle, gfp_t gfp, struct dma_attrs *attrs) | ||||
| { | ||||
|  | @ -1248,7 +1096,8 @@ static void *arm_iommu_alloc_attrs(struct device *dev, size_t size, | |||
| 	if (*handle == DMA_ERROR_CODE) | ||||
| 		goto err_buffer; | ||||
| 
 | ||||
| 	addr = __iommu_alloc_remap(pages, size, gfp, prot); | ||||
| 	addr = __iommu_alloc_remap(pages, size, gfp, prot, | ||||
| 				   __builtin_return_address(0)); | ||||
| 	if (!addr) | ||||
| 		goto err_mapping; | ||||
| 
 | ||||
|  | @ -1265,31 +1114,25 @@ static int arm_iommu_mmap_attrs(struct device *dev, struct vm_area_struct *vma, | |||
| 		    void *cpu_addr, dma_addr_t dma_addr, size_t size, | ||||
| 		    struct dma_attrs *attrs) | ||||
| { | ||||
| 	struct arm_vmregion *c; | ||||
| 	unsigned long uaddr = vma->vm_start; | ||||
| 	unsigned long usize = vma->vm_end - vma->vm_start; | ||||
| 	struct page **pages = __iommu_get_pages(cpu_addr); | ||||
| 
 | ||||
| 	vma->vm_page_prot = __get_dma_pgprot(attrs, vma->vm_page_prot); | ||||
| 	c = arm_vmregion_find(&consistent_head, (unsigned long)cpu_addr); | ||||
| 
 | ||||
| 	if (c) { | ||||
| 		struct page **pages = c->priv; | ||||
| 	if (!pages) | ||||
| 		return -ENXIO; | ||||
| 
 | ||||
| 		unsigned long uaddr = vma->vm_start; | ||||
| 		unsigned long usize = vma->vm_end - vma->vm_start; | ||||
| 		int i = 0; | ||||
| 	do { | ||||
| 		int ret = vm_insert_page(vma, uaddr, *pages++); | ||||
| 		if (ret) { | ||||
| 			pr_err("Remapping memory failed: %d\n", ret); | ||||
| 			return ret; | ||||
| 		} | ||||
| 		uaddr += PAGE_SIZE; | ||||
| 		usize -= PAGE_SIZE; | ||||
| 	} while (usize > 0); | ||||
| 
 | ||||
| 		do { | ||||
| 			int ret; | ||||
| 
 | ||||
| 			ret = vm_insert_page(vma, uaddr, pages[i++]); | ||||
| 			if (ret) { | ||||
| 				pr_err("Remapping memory, error: %d\n", ret); | ||||
| 				return ret; | ||||
| 			} | ||||
| 
 | ||||
| 			uaddr += PAGE_SIZE; | ||||
| 			usize -= PAGE_SIZE; | ||||
| 		} while (usize > 0); | ||||
| 	} | ||||
| 	return 0; | ||||
| } | ||||
| 
 | ||||
|  | @ -1300,16 +1143,19 @@ static int arm_iommu_mmap_attrs(struct device *dev, struct vm_area_struct *vma, | |||
| void arm_iommu_free_attrs(struct device *dev, size_t size, void *cpu_addr, | ||||
| 			  dma_addr_t handle, struct dma_attrs *attrs) | ||||
| { | ||||
| 	struct arm_vmregion *c; | ||||
| 	struct page **pages = __iommu_get_pages(cpu_addr); | ||||
| 	size = PAGE_ALIGN(size); | ||||
| 
 | ||||
| 	c = arm_vmregion_find(&consistent_head, (unsigned long)cpu_addr); | ||||
| 	if (c) { | ||||
| 		struct page **pages = c->priv; | ||||
| 		__dma_free_remap(cpu_addr, size); | ||||
| 		__iommu_remove_mapping(dev, handle, size); | ||||
| 		__iommu_free_buffer(dev, pages, size); | ||||
| 	if (!pages) { | ||||
| 		WARN(1, "trying to free invalid coherent area: %p\n", cpu_addr); | ||||
| 		return; | ||||
| 	} | ||||
| 
 | ||||
| 	unmap_kernel_range((unsigned long)cpu_addr, size); | ||||
| 	vunmap(cpu_addr); | ||||
| 
 | ||||
| 	__iommu_remove_mapping(dev, handle, size); | ||||
| 	__iommu_free_buffer(dev, pages, size); | ||||
| } | ||||
| 
 | ||||
| /*
 | ||||
|  |  | |||
|  | @ -59,6 +59,9 @@ extern void __flush_dcache_page(struct address_space *mapping, struct page *page | |||
| #define VM_ARM_MTYPE(mt)		((mt) << 20) | ||||
| #define VM_ARM_MTYPE_MASK	(0x1f << 20) | ||||
| 
 | ||||
| /* consistent regions used by dma_alloc_attrs() */ | ||||
| #define VM_ARM_DMA_CONSISTENT	0x20000000 | ||||
| 
 | ||||
| #endif | ||||
| 
 | ||||
| #ifdef CONFIG_ZONE_DMA | ||||
|  |  | |||
|  | @ -93,6 +93,7 @@ extern struct vm_struct *__get_vm_area_caller(unsigned long size, | |||
| 					unsigned long start, unsigned long end, | ||||
| 					const void *caller); | ||||
| extern struct vm_struct *remove_vm_area(const void *addr); | ||||
| extern struct vm_struct *find_vm_area(const void *addr); | ||||
| 
 | ||||
| extern int map_vm_area(struct vm_struct *area, pgprot_t prot, | ||||
| 			struct page ***pages); | ||||
|  |  | |||
							
								
								
									
										10
									
								
								mm/vmalloc.c
									
									
									
									
									
								
							
							
						
						
									
										10
									
								
								mm/vmalloc.c
									
									
									
									
									
								
							|  | @ -1403,7 +1403,15 @@ struct vm_struct *get_vm_area_caller(unsigned long size, unsigned long flags, | |||
| 						-1, GFP_KERNEL, caller); | ||||
| } | ||||
| 
 | ||||
| static struct vm_struct *find_vm_area(const void *addr) | ||||
| /**
 | ||||
|  *	find_vm_area  -  find a continuous kernel virtual area | ||||
|  *	@addr:		base address | ||||
|  * | ||||
|  *	Search for the kernel VM area starting at @addr, and return it. | ||||
|  *	It is up to the caller to do all required locking to keep the returned | ||||
|  *	pointer valid. | ||||
|  */ | ||||
| struct vm_struct *find_vm_area(const void *addr) | ||||
| { | ||||
| 	struct vmap_area *va; | ||||
| 
 | ||||
|  |  | |||
		Loading…
	
		Reference in a new issue