mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 10:40:15 +02:00 
			
		
		
		
	x86, PAT: Remove duplicate memtype reserve in devmem mmap
/dev/mem mmap code was doing memtype reserve/free for a while now. Recently we added memtype tracking in remap_pfn_range, and /dev/mem mmap uses it indirectly. So, we don't need seperate tracking in /dev/mem code any more. That means another ~100 lines of code removed :-). Signed-off-by: Suresh Siddha <suresh.b.siddha@intel.com> Signed-off-by: Venkatesh Pallipadi <venkatesh.pallipadi@intel.com> LKML-Reference: <20090409212709.085210000@intel.com> Signed-off-by: Ingo Molnar <mingo@elte.hu>
This commit is contained in:
		
							parent
							
								
									b6ff32d9aa
								
							
						
					
					
						commit
						0c3c8a1836
					
				
					 3 changed files with 2 additions and 89 deletions
				
			
		| 
						 | 
				
			
			@ -18,9 +18,5 @@ extern int free_memtype(u64 start, u64 end);
 | 
			
		|||
 | 
			
		||||
extern int kernel_map_sync_memtype(u64 base, unsigned long size,
 | 
			
		||||
		unsigned long flag);
 | 
			
		||||
extern void map_devmem(unsigned long pfn, unsigned long size,
 | 
			
		||||
		       struct pgprot vma_prot);
 | 
			
		||||
extern void unmap_devmem(unsigned long pfn, unsigned long size,
 | 
			
		||||
			 struct pgprot vma_prot);
 | 
			
		||||
 | 
			
		||||
#endif /* _ASM_X86_PAT_H */
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -536,9 +536,7 @@ static inline int range_is_allowed(unsigned long pfn, unsigned long size)
 | 
			
		|||
int phys_mem_access_prot_allowed(struct file *file, unsigned long pfn,
 | 
			
		||||
				unsigned long size, pgprot_t *vma_prot)
 | 
			
		||||
{
 | 
			
		||||
	u64 offset = ((u64) pfn) << PAGE_SHIFT;
 | 
			
		||||
	unsigned long flags = -1;
 | 
			
		||||
	int retval;
 | 
			
		||||
	unsigned long flags = _PAGE_CACHE_WB;
 | 
			
		||||
 | 
			
		||||
	if (!range_is_allowed(pfn, size))
 | 
			
		||||
		return 0;
 | 
			
		||||
| 
						 | 
				
			
			@ -566,65 +564,11 @@ int phys_mem_access_prot_allowed(struct file *file, unsigned long pfn,
 | 
			
		|||
	}
 | 
			
		||||
#endif
 | 
			
		||||
 | 
			
		||||
	/*
 | 
			
		||||
	 * With O_SYNC, we can only take UC_MINUS mapping. Fail if we cannot.
 | 
			
		||||
	 *
 | 
			
		||||
	 * Without O_SYNC, we want to get
 | 
			
		||||
	 * - WB for WB-able memory and no other conflicting mappings
 | 
			
		||||
	 * - UC_MINUS for non-WB-able memory with no other conflicting mappings
 | 
			
		||||
	 * - Inherit from confliting mappings otherwise
 | 
			
		||||
	 */
 | 
			
		||||
	if (flags != -1) {
 | 
			
		||||
		retval = reserve_memtype(offset, offset + size, flags, NULL);
 | 
			
		||||
	} else {
 | 
			
		||||
		retval = reserve_memtype(offset, offset + size,
 | 
			
		||||
					_PAGE_CACHE_WB, &flags);
 | 
			
		||||
	}
 | 
			
		||||
 | 
			
		||||
	if (retval < 0)
 | 
			
		||||
		return 0;
 | 
			
		||||
 | 
			
		||||
	if (((pfn < max_low_pfn_mapped) ||
 | 
			
		||||
	     (pfn >= (1UL<<(32 - PAGE_SHIFT)) && pfn < max_pfn_mapped)) &&
 | 
			
		||||
	    ioremap_change_attr((unsigned long)__va(offset), size, flags) < 0) {
 | 
			
		||||
		free_memtype(offset, offset + size);
 | 
			
		||||
		printk(KERN_INFO
 | 
			
		||||
		"%s:%d /dev/mem ioremap_change_attr failed %s for %Lx-%Lx\n",
 | 
			
		||||
			current->comm, current->pid,
 | 
			
		||||
			cattr_name(flags),
 | 
			
		||||
			offset, (unsigned long long)(offset + size));
 | 
			
		||||
		return 0;
 | 
			
		||||
	}
 | 
			
		||||
 | 
			
		||||
	*vma_prot = __pgprot((pgprot_val(*vma_prot) & ~_PAGE_CACHE_MASK) |
 | 
			
		||||
			     flags);
 | 
			
		||||
	return 1;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
void map_devmem(unsigned long pfn, unsigned long size, pgprot_t vma_prot)
 | 
			
		||||
{
 | 
			
		||||
	unsigned long want_flags = (pgprot_val(vma_prot) & _PAGE_CACHE_MASK);
 | 
			
		||||
	u64 addr = (u64)pfn << PAGE_SHIFT;
 | 
			
		||||
	unsigned long flags;
 | 
			
		||||
 | 
			
		||||
	reserve_memtype(addr, addr + size, want_flags, &flags);
 | 
			
		||||
	if (flags != want_flags) {
 | 
			
		||||
		printk(KERN_INFO
 | 
			
		||||
		"%s:%d /dev/mem expected mapping type %s for %Lx-%Lx, got %s\n",
 | 
			
		||||
			current->comm, current->pid,
 | 
			
		||||
			cattr_name(want_flags),
 | 
			
		||||
			addr, (unsigned long long)(addr + size),
 | 
			
		||||
			cattr_name(flags));
 | 
			
		||||
	}
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
void unmap_devmem(unsigned long pfn, unsigned long size, pgprot_t vma_prot)
 | 
			
		||||
{
 | 
			
		||||
	u64 addr = (u64)pfn << PAGE_SHIFT;
 | 
			
		||||
 | 
			
		||||
	free_memtype(addr, addr + size);
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
/*
 | 
			
		||||
 * Change the memory type for the physial address range in kernel identity
 | 
			
		||||
 * mapping space if that range is a part of identity map.
 | 
			
		||||
| 
						 | 
				
			
			@ -662,8 +606,8 @@ static int reserve_pfn_range(u64 paddr, unsigned long size, pgprot_t *vma_prot,
 | 
			
		|||
{
 | 
			
		||||
	int is_ram = 0;
 | 
			
		||||
	int ret;
 | 
			
		||||
	unsigned long flags;
 | 
			
		||||
	unsigned long want_flags = (pgprot_val(*vma_prot) & _PAGE_CACHE_MASK);
 | 
			
		||||
	unsigned long flags = want_flags;
 | 
			
		||||
 | 
			
		||||
	is_ram = pat_pagerange_is_ram(paddr, paddr + size);
 | 
			
		||||
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -301,33 +301,7 @@ static inline int private_mapping_ok(struct vm_area_struct *vma)
 | 
			
		|||
}
 | 
			
		||||
#endif
 | 
			
		||||
 | 
			
		||||
void __attribute__((weak))
 | 
			
		||||
map_devmem(unsigned long pfn, unsigned long len, pgprot_t prot)
 | 
			
		||||
{
 | 
			
		||||
	/* nothing. architectures can override. */
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
void __attribute__((weak))
 | 
			
		||||
unmap_devmem(unsigned long pfn, unsigned long len, pgprot_t prot)
 | 
			
		||||
{
 | 
			
		||||
	/* nothing. architectures can override. */
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
static void mmap_mem_open(struct vm_area_struct *vma)
 | 
			
		||||
{
 | 
			
		||||
	map_devmem(vma->vm_pgoff,  vma->vm_end - vma->vm_start,
 | 
			
		||||
			vma->vm_page_prot);
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
static void mmap_mem_close(struct vm_area_struct *vma)
 | 
			
		||||
{
 | 
			
		||||
	unmap_devmem(vma->vm_pgoff,  vma->vm_end - vma->vm_start,
 | 
			
		||||
			vma->vm_page_prot);
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
static struct vm_operations_struct mmap_mem_ops = {
 | 
			
		||||
	.open  = mmap_mem_open,
 | 
			
		||||
	.close = mmap_mem_close,
 | 
			
		||||
#ifdef CONFIG_HAVE_IOREMAP_PROT
 | 
			
		||||
	.access = generic_access_phys
 | 
			
		||||
#endif
 | 
			
		||||
| 
						 | 
				
			
			@ -362,7 +336,6 @@ static int mmap_mem(struct file * file, struct vm_area_struct * vma)
 | 
			
		|||
			    vma->vm_pgoff,
 | 
			
		||||
			    size,
 | 
			
		||||
			    vma->vm_page_prot)) {
 | 
			
		||||
		unmap_devmem(vma->vm_pgoff, size, vma->vm_page_prot);
 | 
			
		||||
		return -EAGAIN;
 | 
			
		||||
	}
 | 
			
		||||
	return 0;
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
		Loading…
	
		Reference in a new issue