mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 02:30:34 +02:00 
			
		
		
		
	dax: update to new mmu_notifier semantic
Replace all mmu_notifier_invalidate_page() calls by *_invalidate_range() and make sure it is bracketed by calls to *_invalidate_range_start()/end(). Note that because we can not presume the pmd value or pte value we have to assume the worst and unconditionaly report an invalidation as happening. Signed-off-by: Jérôme Glisse <jglisse@redhat.com> Cc: Dan Williams <dan.j.williams@intel.com> Cc: Ross Zwisler <ross.zwisler@linux.intel.com> Cc: Bernhard Held <berny156@gmx.de> Cc: Adam Borowski <kilobyte@angband.pl> Cc: Andrea Arcangeli <aarcange@redhat.com> Cc: Radim Krčmář <rkrcmar@redhat.com> Cc: Wanpeng Li <kernellwp@gmail.com> Cc: Paolo Bonzini <pbonzini@redhat.com> Cc: Takashi Iwai <tiwai@suse.de> Cc: Nadav Amit <nadav.amit@gmail.com> Cc: Mike Galbraith <efault@gmx.de> Cc: Kirill A. Shutemov <kirill.shutemov@linux.intel.com> Cc: axie <axie@amd.com> Cc: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
		
							parent
							
								
									42ff72cf27
								
							
						
					
					
						commit
						a4d1a88525
					
				
					 3 changed files with 33 additions and 13 deletions
				
			
		
							
								
								
									
										19
									
								
								fs/dax.c
									
									
									
									
									
								
							
							
						
						
									
										19
									
								
								fs/dax.c
									
									
									
									
									
								
							| 
						 | 
					@ -646,11 +646,10 @@ static void dax_mapping_entry_mkclean(struct address_space *mapping,
 | 
				
			||||||
	pte_t pte, *ptep = NULL;
 | 
						pte_t pte, *ptep = NULL;
 | 
				
			||||||
	pmd_t *pmdp = NULL;
 | 
						pmd_t *pmdp = NULL;
 | 
				
			||||||
	spinlock_t *ptl;
 | 
						spinlock_t *ptl;
 | 
				
			||||||
	bool changed;
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
	i_mmap_lock_read(mapping);
 | 
						i_mmap_lock_read(mapping);
 | 
				
			||||||
	vma_interval_tree_foreach(vma, &mapping->i_mmap, index, index) {
 | 
						vma_interval_tree_foreach(vma, &mapping->i_mmap, index, index) {
 | 
				
			||||||
		unsigned long address;
 | 
							unsigned long address, start, end;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		cond_resched();
 | 
							cond_resched();
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -658,8 +657,13 @@ static void dax_mapping_entry_mkclean(struct address_space *mapping,
 | 
				
			||||||
			continue;
 | 
								continue;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		address = pgoff_address(index, vma);
 | 
							address = pgoff_address(index, vma);
 | 
				
			||||||
		changed = false;
 | 
					
 | 
				
			||||||
		if (follow_pte_pmd(vma->vm_mm, address, &ptep, &pmdp, &ptl))
 | 
							/*
 | 
				
			||||||
 | 
							 * Note because we provide start/end to follow_pte_pmd it will
 | 
				
			||||||
 | 
							 * call mmu_notifier_invalidate_range_start() on our behalf
 | 
				
			||||||
 | 
							 * before taking any lock.
 | 
				
			||||||
 | 
							 */
 | 
				
			||||||
 | 
							if (follow_pte_pmd(vma->vm_mm, address, &start, &end, &ptep, &pmdp, &ptl))
 | 
				
			||||||
			continue;
 | 
								continue;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		if (pmdp) {
 | 
							if (pmdp) {
 | 
				
			||||||
| 
						 | 
					@ -676,7 +680,7 @@ static void dax_mapping_entry_mkclean(struct address_space *mapping,
 | 
				
			||||||
			pmd = pmd_wrprotect(pmd);
 | 
								pmd = pmd_wrprotect(pmd);
 | 
				
			||||||
			pmd = pmd_mkclean(pmd);
 | 
								pmd = pmd_mkclean(pmd);
 | 
				
			||||||
			set_pmd_at(vma->vm_mm, address, pmdp, pmd);
 | 
								set_pmd_at(vma->vm_mm, address, pmdp, pmd);
 | 
				
			||||||
			changed = true;
 | 
								mmu_notifier_invalidate_range(vma->vm_mm, start, end);
 | 
				
			||||||
unlock_pmd:
 | 
					unlock_pmd:
 | 
				
			||||||
			spin_unlock(ptl);
 | 
								spin_unlock(ptl);
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
| 
						 | 
					@ -691,13 +695,12 @@ static void dax_mapping_entry_mkclean(struct address_space *mapping,
 | 
				
			||||||
			pte = pte_wrprotect(pte);
 | 
								pte = pte_wrprotect(pte);
 | 
				
			||||||
			pte = pte_mkclean(pte);
 | 
								pte = pte_mkclean(pte);
 | 
				
			||||||
			set_pte_at(vma->vm_mm, address, ptep, pte);
 | 
								set_pte_at(vma->vm_mm, address, ptep, pte);
 | 
				
			||||||
			changed = true;
 | 
								mmu_notifier_invalidate_range(vma->vm_mm, start, end);
 | 
				
			||||||
unlock_pte:
 | 
					unlock_pte:
 | 
				
			||||||
			pte_unmap_unlock(ptep, ptl);
 | 
								pte_unmap_unlock(ptep, ptl);
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		if (changed)
 | 
							mmu_notifier_invalidate_range_end(vma->vm_mm, start, end);
 | 
				
			||||||
			mmu_notifier_invalidate_page(vma->vm_mm, address);
 | 
					 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
	i_mmap_unlock_read(mapping);
 | 
						i_mmap_unlock_read(mapping);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -1260,6 +1260,7 @@ int copy_page_range(struct mm_struct *dst, struct mm_struct *src,
 | 
				
			||||||
void unmap_mapping_range(struct address_space *mapping,
 | 
					void unmap_mapping_range(struct address_space *mapping,
 | 
				
			||||||
		loff_t const holebegin, loff_t const holelen, int even_cows);
 | 
							loff_t const holebegin, loff_t const holelen, int even_cows);
 | 
				
			||||||
int follow_pte_pmd(struct mm_struct *mm, unsigned long address,
 | 
					int follow_pte_pmd(struct mm_struct *mm, unsigned long address,
 | 
				
			||||||
 | 
								     unsigned long *start, unsigned long *end,
 | 
				
			||||||
			     pte_t **ptepp, pmd_t **pmdpp, spinlock_t **ptlp);
 | 
								     pte_t **ptepp, pmd_t **pmdpp, spinlock_t **ptlp);
 | 
				
			||||||
int follow_pfn(struct vm_area_struct *vma, unsigned long address,
 | 
					int follow_pfn(struct vm_area_struct *vma, unsigned long address,
 | 
				
			||||||
	unsigned long *pfn);
 | 
						unsigned long *pfn);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
							
								
								
									
										24
									
								
								mm/memory.c
									
									
									
									
									
								
							
							
						
						
									
										24
									
								
								mm/memory.c
									
									
									
									
									
								
							| 
						 | 
					@ -4008,6 +4008,7 @@ int __pmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long address)
 | 
				
			||||||
#endif /* __PAGETABLE_PMD_FOLDED */
 | 
					#endif /* __PAGETABLE_PMD_FOLDED */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static int __follow_pte_pmd(struct mm_struct *mm, unsigned long address,
 | 
					static int __follow_pte_pmd(struct mm_struct *mm, unsigned long address,
 | 
				
			||||||
 | 
								    unsigned long *start, unsigned long *end,
 | 
				
			||||||
			    pte_t **ptepp, pmd_t **pmdpp, spinlock_t **ptlp)
 | 
								    pte_t **ptepp, pmd_t **pmdpp, spinlock_t **ptlp)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	pgd_t *pgd;
 | 
						pgd_t *pgd;
 | 
				
			||||||
| 
						 | 
					@ -4035,17 +4036,29 @@ static int __follow_pte_pmd(struct mm_struct *mm, unsigned long address,
 | 
				
			||||||
		if (!pmdpp)
 | 
							if (!pmdpp)
 | 
				
			||||||
			goto out;
 | 
								goto out;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							if (start && end) {
 | 
				
			||||||
 | 
								*start = address & PMD_MASK;
 | 
				
			||||||
 | 
								*end = *start + PMD_SIZE;
 | 
				
			||||||
 | 
								mmu_notifier_invalidate_range_start(mm, *start, *end);
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
		*ptlp = pmd_lock(mm, pmd);
 | 
							*ptlp = pmd_lock(mm, pmd);
 | 
				
			||||||
		if (pmd_huge(*pmd)) {
 | 
							if (pmd_huge(*pmd)) {
 | 
				
			||||||
			*pmdpp = pmd;
 | 
								*pmdpp = pmd;
 | 
				
			||||||
			return 0;
 | 
								return 0;
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
		spin_unlock(*ptlp);
 | 
							spin_unlock(*ptlp);
 | 
				
			||||||
 | 
							if (start && end)
 | 
				
			||||||
 | 
								mmu_notifier_invalidate_range_end(mm, *start, *end);
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (pmd_none(*pmd) || unlikely(pmd_bad(*pmd)))
 | 
						if (pmd_none(*pmd) || unlikely(pmd_bad(*pmd)))
 | 
				
			||||||
		goto out;
 | 
							goto out;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						if (start && end) {
 | 
				
			||||||
 | 
							*start = address & PAGE_MASK;
 | 
				
			||||||
 | 
							*end = *start + PAGE_SIZE;
 | 
				
			||||||
 | 
							mmu_notifier_invalidate_range_start(mm, *start, *end);
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
	ptep = pte_offset_map_lock(mm, pmd, address, ptlp);
 | 
						ptep = pte_offset_map_lock(mm, pmd, address, ptlp);
 | 
				
			||||||
	if (!pte_present(*ptep))
 | 
						if (!pte_present(*ptep))
 | 
				
			||||||
		goto unlock;
 | 
							goto unlock;
 | 
				
			||||||
| 
						 | 
					@ -4053,6 +4066,8 @@ static int __follow_pte_pmd(struct mm_struct *mm, unsigned long address,
 | 
				
			||||||
	return 0;
 | 
						return 0;
 | 
				
			||||||
unlock:
 | 
					unlock:
 | 
				
			||||||
	pte_unmap_unlock(ptep, *ptlp);
 | 
						pte_unmap_unlock(ptep, *ptlp);
 | 
				
			||||||
 | 
						if (start && end)
 | 
				
			||||||
 | 
							mmu_notifier_invalidate_range_end(mm, *start, *end);
 | 
				
			||||||
out:
 | 
					out:
 | 
				
			||||||
	return -EINVAL;
 | 
						return -EINVAL;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
| 
						 | 
					@ -4064,20 +4079,21 @@ static inline int follow_pte(struct mm_struct *mm, unsigned long address,
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* (void) is needed to make gcc happy */
 | 
						/* (void) is needed to make gcc happy */
 | 
				
			||||||
	(void) __cond_lock(*ptlp,
 | 
						(void) __cond_lock(*ptlp,
 | 
				
			||||||
			   !(res = __follow_pte_pmd(mm, address, ptepp, NULL,
 | 
								   !(res = __follow_pte_pmd(mm, address, NULL, NULL,
 | 
				
			||||||
					   ptlp)));
 | 
											    ptepp, NULL, ptlp)));
 | 
				
			||||||
	return res;
 | 
						return res;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int follow_pte_pmd(struct mm_struct *mm, unsigned long address,
 | 
					int follow_pte_pmd(struct mm_struct *mm, unsigned long address,
 | 
				
			||||||
 | 
								     unsigned long *start, unsigned long *end,
 | 
				
			||||||
			     pte_t **ptepp, pmd_t **pmdpp, spinlock_t **ptlp)
 | 
								     pte_t **ptepp, pmd_t **pmdpp, spinlock_t **ptlp)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	int res;
 | 
						int res;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* (void) is needed to make gcc happy */
 | 
						/* (void) is needed to make gcc happy */
 | 
				
			||||||
	(void) __cond_lock(*ptlp,
 | 
						(void) __cond_lock(*ptlp,
 | 
				
			||||||
			   !(res = __follow_pte_pmd(mm, address, ptepp, pmdpp,
 | 
								   !(res = __follow_pte_pmd(mm, address, start, end,
 | 
				
			||||||
					   ptlp)));
 | 
											    ptepp, pmdpp, ptlp)));
 | 
				
			||||||
	return res;
 | 
						return res;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
EXPORT_SYMBOL(follow_pte_pmd);
 | 
					EXPORT_SYMBOL(follow_pte_pmd);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue