forked from mirrors/linux
		
	mm: streamline COW logic in do_swap_page()
Currently we have a different COW logic when: * triggering a read-fault to swapin first and then trigger a write-fault -> do_swap_page() + do_wp_page() * triggering a write-fault to swapin -> do_swap_page() + do_wp_page() only if we fail reuse in do_swap_page() The COW logic in do_swap_page() is different than our reuse logic in do_wp_page(). The COW logic in do_wp_page() -- page_count() == 1 -- makes currently sure that we certainly don't have a remaining reference, e.g., via GUP, on the target page we want to reuse: if there is any unexpected reference, we have to copy to avoid information leaks. As do_swap_page() behaves differently, in environments with swap enabled we can currently have an unintended information leak from the parent to the child, similar as known from CVE-2020-29374: 1. Parent writes to anonymous page -> Page is mapped writable and modified 2. Page is swapped out -> Page is unmapped and replaced by swap entry 3. fork() -> Swap entries are copied to child 4. Child pins page R/O -> Page is mapped R/O into child 5. Child unmaps page -> Child still holds GUP reference 6. Parent writes to page -> Page is reused in do_swap_page() -> Child can observe changes Exchanging 2. and 3. should have the same effect. Let's apply the same COW logic as in do_wp_page(), conditionally trying to remove the page from the swapcache after freeing the swap entry, however, before actually mapping our page. We can change the order now that we use try_to_free_swap(), which doesn't care about the mapcount, instead of reuse_swap_page(). To handle references from the LRU pagevecs, conditionally drain the local LRU pagevecs when required, however, don't consider the page_count() when deciding whether to drain to keep it simple for now. Link: https://lkml.kernel.org/r/20220131162940.210846-5-david@redhat.com Signed-off-by: David Hildenbrand <david@redhat.com> Acked-by: Vlastimil Babka <vbabka@suse.cz> Cc: Andrea Arcangeli <aarcange@redhat.com> Cc: Christoph Hellwig <hch@lst.de> Cc: David Rientjes <rientjes@google.com> Cc: Don Dutile <ddutile@redhat.com> Cc: Hugh Dickins <hughd@google.com> Cc: Jan Kara <jack@suse.cz> Cc: Jann Horn <jannh@google.com> Cc: Jason Gunthorpe <jgg@nvidia.com> Cc: John Hubbard <jhubbard@nvidia.com> Cc: Kirill A. Shutemov <kirill.shutemov@linux.intel.com> Cc: Liang Zhang <zhangliang5@huawei.com> Cc: Matthew Wilcox (Oracle) <willy@infradead.org> Cc: Michal Hocko <mhocko@kernel.org> Cc: Mike Kravetz <mike.kravetz@oracle.com> Cc: Mike Rapoport <rppt@linux.ibm.com> Cc: Nadav Amit <nadav.amit@gmail.com> Cc: Oleg Nesterov <oleg@redhat.com> Cc: Peter Xu <peterx@redhat.com> Cc: Rik van Riel <riel@surriel.com> Cc: Roman Gushchin <roman.gushchin@linux.dev> Cc: Shakeel Butt <shakeelb@google.com> Cc: Yang Shi <shy828301@gmail.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
		
							parent
							
								
									84d60fdd37
								
							
						
					
					
						commit
						c145e0b47c
					
				
					 1 changed files with 43 additions and 12 deletions
				
			
		
							
								
								
									
										55
									
								
								mm/memory.c
									
									
									
									
									
								
							
							
						
						
									
										55
									
								
								mm/memory.c
									
									
									
									
									
								
							|  | @ -3489,6 +3489,25 @@ static vm_fault_t remove_device_exclusive_entry(struct vm_fault *vmf) | |||
| 	return 0; | ||||
| } | ||||
| 
 | ||||
| static inline bool should_try_to_free_swap(struct page *page, | ||||
| 					   struct vm_area_struct *vma, | ||||
| 					   unsigned int fault_flags) | ||||
| { | ||||
| 	if (!PageSwapCache(page)) | ||||
| 		return false; | ||||
| 	if (mem_cgroup_swap_full(page) || (vma->vm_flags & VM_LOCKED) || | ||||
| 	    PageMlocked(page)) | ||||
| 		return true; | ||||
| 	/*
 | ||||
| 	 * If we want to map a page that's in the swapcache writable, we | ||||
| 	 * have to detect via the refcount if we're really the exclusive | ||||
| 	 * user. Try freeing the swapcache to get rid of the swapcache | ||||
| 	 * reference only in case it's likely that we'll be the exlusive user. | ||||
| 	 */ | ||||
| 	return (fault_flags & FAULT_FLAG_WRITE) && !PageKsm(page) && | ||||
| 		page_count(page) == 2; | ||||
| } | ||||
| 
 | ||||
| /*
 | ||||
|  * We enter with non-exclusive mmap_lock (to exclude vma changes, | ||||
|  * but allow concurrent faults), and pte mapped but not yet locked. | ||||
|  | @ -3630,6 +3649,16 @@ vm_fault_t do_swap_page(struct vm_fault *vmf) | |||
| 			page = swapcache; | ||||
| 			goto out_page; | ||||
| 		} | ||||
| 
 | ||||
| 		/*
 | ||||
| 		 * If we want to map a page that's in the swapcache writable, we | ||||
| 		 * have to detect via the refcount if we're really the exclusive | ||||
| 		 * owner. Try removing the extra reference from the local LRU | ||||
| 		 * pagevecs if required. | ||||
| 		 */ | ||||
| 		if ((vmf->flags & FAULT_FLAG_WRITE) && page == swapcache && | ||||
| 		    !PageKsm(page) && !PageLRU(page)) | ||||
| 			lru_add_drain(); | ||||
| 	} | ||||
| 
 | ||||
| 	cgroup_throttle_swaprate(page, GFP_KERNEL); | ||||
|  | @ -3648,19 +3677,25 @@ vm_fault_t do_swap_page(struct vm_fault *vmf) | |||
| 	} | ||||
| 
 | ||||
| 	/*
 | ||||
| 	 * The page isn't present yet, go ahead with the fault. | ||||
| 	 * | ||||
| 	 * Be careful about the sequence of operations here. | ||||
| 	 * To get its accounting right, reuse_swap_page() must be called | ||||
| 	 * while the page is counted on swap but not yet in mapcount i.e. | ||||
| 	 * before page_add_anon_rmap() and swap_free(); try_to_free_swap() | ||||
| 	 * must be called after the swap_free(), or it will never succeed. | ||||
| 	 * Remove the swap entry and conditionally try to free up the swapcache. | ||||
| 	 * We're already holding a reference on the page but haven't mapped it | ||||
| 	 * yet. | ||||
| 	 */ | ||||
| 	swap_free(entry); | ||||
| 	if (should_try_to_free_swap(page, vma, vmf->flags)) | ||||
| 		try_to_free_swap(page); | ||||
| 
 | ||||
| 	inc_mm_counter_fast(vma->vm_mm, MM_ANONPAGES); | ||||
| 	dec_mm_counter_fast(vma->vm_mm, MM_SWAPENTS); | ||||
| 	pte = mk_pte(page, vma->vm_page_prot); | ||||
| 	if ((vmf->flags & FAULT_FLAG_WRITE) && reuse_swap_page(page)) { | ||||
| 
 | ||||
| 	/*
 | ||||
| 	 * Same logic as in do_wp_page(); however, optimize for fresh pages | ||||
| 	 * that are certainly not shared because we just allocated them without | ||||
| 	 * exposing them to the swapcache. | ||||
| 	 */ | ||||
| 	if ((vmf->flags & FAULT_FLAG_WRITE) && !PageKsm(page) && | ||||
| 	    (page != swapcache || page_count(page) == 1)) { | ||||
| 		pte = maybe_mkwrite(pte_mkdirty(pte), vma); | ||||
| 		vmf->flags &= ~FAULT_FLAG_WRITE; | ||||
| 		ret |= VM_FAULT_WRITE; | ||||
|  | @ -3686,10 +3721,6 @@ vm_fault_t do_swap_page(struct vm_fault *vmf) | |||
| 	set_pte_at(vma->vm_mm, vmf->address, vmf->pte, pte); | ||||
| 	arch_do_swap_page(vma->vm_mm, vma, vmf->address, pte, vmf->orig_pte); | ||||
| 
 | ||||
| 	swap_free(entry); | ||||
| 	if (mem_cgroup_swap_full(page) || | ||||
| 	    (vma->vm_flags & VM_LOCKED) || PageMlocked(page)) | ||||
| 		try_to_free_swap(page); | ||||
| 	unlock_page(page); | ||||
| 	if (page != swapcache && swapcache) { | ||||
| 		/*
 | ||||
|  |  | |||
		Loading…
	
		Reference in a new issue
	
	 David Hildenbrand
						David Hildenbrand