mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 02:30:34 +02:00 
			
		
		
		
	swap: remove remnants of polling from read_swap_cache_async
Patch series "Per-VMA lock support for swap and userfaults", v7. When per-VMA locks were introduced in [1] several types of page faults would still fall back to mmap_lock to keep the patchset simple. Among them are swap and userfault pages. The main reason for skipping those cases was the fact that mmap_lock could be dropped while handling these faults and that required additional logic to be implemented. Implement the mechanism to allow per-VMA locks to be dropped for these cases. First, change handle_mm_fault to drop per-VMA locks when returning VM_FAULT_RETRY or VM_FAULT_COMPLETED to be consistent with the way mmap_lock is handled. Then change folio_lock_or_retry to accept vm_fault and return vm_fault_t which simplifies later patches. Finally allow swap and uffd page faults to be handled under per-VMA locks by dropping per-VMA and retrying, the same way it's done under mmap_lock. Naturally, once VMA lock is dropped that VMA should be assumed unstable and can't be used. This patch (of 6): Commit [1] introduced IO polling support duding swapin to reduce swap read latency for block devices that can be polled. However later commit [2] removed polling support. Therefore it seems safe to remove do_poll parameter in read_swap_cache_async and always call swap_readpage with synchronous=false waiting for IO completion in folio_lock_or_retry. [1] commit23955622ff("swap: add block io poll in swapin path") [2] commit9650b453a3("block: ignore RWF_HIPRI hint for sync dio") Link: https://lkml.kernel.org/r/20230630211957.1341547-1-surenb@google.com Link: https://lkml.kernel.org/r/20230630211957.1341547-2-surenb@google.com Signed-off-by: Suren Baghdasaryan <surenb@google.com> Suggested-by: "Huang, Ying" <ying.huang@intel.com> Reviewed-by: "Huang, Ying" <ying.huang@intel.com> Reviewed-by: Christoph Hellwig <hch@lst.de> Cc: Alistair Popple <apopple@nvidia.com> Cc: Al Viro <viro@zeniv.linux.org.uk> Cc: Christian Brauner <brauner@kernel.org> Cc: David Hildenbrand <david@redhat.com> Cc: David Howells <dhowells@redhat.com> Cc: Davidlohr Bueso <dave@stgolabs.net> Cc: Hillf Danton <hdanton@sina.com> Cc: Hugh Dickins <hughd@google.com> Cc: Jan Kara <jack@suse.cz> Cc: Johannes Weiner <hannes@cmpxchg.org> Cc: Josef Bacik <josef@toxicpanda.com> Cc: Laurent Dufour <ldufour@linux.ibm.com> Cc: Liam R. Howlett <Liam.Howlett@oracle.com> Cc: Lorenzo Stoakes <lstoakes@gmail.com> Cc: Matthew Wilcox <willy@infradead.org> Cc: Michal Hocko <mhocko@suse.com> Cc: Michel Lespinasse <michel@lespinasse.org> Cc: Minchan Kim <minchan@google.com> Cc: Pavel Tatashin <pasha.tatashin@soleen.com> Cc: Peter Xu <peterx@redhat.com> Cc: Punit Agrawal <punit.agrawal@bytedance.com> Cc: Vlastimil Babka <vbabka@suse.cz> Cc: Yu Zhao <yuzhao@google.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
		
							parent
							
								
									d51b68469b
								
							
						
					
					
						commit
						b243dcbf2f
					
				
					 3 changed files with 7 additions and 10 deletions
				
			
		| 
						 | 
					@ -217,7 +217,7 @@ static int swapin_walk_pmd_entry(pmd_t *pmd, unsigned long start,
 | 
				
			||||||
		ptep = NULL;
 | 
							ptep = NULL;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		page = read_swap_cache_async(entry, GFP_HIGHUSER_MOVABLE,
 | 
							page = read_swap_cache_async(entry, GFP_HIGHUSER_MOVABLE,
 | 
				
			||||||
					     vma, addr, false, &splug);
 | 
										     vma, addr, &splug);
 | 
				
			||||||
		if (page)
 | 
							if (page)
 | 
				
			||||||
			put_page(page);
 | 
								put_page(page);
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
| 
						 | 
					@ -262,7 +262,7 @@ static void shmem_swapin_range(struct vm_area_struct *vma,
 | 
				
			||||||
		rcu_read_unlock();
 | 
							rcu_read_unlock();
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		page = read_swap_cache_async(entry, mapping_gfp_mask(mapping),
 | 
							page = read_swap_cache_async(entry, mapping_gfp_mask(mapping),
 | 
				
			||||||
					     vma, addr, false, &splug);
 | 
										     vma, addr, &splug);
 | 
				
			||||||
		if (page)
 | 
							if (page)
 | 
				
			||||||
			put_page(page);
 | 
								put_page(page);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -46,7 +46,6 @@ struct folio *filemap_get_incore_folio(struct address_space *mapping,
 | 
				
			||||||
struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask,
 | 
					struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask,
 | 
				
			||||||
				   struct vm_area_struct *vma,
 | 
									   struct vm_area_struct *vma,
 | 
				
			||||||
				   unsigned long addr,
 | 
									   unsigned long addr,
 | 
				
			||||||
				   bool do_poll,
 | 
					 | 
				
			||||||
				   struct swap_iocb **plug);
 | 
									   struct swap_iocb **plug);
 | 
				
			||||||
struct page *__read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask,
 | 
					struct page *__read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask,
 | 
				
			||||||
				     struct vm_area_struct *vma,
 | 
									     struct vm_area_struct *vma,
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -526,15 +526,14 @@ struct page *__read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask,
 | 
				
			||||||
 */
 | 
					 */
 | 
				
			||||||
struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask,
 | 
					struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask,
 | 
				
			||||||
				   struct vm_area_struct *vma,
 | 
									   struct vm_area_struct *vma,
 | 
				
			||||||
				   unsigned long addr, bool do_poll,
 | 
									   unsigned long addr, struct swap_iocb **plug)
 | 
				
			||||||
				   struct swap_iocb **plug)
 | 
					 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	bool page_was_allocated;
 | 
						bool page_was_allocated;
 | 
				
			||||||
	struct page *retpage = __read_swap_cache_async(entry, gfp_mask,
 | 
						struct page *retpage = __read_swap_cache_async(entry, gfp_mask,
 | 
				
			||||||
			vma, addr, &page_was_allocated);
 | 
								vma, addr, &page_was_allocated);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (page_was_allocated)
 | 
						if (page_was_allocated)
 | 
				
			||||||
		swap_readpage(retpage, do_poll, plug);
 | 
							swap_readpage(retpage, false, plug);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	return retpage;
 | 
						return retpage;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
| 
						 | 
					@ -629,7 +628,7 @@ struct page *swap_cluster_readahead(swp_entry_t entry, gfp_t gfp_mask,
 | 
				
			||||||
	struct swap_info_struct *si = swp_swap_info(entry);
 | 
						struct swap_info_struct *si = swp_swap_info(entry);
 | 
				
			||||||
	struct blk_plug plug;
 | 
						struct blk_plug plug;
 | 
				
			||||||
	struct swap_iocb *splug = NULL;
 | 
						struct swap_iocb *splug = NULL;
 | 
				
			||||||
	bool do_poll = true, page_allocated;
 | 
						bool page_allocated;
 | 
				
			||||||
	struct vm_area_struct *vma = vmf->vma;
 | 
						struct vm_area_struct *vma = vmf->vma;
 | 
				
			||||||
	unsigned long addr = vmf->address;
 | 
						unsigned long addr = vmf->address;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -637,7 +636,6 @@ struct page *swap_cluster_readahead(swp_entry_t entry, gfp_t gfp_mask,
 | 
				
			||||||
	if (!mask)
 | 
						if (!mask)
 | 
				
			||||||
		goto skip;
 | 
							goto skip;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	do_poll = false;
 | 
					 | 
				
			||||||
	/* Read a page_cluster sized and aligned cluster around offset. */
 | 
						/* Read a page_cluster sized and aligned cluster around offset. */
 | 
				
			||||||
	start_offset = offset & ~mask;
 | 
						start_offset = offset & ~mask;
 | 
				
			||||||
	end_offset = offset | mask;
 | 
						end_offset = offset | mask;
 | 
				
			||||||
| 
						 | 
					@ -669,7 +667,7 @@ struct page *swap_cluster_readahead(swp_entry_t entry, gfp_t gfp_mask,
 | 
				
			||||||
	lru_add_drain();	/* Push any new pages onto the LRU now */
 | 
						lru_add_drain();	/* Push any new pages onto the LRU now */
 | 
				
			||||||
skip:
 | 
					skip:
 | 
				
			||||||
	/* The page was likely read above, so no need for plugging here */
 | 
						/* The page was likely read above, so no need for plugging here */
 | 
				
			||||||
	return read_swap_cache_async(entry, gfp_mask, vma, addr, do_poll, NULL);
 | 
						return read_swap_cache_async(entry, gfp_mask, vma, addr, NULL);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int init_swap_address_space(unsigned int type, unsigned long nr_pages)
 | 
					int init_swap_address_space(unsigned int type, unsigned long nr_pages)
 | 
				
			||||||
| 
						 | 
					@ -837,7 +835,7 @@ static struct page *swap_vma_readahead(swp_entry_t fentry, gfp_t gfp_mask,
 | 
				
			||||||
skip:
 | 
					skip:
 | 
				
			||||||
	/* The page was likely read above, so no need for plugging here */
 | 
						/* The page was likely read above, so no need for plugging here */
 | 
				
			||||||
	return read_swap_cache_async(fentry, gfp_mask, vma, vmf->address,
 | 
						return read_swap_cache_async(fentry, gfp_mask, vma, vmf->address,
 | 
				
			||||||
				     ra_info.win == 1, NULL);
 | 
									     NULL);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/**
 | 
					/**
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue