mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 02:30:34 +02:00 
			
		
		
		
	mm: hugetlb: cleanup using paeg_huge_active()
Now we have an easy access to hugepages' activeness, so existing helpers to get the information can be cleaned up. [akpm@linux-foundation.org: s/PageHugeActive/page_huge_active/] Signed-off-by: Naoya Horiguchi <n-horiguchi@ah.jp.nec.com> Cc: Hugh Dickins <hughd@google.com> Reviewed-by: Michal Hocko <mhocko@suse.cz> Cc: Mel Gorman <mgorman@suse.de> Cc: Johannes Weiner <hannes@cmpxchg.org> Cc: David Rientjes <rientjes@google.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
		
							parent
							
								
									bcc5422230
								
							
						
					
					
						commit
						7e1f049efb
					
				
					 4 changed files with 13 additions and 40 deletions
				
			
		| 
						 | 
					@ -84,7 +84,6 @@ void hugetlb_unreserve_pages(struct inode *inode, long offset, long freed);
 | 
				
			||||||
int dequeue_hwpoisoned_huge_page(struct page *page);
 | 
					int dequeue_hwpoisoned_huge_page(struct page *page);
 | 
				
			||||||
bool isolate_huge_page(struct page *page, struct list_head *list);
 | 
					bool isolate_huge_page(struct page *page, struct list_head *list);
 | 
				
			||||||
void putback_active_hugepage(struct page *page);
 | 
					void putback_active_hugepage(struct page *page);
 | 
				
			||||||
bool is_hugepage_active(struct page *page);
 | 
					 | 
				
			||||||
void free_huge_page(struct page *page);
 | 
					void free_huge_page(struct page *page);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#ifdef CONFIG_ARCH_WANT_HUGE_PMD_SHARE
 | 
					#ifdef CONFIG_ARCH_WANT_HUGE_PMD_SHARE
 | 
				
			||||||
| 
						 | 
					@ -152,7 +151,6 @@ static inline bool isolate_huge_page(struct page *page, struct list_head *list)
 | 
				
			||||||
	return false;
 | 
						return false;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
#define putback_active_hugepage(p)	do {} while (0)
 | 
					#define putback_active_hugepage(p)	do {} while (0)
 | 
				
			||||||
#define is_hugepage_active(x)	false
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
static inline unsigned long hugetlb_change_protection(struct vm_area_struct *vma,
 | 
					static inline unsigned long hugetlb_change_protection(struct vm_area_struct *vma,
 | 
				
			||||||
		unsigned long address, unsigned long end, pgprot_t newprot)
 | 
							unsigned long address, unsigned long end, pgprot_t newprot)
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -470,11 +470,18 @@ static inline void ClearPageCompound(struct page *page)
 | 
				
			||||||
#ifdef CONFIG_HUGETLB_PAGE
 | 
					#ifdef CONFIG_HUGETLB_PAGE
 | 
				
			||||||
int PageHuge(struct page *page);
 | 
					int PageHuge(struct page *page);
 | 
				
			||||||
int PageHeadHuge(struct page *page);
 | 
					int PageHeadHuge(struct page *page);
 | 
				
			||||||
 | 
					bool page_huge_active(struct page *page);
 | 
				
			||||||
#else
 | 
					#else
 | 
				
			||||||
TESTPAGEFLAG_FALSE(Huge)
 | 
					TESTPAGEFLAG_FALSE(Huge)
 | 
				
			||||||
TESTPAGEFLAG_FALSE(HeadHuge)
 | 
					TESTPAGEFLAG_FALSE(HeadHuge)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					static inline bool page_huge_active(struct page *page)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						return 0;
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#ifdef CONFIG_TRANSPARENT_HUGEPAGE
 | 
					#ifdef CONFIG_TRANSPARENT_HUGEPAGE
 | 
				
			||||||
/*
 | 
					/*
 | 
				
			||||||
 * PageHuge() only returns true for hugetlbfs pages, but not for
 | 
					 * PageHuge() only returns true for hugetlbfs pages, but not for
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
							
								
								
									
										42
									
								
								mm/hugetlb.c
									
									
									
									
									
								
							
							
						
						
									
										42
									
								
								mm/hugetlb.c
									
									
									
									
									
								
							| 
						 | 
					@ -3896,20 +3896,6 @@ follow_huge_pud(struct mm_struct *mm, unsigned long address,
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#ifdef CONFIG_MEMORY_FAILURE
 | 
					#ifdef CONFIG_MEMORY_FAILURE
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/* Should be called in hugetlb_lock */
 | 
					 | 
				
			||||||
static int is_hugepage_on_freelist(struct page *hpage)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	struct page *page;
 | 
					 | 
				
			||||||
	struct page *tmp;
 | 
					 | 
				
			||||||
	struct hstate *h = page_hstate(hpage);
 | 
					 | 
				
			||||||
	int nid = page_to_nid(hpage);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	list_for_each_entry_safe(page, tmp, &h->hugepage_freelists[nid], lru)
 | 
					 | 
				
			||||||
		if (page == hpage)
 | 
					 | 
				
			||||||
			return 1;
 | 
					 | 
				
			||||||
	return 0;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/*
 | 
					/*
 | 
				
			||||||
 * This function is called from memory failure code.
 | 
					 * This function is called from memory failure code.
 | 
				
			||||||
 * Assume the caller holds page lock of the head page.
 | 
					 * Assume the caller holds page lock of the head page.
 | 
				
			||||||
| 
						 | 
					@ -3921,7 +3907,11 @@ int dequeue_hwpoisoned_huge_page(struct page *hpage)
 | 
				
			||||||
	int ret = -EBUSY;
 | 
						int ret = -EBUSY;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	spin_lock(&hugetlb_lock);
 | 
						spin_lock(&hugetlb_lock);
 | 
				
			||||||
	if (is_hugepage_on_freelist(hpage)) {
 | 
						/*
 | 
				
			||||||
 | 
						 * Just checking !page_huge_active is not enough, because that could be
 | 
				
			||||||
 | 
						 * an isolated/hwpoisoned hugepage (which have >0 refcount).
 | 
				
			||||||
 | 
						 */
 | 
				
			||||||
 | 
						if (!page_huge_active(hpage) && !page_count(hpage)) {
 | 
				
			||||||
		/*
 | 
							/*
 | 
				
			||||||
		 * Hwpoisoned hugepage isn't linked to activelist or freelist,
 | 
							 * Hwpoisoned hugepage isn't linked to activelist or freelist,
 | 
				
			||||||
		 * but dangling hpage->lru can trigger list-debug warnings
 | 
							 * but dangling hpage->lru can trigger list-debug warnings
 | 
				
			||||||
| 
						 | 
					@ -3965,25 +3955,3 @@ void putback_active_hugepage(struct page *page)
 | 
				
			||||||
	spin_unlock(&hugetlb_lock);
 | 
						spin_unlock(&hugetlb_lock);
 | 
				
			||||||
	put_page(page);
 | 
						put_page(page);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					 | 
				
			||||||
bool is_hugepage_active(struct page *page)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	VM_BUG_ON_PAGE(!PageHuge(page), page);
 | 
					 | 
				
			||||||
	/*
 | 
					 | 
				
			||||||
	 * This function can be called for a tail page because the caller,
 | 
					 | 
				
			||||||
	 * scan_movable_pages, scans through a given pfn-range which typically
 | 
					 | 
				
			||||||
	 * covers one memory block. In systems using gigantic hugepage (1GB
 | 
					 | 
				
			||||||
	 * for x86_64,) a hugepage is larger than a memory block, and we don't
 | 
					 | 
				
			||||||
	 * support migrating such large hugepages for now, so return false
 | 
					 | 
				
			||||||
	 * when called for tail pages.
 | 
					 | 
				
			||||||
	 */
 | 
					 | 
				
			||||||
	if (PageTail(page))
 | 
					 | 
				
			||||||
		return false;
 | 
					 | 
				
			||||||
	/*
 | 
					 | 
				
			||||||
	 * Refcount of a hwpoisoned hugepages is 1, but they are not active,
 | 
					 | 
				
			||||||
	 * so we should return false for them.
 | 
					 | 
				
			||||||
	 */
 | 
					 | 
				
			||||||
	if (unlikely(PageHWPoison(page)))
 | 
					 | 
				
			||||||
		return false;
 | 
					 | 
				
			||||||
	return page_count(page) > 0;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -1373,7 +1373,7 @@ static unsigned long scan_movable_pages(unsigned long start, unsigned long end)
 | 
				
			||||||
			if (PageLRU(page))
 | 
								if (PageLRU(page))
 | 
				
			||||||
				return pfn;
 | 
									return pfn;
 | 
				
			||||||
			if (PageHuge(page)) {
 | 
								if (PageHuge(page)) {
 | 
				
			||||||
				if (is_hugepage_active(page))
 | 
									if (page_huge_active(page))
 | 
				
			||||||
					return pfn;
 | 
										return pfn;
 | 
				
			||||||
				else
 | 
									else
 | 
				
			||||||
					pfn = round_up(pfn + 1,
 | 
										pfn = round_up(pfn + 1,
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue