mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-03 18:20:25 +02:00 
			
		
		
		
	hugetlb: parallelize 1G hugetlb initialization
Optimizing the initialization speed of 1G huge pages through
parallelization.
1G hugetlbs are allocated from bootmem, a process that is already very
fast and does not currently require optimization.  Therefore, we focus on
parallelizing only the initialization phase in `gather_bootmem_prealloc`.
Here are some test results:
      test case       no patch(ms)   patched(ms)   saved
 ------------------- -------------- ------------- --------
  256c2T(4 node) 1G           4745          2024   57.34%
  128c1T(2 node) 1G           3358          1712   49.02%
     12T         1G          77000         18300   76.23%
[akpm@linux-foundation.org: s/initialied/initialized/, per Alexey]
Link: https://lkml.kernel.org/r/20240222140422.393911-9-gang.li@linux.dev
Signed-off-by: Gang Li <ligang.bdlg@bytedance.com>
Tested-by: David Rientjes <rientjes@google.com>
Reviewed-by: Muchun Song <muchun.song@linux.dev>
Cc: Alexey Dobriyan <adobriyan@gmail.com>
Cc: Daniel Jordan <daniel.m.jordan@oracle.com>
Cc: David Hildenbrand <david@redhat.com>
Cc: Jane Chu <jane.chu@oracle.com>
Cc: Mike Kravetz <mike.kravetz@oracle.com>
Cc: Paul E. McKenney <paulmck@kernel.org>
Cc: Randy Dunlap <rdunlap@infradead.org>
Cc: Steffen Klassert <steffen.klassert@secunet.com>
Cc: Tim Chen <tim.c.chen@linux.intel.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
			
			
This commit is contained in:
		
							parent
							
								
									c6c21c31d0
								
							
						
					
					
						commit
						b78b27d029
					
				
					 3 changed files with 45 additions and 10 deletions
				
			
		| 
						 | 
					@ -226,7 +226,7 @@ static int __init pseries_alloc_bootmem_huge_page(struct hstate *hstate)
 | 
				
			||||||
		return 0;
 | 
							return 0;
 | 
				
			||||||
	m = phys_to_virt(gpage_freearray[--nr_gpages]);
 | 
						m = phys_to_virt(gpage_freearray[--nr_gpages]);
 | 
				
			||||||
	gpage_freearray[nr_gpages] = 0;
 | 
						gpage_freearray[nr_gpages] = 0;
 | 
				
			||||||
	list_add(&m->list, &huge_boot_pages);
 | 
						list_add(&m->list, &huge_boot_pages[0]);
 | 
				
			||||||
	m->hstate = hstate;
 | 
						m->hstate = hstate;
 | 
				
			||||||
	return 1;
 | 
						return 1;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -178,7 +178,7 @@ pte_t *huge_pmd_share(struct mm_struct *mm, struct vm_area_struct *vma,
 | 
				
			||||||
struct address_space *hugetlb_page_mapping_lock_write(struct page *hpage);
 | 
					struct address_space *hugetlb_page_mapping_lock_write(struct page *hpage);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
extern int sysctl_hugetlb_shm_group;
 | 
					extern int sysctl_hugetlb_shm_group;
 | 
				
			||||||
extern struct list_head huge_boot_pages;
 | 
					extern struct list_head huge_boot_pages[MAX_NUMNODES];
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/* arch callbacks */
 | 
					/* arch callbacks */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
							
								
								
									
										51
									
								
								mm/hugetlb.c
									
									
									
									
									
								
							
							
						
						
									
										51
									
								
								mm/hugetlb.c
									
									
									
									
									
								
							| 
						 | 
					@ -69,7 +69,7 @@ static bool hugetlb_cma_folio(struct folio *folio, unsigned int order)
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
static unsigned long hugetlb_cma_size __initdata;
 | 
					static unsigned long hugetlb_cma_size __initdata;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
__initdata LIST_HEAD(huge_boot_pages);
 | 
					__initdata struct list_head huge_boot_pages[MAX_NUMNODES];
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/* for command line parsing */
 | 
					/* for command line parsing */
 | 
				
			||||||
static struct hstate * __initdata parsed_hstate;
 | 
					static struct hstate * __initdata parsed_hstate;
 | 
				
			||||||
| 
						 | 
					@ -3301,7 +3301,7 @@ int alloc_bootmem_huge_page(struct hstate *h, int nid)
 | 
				
			||||||
int __alloc_bootmem_huge_page(struct hstate *h, int nid)
 | 
					int __alloc_bootmem_huge_page(struct hstate *h, int nid)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct huge_bootmem_page *m = NULL; /* initialize for clang */
 | 
						struct huge_bootmem_page *m = NULL; /* initialize for clang */
 | 
				
			||||||
	int nr_nodes, node;
 | 
						int nr_nodes, node = nid;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* do node specific alloc */
 | 
						/* do node specific alloc */
 | 
				
			||||||
	if (nid != NUMA_NO_NODE) {
 | 
						if (nid != NUMA_NO_NODE) {
 | 
				
			||||||
| 
						 | 
					@ -3339,7 +3339,7 @@ int __alloc_bootmem_huge_page(struct hstate *h, int nid)
 | 
				
			||||||
		huge_page_size(h) - PAGE_SIZE);
 | 
							huge_page_size(h) - PAGE_SIZE);
 | 
				
			||||||
	/* Put them into a private list first because mem_map is not up yet */
 | 
						/* Put them into a private list first because mem_map is not up yet */
 | 
				
			||||||
	INIT_LIST_HEAD(&m->list);
 | 
						INIT_LIST_HEAD(&m->list);
 | 
				
			||||||
	list_add(&m->list, &huge_boot_pages);
 | 
						list_add(&m->list, &huge_boot_pages[node]);
 | 
				
			||||||
	m->hstate = h;
 | 
						m->hstate = h;
 | 
				
			||||||
	return 1;
 | 
						return 1;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
| 
						 | 
					@ -3390,8 +3390,6 @@ static void __init prep_and_add_bootmem_folios(struct hstate *h,
 | 
				
			||||||
	/* Send list for bulk vmemmap optimization processing */
 | 
						/* Send list for bulk vmemmap optimization processing */
 | 
				
			||||||
	hugetlb_vmemmap_optimize_folios(h, folio_list);
 | 
						hugetlb_vmemmap_optimize_folios(h, folio_list);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* Add all new pool pages to free lists in one lock cycle */
 | 
					 | 
				
			||||||
	spin_lock_irqsave(&hugetlb_lock, flags);
 | 
					 | 
				
			||||||
	list_for_each_entry_safe(folio, tmp_f, folio_list, lru) {
 | 
						list_for_each_entry_safe(folio, tmp_f, folio_list, lru) {
 | 
				
			||||||
		if (!folio_test_hugetlb_vmemmap_optimized(folio)) {
 | 
							if (!folio_test_hugetlb_vmemmap_optimized(folio)) {
 | 
				
			||||||
			/*
 | 
								/*
 | 
				
			||||||
| 
						 | 
					@ -3404,23 +3402,25 @@ static void __init prep_and_add_bootmem_folios(struct hstate *h,
 | 
				
			||||||
					HUGETLB_VMEMMAP_RESERVE_PAGES,
 | 
										HUGETLB_VMEMMAP_RESERVE_PAGES,
 | 
				
			||||||
					pages_per_huge_page(h));
 | 
										pages_per_huge_page(h));
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
 | 
							/* Subdivide locks to achieve better parallel performance */
 | 
				
			||||||
 | 
							spin_lock_irqsave(&hugetlb_lock, flags);
 | 
				
			||||||
		__prep_account_new_huge_page(h, folio_nid(folio));
 | 
							__prep_account_new_huge_page(h, folio_nid(folio));
 | 
				
			||||||
		enqueue_hugetlb_folio(h, folio);
 | 
							enqueue_hugetlb_folio(h, folio);
 | 
				
			||||||
 | 
							spin_unlock_irqrestore(&hugetlb_lock, flags);
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
	spin_unlock_irqrestore(&hugetlb_lock, flags);
 | 
					 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/*
 | 
					/*
 | 
				
			||||||
 * Put bootmem huge pages into the standard lists after mem_map is up.
 | 
					 * Put bootmem huge pages into the standard lists after mem_map is up.
 | 
				
			||||||
 * Note: This only applies to gigantic (order > MAX_PAGE_ORDER) pages.
 | 
					 * Note: This only applies to gigantic (order > MAX_PAGE_ORDER) pages.
 | 
				
			||||||
 */
 | 
					 */
 | 
				
			||||||
static void __init gather_bootmem_prealloc(void)
 | 
					static void __init gather_bootmem_prealloc_node(unsigned long nid)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	LIST_HEAD(folio_list);
 | 
						LIST_HEAD(folio_list);
 | 
				
			||||||
	struct huge_bootmem_page *m;
 | 
						struct huge_bootmem_page *m;
 | 
				
			||||||
	struct hstate *h = NULL, *prev_h = NULL;
 | 
						struct hstate *h = NULL, *prev_h = NULL;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	list_for_each_entry(m, &huge_boot_pages, list) {
 | 
						list_for_each_entry(m, &huge_boot_pages[nid], list) {
 | 
				
			||||||
		struct page *page = virt_to_page(m);
 | 
							struct page *page = virt_to_page(m);
 | 
				
			||||||
		struct folio *folio = (void *)page;
 | 
							struct folio *folio = (void *)page;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -3453,6 +3453,31 @@ static void __init gather_bootmem_prealloc(void)
 | 
				
			||||||
	prep_and_add_bootmem_folios(h, &folio_list);
 | 
						prep_and_add_bootmem_folios(h, &folio_list);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					static void __init gather_bootmem_prealloc_parallel(unsigned long start,
 | 
				
			||||||
 | 
											    unsigned long end, void *arg)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						int nid;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						for (nid = start; nid < end; nid++)
 | 
				
			||||||
 | 
							gather_bootmem_prealloc_node(nid);
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					static void __init gather_bootmem_prealloc(void)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						struct padata_mt_job job = {
 | 
				
			||||||
 | 
							.thread_fn	= gather_bootmem_prealloc_parallel,
 | 
				
			||||||
 | 
							.fn_arg		= NULL,
 | 
				
			||||||
 | 
							.start		= 0,
 | 
				
			||||||
 | 
							.size		= num_node_state(N_MEMORY),
 | 
				
			||||||
 | 
							.align		= 1,
 | 
				
			||||||
 | 
							.min_chunk	= 1,
 | 
				
			||||||
 | 
							.max_threads	= num_node_state(N_MEMORY),
 | 
				
			||||||
 | 
							.numa_aware	= true,
 | 
				
			||||||
 | 
						};
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						padata_do_multithreaded(&job);
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static void __init hugetlb_hstate_alloc_pages_onenode(struct hstate *h, int nid)
 | 
					static void __init hugetlb_hstate_alloc_pages_onenode(struct hstate *h, int nid)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	unsigned long i;
 | 
						unsigned long i;
 | 
				
			||||||
| 
						 | 
					@ -3600,6 +3625,7 @@ static unsigned long __init hugetlb_pages_alloc_boot(struct hstate *h)
 | 
				
			||||||
static void __init hugetlb_hstate_alloc_pages(struct hstate *h)
 | 
					static void __init hugetlb_hstate_alloc_pages(struct hstate *h)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	unsigned long allocated;
 | 
						unsigned long allocated;
 | 
				
			||||||
 | 
						static bool initialized __initdata;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* skip gigantic hugepages allocation if hugetlb_cma enabled */
 | 
						/* skip gigantic hugepages allocation if hugetlb_cma enabled */
 | 
				
			||||||
	if (hstate_is_gigantic(h) && hugetlb_cma_size) {
 | 
						if (hstate_is_gigantic(h) && hugetlb_cma_size) {
 | 
				
			||||||
| 
						 | 
					@ -3607,6 +3633,15 @@ static void __init hugetlb_hstate_alloc_pages(struct hstate *h)
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						/* hugetlb_hstate_alloc_pages will be called many times, initialize huge_boot_pages once */
 | 
				
			||||||
 | 
						if (!initialized) {
 | 
				
			||||||
 | 
							int i = 0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							for (i = 0; i < MAX_NUMNODES; i++)
 | 
				
			||||||
 | 
								INIT_LIST_HEAD(&huge_boot_pages[i]);
 | 
				
			||||||
 | 
							initialized = true;
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* do node specific alloc */
 | 
						/* do node specific alloc */
 | 
				
			||||||
	if (hugetlb_hstate_alloc_pages_specific_nodes(h))
 | 
						if (hugetlb_hstate_alloc_pages_specific_nodes(h))
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue