mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 10:40:15 +02:00 
			
		
		
		
	mm, hotplug: avoid compiling memory hotremove functions when disabled
__remove_pages() is only necessary for CONFIG_MEMORY_HOTREMOVE. PowerPC pseries will return -EOPNOTSUPP if unsupported. Adding an #ifdef causes several other functions it depends on to also become unnecessary, which saves in .text when disabled (it's disabled in most defconfigs besides powerpc, including x86). remove_memory_block() becomes static since it is not referenced outside of drivers/base/memory.c. Build tested on x86 and powerpc with CONFIG_MEMORY_HOTREMOVE both enabled and disabled. Signed-off-by: David Rientjes <rientjes@google.com> Acked-by: Toshi Kani <toshi.kani@hp.com> Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org> Cc: Paul Mackerras <paulus@samba.org> Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org> Cc: Wen Congyang <wency@cn.fujitsu.com> Cc: Tang Chen <tangchen@cn.fujitsu.com> Cc: Yasuaki Ishimatsu <isimatu.yasuaki@jp.fujitsu.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
		
							parent
							
								
									fe74ebb106
								
							
						
					
					
						commit
						4edd7ceff0
					
				
					 6 changed files with 113 additions and 90 deletions
				
			
		| 
						 | 
					@ -72,6 +72,7 @@ unsigned long memory_block_size_bytes(void)
 | 
				
			||||||
	return get_memblock_size();
 | 
						return get_memblock_size();
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					#ifdef CONFIG_MEMORY_HOTREMOVE
 | 
				
			||||||
static int pseries_remove_memblock(unsigned long base, unsigned int memblock_size)
 | 
					static int pseries_remove_memblock(unsigned long base, unsigned int memblock_size)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	unsigned long start, start_pfn;
 | 
						unsigned long start, start_pfn;
 | 
				
			||||||
| 
						 | 
					@ -153,6 +154,17 @@ static int pseries_remove_memory(struct device_node *np)
 | 
				
			||||||
	ret = pseries_remove_memblock(base, lmb_size);
 | 
						ret = pseries_remove_memblock(base, lmb_size);
 | 
				
			||||||
	return ret;
 | 
						return ret;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					#else
 | 
				
			||||||
 | 
					static inline int pseries_remove_memblock(unsigned long base,
 | 
				
			||||||
 | 
										  unsigned int memblock_size)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						return -EOPNOTSUPP;
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					static inline int pseries_remove_memory(struct device_node *np)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						return -EOPNOTSUPP;
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					#endif /* CONFIG_MEMORY_HOTREMOVE */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static int pseries_add_memory(struct device_node *np)
 | 
					static int pseries_add_memory(struct device_node *np)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -93,16 +93,6 @@ int register_memory(struct memory_block *memory)
 | 
				
			||||||
	return error;
 | 
						return error;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static void
 | 
					 | 
				
			||||||
unregister_memory(struct memory_block *memory)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	BUG_ON(memory->dev.bus != &memory_subsys);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/* drop the ref. we got in remove_memory_block() */
 | 
					 | 
				
			||||||
	kobject_put(&memory->dev.kobj);
 | 
					 | 
				
			||||||
	device_unregister(&memory->dev);
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
unsigned long __weak memory_block_size_bytes(void)
 | 
					unsigned long __weak memory_block_size_bytes(void)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	return MIN_MEMORY_BLOCK_SIZE;
 | 
						return MIN_MEMORY_BLOCK_SIZE;
 | 
				
			||||||
| 
						 | 
					@ -637,8 +627,28 @@ static int add_memory_section(int nid, struct mem_section *section,
 | 
				
			||||||
	return ret;
 | 
						return ret;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int remove_memory_block(unsigned long node_id, struct mem_section *section,
 | 
					/*
 | 
				
			||||||
		int phys_device)
 | 
					 * need an interface for the VM to add new memory regions,
 | 
				
			||||||
 | 
					 * but without onlining it.
 | 
				
			||||||
 | 
					 */
 | 
				
			||||||
 | 
					int register_new_memory(int nid, struct mem_section *section)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						return add_memory_section(nid, section, NULL, MEM_OFFLINE, HOTPLUG);
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					#ifdef CONFIG_MEMORY_HOTREMOVE
 | 
				
			||||||
 | 
					static void
 | 
				
			||||||
 | 
					unregister_memory(struct memory_block *memory)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						BUG_ON(memory->dev.bus != &memory_subsys);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						/* drop the ref. we got in remove_memory_block() */
 | 
				
			||||||
 | 
						kobject_put(&memory->dev.kobj);
 | 
				
			||||||
 | 
						device_unregister(&memory->dev);
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					static int remove_memory_block(unsigned long node_id,
 | 
				
			||||||
 | 
								       struct mem_section *section, int phys_device)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct memory_block *mem;
 | 
						struct memory_block *mem;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -661,15 +671,6 @@ int remove_memory_block(unsigned long node_id, struct mem_section *section,
 | 
				
			||||||
	return 0;
 | 
						return 0;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/*
 | 
					 | 
				
			||||||
 * need an interface for the VM to add new memory regions,
 | 
					 | 
				
			||||||
 * but without onlining it.
 | 
					 | 
				
			||||||
 */
 | 
					 | 
				
			||||||
int register_new_memory(int nid, struct mem_section *section)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	return add_memory_section(nid, section, NULL, MEM_OFFLINE, HOTPLUG);
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
int unregister_memory_section(struct mem_section *section)
 | 
					int unregister_memory_section(struct mem_section *section)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	if (!present_section(section))
 | 
						if (!present_section(section))
 | 
				
			||||||
| 
						 | 
					@ -677,6 +678,7 @@ int unregister_memory_section(struct mem_section *section)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	return remove_memory_block(0, section, 0);
 | 
						return remove_memory_block(0, section, 0);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					#endif /* CONFIG_MEMORY_HOTREMOVE */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/*
 | 
					/*
 | 
				
			||||||
 * offline one memory block. If the memory block has been offlined, do nothing.
 | 
					 * offline one memory block. If the memory block has been offlined, do nothing.
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -115,9 +115,10 @@ extern void unregister_memory_notifier(struct notifier_block *nb);
 | 
				
			||||||
extern int register_memory_isolate_notifier(struct notifier_block *nb);
 | 
					extern int register_memory_isolate_notifier(struct notifier_block *nb);
 | 
				
			||||||
extern void unregister_memory_isolate_notifier(struct notifier_block *nb);
 | 
					extern void unregister_memory_isolate_notifier(struct notifier_block *nb);
 | 
				
			||||||
extern int register_new_memory(int, struct mem_section *);
 | 
					extern int register_new_memory(int, struct mem_section *);
 | 
				
			||||||
 | 
					#ifdef CONFIG_MEMORY_HOTREMOVE
 | 
				
			||||||
extern int unregister_memory_section(struct mem_section *);
 | 
					extern int unregister_memory_section(struct mem_section *);
 | 
				
			||||||
 | 
					#endif
 | 
				
			||||||
extern int memory_dev_init(void);
 | 
					extern int memory_dev_init(void);
 | 
				
			||||||
extern int remove_memory_block(unsigned long, struct mem_section *, int);
 | 
					 | 
				
			||||||
extern int memory_notify(unsigned long val, void *v);
 | 
					extern int memory_notify(unsigned long val, void *v);
 | 
				
			||||||
extern int memory_isolate_notify(unsigned long val, void *v);
 | 
					extern int memory_isolate_notify(unsigned long val, void *v);
 | 
				
			||||||
extern struct memory_block *find_memory_block_hinted(struct mem_section *,
 | 
					extern struct memory_block *find_memory_block_hinted(struct mem_section *,
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -97,13 +97,13 @@ extern void __online_page_free(struct page *page);
 | 
				
			||||||
#ifdef CONFIG_MEMORY_HOTREMOVE
 | 
					#ifdef CONFIG_MEMORY_HOTREMOVE
 | 
				
			||||||
extern bool is_pageblock_removable_nolock(struct page *page);
 | 
					extern bool is_pageblock_removable_nolock(struct page *page);
 | 
				
			||||||
extern int arch_remove_memory(u64 start, u64 size);
 | 
					extern int arch_remove_memory(u64 start, u64 size);
 | 
				
			||||||
 | 
					extern int __remove_pages(struct zone *zone, unsigned long start_pfn,
 | 
				
			||||||
 | 
						unsigned long nr_pages);
 | 
				
			||||||
#endif /* CONFIG_MEMORY_HOTREMOVE */
 | 
					#endif /* CONFIG_MEMORY_HOTREMOVE */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/* reasonably generic interface to expand the physical pages in a zone  */
 | 
					/* reasonably generic interface to expand the physical pages in a zone  */
 | 
				
			||||||
extern int __add_pages(int nid, struct zone *zone, unsigned long start_pfn,
 | 
					extern int __add_pages(int nid, struct zone *zone, unsigned long start_pfn,
 | 
				
			||||||
	unsigned long nr_pages);
 | 
						unsigned long nr_pages);
 | 
				
			||||||
extern int __remove_pages(struct zone *zone, unsigned long start_pfn,
 | 
					 | 
				
			||||||
	unsigned long nr_pages);
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
#ifdef CONFIG_NUMA
 | 
					#ifdef CONFIG_NUMA
 | 
				
			||||||
extern int memory_add_physaddr_to_nid(u64 start);
 | 
					extern int memory_add_physaddr_to_nid(u64 start);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -436,6 +436,40 @@ static int __meminit __add_section(int nid, struct zone *zone,
 | 
				
			||||||
	return register_new_memory(nid, __pfn_to_section(phys_start_pfn));
 | 
						return register_new_memory(nid, __pfn_to_section(phys_start_pfn));
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					/*
 | 
				
			||||||
 | 
					 * Reasonably generic function for adding memory.  It is
 | 
				
			||||||
 | 
					 * expected that archs that support memory hotplug will
 | 
				
			||||||
 | 
					 * call this function after deciding the zone to which to
 | 
				
			||||||
 | 
					 * add the new pages.
 | 
				
			||||||
 | 
					 */
 | 
				
			||||||
 | 
					int __ref __add_pages(int nid, struct zone *zone, unsigned long phys_start_pfn,
 | 
				
			||||||
 | 
								unsigned long nr_pages)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						unsigned long i;
 | 
				
			||||||
 | 
						int err = 0;
 | 
				
			||||||
 | 
						int start_sec, end_sec;
 | 
				
			||||||
 | 
						/* during initialize mem_map, align hot-added range to section */
 | 
				
			||||||
 | 
						start_sec = pfn_to_section_nr(phys_start_pfn);
 | 
				
			||||||
 | 
						end_sec = pfn_to_section_nr(phys_start_pfn + nr_pages - 1);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						for (i = start_sec; i <= end_sec; i++) {
 | 
				
			||||||
 | 
							err = __add_section(nid, zone, i << PFN_SECTION_SHIFT);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							/*
 | 
				
			||||||
 | 
							 * EEXIST is finally dealt with by ioresource collision
 | 
				
			||||||
 | 
							 * check. see add_memory() => register_memory_resource()
 | 
				
			||||||
 | 
							 * Warning will be printed if there is collision.
 | 
				
			||||||
 | 
							 */
 | 
				
			||||||
 | 
							if (err && (err != -EEXIST))
 | 
				
			||||||
 | 
								break;
 | 
				
			||||||
 | 
							err = 0;
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						return err;
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					EXPORT_SYMBOL_GPL(__add_pages);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					#ifdef CONFIG_MEMORY_HOTREMOVE
 | 
				
			||||||
/* find the smallest valid pfn in the range [start_pfn, end_pfn) */
 | 
					/* find the smallest valid pfn in the range [start_pfn, end_pfn) */
 | 
				
			||||||
static int find_smallest_section_pfn(int nid, struct zone *zone,
 | 
					static int find_smallest_section_pfn(int nid, struct zone *zone,
 | 
				
			||||||
				     unsigned long start_pfn,
 | 
									     unsigned long start_pfn,
 | 
				
			||||||
| 
						 | 
					@ -658,39 +692,6 @@ static int __remove_section(struct zone *zone, struct mem_section *ms)
 | 
				
			||||||
	return 0;
 | 
						return 0;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/*
 | 
					 | 
				
			||||||
 * Reasonably generic function for adding memory.  It is
 | 
					 | 
				
			||||||
 * expected that archs that support memory hotplug will
 | 
					 | 
				
			||||||
 * call this function after deciding the zone to which to
 | 
					 | 
				
			||||||
 * add the new pages.
 | 
					 | 
				
			||||||
 */
 | 
					 | 
				
			||||||
int __ref __add_pages(int nid, struct zone *zone, unsigned long phys_start_pfn,
 | 
					 | 
				
			||||||
			unsigned long nr_pages)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	unsigned long i;
 | 
					 | 
				
			||||||
	int err = 0;
 | 
					 | 
				
			||||||
	int start_sec, end_sec;
 | 
					 | 
				
			||||||
	/* during initialize mem_map, align hot-added range to section */
 | 
					 | 
				
			||||||
	start_sec = pfn_to_section_nr(phys_start_pfn);
 | 
					 | 
				
			||||||
	end_sec = pfn_to_section_nr(phys_start_pfn + nr_pages - 1);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	for (i = start_sec; i <= end_sec; i++) {
 | 
					 | 
				
			||||||
		err = __add_section(nid, zone, i << PFN_SECTION_SHIFT);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		/*
 | 
					 | 
				
			||||||
		 * EEXIST is finally dealt with by ioresource collision
 | 
					 | 
				
			||||||
		 * check. see add_memory() => register_memory_resource()
 | 
					 | 
				
			||||||
		 * Warning will be printed if there is collision.
 | 
					 | 
				
			||||||
		 */
 | 
					 | 
				
			||||||
		if (err && (err != -EEXIST))
 | 
					 | 
				
			||||||
			break;
 | 
					 | 
				
			||||||
		err = 0;
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	return err;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
EXPORT_SYMBOL_GPL(__add_pages);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/**
 | 
					/**
 | 
				
			||||||
 * __remove_pages() - remove sections of pages from a zone
 | 
					 * __remove_pages() - remove sections of pages from a zone
 | 
				
			||||||
 * @zone: zone from which pages need to be removed
 | 
					 * @zone: zone from which pages need to be removed
 | 
				
			||||||
| 
						 | 
					@ -733,6 +734,7 @@ int __remove_pages(struct zone *zone, unsigned long phys_start_pfn,
 | 
				
			||||||
	return ret;
 | 
						return ret;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
EXPORT_SYMBOL_GPL(__remove_pages);
 | 
					EXPORT_SYMBOL_GPL(__remove_pages);
 | 
				
			||||||
 | 
					#endif /* CONFIG_MEMORY_HOTREMOVE */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int set_online_page_callback(online_page_callback_t callback)
 | 
					int set_online_page_callback(online_page_callback_t callback)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
							
								
								
									
										72
									
								
								mm/sparse.c
									
									
									
									
									
								
							
							
						
						
									
										72
									
								
								mm/sparse.c
									
									
									
									
									
								
							| 
						 | 
					@ -620,6 +620,7 @@ static void __kfree_section_memmap(struct page *memmap, unsigned long nr_pages)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	vmemmap_free(start, end);
 | 
						vmemmap_free(start, end);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					#ifdef CONFIG_MEMORY_HOTREMOVE
 | 
				
			||||||
static void free_map_bootmem(struct page *memmap, unsigned long nr_pages)
 | 
					static void free_map_bootmem(struct page *memmap, unsigned long nr_pages)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	unsigned long start = (unsigned long)memmap;
 | 
						unsigned long start = (unsigned long)memmap;
 | 
				
			||||||
| 
						 | 
					@ -627,6 +628,7 @@ static void free_map_bootmem(struct page *memmap, unsigned long nr_pages)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	vmemmap_free(start, end);
 | 
						vmemmap_free(start, end);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					#endif /* CONFIG_MEMORY_HOTREMOVE */
 | 
				
			||||||
#else
 | 
					#else
 | 
				
			||||||
static struct page *__kmalloc_section_memmap(unsigned long nr_pages)
 | 
					static struct page *__kmalloc_section_memmap(unsigned long nr_pages)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
| 
						 | 
					@ -664,6 +666,7 @@ static void __kfree_section_memmap(struct page *memmap, unsigned long nr_pages)
 | 
				
			||||||
			   get_order(sizeof(struct page) * nr_pages));
 | 
								   get_order(sizeof(struct page) * nr_pages));
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					#ifdef CONFIG_MEMORY_HOTREMOVE
 | 
				
			||||||
static void free_map_bootmem(struct page *memmap, unsigned long nr_pages)
 | 
					static void free_map_bootmem(struct page *memmap, unsigned long nr_pages)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	unsigned long maps_section_nr, removing_section_nr, i;
 | 
						unsigned long maps_section_nr, removing_section_nr, i;
 | 
				
			||||||
| 
						 | 
					@ -690,40 +693,9 @@ static void free_map_bootmem(struct page *memmap, unsigned long nr_pages)
 | 
				
			||||||
			put_page_bootmem(page);
 | 
								put_page_bootmem(page);
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					#endif /* CONFIG_MEMORY_HOTREMOVE */
 | 
				
			||||||
#endif /* CONFIG_SPARSEMEM_VMEMMAP */
 | 
					#endif /* CONFIG_SPARSEMEM_VMEMMAP */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static void free_section_usemap(struct page *memmap, unsigned long *usemap)
 | 
					 | 
				
			||||||
{
 | 
					 | 
				
			||||||
	struct page *usemap_page;
 | 
					 | 
				
			||||||
	unsigned long nr_pages;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	if (!usemap)
 | 
					 | 
				
			||||||
		return;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	usemap_page = virt_to_page(usemap);
 | 
					 | 
				
			||||||
	/*
 | 
					 | 
				
			||||||
	 * Check to see if allocation came from hot-plug-add
 | 
					 | 
				
			||||||
	 */
 | 
					 | 
				
			||||||
	if (PageSlab(usemap_page) || PageCompound(usemap_page)) {
 | 
					 | 
				
			||||||
		kfree(usemap);
 | 
					 | 
				
			||||||
		if (memmap)
 | 
					 | 
				
			||||||
			__kfree_section_memmap(memmap, PAGES_PER_SECTION);
 | 
					 | 
				
			||||||
		return;
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	/*
 | 
					 | 
				
			||||||
	 * The usemap came from bootmem. This is packed with other usemaps
 | 
					 | 
				
			||||||
	 * on the section which has pgdat at boot time. Just keep it as is now.
 | 
					 | 
				
			||||||
	 */
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	if (memmap) {
 | 
					 | 
				
			||||||
		nr_pages = PAGE_ALIGN(PAGES_PER_SECTION * sizeof(struct page))
 | 
					 | 
				
			||||||
			>> PAGE_SHIFT;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
		free_map_bootmem(memmap, nr_pages);
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
/*
 | 
					/*
 | 
				
			||||||
 * returns the number of sections whose mem_maps were properly
 | 
					 * returns the number of sections whose mem_maps were properly
 | 
				
			||||||
 * set.  If this is <=0, then that means that the passed-in
 | 
					 * set.  If this is <=0, then that means that the passed-in
 | 
				
			||||||
| 
						 | 
					@ -800,6 +772,39 @@ static inline void clear_hwpoisoned_pages(struct page *memmap, int nr_pages)
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					#ifdef CONFIG_MEMORY_HOTREMOVE
 | 
				
			||||||
 | 
					static void free_section_usemap(struct page *memmap, unsigned long *usemap)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						struct page *usemap_page;
 | 
				
			||||||
 | 
						unsigned long nr_pages;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						if (!usemap)
 | 
				
			||||||
 | 
							return;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						usemap_page = virt_to_page(usemap);
 | 
				
			||||||
 | 
						/*
 | 
				
			||||||
 | 
						 * Check to see if allocation came from hot-plug-add
 | 
				
			||||||
 | 
						 */
 | 
				
			||||||
 | 
						if (PageSlab(usemap_page) || PageCompound(usemap_page)) {
 | 
				
			||||||
 | 
							kfree(usemap);
 | 
				
			||||||
 | 
							if (memmap)
 | 
				
			||||||
 | 
								__kfree_section_memmap(memmap, PAGES_PER_SECTION);
 | 
				
			||||||
 | 
							return;
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						/*
 | 
				
			||||||
 | 
						 * The usemap came from bootmem. This is packed with other usemaps
 | 
				
			||||||
 | 
						 * on the section which has pgdat at boot time. Just keep it as is now.
 | 
				
			||||||
 | 
						 */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						if (memmap) {
 | 
				
			||||||
 | 
							nr_pages = PAGE_ALIGN(PAGES_PER_SECTION * sizeof(struct page))
 | 
				
			||||||
 | 
								>> PAGE_SHIFT;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							free_map_bootmem(memmap, nr_pages);
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void sparse_remove_one_section(struct zone *zone, struct mem_section *ms)
 | 
					void sparse_remove_one_section(struct zone *zone, struct mem_section *ms)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct page *memmap = NULL;
 | 
						struct page *memmap = NULL;
 | 
				
			||||||
| 
						 | 
					@ -819,4 +824,5 @@ void sparse_remove_one_section(struct zone *zone, struct mem_section *ms)
 | 
				
			||||||
	clear_hwpoisoned_pages(memmap, PAGES_PER_SECTION);
 | 
						clear_hwpoisoned_pages(memmap, PAGES_PER_SECTION);
 | 
				
			||||||
	free_section_usemap(memmap, usemap);
 | 
						free_section_usemap(memmap, usemap);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
#endif
 | 
					#endif /* CONFIG_MEMORY_HOTREMOVE */
 | 
				
			||||||
 | 
					#endif /* CONFIG_MEMORY_HOTPLUG */
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue