forked from mirrors/linux
		
	[PATCH] Fix sparsemem on Cell
Fix an oops experienced on the Cell architecture when init-time functions, early_*(), are called at runtime. It alters the call paths to make sure that the callers explicitly say whether the call is being made on behalf of a hotplug even, or happening at boot-time. It has been compile tested on ppc64, ia64, s390, i386 and x86_64. Acked-by: Arnd Bergmann <arndb@de.ibm.com> Signed-off-by: Dave Hansen <haveblue@us.ibm.com> Cc: Yasunori Goto <y-goto@jp.fujitsu.com> Acked-by: Andy Whitcroft <apw@shadowen.org> Cc: Christoph Lameter <clameter@engr.sgi.com> Cc: Martin Schwidefsky <schwidefsky@de.ibm.com> Acked-by: Heiko Carstens <heiko.carstens@de.ibm.com> Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org> Cc: Paul Mackerras <paulus@samba.org> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
This commit is contained in:
		
							parent
							
								
									47a4d5be7c
								
							
						
					
					
						commit
						a2f3aa0257
					
				
					 6 changed files with 34 additions and 16 deletions
				
			
		|  | @ -543,7 +543,8 @@ virtual_memmap_init (u64 start, u64 end, void *arg) | ||||||
| 
 | 
 | ||||||
| 	if (map_start < map_end) | 	if (map_start < map_end) | ||||||
| 		memmap_init_zone((unsigned long)(map_end - map_start), | 		memmap_init_zone((unsigned long)(map_end - map_start), | ||||||
| 				 args->nid, args->zone, page_to_pfn(map_start)); | 				 args->nid, args->zone, page_to_pfn(map_start), | ||||||
|  | 				 MEMMAP_EARLY); | ||||||
| 	return 0; | 	return 0; | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
|  | @ -552,7 +553,7 @@ memmap_init (unsigned long size, int nid, unsigned long zone, | ||||||
| 	     unsigned long start_pfn) | 	     unsigned long start_pfn) | ||||||
| { | { | ||||||
| 	if (!vmem_map) | 	if (!vmem_map) | ||||||
| 		memmap_init_zone(size, nid, zone, start_pfn); | 		memmap_init_zone(size, nid, zone, start_pfn, MEMMAP_EARLY); | ||||||
| 	else { | 	else { | ||||||
| 		struct page *start; | 		struct page *start; | ||||||
| 		struct memmap_init_callback_data args; | 		struct memmap_init_callback_data args; | ||||||
|  |  | ||||||
|  | @ -61,7 +61,8 @@ void memmap_init(unsigned long size, int nid, unsigned long zone, | ||||||
| 
 | 
 | ||||||
| 		if (map_start < map_end) | 		if (map_start < map_end) | ||||||
| 			memmap_init_zone((unsigned long)(map_end - map_start), | 			memmap_init_zone((unsigned long)(map_end - map_start), | ||||||
| 					 nid, zone, page_to_pfn(map_start)); | 					 nid, zone, page_to_pfn(map_start), | ||||||
|  | 					 MEMMAP_EARLY); | ||||||
| 	} | 	} | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
|  |  | ||||||
|  | @ -978,7 +978,8 @@ extern int early_pfn_to_nid(unsigned long pfn); | ||||||
| #endif /* CONFIG_HAVE_ARCH_EARLY_PFN_TO_NID */ | #endif /* CONFIG_HAVE_ARCH_EARLY_PFN_TO_NID */ | ||||||
| #endif /* CONFIG_ARCH_POPULATES_NODE_MAP */ | #endif /* CONFIG_ARCH_POPULATES_NODE_MAP */ | ||||||
| extern void set_dma_reserve(unsigned long new_dma_reserve); | extern void set_dma_reserve(unsigned long new_dma_reserve); | ||||||
| extern void memmap_init_zone(unsigned long, int, unsigned long, unsigned long); | extern void memmap_init_zone(unsigned long, int, unsigned long, | ||||||
|  | 				unsigned long, enum memmap_context); | ||||||
| extern void setup_per_zone_pages_min(void); | extern void setup_per_zone_pages_min(void); | ||||||
| extern void mem_init(void); | extern void mem_init(void); | ||||||
| extern void show_mem(void); | extern void show_mem(void); | ||||||
|  |  | ||||||
|  | @ -450,9 +450,13 @@ void build_all_zonelists(void); | ||||||
| void wakeup_kswapd(struct zone *zone, int order); | void wakeup_kswapd(struct zone *zone, int order); | ||||||
| int zone_watermark_ok(struct zone *z, int order, unsigned long mark, | int zone_watermark_ok(struct zone *z, int order, unsigned long mark, | ||||||
| 		int classzone_idx, int alloc_flags); | 		int classzone_idx, int alloc_flags); | ||||||
| 
 | enum memmap_context { | ||||||
|  | 	MEMMAP_EARLY, | ||||||
|  | 	MEMMAP_HOTPLUG, | ||||||
|  | }; | ||||||
| extern int init_currently_empty_zone(struct zone *zone, unsigned long start_pfn, | extern int init_currently_empty_zone(struct zone *zone, unsigned long start_pfn, | ||||||
| 				     unsigned long size); | 				     unsigned long size, | ||||||
|  | 				     enum memmap_context context); | ||||||
| 
 | 
 | ||||||
| #ifdef CONFIG_HAVE_MEMORY_PRESENT | #ifdef CONFIG_HAVE_MEMORY_PRESENT | ||||||
| void memory_present(int nid, unsigned long start, unsigned long end); | void memory_present(int nid, unsigned long start, unsigned long end); | ||||||
|  |  | ||||||
|  | @ -67,11 +67,13 @@ static int __add_zone(struct zone *zone, unsigned long phys_start_pfn) | ||||||
| 	zone_type = zone - pgdat->node_zones; | 	zone_type = zone - pgdat->node_zones; | ||||||
| 	if (!populated_zone(zone)) { | 	if (!populated_zone(zone)) { | ||||||
| 		int ret = 0; | 		int ret = 0; | ||||||
| 		ret = init_currently_empty_zone(zone, phys_start_pfn, nr_pages); | 		ret = init_currently_empty_zone(zone, phys_start_pfn, | ||||||
|  | 						nr_pages, MEMMAP_HOTPLUG); | ||||||
| 		if (ret < 0) | 		if (ret < 0) | ||||||
| 			return ret; | 			return ret; | ||||||
| 	} | 	} | ||||||
| 	memmap_init_zone(nr_pages, nid, zone_type, phys_start_pfn); | 	memmap_init_zone(nr_pages, nid, zone_type, | ||||||
|  | 			 phys_start_pfn, MEMMAP_HOTPLUG); | ||||||
| 	return 0; | 	return 0; | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
|  |  | ||||||
|  | @ -1956,17 +1956,24 @@ static inline unsigned long wait_table_bits(unsigned long size) | ||||||
|  * done. Non-atomic initialization, single-pass. |  * done. Non-atomic initialization, single-pass. | ||||||
|  */ |  */ | ||||||
| void __meminit memmap_init_zone(unsigned long size, int nid, unsigned long zone, | void __meminit memmap_init_zone(unsigned long size, int nid, unsigned long zone, | ||||||
| 		unsigned long start_pfn) | 		unsigned long start_pfn, enum memmap_context context) | ||||||
| { | { | ||||||
| 	struct page *page; | 	struct page *page; | ||||||
| 	unsigned long end_pfn = start_pfn + size; | 	unsigned long end_pfn = start_pfn + size; | ||||||
| 	unsigned long pfn; | 	unsigned long pfn; | ||||||
| 
 | 
 | ||||||
| 	for (pfn = start_pfn; pfn < end_pfn; pfn++) { | 	for (pfn = start_pfn; pfn < end_pfn; pfn++) { | ||||||
|  | 		/*
 | ||||||
|  | 		 * There can be holes in boot-time mem_map[]s | ||||||
|  | 		 * handed to this function.  They do not | ||||||
|  | 		 * exist on hotplugged memory. | ||||||
|  | 		 */ | ||||||
|  | 		if (context == MEMMAP_EARLY) { | ||||||
| 			if (!early_pfn_valid(pfn)) | 			if (!early_pfn_valid(pfn)) | ||||||
| 				continue; | 				continue; | ||||||
| 			if (!early_pfn_in_nid(pfn, nid)) | 			if (!early_pfn_in_nid(pfn, nid)) | ||||||
| 				continue; | 				continue; | ||||||
|  | 		} | ||||||
| 		page = pfn_to_page(pfn); | 		page = pfn_to_page(pfn); | ||||||
| 		set_page_links(page, zone, nid, pfn); | 		set_page_links(page, zone, nid, pfn); | ||||||
| 		init_page_count(page); | 		init_page_count(page); | ||||||
|  | @ -1993,7 +2000,7 @@ void zone_init_free_lists(struct pglist_data *pgdat, struct zone *zone, | ||||||
| 
 | 
 | ||||||
| #ifndef __HAVE_ARCH_MEMMAP_INIT | #ifndef __HAVE_ARCH_MEMMAP_INIT | ||||||
| #define memmap_init(size, nid, zone, start_pfn) \ | #define memmap_init(size, nid, zone, start_pfn) \ | ||||||
| 	memmap_init_zone((size), (nid), (zone), (start_pfn)) | 	memmap_init_zone((size), (nid), (zone), (start_pfn), MEMMAP_EARLY) | ||||||
| #endif | #endif | ||||||
| 
 | 
 | ||||||
| static int __cpuinit zone_batchsize(struct zone *zone) | static int __cpuinit zone_batchsize(struct zone *zone) | ||||||
|  | @ -2239,7 +2246,8 @@ static __meminit void zone_pcp_init(struct zone *zone) | ||||||
| 
 | 
 | ||||||
| __meminit int init_currently_empty_zone(struct zone *zone, | __meminit int init_currently_empty_zone(struct zone *zone, | ||||||
| 					unsigned long zone_start_pfn, | 					unsigned long zone_start_pfn, | ||||||
| 					unsigned long size) | 					unsigned long size, | ||||||
|  | 					enum memmap_context context) | ||||||
| { | { | ||||||
| 	struct pglist_data *pgdat = zone->zone_pgdat; | 	struct pglist_data *pgdat = zone->zone_pgdat; | ||||||
| 	int ret; | 	int ret; | ||||||
|  | @ -2683,7 +2691,8 @@ static void __meminit free_area_init_core(struct pglist_data *pgdat, | ||||||
| 		if (!size) | 		if (!size) | ||||||
| 			continue; | 			continue; | ||||||
| 
 | 
 | ||||||
| 		ret = init_currently_empty_zone(zone, zone_start_pfn, size); | 		ret = init_currently_empty_zone(zone, zone_start_pfn, | ||||||
|  | 						size, MEMMAP_EARLY); | ||||||
| 		BUG_ON(ret); | 		BUG_ON(ret); | ||||||
| 		zone_start_pfn += size; | 		zone_start_pfn += size; | ||||||
| 	} | 	} | ||||||
|  |  | ||||||
		Loading…
	
		Reference in a new issue
	
	 Dave Hansen
						Dave Hansen