mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 02:30:34 +02:00 
			
		
		
		
	swsusp should not use memcpy for snapshotting memory, because on some architectures memcpy may increase preempt_count (i386 does this when CONFIG_X86_USE_3DNOW is set). Then, as a result, wrong value of preempt_count is stored in the image. Replace memcpy in copy_data_pages with an open-coded loop. Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl> Acked-by: Pavel Machek <pavel@ucw.cz> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
		
			
				
	
	
		
			868 lines
		
	
	
	
		
			21 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			868 lines
		
	
	
	
		
			21 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * linux/kernel/power/snapshot.c
 | 
						|
 *
 | 
						|
 * This file provide system snapshot/restore functionality.
 | 
						|
 *
 | 
						|
 * Copyright (C) 1998-2005 Pavel Machek <pavel@suse.cz>
 | 
						|
 *
 | 
						|
 * This file is released under the GPLv2, and is based on swsusp.c.
 | 
						|
 *
 | 
						|
 */
 | 
						|
 | 
						|
 | 
						|
#include <linux/version.h>
 | 
						|
#include <linux/module.h>
 | 
						|
#include <linux/mm.h>
 | 
						|
#include <linux/suspend.h>
 | 
						|
#include <linux/smp_lock.h>
 | 
						|
#include <linux/delay.h>
 | 
						|
#include <linux/bitops.h>
 | 
						|
#include <linux/spinlock.h>
 | 
						|
#include <linux/kernel.h>
 | 
						|
#include <linux/pm.h>
 | 
						|
#include <linux/device.h>
 | 
						|
#include <linux/bootmem.h>
 | 
						|
#include <linux/syscalls.h>
 | 
						|
#include <linux/console.h>
 | 
						|
#include <linux/highmem.h>
 | 
						|
 | 
						|
#include <asm/uaccess.h>
 | 
						|
#include <asm/mmu_context.h>
 | 
						|
#include <asm/pgtable.h>
 | 
						|
#include <asm/tlbflush.h>
 | 
						|
#include <asm/io.h>
 | 
						|
 | 
						|
#include "power.h"
 | 
						|
 | 
						|
struct pbe *pagedir_nosave;
 | 
						|
static unsigned int nr_copy_pages;
 | 
						|
static unsigned int nr_meta_pages;
 | 
						|
static unsigned long *buffer;
 | 
						|
 | 
						|
#ifdef CONFIG_HIGHMEM
 | 
						|
unsigned int count_highmem_pages(void)
 | 
						|
{
 | 
						|
	struct zone *zone;
 | 
						|
	unsigned long zone_pfn;
 | 
						|
	unsigned int n = 0;
 | 
						|
 | 
						|
	for_each_zone (zone)
 | 
						|
		if (is_highmem(zone)) {
 | 
						|
			mark_free_pages(zone);
 | 
						|
			for (zone_pfn = 0; zone_pfn < zone->spanned_pages; zone_pfn++) {
 | 
						|
				struct page *page;
 | 
						|
				unsigned long pfn = zone_pfn + zone->zone_start_pfn;
 | 
						|
				if (!pfn_valid(pfn))
 | 
						|
					continue;
 | 
						|
				page = pfn_to_page(pfn);
 | 
						|
				if (PageReserved(page))
 | 
						|
					continue;
 | 
						|
				if (PageNosaveFree(page))
 | 
						|
					continue;
 | 
						|
				n++;
 | 
						|
			}
 | 
						|
		}
 | 
						|
	return n;
 | 
						|
}
 | 
						|
 | 
						|
struct highmem_page {
 | 
						|
	char *data;
 | 
						|
	struct page *page;
 | 
						|
	struct highmem_page *next;
 | 
						|
};
 | 
						|
 | 
						|
static struct highmem_page *highmem_copy;
 | 
						|
 | 
						|
static int save_highmem_zone(struct zone *zone)
 | 
						|
{
 | 
						|
	unsigned long zone_pfn;
 | 
						|
	mark_free_pages(zone);
 | 
						|
	for (zone_pfn = 0; zone_pfn < zone->spanned_pages; ++zone_pfn) {
 | 
						|
		struct page *page;
 | 
						|
		struct highmem_page *save;
 | 
						|
		void *kaddr;
 | 
						|
		unsigned long pfn = zone_pfn + zone->zone_start_pfn;
 | 
						|
 | 
						|
		if (!(pfn%10000))
 | 
						|
			printk(".");
 | 
						|
		if (!pfn_valid(pfn))
 | 
						|
			continue;
 | 
						|
		page = pfn_to_page(pfn);
 | 
						|
		/*
 | 
						|
		 * This condition results from rvmalloc() sans vmalloc_32()
 | 
						|
		 * and architectural memory reservations. This should be
 | 
						|
		 * corrected eventually when the cases giving rise to this
 | 
						|
		 * are better understood.
 | 
						|
		 */
 | 
						|
		if (PageReserved(page))
 | 
						|
			continue;
 | 
						|
		BUG_ON(PageNosave(page));
 | 
						|
		if (PageNosaveFree(page))
 | 
						|
			continue;
 | 
						|
		save = kmalloc(sizeof(struct highmem_page), GFP_ATOMIC);
 | 
						|
		if (!save)
 | 
						|
			return -ENOMEM;
 | 
						|
		save->next = highmem_copy;
 | 
						|
		save->page = page;
 | 
						|
		save->data = (void *) get_zeroed_page(GFP_ATOMIC);
 | 
						|
		if (!save->data) {
 | 
						|
			kfree(save);
 | 
						|
			return -ENOMEM;
 | 
						|
		}
 | 
						|
		kaddr = kmap_atomic(page, KM_USER0);
 | 
						|
		memcpy(save->data, kaddr, PAGE_SIZE);
 | 
						|
		kunmap_atomic(kaddr, KM_USER0);
 | 
						|
		highmem_copy = save;
 | 
						|
	}
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
int save_highmem(void)
 | 
						|
{
 | 
						|
	struct zone *zone;
 | 
						|
	int res = 0;
 | 
						|
 | 
						|
	pr_debug("swsusp: Saving Highmem");
 | 
						|
	drain_local_pages();
 | 
						|
	for_each_zone (zone) {
 | 
						|
		if (is_highmem(zone))
 | 
						|
			res = save_highmem_zone(zone);
 | 
						|
		if (res)
 | 
						|
			return res;
 | 
						|
	}
 | 
						|
	printk("\n");
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
int restore_highmem(void)
 | 
						|
{
 | 
						|
	printk("swsusp: Restoring Highmem\n");
 | 
						|
	while (highmem_copy) {
 | 
						|
		struct highmem_page *save = highmem_copy;
 | 
						|
		void *kaddr;
 | 
						|
		highmem_copy = save->next;
 | 
						|
 | 
						|
		kaddr = kmap_atomic(save->page, KM_USER0);
 | 
						|
		memcpy(kaddr, save->data, PAGE_SIZE);
 | 
						|
		kunmap_atomic(kaddr, KM_USER0);
 | 
						|
		free_page((long) save->data);
 | 
						|
		kfree(save);
 | 
						|
	}
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
#else
 | 
						|
static inline unsigned int count_highmem_pages(void) {return 0;}
 | 
						|
static inline int save_highmem(void) {return 0;}
 | 
						|
static inline int restore_highmem(void) {return 0;}
 | 
						|
#endif
 | 
						|
 | 
						|
static int pfn_is_nosave(unsigned long pfn)
 | 
						|
{
 | 
						|
	unsigned long nosave_begin_pfn = __pa(&__nosave_begin) >> PAGE_SHIFT;
 | 
						|
	unsigned long nosave_end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT;
 | 
						|
	return (pfn >= nosave_begin_pfn) && (pfn < nosave_end_pfn);
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
 *	saveable - Determine whether a page should be cloned or not.
 | 
						|
 *	@pfn:	The page
 | 
						|
 *
 | 
						|
 *	We save a page if it's Reserved, and not in the range of pages
 | 
						|
 *	statically defined as 'unsaveable', or if it isn't reserved, and
 | 
						|
 *	isn't part of a free chunk of pages.
 | 
						|
 */
 | 
						|
 | 
						|
static int saveable(struct zone *zone, unsigned long *zone_pfn)
 | 
						|
{
 | 
						|
	unsigned long pfn = *zone_pfn + zone->zone_start_pfn;
 | 
						|
	struct page *page;
 | 
						|
 | 
						|
	if (!pfn_valid(pfn))
 | 
						|
		return 0;
 | 
						|
 | 
						|
	page = pfn_to_page(pfn);
 | 
						|
	BUG_ON(PageReserved(page) && PageNosave(page));
 | 
						|
	if (PageNosave(page))
 | 
						|
		return 0;
 | 
						|
	if (PageReserved(page) && pfn_is_nosave(pfn))
 | 
						|
		return 0;
 | 
						|
	if (PageNosaveFree(page))
 | 
						|
		return 0;
 | 
						|
 | 
						|
	return 1;
 | 
						|
}
 | 
						|
 | 
						|
unsigned int count_data_pages(void)
 | 
						|
{
 | 
						|
	struct zone *zone;
 | 
						|
	unsigned long zone_pfn;
 | 
						|
	unsigned int n = 0;
 | 
						|
 | 
						|
	for_each_zone (zone) {
 | 
						|
		if (is_highmem(zone))
 | 
						|
			continue;
 | 
						|
		mark_free_pages(zone);
 | 
						|
		for (zone_pfn = 0; zone_pfn < zone->spanned_pages; ++zone_pfn)
 | 
						|
			n += saveable(zone, &zone_pfn);
 | 
						|
	}
 | 
						|
	return n;
 | 
						|
}
 | 
						|
 | 
						|
static void copy_data_pages(struct pbe *pblist)
 | 
						|
{
 | 
						|
	struct zone *zone;
 | 
						|
	unsigned long zone_pfn;
 | 
						|
	struct pbe *pbe, *p;
 | 
						|
 | 
						|
	pbe = pblist;
 | 
						|
	for_each_zone (zone) {
 | 
						|
		if (is_highmem(zone))
 | 
						|
			continue;
 | 
						|
		mark_free_pages(zone);
 | 
						|
		/* This is necessary for swsusp_free() */
 | 
						|
		for_each_pb_page (p, pblist)
 | 
						|
			SetPageNosaveFree(virt_to_page(p));
 | 
						|
		for_each_pbe (p, pblist)
 | 
						|
			SetPageNosaveFree(virt_to_page(p->address));
 | 
						|
		for (zone_pfn = 0; zone_pfn < zone->spanned_pages; ++zone_pfn) {
 | 
						|
			if (saveable(zone, &zone_pfn)) {
 | 
						|
				struct page *page;
 | 
						|
				long *src, *dst;
 | 
						|
				int n;
 | 
						|
 | 
						|
				page = pfn_to_page(zone_pfn + zone->zone_start_pfn);
 | 
						|
				BUG_ON(!pbe);
 | 
						|
				pbe->orig_address = (unsigned long)page_address(page);
 | 
						|
				/* copy_page and memcpy are not usable for copying task structs. */
 | 
						|
				dst = (long *)pbe->address;
 | 
						|
				src = (long *)pbe->orig_address;
 | 
						|
				for (n = PAGE_SIZE / sizeof(long); n; n--)
 | 
						|
					*dst++ = *src++;
 | 
						|
				pbe = pbe->next;
 | 
						|
			}
 | 
						|
		}
 | 
						|
	}
 | 
						|
	BUG_ON(pbe);
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
/**
 | 
						|
 *	free_pagedir - free pages allocated with alloc_pagedir()
 | 
						|
 */
 | 
						|
 | 
						|
static void free_pagedir(struct pbe *pblist, int clear_nosave_free)
 | 
						|
{
 | 
						|
	struct pbe *pbe;
 | 
						|
 | 
						|
	while (pblist) {
 | 
						|
		pbe = (pblist + PB_PAGE_SKIP)->next;
 | 
						|
		ClearPageNosave(virt_to_page(pblist));
 | 
						|
		if (clear_nosave_free)
 | 
						|
			ClearPageNosaveFree(virt_to_page(pblist));
 | 
						|
		free_page((unsigned long)pblist);
 | 
						|
		pblist = pbe;
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
 *	fill_pb_page - Create a list of PBEs on a given memory page
 | 
						|
 */
 | 
						|
 | 
						|
static inline void fill_pb_page(struct pbe *pbpage)
 | 
						|
{
 | 
						|
	struct pbe *p;
 | 
						|
 | 
						|
	p = pbpage;
 | 
						|
	pbpage += PB_PAGE_SKIP;
 | 
						|
	do
 | 
						|
		p->next = p + 1;
 | 
						|
	while (++p < pbpage);
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
 *	create_pbe_list - Create a list of PBEs on top of a given chain
 | 
						|
 *	of memory pages allocated with alloc_pagedir()
 | 
						|
 */
 | 
						|
 | 
						|
static inline void create_pbe_list(struct pbe *pblist, unsigned int nr_pages)
 | 
						|
{
 | 
						|
	struct pbe *pbpage, *p;
 | 
						|
	unsigned int num = PBES_PER_PAGE;
 | 
						|
 | 
						|
	for_each_pb_page (pbpage, pblist) {
 | 
						|
		if (num >= nr_pages)
 | 
						|
			break;
 | 
						|
 | 
						|
		fill_pb_page(pbpage);
 | 
						|
		num += PBES_PER_PAGE;
 | 
						|
	}
 | 
						|
	if (pbpage) {
 | 
						|
		for (num -= PBES_PER_PAGE - 1, p = pbpage; num < nr_pages; p++, num++)
 | 
						|
			p->next = p + 1;
 | 
						|
		p->next = NULL;
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static unsigned int unsafe_pages;
 | 
						|
 | 
						|
/**
 | 
						|
 *	@safe_needed - on resume, for storing the PBE list and the image,
 | 
						|
 *	we can only use memory pages that do not conflict with the pages
 | 
						|
 *	used before suspend.
 | 
						|
 *
 | 
						|
 *	The unsafe pages are marked with the PG_nosave_free flag
 | 
						|
 *	and we count them using unsafe_pages
 | 
						|
 */
 | 
						|
 | 
						|
static inline void *alloc_image_page(gfp_t gfp_mask, int safe_needed)
 | 
						|
{
 | 
						|
	void *res;
 | 
						|
 | 
						|
	res = (void *)get_zeroed_page(gfp_mask);
 | 
						|
	if (safe_needed)
 | 
						|
		while (res && PageNosaveFree(virt_to_page(res))) {
 | 
						|
			/* The page is unsafe, mark it for swsusp_free() */
 | 
						|
			SetPageNosave(virt_to_page(res));
 | 
						|
			unsafe_pages++;
 | 
						|
			res = (void *)get_zeroed_page(gfp_mask);
 | 
						|
		}
 | 
						|
	if (res) {
 | 
						|
		SetPageNosave(virt_to_page(res));
 | 
						|
		SetPageNosaveFree(virt_to_page(res));
 | 
						|
	}
 | 
						|
	return res;
 | 
						|
}
 | 
						|
 | 
						|
unsigned long get_safe_page(gfp_t gfp_mask)
 | 
						|
{
 | 
						|
	return (unsigned long)alloc_image_page(gfp_mask, 1);
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
 *	alloc_pagedir - Allocate the page directory.
 | 
						|
 *
 | 
						|
 *	First, determine exactly how many pages we need and
 | 
						|
 *	allocate them.
 | 
						|
 *
 | 
						|
 *	We arrange the pages in a chain: each page is an array of PBES_PER_PAGE
 | 
						|
 *	struct pbe elements (pbes) and the last element in the page points
 | 
						|
 *	to the next page.
 | 
						|
 *
 | 
						|
 *	On each page we set up a list of struct_pbe elements.
 | 
						|
 */
 | 
						|
 | 
						|
static struct pbe *alloc_pagedir(unsigned int nr_pages, gfp_t gfp_mask,
 | 
						|
				 int safe_needed)
 | 
						|
{
 | 
						|
	unsigned int num;
 | 
						|
	struct pbe *pblist, *pbe;
 | 
						|
 | 
						|
	if (!nr_pages)
 | 
						|
		return NULL;
 | 
						|
 | 
						|
	pblist = alloc_image_page(gfp_mask, safe_needed);
 | 
						|
	/* FIXME: rewrite this ugly loop */
 | 
						|
	for (pbe = pblist, num = PBES_PER_PAGE; pbe && num < nr_pages;
 | 
						|
        		pbe = pbe->next, num += PBES_PER_PAGE) {
 | 
						|
		pbe += PB_PAGE_SKIP;
 | 
						|
		pbe->next = alloc_image_page(gfp_mask, safe_needed);
 | 
						|
	}
 | 
						|
	if (!pbe) { /* get_zeroed_page() failed */
 | 
						|
		free_pagedir(pblist, 1);
 | 
						|
		pblist = NULL;
 | 
						|
        } else
 | 
						|
		create_pbe_list(pblist, nr_pages);
 | 
						|
	return pblist;
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
 * Free pages we allocated for suspend. Suspend pages are alocated
 | 
						|
 * before atomic copy, so we need to free them after resume.
 | 
						|
 */
 | 
						|
 | 
						|
void swsusp_free(void)
 | 
						|
{
 | 
						|
	struct zone *zone;
 | 
						|
	unsigned long zone_pfn;
 | 
						|
 | 
						|
	for_each_zone(zone) {
 | 
						|
		for (zone_pfn = 0; zone_pfn < zone->spanned_pages; ++zone_pfn)
 | 
						|
			if (pfn_valid(zone_pfn + zone->zone_start_pfn)) {
 | 
						|
				struct page *page;
 | 
						|
				page = pfn_to_page(zone_pfn + zone->zone_start_pfn);
 | 
						|
				if (PageNosave(page) && PageNosaveFree(page)) {
 | 
						|
					ClearPageNosave(page);
 | 
						|
					ClearPageNosaveFree(page);
 | 
						|
					free_page((long) page_address(page));
 | 
						|
				}
 | 
						|
			}
 | 
						|
	}
 | 
						|
	nr_copy_pages = 0;
 | 
						|
	nr_meta_pages = 0;
 | 
						|
	pagedir_nosave = NULL;
 | 
						|
	buffer = NULL;
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
/**
 | 
						|
 *	enough_free_mem - Make sure we enough free memory to snapshot.
 | 
						|
 *
 | 
						|
 *	Returns TRUE or FALSE after checking the number of available
 | 
						|
 *	free pages.
 | 
						|
 */
 | 
						|
 | 
						|
static int enough_free_mem(unsigned int nr_pages)
 | 
						|
{
 | 
						|
	struct zone *zone;
 | 
						|
	unsigned int n = 0;
 | 
						|
 | 
						|
	for_each_zone (zone)
 | 
						|
		if (!is_highmem(zone))
 | 
						|
			n += zone->free_pages;
 | 
						|
	pr_debug("swsusp: available memory: %u pages\n", n);
 | 
						|
	return n > (nr_pages + PAGES_FOR_IO +
 | 
						|
		(nr_pages + PBES_PER_PAGE - 1) / PBES_PER_PAGE);
 | 
						|
}
 | 
						|
 | 
						|
static int alloc_data_pages(struct pbe *pblist, gfp_t gfp_mask, int safe_needed)
 | 
						|
{
 | 
						|
	struct pbe *p;
 | 
						|
 | 
						|
	for_each_pbe (p, pblist) {
 | 
						|
		p->address = (unsigned long)alloc_image_page(gfp_mask, safe_needed);
 | 
						|
		if (!p->address)
 | 
						|
			return -ENOMEM;
 | 
						|
	}
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static struct pbe *swsusp_alloc(unsigned int nr_pages)
 | 
						|
{
 | 
						|
	struct pbe *pblist;
 | 
						|
 | 
						|
	if (!(pblist = alloc_pagedir(nr_pages, GFP_ATOMIC | __GFP_COLD, 0))) {
 | 
						|
		printk(KERN_ERR "suspend: Allocating pagedir failed.\n");
 | 
						|
		return NULL;
 | 
						|
	}
 | 
						|
 | 
						|
	if (alloc_data_pages(pblist, GFP_ATOMIC | __GFP_COLD, 0)) {
 | 
						|
		printk(KERN_ERR "suspend: Allocating image pages failed.\n");
 | 
						|
		swsusp_free();
 | 
						|
		return NULL;
 | 
						|
	}
 | 
						|
 | 
						|
	return pblist;
 | 
						|
}
 | 
						|
 | 
						|
asmlinkage int swsusp_save(void)
 | 
						|
{
 | 
						|
	unsigned int nr_pages;
 | 
						|
 | 
						|
	pr_debug("swsusp: critical section: \n");
 | 
						|
 | 
						|
	drain_local_pages();
 | 
						|
	nr_pages = count_data_pages();
 | 
						|
	printk("swsusp: Need to copy %u pages\n", nr_pages);
 | 
						|
 | 
						|
	pr_debug("swsusp: pages needed: %u + %lu + %u, free: %u\n",
 | 
						|
		 nr_pages,
 | 
						|
		 (nr_pages + PBES_PER_PAGE - 1) / PBES_PER_PAGE,
 | 
						|
		 PAGES_FOR_IO, nr_free_pages());
 | 
						|
 | 
						|
	if (!enough_free_mem(nr_pages)) {
 | 
						|
		printk(KERN_ERR "swsusp: Not enough free memory\n");
 | 
						|
		return -ENOMEM;
 | 
						|
	}
 | 
						|
 | 
						|
	pagedir_nosave = swsusp_alloc(nr_pages);
 | 
						|
	if (!pagedir_nosave)
 | 
						|
		return -ENOMEM;
 | 
						|
 | 
						|
	/* During allocating of suspend pagedir, new cold pages may appear.
 | 
						|
	 * Kill them.
 | 
						|
	 */
 | 
						|
	drain_local_pages();
 | 
						|
	copy_data_pages(pagedir_nosave);
 | 
						|
 | 
						|
	/*
 | 
						|
	 * End of critical section. From now on, we can write to memory,
 | 
						|
	 * but we should not touch disk. This specially means we must _not_
 | 
						|
	 * touch swap space! Except we must write out our image of course.
 | 
						|
	 */
 | 
						|
 | 
						|
	nr_copy_pages = nr_pages;
 | 
						|
	nr_meta_pages = (nr_pages * sizeof(long) + PAGE_SIZE - 1) >> PAGE_SHIFT;
 | 
						|
 | 
						|
	printk("swsusp: critical section/: done (%d pages copied)\n", nr_pages);
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static void init_header(struct swsusp_info *info)
 | 
						|
{
 | 
						|
	memset(info, 0, sizeof(struct swsusp_info));
 | 
						|
	info->version_code = LINUX_VERSION_CODE;
 | 
						|
	info->num_physpages = num_physpages;
 | 
						|
	memcpy(&info->uts, &system_utsname, sizeof(system_utsname));
 | 
						|
	info->cpus = num_online_cpus();
 | 
						|
	info->image_pages = nr_copy_pages;
 | 
						|
	info->pages = nr_copy_pages + nr_meta_pages + 1;
 | 
						|
	info->size = info->pages;
 | 
						|
	info->size <<= PAGE_SHIFT;
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
 *	pack_orig_addresses - the .orig_address fields of the PBEs from the
 | 
						|
 *	list starting at @pbe are stored in the array @buf[] (1 page)
 | 
						|
 */
 | 
						|
 | 
						|
static inline struct pbe *pack_orig_addresses(unsigned long *buf, struct pbe *pbe)
 | 
						|
{
 | 
						|
	int j;
 | 
						|
 | 
						|
	for (j = 0; j < PAGE_SIZE / sizeof(long) && pbe; j++) {
 | 
						|
		buf[j] = pbe->orig_address;
 | 
						|
		pbe = pbe->next;
 | 
						|
	}
 | 
						|
	if (!pbe)
 | 
						|
		for (; j < PAGE_SIZE / sizeof(long); j++)
 | 
						|
			buf[j] = 0;
 | 
						|
	return pbe;
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
 *	snapshot_read_next - used for reading the system memory snapshot.
 | 
						|
 *
 | 
						|
 *	On the first call to it @handle should point to a zeroed
 | 
						|
 *	snapshot_handle structure.  The structure gets updated and a pointer
 | 
						|
 *	to it should be passed to this function every next time.
 | 
						|
 *
 | 
						|
 *	The @count parameter should contain the number of bytes the caller
 | 
						|
 *	wants to read from the snapshot.  It must not be zero.
 | 
						|
 *
 | 
						|
 *	On success the function returns a positive number.  Then, the caller
 | 
						|
 *	is allowed to read up to the returned number of bytes from the memory
 | 
						|
 *	location computed by the data_of() macro.  The number returned
 | 
						|
 *	may be smaller than @count, but this only happens if the read would
 | 
						|
 *	cross a page boundary otherwise.
 | 
						|
 *
 | 
						|
 *	The function returns 0 to indicate the end of data stream condition,
 | 
						|
 *	and a negative number is returned on error.  In such cases the
 | 
						|
 *	structure pointed to by @handle is not updated and should not be used
 | 
						|
 *	any more.
 | 
						|
 */
 | 
						|
 | 
						|
int snapshot_read_next(struct snapshot_handle *handle, size_t count)
 | 
						|
{
 | 
						|
	if (handle->page > nr_meta_pages + nr_copy_pages)
 | 
						|
		return 0;
 | 
						|
	if (!buffer) {
 | 
						|
		/* This makes the buffer be freed by swsusp_free() */
 | 
						|
		buffer = alloc_image_page(GFP_ATOMIC, 0);
 | 
						|
		if (!buffer)
 | 
						|
			return -ENOMEM;
 | 
						|
	}
 | 
						|
	if (!handle->offset) {
 | 
						|
		init_header((struct swsusp_info *)buffer);
 | 
						|
		handle->buffer = buffer;
 | 
						|
		handle->pbe = pagedir_nosave;
 | 
						|
	}
 | 
						|
	if (handle->prev < handle->page) {
 | 
						|
		if (handle->page <= nr_meta_pages) {
 | 
						|
			handle->pbe = pack_orig_addresses(buffer, handle->pbe);
 | 
						|
			if (!handle->pbe)
 | 
						|
				handle->pbe = pagedir_nosave;
 | 
						|
		} else {
 | 
						|
			handle->buffer = (void *)handle->pbe->address;
 | 
						|
			handle->pbe = handle->pbe->next;
 | 
						|
		}
 | 
						|
		handle->prev = handle->page;
 | 
						|
	}
 | 
						|
	handle->buf_offset = handle->page_offset;
 | 
						|
	if (handle->page_offset + count >= PAGE_SIZE) {
 | 
						|
		count = PAGE_SIZE - handle->page_offset;
 | 
						|
		handle->page_offset = 0;
 | 
						|
		handle->page++;
 | 
						|
	} else {
 | 
						|
		handle->page_offset += count;
 | 
						|
	}
 | 
						|
	handle->offset += count;
 | 
						|
	return count;
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
 *	mark_unsafe_pages - mark the pages that cannot be used for storing
 | 
						|
 *	the image during resume, because they conflict with the pages that
 | 
						|
 *	had been used before suspend
 | 
						|
 */
 | 
						|
 | 
						|
static int mark_unsafe_pages(struct pbe *pblist)
 | 
						|
{
 | 
						|
	struct zone *zone;
 | 
						|
	unsigned long zone_pfn;
 | 
						|
	struct pbe *p;
 | 
						|
 | 
						|
	if (!pblist) /* a sanity check */
 | 
						|
		return -EINVAL;
 | 
						|
 | 
						|
	/* Clear page flags */
 | 
						|
	for_each_zone (zone) {
 | 
						|
		for (zone_pfn = 0; zone_pfn < zone->spanned_pages; ++zone_pfn)
 | 
						|
			if (pfn_valid(zone_pfn + zone->zone_start_pfn))
 | 
						|
				ClearPageNosaveFree(pfn_to_page(zone_pfn +
 | 
						|
					zone->zone_start_pfn));
 | 
						|
	}
 | 
						|
 | 
						|
	/* Mark orig addresses */
 | 
						|
	for_each_pbe (p, pblist) {
 | 
						|
		if (virt_addr_valid(p->orig_address))
 | 
						|
			SetPageNosaveFree(virt_to_page(p->orig_address));
 | 
						|
		else
 | 
						|
			return -EFAULT;
 | 
						|
	}
 | 
						|
 | 
						|
	unsafe_pages = 0;
 | 
						|
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
static void copy_page_backup_list(struct pbe *dst, struct pbe *src)
 | 
						|
{
 | 
						|
	/* We assume both lists contain the same number of elements */
 | 
						|
	while (src) {
 | 
						|
		dst->orig_address = src->orig_address;
 | 
						|
		dst = dst->next;
 | 
						|
		src = src->next;
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
static int check_header(struct swsusp_info *info)
 | 
						|
{
 | 
						|
	char *reason = NULL;
 | 
						|
 | 
						|
	if (info->version_code != LINUX_VERSION_CODE)
 | 
						|
		reason = "kernel version";
 | 
						|
	if (info->num_physpages != num_physpages)
 | 
						|
		reason = "memory size";
 | 
						|
	if (strcmp(info->uts.sysname,system_utsname.sysname))
 | 
						|
		reason = "system type";
 | 
						|
	if (strcmp(info->uts.release,system_utsname.release))
 | 
						|
		reason = "kernel release";
 | 
						|
	if (strcmp(info->uts.version,system_utsname.version))
 | 
						|
		reason = "version";
 | 
						|
	if (strcmp(info->uts.machine,system_utsname.machine))
 | 
						|
		reason = "machine";
 | 
						|
	if (reason) {
 | 
						|
		printk(KERN_ERR "swsusp: Resume mismatch: %s\n", reason);
 | 
						|
		return -EPERM;
 | 
						|
	}
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
 *	load header - check the image header and copy data from it
 | 
						|
 */
 | 
						|
 | 
						|
static int load_header(struct snapshot_handle *handle,
 | 
						|
                              struct swsusp_info *info)
 | 
						|
{
 | 
						|
	int error;
 | 
						|
	struct pbe *pblist;
 | 
						|
 | 
						|
	error = check_header(info);
 | 
						|
	if (!error) {
 | 
						|
		pblist = alloc_pagedir(info->image_pages, GFP_ATOMIC, 0);
 | 
						|
		if (!pblist)
 | 
						|
			return -ENOMEM;
 | 
						|
		pagedir_nosave = pblist;
 | 
						|
		handle->pbe = pblist;
 | 
						|
		nr_copy_pages = info->image_pages;
 | 
						|
		nr_meta_pages = info->pages - info->image_pages - 1;
 | 
						|
	}
 | 
						|
	return error;
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
 *	unpack_orig_addresses - copy the elements of @buf[] (1 page) to
 | 
						|
 *	the PBEs in the list starting at @pbe
 | 
						|
 */
 | 
						|
 | 
						|
static inline struct pbe *unpack_orig_addresses(unsigned long *buf,
 | 
						|
                                                struct pbe *pbe)
 | 
						|
{
 | 
						|
	int j;
 | 
						|
 | 
						|
	for (j = 0; j < PAGE_SIZE / sizeof(long) && pbe; j++) {
 | 
						|
		pbe->orig_address = buf[j];
 | 
						|
		pbe = pbe->next;
 | 
						|
	}
 | 
						|
	return pbe;
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
 *	prepare_image - use metadata contained in the PBE list
 | 
						|
 *	pointed to by pagedir_nosave to mark the pages that will
 | 
						|
 *	be overwritten in the process of restoring the system
 | 
						|
 *	memory state from the image ("unsafe" pages) and allocate
 | 
						|
 *	memory for the image
 | 
						|
 *
 | 
						|
 *	The idea is to allocate the PBE list first and then
 | 
						|
 *	allocate as many pages as it's needed for the image data,
 | 
						|
 *	but not to assign these pages to the PBEs initially.
 | 
						|
 *	Instead, we just mark them as allocated and create a list
 | 
						|
 *	of "safe" which will be used later
 | 
						|
 */
 | 
						|
 | 
						|
struct safe_page {
 | 
						|
	struct safe_page *next;
 | 
						|
	char padding[PAGE_SIZE - sizeof(void *)];
 | 
						|
};
 | 
						|
 | 
						|
static struct safe_page *safe_pages;
 | 
						|
 | 
						|
static int prepare_image(struct snapshot_handle *handle)
 | 
						|
{
 | 
						|
	int error = 0;
 | 
						|
	unsigned int nr_pages = nr_copy_pages;
 | 
						|
	struct pbe *p, *pblist = NULL;
 | 
						|
 | 
						|
	p = pagedir_nosave;
 | 
						|
	error = mark_unsafe_pages(p);
 | 
						|
	if (!error) {
 | 
						|
		pblist = alloc_pagedir(nr_pages, GFP_ATOMIC, 1);
 | 
						|
		if (pblist)
 | 
						|
			copy_page_backup_list(pblist, p);
 | 
						|
		free_pagedir(p, 0);
 | 
						|
		if (!pblist)
 | 
						|
			error = -ENOMEM;
 | 
						|
	}
 | 
						|
	safe_pages = NULL;
 | 
						|
	if (!error && nr_pages > unsafe_pages) {
 | 
						|
		nr_pages -= unsafe_pages;
 | 
						|
		while (nr_pages--) {
 | 
						|
			struct safe_page *ptr;
 | 
						|
 | 
						|
			ptr = (struct safe_page *)get_zeroed_page(GFP_ATOMIC);
 | 
						|
			if (!ptr) {
 | 
						|
				error = -ENOMEM;
 | 
						|
				break;
 | 
						|
			}
 | 
						|
			if (!PageNosaveFree(virt_to_page(ptr))) {
 | 
						|
				/* The page is "safe", add it to the list */
 | 
						|
				ptr->next = safe_pages;
 | 
						|
				safe_pages = ptr;
 | 
						|
			}
 | 
						|
			/* Mark the page as allocated */
 | 
						|
			SetPageNosave(virt_to_page(ptr));
 | 
						|
			SetPageNosaveFree(virt_to_page(ptr));
 | 
						|
		}
 | 
						|
	}
 | 
						|
	if (!error) {
 | 
						|
		pagedir_nosave = pblist;
 | 
						|
	} else {
 | 
						|
		handle->pbe = NULL;
 | 
						|
		swsusp_free();
 | 
						|
	}
 | 
						|
	return error;
 | 
						|
}
 | 
						|
 | 
						|
static void *get_buffer(struct snapshot_handle *handle)
 | 
						|
{
 | 
						|
	struct pbe *pbe = handle->pbe, *last = handle->last_pbe;
 | 
						|
	struct page *page = virt_to_page(pbe->orig_address);
 | 
						|
 | 
						|
	if (PageNosave(page) && PageNosaveFree(page)) {
 | 
						|
		/*
 | 
						|
		 * We have allocated the "original" page frame and we can
 | 
						|
		 * use it directly to store the read page
 | 
						|
		 */
 | 
						|
		pbe->address = 0;
 | 
						|
		if (last && last->next)
 | 
						|
			last->next = NULL;
 | 
						|
		return (void *)pbe->orig_address;
 | 
						|
	}
 | 
						|
	/*
 | 
						|
	 * The "original" page frame has not been allocated and we have to
 | 
						|
	 * use a "safe" page frame to store the read page
 | 
						|
	 */
 | 
						|
	pbe->address = (unsigned long)safe_pages;
 | 
						|
	safe_pages = safe_pages->next;
 | 
						|
	if (last)
 | 
						|
		last->next = pbe;
 | 
						|
	handle->last_pbe = pbe;
 | 
						|
	return (void *)pbe->address;
 | 
						|
}
 | 
						|
 | 
						|
/**
 | 
						|
 *	snapshot_write_next - used for writing the system memory snapshot.
 | 
						|
 *
 | 
						|
 *	On the first call to it @handle should point to a zeroed
 | 
						|
 *	snapshot_handle structure.  The structure gets updated and a pointer
 | 
						|
 *	to it should be passed to this function every next time.
 | 
						|
 *
 | 
						|
 *	The @count parameter should contain the number of bytes the caller
 | 
						|
 *	wants to write to the image.  It must not be zero.
 | 
						|
 *
 | 
						|
 *	On success the function returns a positive number.  Then, the caller
 | 
						|
 *	is allowed to write up to the returned number of bytes to the memory
 | 
						|
 *	location computed by the data_of() macro.  The number returned
 | 
						|
 *	may be smaller than @count, but this only happens if the write would
 | 
						|
 *	cross a page boundary otherwise.
 | 
						|
 *
 | 
						|
 *	The function returns 0 to indicate the "end of file" condition,
 | 
						|
 *	and a negative number is returned on error.  In such cases the
 | 
						|
 *	structure pointed to by @handle is not updated and should not be used
 | 
						|
 *	any more.
 | 
						|
 */
 | 
						|
 | 
						|
int snapshot_write_next(struct snapshot_handle *handle, size_t count)
 | 
						|
{
 | 
						|
	int error = 0;
 | 
						|
 | 
						|
	if (handle->prev && handle->page > nr_meta_pages + nr_copy_pages)
 | 
						|
		return 0;
 | 
						|
	if (!buffer) {
 | 
						|
		/* This makes the buffer be freed by swsusp_free() */
 | 
						|
		buffer = alloc_image_page(GFP_ATOMIC, 0);
 | 
						|
		if (!buffer)
 | 
						|
			return -ENOMEM;
 | 
						|
	}
 | 
						|
	if (!handle->offset)
 | 
						|
		handle->buffer = buffer;
 | 
						|
	if (handle->prev < handle->page) {
 | 
						|
		if (!handle->prev) {
 | 
						|
			error = load_header(handle, (struct swsusp_info *)buffer);
 | 
						|
			if (error)
 | 
						|
				return error;
 | 
						|
		} else if (handle->prev <= nr_meta_pages) {
 | 
						|
			handle->pbe = unpack_orig_addresses(buffer, handle->pbe);
 | 
						|
			if (!handle->pbe) {
 | 
						|
				error = prepare_image(handle);
 | 
						|
				if (error)
 | 
						|
					return error;
 | 
						|
				handle->pbe = pagedir_nosave;
 | 
						|
				handle->last_pbe = NULL;
 | 
						|
				handle->buffer = get_buffer(handle);
 | 
						|
			}
 | 
						|
		} else {
 | 
						|
			handle->pbe = handle->pbe->next;
 | 
						|
			handle->buffer = get_buffer(handle);
 | 
						|
		}
 | 
						|
		handle->prev = handle->page;
 | 
						|
	}
 | 
						|
	handle->buf_offset = handle->page_offset;
 | 
						|
	if (handle->page_offset + count >= PAGE_SIZE) {
 | 
						|
		count = PAGE_SIZE - handle->page_offset;
 | 
						|
		handle->page_offset = 0;
 | 
						|
		handle->page++;
 | 
						|
	} else {
 | 
						|
		handle->page_offset += count;
 | 
						|
	}
 | 
						|
	handle->offset += count;
 | 
						|
	return count;
 | 
						|
}
 | 
						|
 | 
						|
int snapshot_image_loaded(struct snapshot_handle *handle)
 | 
						|
{
 | 
						|
	return !(!handle->pbe || handle->pbe->next || !nr_copy_pages ||
 | 
						|
		handle->page <= nr_meta_pages + nr_copy_pages);
 | 
						|
}
 |