2005-10-31 01:59:56 +03:00
/*
2005-10-31 01:59:58 +03:00
* linux / kernel / power / snapshot . c
2005-10-31 01:59:56 +03:00
*
2005-10-31 01:59:58 +03:00
* This file provide system snapshot / restore functionality .
2005-10-31 01:59:56 +03:00
*
* Copyright ( C ) 1998 - 2005 Pavel Machek < pavel @ suse . cz >
*
* This file is released under the GPLv2 , and is based on swsusp . c .
*
*/
2006-03-23 13:59:59 +03:00
# include <linux/version.h>
2005-10-31 01:59:56 +03:00
# include <linux/module.h>
# include <linux/mm.h>
# include <linux/suspend.h>
# include <linux/smp_lock.h>
# include <linux/delay.h>
# include <linux/bitops.h>
# include <linux/spinlock.h>
# include <linux/kernel.h>
# include <linux/pm.h>
# include <linux/device.h>
# include <linux/bootmem.h>
# include <linux/syscalls.h>
# include <linux/console.h>
# include <linux/highmem.h>
# include <asm/uaccess.h>
# include <asm/mmu_context.h>
# include <asm/pgtable.h>
# include <asm/tlbflush.h>
# include <asm/io.h>
# include "power.h"
2006-01-06 11:13:05 +03:00
struct pbe * pagedir_nosave ;
2006-03-23 13:59:59 +03:00
static unsigned int nr_copy_pages ;
static unsigned int nr_meta_pages ;
2006-03-23 14:00:03 +03:00
static unsigned long * buffer ;
2006-01-06 11:13:05 +03:00
2005-10-31 01:59:56 +03:00
# ifdef CONFIG_HIGHMEM
2006-06-26 05:41:00 +04:00
unsigned int count_highmem_pages ( void )
2006-01-06 11:13:46 +03:00
{
struct zone * zone ;
unsigned long zone_pfn ;
unsigned int n = 0 ;
for_each_zone ( zone )
if ( is_highmem ( zone ) ) {
mark_free_pages ( zone ) ;
for ( zone_pfn = 0 ; zone_pfn < zone - > spanned_pages ; zone_pfn + + ) {
struct page * page ;
unsigned long pfn = zone_pfn + zone - > zone_start_pfn ;
if ( ! pfn_valid ( pfn ) )
continue ;
page = pfn_to_page ( pfn ) ;
if ( PageReserved ( page ) )
continue ;
if ( PageNosaveFree ( page ) )
continue ;
n + + ;
}
}
return n ;
}
2005-10-31 01:59:56 +03:00
struct highmem_page {
char * data ;
struct page * page ;
struct highmem_page * next ;
} ;
static struct highmem_page * highmem_copy ;
static int save_highmem_zone ( struct zone * zone )
{
unsigned long zone_pfn ;
mark_free_pages ( zone ) ;
for ( zone_pfn = 0 ; zone_pfn < zone - > spanned_pages ; + + zone_pfn ) {
struct page * page ;
struct highmem_page * save ;
void * kaddr ;
unsigned long pfn = zone_pfn + zone - > zone_start_pfn ;
2006-03-23 14:00:05 +03:00
if ( ! ( pfn % 10000 ) )
2005-10-31 01:59:56 +03:00
printk ( " . " ) ;
if ( ! pfn_valid ( pfn ) )
continue ;
page = pfn_to_page ( pfn ) ;
/*
* This condition results from rvmalloc ( ) sans vmalloc_32 ( )
* and architectural memory reservations . This should be
* corrected eventually when the cases giving rise to this
* are better understood .
*/
2006-02-16 02:17:42 +03:00
if ( PageReserved ( page ) )
2005-10-31 01:59:56 +03:00
continue ;
BUG_ON ( PageNosave ( page ) ) ;
if ( PageNosaveFree ( page ) )
continue ;
save = kmalloc ( sizeof ( struct highmem_page ) , GFP_ATOMIC ) ;
if ( ! save )
return - ENOMEM ;
save - > next = highmem_copy ;
save - > page = page ;
save - > data = ( void * ) get_zeroed_page ( GFP_ATOMIC ) ;
if ( ! save - > data ) {
kfree ( save ) ;
return - ENOMEM ;
}
kaddr = kmap_atomic ( page , KM_USER0 ) ;
memcpy ( save - > data , kaddr , PAGE_SIZE ) ;
kunmap_atomic ( kaddr , KM_USER0 ) ;
highmem_copy = save ;
}
return 0 ;
}
2006-06-26 05:41:00 +04:00
int save_highmem ( void )
2005-10-31 01:59:56 +03:00
{
struct zone * zone ;
int res = 0 ;
2006-03-23 14:00:05 +03:00
pr_debug ( " swsusp: Saving Highmem " ) ;
2006-03-23 14:00:06 +03:00
drain_local_pages ( ) ;
2005-10-31 01:59:56 +03:00
for_each_zone ( zone ) {
if ( is_highmem ( zone ) )
res = save_highmem_zone ( zone ) ;
if ( res )
return res ;
}
2006-03-23 14:00:05 +03:00
printk ( " \n " ) ;
2005-10-31 01:59:56 +03:00
return 0 ;
}
2006-06-26 05:41:00 +04:00
int restore_highmem ( void )
2005-10-31 01:59:56 +03:00
{
printk ( " swsusp: Restoring Highmem \n " ) ;
while ( highmem_copy ) {
struct highmem_page * save = highmem_copy ;
void * kaddr ;
highmem_copy = save - > next ;
kaddr = kmap_atomic ( save - > page , KM_USER0 ) ;
memcpy ( kaddr , save - > data , PAGE_SIZE ) ;
kunmap_atomic ( kaddr , KM_USER0 ) ;
free_page ( ( long ) save - > data ) ;
kfree ( save ) ;
}
return 0 ;
}
2006-06-23 13:04:44 +04:00
# else
2006-06-23 13:04:47 +04:00
static inline unsigned int count_highmem_pages ( void ) { return 0 ; }
static inline int save_highmem ( void ) { return 0 ; }
static inline int restore_highmem ( void ) { return 0 ; }
2005-11-09 08:34:41 +03:00
# endif
2005-10-31 01:59:56 +03:00
2006-09-26 10:32:50 +04:00
/**
* @ safe_needed - on resume , for storing the PBE list and the image ,
* we can only use memory pages that do not conflict with the pages
* used before suspend .
*
* The unsafe pages are marked with the PG_nosave_free flag
* and we count them using unsafe_pages
*/
static unsigned int unsafe_pages ;
static void * alloc_image_page ( gfp_t gfp_mask , int safe_needed )
{
void * res ;
res = ( void * ) get_zeroed_page ( gfp_mask ) ;
if ( safe_needed )
while ( res & & PageNosaveFree ( virt_to_page ( res ) ) ) {
/* The page is unsafe, mark it for swsusp_free() */
SetPageNosave ( virt_to_page ( res ) ) ;
unsafe_pages + + ;
res = ( void * ) get_zeroed_page ( gfp_mask ) ;
}
if ( res ) {
SetPageNosave ( virt_to_page ( res ) ) ;
SetPageNosaveFree ( virt_to_page ( res ) ) ;
}
return res ;
}
unsigned long get_safe_page ( gfp_t gfp_mask )
{
return ( unsigned long ) alloc_image_page ( gfp_mask , 1 ) ;
}
/**
* free_image_page - free page represented by @ addr , allocated with
* alloc_image_page ( page flags set by it must be cleared )
*/
static inline void free_image_page ( void * addr , int clear_nosave_free )
{
ClearPageNosave ( virt_to_page ( addr ) ) ;
if ( clear_nosave_free )
ClearPageNosaveFree ( virt_to_page ( addr ) ) ;
free_page ( ( unsigned long ) addr ) ;
}
/**
* pfn_is_nosave - check if given pfn is in the ' nosave ' section
*/
2006-09-26 10:32:45 +04:00
static inline int pfn_is_nosave ( unsigned long pfn )
2005-10-31 01:59:56 +03:00
{
unsigned long nosave_begin_pfn = __pa ( & __nosave_begin ) > > PAGE_SHIFT ;
unsigned long nosave_end_pfn = PAGE_ALIGN ( __pa ( & __nosave_end ) ) > > PAGE_SHIFT ;
return ( pfn > = nosave_begin_pfn ) & & ( pfn < nosave_end_pfn ) ;
}
/**
* saveable - Determine whether a page should be cloned or not .
* @ pfn : The page
*
2006-09-26 10:32:45 +04:00
* We save a page if it isn ' t Nosave , and is not in the range of pages
* statically defined as ' unsaveable ' , and it
* isn ' t a part of a free chunk of pages .
2005-10-31 01:59:56 +03:00
*/
2006-09-26 10:32:45 +04:00
static struct page * saveable_page ( unsigned long pfn )
2005-10-31 01:59:56 +03:00
{
2005-10-31 01:59:59 +03:00
struct page * page ;
2005-10-31 01:59:56 +03:00
if ( ! pfn_valid ( pfn ) )
2006-09-26 10:32:45 +04:00
return NULL ;
2005-10-31 01:59:56 +03:00
page = pfn_to_page ( pfn ) ;
2006-09-26 10:32:45 +04:00
2005-10-31 01:59:56 +03:00
if ( PageNosave ( page ) )
2006-09-26 10:32:45 +04:00
return NULL ;
2006-01-06 11:13:46 +03:00
if ( PageReserved ( page ) & & pfn_is_nosave ( pfn ) )
2006-09-26 10:32:45 +04:00
return NULL ;
2005-10-31 01:59:56 +03:00
if ( PageNosaveFree ( page ) )
2006-09-26 10:32:45 +04:00
return NULL ;
2005-10-31 01:59:56 +03:00
2006-09-26 10:32:45 +04:00
return page ;
2005-10-31 01:59:56 +03:00
}
2006-01-06 11:13:46 +03:00
unsigned int count_data_pages ( void )
2005-10-31 01:59:56 +03:00
{
struct zone * zone ;
2006-09-26 10:32:45 +04:00
unsigned long pfn , max_zone_pfn ;
2005-11-07 11:58:40 +03:00
unsigned int n = 0 ;
2005-10-31 01:59:56 +03:00
for_each_zone ( zone ) {
if ( is_highmem ( zone ) )
continue ;
mark_free_pages ( zone ) ;
2006-09-26 10:32:45 +04:00
max_zone_pfn = zone - > zone_start_pfn + zone - > spanned_pages ;
for ( pfn = zone - > zone_start_pfn ; pfn < max_zone_pfn ; pfn + + )
n + = ! ! saveable_page ( pfn ) ;
2005-10-31 01:59:56 +03:00
}
2005-10-31 01:59:57 +03:00
return n ;
2005-10-31 01:59:56 +03:00
}
2006-09-26 10:32:49 +04:00
static inline void copy_data_page ( long * dst , long * src )
{
int n ;
/* copy_page and memcpy are not usable for copying task structs. */
for ( n = PAGE_SIZE / sizeof ( long ) ; n ; n - - )
* dst + + = * src + + ;
}
2005-10-31 01:59:57 +03:00
static void copy_data_pages ( struct pbe * pblist )
2005-10-31 01:59:56 +03:00
{
struct zone * zone ;
2006-09-26 10:32:45 +04:00
unsigned long pfn , max_zone_pfn ;
2006-09-26 10:32:49 +04:00
struct pbe * pbe ;
2005-10-31 01:59:56 +03:00
2005-10-31 01:59:57 +03:00
pbe = pblist ;
2005-10-31 01:59:56 +03:00
for_each_zone ( zone ) {
if ( is_highmem ( zone ) )
continue ;
mark_free_pages ( zone ) ;
2006-09-26 10:32:45 +04:00
max_zone_pfn = zone - > zone_start_pfn + zone - > spanned_pages ;
for ( pfn = zone - > zone_start_pfn ; pfn < max_zone_pfn ; pfn + + ) {
struct page * page = saveable_page ( pfn ) ;
if ( page ) {
2006-09-26 10:32:49 +04:00
void * ptr = page_address ( page ) ;
2006-07-10 15:45:00 +04:00
2005-10-31 01:59:56 +03:00
BUG_ON ( ! pbe ) ;
2006-09-26 10:32:49 +04:00
copy_data_page ( ( void * ) pbe - > address , ptr ) ;
pbe - > orig_address = ( unsigned long ) ptr ;
2005-10-31 01:59:56 +03:00
pbe = pbe - > next ;
}
}
}
BUG_ON ( pbe ) ;
}
/**
* free_pagedir - free pages allocated with alloc_pagedir ( )
*/
2006-04-19 09:20:29 +04:00
static void free_pagedir ( struct pbe * pblist , int clear_nosave_free )
2005-10-31 01:59:56 +03:00
{
struct pbe * pbe ;
while ( pblist ) {
pbe = ( pblist + PB_PAGE_SKIP ) - > next ;
2006-09-26 10:32:50 +04:00
free_image_page ( pblist , clear_nosave_free ) ;
2005-10-31 01:59:56 +03:00
pblist = pbe ;
}
}
/**
* fill_pb_page - Create a list of PBEs on a given memory page
*/
static inline void fill_pb_page ( struct pbe * pbpage )
{
struct pbe * p ;
p = pbpage ;
pbpage + = PB_PAGE_SKIP ;
do
p - > next = p + 1 ;
while ( + + p < pbpage ) ;
}
/**
* create_pbe_list - Create a list of PBEs on top of a given chain
* of memory pages allocated with alloc_pagedir ( )
*/
2006-01-06 11:13:05 +03:00
static inline void create_pbe_list ( struct pbe * pblist , unsigned int nr_pages )
2005-10-31 01:59:56 +03:00
{
struct pbe * pbpage , * p ;
2005-11-07 11:58:40 +03:00
unsigned int num = PBES_PER_PAGE ;
2005-10-31 01:59:56 +03:00
for_each_pb_page ( pbpage , pblist ) {
if ( num > = nr_pages )
break ;
fill_pb_page ( pbpage ) ;
num + = PBES_PER_PAGE ;
}
if ( pbpage ) {
for ( num - = PBES_PER_PAGE - 1 , p = pbpage ; num < nr_pages ; p + + , num + + )
p - > next = p + 1 ;
p - > next = NULL ;
}
}
/**
* alloc_pagedir - Allocate the page directory .
*
* First , determine exactly how many pages we need and
* allocate them .
*
* We arrange the pages in a chain : each page is an array of PBES_PER_PAGE
* struct pbe elements ( pbes ) and the last element in the page points
* to the next page .
*
* On each page we set up a list of struct_pbe elements .
*/
2006-06-23 13:04:47 +04:00
static struct pbe * alloc_pagedir ( unsigned int nr_pages , gfp_t gfp_mask ,
int safe_needed )
2005-10-31 01:59:56 +03:00
{
2005-11-07 11:58:40 +03:00
unsigned int num ;
2005-10-31 01:59:56 +03:00
struct pbe * pblist , * pbe ;
if ( ! nr_pages )
return NULL ;
2005-11-09 08:34:39 +03:00
pblist = alloc_image_page ( gfp_mask , safe_needed ) ;
2005-10-31 01:59:56 +03:00
/* FIXME: rewrite this ugly loop */
for ( pbe = pblist , num = PBES_PER_PAGE ; pbe & & num < nr_pages ;
pbe = pbe - > next , num + = PBES_PER_PAGE ) {
pbe + = PB_PAGE_SKIP ;
2005-11-09 08:34:39 +03:00
pbe - > next = alloc_image_page ( gfp_mask , safe_needed ) ;
2005-10-31 01:59:56 +03:00
}
if ( ! pbe ) { /* get_zeroed_page() failed */
2006-04-19 09:20:29 +04:00
free_pagedir ( pblist , 1 ) ;
2005-10-31 01:59:56 +03:00
pblist = NULL ;
2006-01-06 11:13:05 +03:00
} else
2006-03-23 14:00:03 +03:00
create_pbe_list ( pblist , nr_pages ) ;
2005-10-31 01:59:56 +03:00
return pblist ;
}
/**
* Free pages we allocated for suspend . Suspend pages are alocated
* before atomic copy , so we need to free them after resume .
*/
void swsusp_free ( void )
{
struct zone * zone ;
2006-09-26 10:32:45 +04:00
unsigned long pfn , max_zone_pfn ;
2005-10-31 01:59:56 +03:00
for_each_zone ( zone ) {
2006-09-26 10:32:45 +04:00
max_zone_pfn = zone - > zone_start_pfn + zone - > spanned_pages ;
for ( pfn = zone - > zone_start_pfn ; pfn < max_zone_pfn ; pfn + + )
if ( pfn_valid ( pfn ) ) {
struct page * page = pfn_to_page ( pfn ) ;
2005-10-31 01:59:56 +03:00
if ( PageNosave ( page ) & & PageNosaveFree ( page ) ) {
ClearPageNosave ( page ) ;
ClearPageNosaveFree ( page ) ;
free_page ( ( long ) page_address ( page ) ) ;
}
}
}
2006-03-23 13:59:59 +03:00
nr_copy_pages = 0 ;
nr_meta_pages = 0 ;
pagedir_nosave = NULL ;
2006-03-23 14:00:03 +03:00
buffer = NULL ;
2005-10-31 01:59:56 +03:00
}
/**
* enough_free_mem - Make sure we enough free memory to snapshot .
*
* Returns TRUE or FALSE after checking the number of available
* free pages .
*/
2005-11-07 11:58:40 +03:00
static int enough_free_mem ( unsigned int nr_pages )
2005-10-31 01:59:56 +03:00
{
2006-01-06 11:14:20 +03:00
struct zone * zone ;
unsigned int n = 0 ;
for_each_zone ( zone )
if ( ! is_highmem ( zone ) )
n + = zone - > free_pages ;
pr_debug ( " swsusp: available memory: %u pages \n " , n ) ;
return n > ( nr_pages + PAGES_FOR_IO +
2005-10-31 01:59:57 +03:00
( nr_pages + PBES_PER_PAGE - 1 ) / PBES_PER_PAGE ) ;
2005-10-31 01:59:56 +03:00
}
2006-03-23 13:59:59 +03:00
static int alloc_data_pages ( struct pbe * pblist , gfp_t gfp_mask , int safe_needed )
2005-11-09 08:34:39 +03:00
{
struct pbe * p ;
for_each_pbe ( p , pblist ) {
p - > address = ( unsigned long ) alloc_image_page ( gfp_mask , safe_needed ) ;
if ( ! p - > address )
return - ENOMEM ;
}
return 0 ;
}
2005-10-31 01:59:56 +03:00
2005-11-07 11:58:40 +03:00
static struct pbe * swsusp_alloc ( unsigned int nr_pages )
2005-10-31 01:59:56 +03:00
{
2005-11-09 08:34:39 +03:00
struct pbe * pblist ;
2005-10-31 01:59:56 +03:00
2005-11-09 08:34:39 +03:00
if ( ! ( pblist = alloc_pagedir ( nr_pages , GFP_ATOMIC | __GFP_COLD , 0 ) ) ) {
2005-10-31 01:59:56 +03:00
printk ( KERN_ERR " suspend: Allocating pagedir failed. \n " ) ;
2005-10-31 01:59:57 +03:00
return NULL ;
2005-10-31 01:59:56 +03:00
}
2005-11-09 08:34:39 +03:00
if ( alloc_data_pages ( pblist , GFP_ATOMIC | __GFP_COLD , 0 ) ) {
printk ( KERN_ERR " suspend: Allocating image pages failed. \n " ) ;
swsusp_free ( ) ;
return NULL ;
2005-10-31 01:59:56 +03:00
}
2005-10-31 01:59:57 +03:00
return pblist ;
2005-10-31 01:59:56 +03:00
}
2005-10-31 02:00:00 +03:00
asmlinkage int swsusp_save ( void )
2005-10-31 01:59:56 +03:00
{
2005-11-07 11:58:40 +03:00
unsigned int nr_pages ;
2005-10-31 01:59:56 +03:00
pr_debug ( " swsusp: critical section: \n " ) ;
drain_local_pages ( ) ;
2005-10-31 01:59:57 +03:00
nr_pages = count_data_pages ( ) ;
printk ( " swsusp: Need to copy %u pages \n " , nr_pages ) ;
2005-10-31 01:59:56 +03:00
pr_debug ( " swsusp: pages needed: %u + %lu + %u, free: %u \n " ,
2005-10-31 01:59:57 +03:00
nr_pages ,
( nr_pages + PBES_PER_PAGE - 1 ) / PBES_PER_PAGE ,
2005-10-31 01:59:56 +03:00
PAGES_FOR_IO , nr_free_pages ( ) ) ;
2005-10-31 01:59:57 +03:00
if ( ! enough_free_mem ( nr_pages ) ) {
2005-10-31 01:59:56 +03:00
printk ( KERN_ERR " swsusp: Not enough free memory \n " ) ;
return - ENOMEM ;
}
2005-10-31 01:59:57 +03:00
pagedir_nosave = swsusp_alloc ( nr_pages ) ;
if ( ! pagedir_nosave )
return - ENOMEM ;
2005-10-31 01:59:56 +03:00
/* During allocating of suspend pagedir, new cold pages may appear.
* Kill them .
*/
drain_local_pages ( ) ;
2005-10-31 01:59:57 +03:00
copy_data_pages ( pagedir_nosave ) ;
2005-10-31 01:59:56 +03:00
/*
* End of critical section . From now on , we can write to memory ,
* but we should not touch disk . This specially means we must _not_
* touch swap space ! Except we must write out our image of course .
*/
2005-10-31 01:59:57 +03:00
nr_copy_pages = nr_pages ;
2006-03-23 13:59:59 +03:00
nr_meta_pages = ( nr_pages * sizeof ( long ) + PAGE_SIZE - 1 ) > > PAGE_SHIFT ;
2005-10-31 01:59:57 +03:00
printk ( " swsusp: critical section/: done (%d pages copied) \n " , nr_pages ) ;
2005-10-31 01:59:56 +03:00
return 0 ;
}
2006-03-23 13:59:59 +03:00
static void init_header ( struct swsusp_info * info )
{
memset ( info , 0 , sizeof ( struct swsusp_info ) ) ;
info - > version_code = LINUX_VERSION_CODE ;
info - > num_physpages = num_physpages ;
memcpy ( & info - > uts , & system_utsname , sizeof ( system_utsname ) ) ;
info - > cpus = num_online_cpus ( ) ;
info - > image_pages = nr_copy_pages ;
info - > pages = nr_copy_pages + nr_meta_pages + 1 ;
2006-03-23 14:00:03 +03:00
info - > size = info - > pages ;
info - > size < < = PAGE_SHIFT ;
2006-03-23 13:59:59 +03:00
}
/**
* pack_orig_addresses - the . orig_address fields of the PBEs from the
* list starting at @ pbe are stored in the array @ buf [ ] ( 1 page )
*/
static inline struct pbe * pack_orig_addresses ( unsigned long * buf , struct pbe * pbe )
{
int j ;
for ( j = 0 ; j < PAGE_SIZE / sizeof ( long ) & & pbe ; j + + ) {
buf [ j ] = pbe - > orig_address ;
pbe = pbe - > next ;
}
if ( ! pbe )
for ( ; j < PAGE_SIZE / sizeof ( long ) ; j + + )
buf [ j ] = 0 ;
return pbe ;
}
/**
* snapshot_read_next - used for reading the system memory snapshot .
*
* On the first call to it @ handle should point to a zeroed
* snapshot_handle structure . The structure gets updated and a pointer
* to it should be passed to this function every next time .
*
* The @ count parameter should contain the number of bytes the caller
* wants to read from the snapshot . It must not be zero .
*
* On success the function returns a positive number . Then , the caller
* is allowed to read up to the returned number of bytes from the memory
* location computed by the data_of ( ) macro . The number returned
* may be smaller than @ count , but this only happens if the read would
* cross a page boundary otherwise .
*
* The function returns 0 to indicate the end of data stream condition ,
* and a negative number is returned on error . In such cases the
* structure pointed to by @ handle is not updated and should not be used
* any more .
*/
int snapshot_read_next ( struct snapshot_handle * handle , size_t count )
{
2006-09-26 10:32:46 +04:00
if ( handle - > cur > nr_meta_pages + nr_copy_pages )
2006-03-23 13:59:59 +03:00
return 0 ;
if ( ! buffer ) {
/* This makes the buffer be freed by swsusp_free() */
buffer = alloc_image_page ( GFP_ATOMIC , 0 ) ;
if ( ! buffer )
return - ENOMEM ;
}
if ( ! handle - > offset ) {
init_header ( ( struct swsusp_info * ) buffer ) ;
handle - > buffer = buffer ;
handle - > pbe = pagedir_nosave ;
}
2006-09-26 10:32:46 +04:00
if ( handle - > prev < handle - > cur ) {
if ( handle - > cur < = nr_meta_pages ) {
2006-03-23 13:59:59 +03:00
handle - > pbe = pack_orig_addresses ( buffer , handle - > pbe ) ;
if ( ! handle - > pbe )
handle - > pbe = pagedir_nosave ;
} else {
handle - > buffer = ( void * ) handle - > pbe - > address ;
handle - > pbe = handle - > pbe - > next ;
}
2006-09-26 10:32:46 +04:00
handle - > prev = handle - > cur ;
2006-03-23 13:59:59 +03:00
}
2006-09-26 10:32:46 +04:00
handle - > buf_offset = handle - > cur_offset ;
if ( handle - > cur_offset + count > = PAGE_SIZE ) {
count = PAGE_SIZE - handle - > cur_offset ;
handle - > cur_offset = 0 ;
handle - > cur + + ;
2006-03-23 13:59:59 +03:00
} else {
2006-09-26 10:32:46 +04:00
handle - > cur_offset + = count ;
2006-03-23 13:59:59 +03:00
}
handle - > offset + = count ;
return count ;
}
/**
* mark_unsafe_pages - mark the pages that cannot be used for storing
* the image during resume , because they conflict with the pages that
* had been used before suspend
*/
static int mark_unsafe_pages ( struct pbe * pblist )
{
struct zone * zone ;
2006-09-26 10:32:45 +04:00
unsigned long pfn , max_zone_pfn ;
2006-03-23 13:59:59 +03:00
struct pbe * p ;
if ( ! pblist ) /* a sanity check */
return - EINVAL ;
/* Clear page flags */
for_each_zone ( zone ) {
2006-09-26 10:32:45 +04:00
max_zone_pfn = zone - > zone_start_pfn + zone - > spanned_pages ;
for ( pfn = zone - > zone_start_pfn ; pfn < max_zone_pfn ; pfn + + )
if ( pfn_valid ( pfn ) )
ClearPageNosaveFree ( pfn_to_page ( pfn ) ) ;
2006-03-23 13:59:59 +03:00
}
/* Mark orig addresses */
for_each_pbe ( p , pblist ) {
if ( virt_addr_valid ( p - > orig_address ) )
SetPageNosaveFree ( virt_to_page ( p - > orig_address ) ) ;
else
return - EFAULT ;
}
2006-06-23 13:04:48 +04:00
unsafe_pages = 0 ;
2006-03-23 13:59:59 +03:00
return 0 ;
}
static void copy_page_backup_list ( struct pbe * dst , struct pbe * src )
{
/* We assume both lists contain the same number of elements */
while ( src ) {
dst - > orig_address = src - > orig_address ;
dst = dst - > next ;
src = src - > next ;
}
}
static int check_header ( struct swsusp_info * info )
{
char * reason = NULL ;
if ( info - > version_code ! = LINUX_VERSION_CODE )
reason = " kernel version " ;
if ( info - > num_physpages ! = num_physpages )
reason = " memory size " ;
if ( strcmp ( info - > uts . sysname , system_utsname . sysname ) )
reason = " system type " ;
if ( strcmp ( info - > uts . release , system_utsname . release ) )
reason = " kernel release " ;
if ( strcmp ( info - > uts . version , system_utsname . version ) )
reason = " version " ;
if ( strcmp ( info - > uts . machine , system_utsname . machine ) )
reason = " machine " ;
if ( reason ) {
printk ( KERN_ERR " swsusp: Resume mismatch: %s \n " , reason ) ;
return - EPERM ;
}
return 0 ;
}
/**
* load header - check the image header and copy data from it
*/
static int load_header ( struct snapshot_handle * handle ,
struct swsusp_info * info )
{
int error ;
struct pbe * pblist ;
error = check_header ( info ) ;
if ( ! error ) {
pblist = alloc_pagedir ( info - > image_pages , GFP_ATOMIC , 0 ) ;
if ( ! pblist )
return - ENOMEM ;
pagedir_nosave = pblist ;
handle - > pbe = pblist ;
nr_copy_pages = info - > image_pages ;
nr_meta_pages = info - > pages - info - > image_pages - 1 ;
}
return error ;
}
/**
* unpack_orig_addresses - copy the elements of @ buf [ ] ( 1 page ) to
* the PBEs in the list starting at @ pbe
*/
static inline struct pbe * unpack_orig_addresses ( unsigned long * buf ,
struct pbe * pbe )
{
int j ;
for ( j = 0 ; j < PAGE_SIZE / sizeof ( long ) & & pbe ; j + + ) {
pbe - > orig_address = buf [ j ] ;
pbe = pbe - > next ;
}
return pbe ;
}
/**
2006-06-23 13:04:48 +04:00
* prepare_image - use metadata contained in the PBE list
2006-03-23 13:59:59 +03:00
* pointed to by pagedir_nosave to mark the pages that will
* be overwritten in the process of restoring the system
2006-06-23 13:04:48 +04:00
* memory state from the image ( " unsafe " pages ) and allocate
* memory for the image
*
* The idea is to allocate the PBE list first and then
* allocate as many pages as it ' s needed for the image data ,
* but not to assign these pages to the PBEs initially .
* Instead , we just mark them as allocated and create a list
* of " safe " which will be used later
2006-03-23 13:59:59 +03:00
*/
2006-06-23 13:04:48 +04:00
struct safe_page {
struct safe_page * next ;
char padding [ PAGE_SIZE - sizeof ( void * ) ] ;
} ;
static struct safe_page * safe_pages ;
static int prepare_image ( struct snapshot_handle * handle )
2006-03-23 13:59:59 +03:00
{
int error = 0 ;
2006-06-23 13:04:48 +04:00
unsigned int nr_pages = nr_copy_pages ;
struct pbe * p , * pblist = NULL ;
2006-03-23 13:59:59 +03:00
p = pagedir_nosave ;
error = mark_unsafe_pages ( p ) ;
if ( ! error ) {
2006-06-23 13:04:48 +04:00
pblist = alloc_pagedir ( nr_pages , GFP_ATOMIC , 1 ) ;
2006-03-23 13:59:59 +03:00
if ( pblist )
copy_page_backup_list ( pblist , p ) ;
2006-04-19 09:20:29 +04:00
free_pagedir ( p , 0 ) ;
2006-03-23 13:59:59 +03:00
if ( ! pblist )
error = - ENOMEM ;
}
2006-06-23 13:04:48 +04:00
safe_pages = NULL ;
if ( ! error & & nr_pages > unsafe_pages ) {
nr_pages - = unsafe_pages ;
while ( nr_pages - - ) {
struct safe_page * ptr ;
ptr = ( struct safe_page * ) get_zeroed_page ( GFP_ATOMIC ) ;
if ( ! ptr ) {
error = - ENOMEM ;
break ;
}
if ( ! PageNosaveFree ( virt_to_page ( ptr ) ) ) {
/* The page is "safe", add it to the list */
ptr - > next = safe_pages ;
safe_pages = ptr ;
}
/* Mark the page as allocated */
SetPageNosave ( virt_to_page ( ptr ) ) ;
SetPageNosaveFree ( virt_to_page ( ptr ) ) ;
}
}
2006-03-23 13:59:59 +03:00
if ( ! error ) {
pagedir_nosave = pblist ;
} else {
handle - > pbe = NULL ;
2006-06-23 13:04:48 +04:00
swsusp_free ( ) ;
2006-03-23 13:59:59 +03:00
}
return error ;
}
2006-06-23 13:04:48 +04:00
static void * get_buffer ( struct snapshot_handle * handle )
{
struct pbe * pbe = handle - > pbe , * last = handle - > last_pbe ;
struct page * page = virt_to_page ( pbe - > orig_address ) ;
if ( PageNosave ( page ) & & PageNosaveFree ( page ) ) {
/*
* We have allocated the " original " page frame and we can
* use it directly to store the read page
*/
pbe - > address = 0 ;
if ( last & & last - > next )
last - > next = NULL ;
return ( void * ) pbe - > orig_address ;
}
/*
* The " original " page frame has not been allocated and we have to
* use a " safe " page frame to store the read page
*/
pbe - > address = ( unsigned long ) safe_pages ;
safe_pages = safe_pages - > next ;
if ( last )
last - > next = pbe ;
handle - > last_pbe = pbe ;
return ( void * ) pbe - > address ;
}
2006-03-23 13:59:59 +03:00
/**
* snapshot_write_next - used for writing the system memory snapshot .
*
* On the first call to it @ handle should point to a zeroed
* snapshot_handle structure . The structure gets updated and a pointer
* to it should be passed to this function every next time .
*
* The @ count parameter should contain the number of bytes the caller
* wants to write to the image . It must not be zero .
*
* On success the function returns a positive number . Then , the caller
* is allowed to write up to the returned number of bytes to the memory
* location computed by the data_of ( ) macro . The number returned
* may be smaller than @ count , but this only happens if the write would
* cross a page boundary otherwise .
*
* The function returns 0 to indicate the " end of file " condition ,
* and a negative number is returned on error . In such cases the
* structure pointed to by @ handle is not updated and should not be used
* any more .
*/
int snapshot_write_next ( struct snapshot_handle * handle , size_t count )
{
int error = 0 ;
2006-09-26 10:32:46 +04:00
if ( handle - > prev & & handle - > cur > nr_meta_pages + nr_copy_pages )
2006-03-23 13:59:59 +03:00
return 0 ;
if ( ! buffer ) {
/* This makes the buffer be freed by swsusp_free() */
buffer = alloc_image_page ( GFP_ATOMIC , 0 ) ;
if ( ! buffer )
return - ENOMEM ;
}
if ( ! handle - > offset )
handle - > buffer = buffer ;
2006-09-26 10:32:44 +04:00
handle - > sync_read = 1 ;
2006-09-26 10:32:46 +04:00
if ( handle - > prev < handle - > cur ) {
2006-03-23 13:59:59 +03:00
if ( ! handle - > prev ) {
2006-09-26 10:32:44 +04:00
error = load_header ( handle ,
( struct swsusp_info * ) buffer ) ;
2006-03-23 13:59:59 +03:00
if ( error )
return error ;
} else if ( handle - > prev < = nr_meta_pages ) {
2006-09-26 10:32:44 +04:00
handle - > pbe = unpack_orig_addresses ( buffer ,
handle - > pbe ) ;
2006-03-23 13:59:59 +03:00
if ( ! handle - > pbe ) {
2006-06-23 13:04:48 +04:00
error = prepare_image ( handle ) ;
2006-03-23 13:59:59 +03:00
if ( error )
return error ;
handle - > pbe = pagedir_nosave ;
2006-06-23 13:04:48 +04:00
handle - > last_pbe = NULL ;
handle - > buffer = get_buffer ( handle ) ;
2006-09-26 10:32:44 +04:00
handle - > sync_read = 0 ;
2006-03-23 13:59:59 +03:00
}
} else {
handle - > pbe = handle - > pbe - > next ;
2006-06-23 13:04:48 +04:00
handle - > buffer = get_buffer ( handle ) ;
2006-09-26 10:32:44 +04:00
handle - > sync_read = 0 ;
2006-03-23 13:59:59 +03:00
}
2006-09-26 10:32:46 +04:00
handle - > prev = handle - > cur ;
2006-03-23 13:59:59 +03:00
}
2006-09-26 10:32:46 +04:00
handle - > buf_offset = handle - > cur_offset ;
if ( handle - > cur_offset + count > = PAGE_SIZE ) {
count = PAGE_SIZE - handle - > cur_offset ;
handle - > cur_offset = 0 ;
handle - > cur + + ;
2006-03-23 13:59:59 +03:00
} else {
2006-09-26 10:32:46 +04:00
handle - > cur_offset + = count ;
2006-03-23 13:59:59 +03:00
}
handle - > offset + = count ;
return count ;
}
int snapshot_image_loaded ( struct snapshot_handle * handle )
{
return ! ( ! handle - > pbe | | handle - > pbe - > next | | ! nr_copy_pages | |
2006-09-26 10:32:46 +04:00
handle - > cur < = nr_meta_pages + nr_copy_pages ) ;
2006-03-23 13:59:59 +03:00
}