2006-01-16 19:50:04 +03:00
/*
* Copyright ( C ) Sistina Software , Inc . 1997 - 2003 All rights reserved .
2006-05-18 23:09:15 +04:00
* Copyright ( C ) 2004 - 2006 Red Hat , Inc . All rights reserved .
2006-01-16 19:50:04 +03:00
*
* This copyrighted material is made available to anyone wishing to use ,
* modify , copy , or redistribute it subject to the terms and conditions
* of the GNU General Public License v .2 .
*/
# include <linux/sched.h>
# include <linux/slab.h>
# include <linux/spinlock.h>
# include <linux/completion.h>
# include <linux/buffer_head.h>
2006-02-28 01:23:27 +03:00
# include <linux/gfs2_ondisk.h>
2006-01-16 19:50:04 +03:00
# include "gfs2.h"
2006-02-28 01:23:27 +03:00
# include "lm_interface.h"
# include "incore.h"
2006-01-16 19:50:04 +03:00
# include "bmap.h"
# include "glock.h"
# include "glops.h"
# include "inode.h"
# include "log.h"
# include "meta_io.h"
# include "recovery.h"
# include "rgrp.h"
2006-02-28 01:23:27 +03:00
# include "util.h"
2006-01-16 19:50:04 +03:00
2006-07-26 19:27:10 +04:00
/**
* gfs2_pte_inval - Sync and invalidate all PTEs associated with a glock
* @ gl : the glock
*
*/
static void gfs2_pte_inval ( struct gfs2_glock * gl )
{
struct gfs2_inode * ip ;
struct inode * inode ;
ip = gl - > gl_object ;
inode = & ip - > i_inode ;
if ( ! ip | | ! S_ISREG ( ip - > i_di . di_mode ) )
return ;
if ( ! test_bit ( GIF_PAGED , & ip - > i_flags ) )
return ;
unmap_shared_mapping_range ( inode - > i_mapping , 0 , 0 ) ;
if ( test_bit ( GIF_SW_PAGED , & ip - > i_flags ) )
set_bit ( GLF_DIRTY , & gl - > gl_flags ) ;
clear_bit ( GIF_SW_PAGED , & ip - > i_flags ) ;
}
/**
* gfs2_page_inval - Invalidate all pages associated with a glock
* @ gl : the glock
*
*/
static void gfs2_page_inval ( struct gfs2_glock * gl )
{
struct gfs2_inode * ip ;
struct inode * inode ;
ip = gl - > gl_object ;
inode = & ip - > i_inode ;
if ( ! ip | | ! S_ISREG ( ip - > i_di . di_mode ) )
return ;
truncate_inode_pages ( inode - > i_mapping , 0 ) ;
gfs2_assert_withdraw ( GFS2_SB ( & ip - > i_inode ) , ! inode - > i_mapping - > nrpages ) ;
clear_bit ( GIF_PAGED , & ip - > i_flags ) ;
}
/**
* gfs2_page_sync - Sync the data pages ( not metadata ) associated with a glock
* @ gl : the glock
* @ flags : DIO_START | DIO_WAIT
*
* Syncs data ( not metadata ) for a regular file .
* No - op for all other types .
*/
static void gfs2_page_sync ( struct gfs2_glock * gl , int flags )
{
struct gfs2_inode * ip ;
struct inode * inode ;
struct address_space * mapping ;
int error = 0 ;
ip = gl - > gl_object ;
inode = & ip - > i_inode ;
if ( ! ip | | ! S_ISREG ( ip - > i_di . di_mode ) )
return ;
mapping = inode - > i_mapping ;
if ( flags & DIO_START )
filemap_fdatawrite ( mapping ) ;
if ( ! error & & ( flags & DIO_WAIT ) )
error = filemap_fdatawait ( mapping ) ;
/* Put back any errors cleared by filemap_fdatawait()
so they can be caught by someone who can pass them
up to user space . */
if ( error = = - ENOSPC )
set_bit ( AS_ENOSPC , & mapping - > flags ) ;
else if ( error )
set_bit ( AS_EIO , & mapping - > flags ) ;
}
2006-01-16 19:50:04 +03:00
/**
* meta_go_sync - sync out the metadata for this glock
* @ gl : the glock
* @ flags : DIO_ *
*
* Called when demoting or unlocking an EX glock . We must flush
* to disk all dirty buffers / pages relating to this glock , and must not
* not return to caller to demote / unlock the glock until I / O is complete .
*/
static void meta_go_sync ( struct gfs2_glock * gl , int flags )
{
if ( ! ( flags & DIO_METADATA ) )
return ;
if ( test_and_clear_bit ( GLF_DIRTY , & gl - > gl_flags ) ) {
2006-04-07 19:17:32 +04:00
gfs2_log_flush ( gl - > gl_sbd , gl ) ;
2006-01-16 19:50:04 +03:00
gfs2_meta_sync ( gl , flags | DIO_START | DIO_WAIT ) ;
if ( flags & DIO_RELEASE )
gfs2_ail_empty_gl ( gl ) ;
}
clear_bit ( GLF_SYNC , & gl - > gl_flags ) ;
}
/**
* meta_go_inval - invalidate the metadata for this glock
* @ gl : the glock
* @ flags :
*
*/
static void meta_go_inval ( struct gfs2_glock * gl , int flags )
{
if ( ! ( flags & DIO_METADATA ) )
return ;
gfs2_meta_inval ( gl ) ;
gl - > gl_vn + + ;
}
/**
* meta_go_demote_ok - Check to see if it ' s ok to unlock a glock
* @ gl : the glock
*
* Returns : 1 if we have no cached data ; ok to demote meta glock
*/
static int meta_go_demote_ok ( struct gfs2_glock * gl )
{
return ! gl - > gl_aspace - > i_mapping - > nrpages ;
}
/**
* inode_go_xmote_th - promote / demote a glock
* @ gl : the glock
* @ state : the requested state
* @ flags :
*
*/
static void inode_go_xmote_th ( struct gfs2_glock * gl , unsigned int state ,
int flags )
{
if ( gl - > gl_state ! = LM_ST_UNLOCKED )
gfs2_pte_inval ( gl ) ;
gfs2_glock_xmote_th ( gl , state , flags ) ;
}
/**
* inode_go_xmote_bh - After promoting / demoting a glock
* @ gl : the glock
*
*/
static void inode_go_xmote_bh ( struct gfs2_glock * gl )
{
struct gfs2_holder * gh = gl - > gl_req_gh ;
struct buffer_head * bh ;
int error ;
if ( gl - > gl_state ! = LM_ST_UNLOCKED & &
( ! gh | | ! ( gh - > gh_flags & GL_SKIP ) ) ) {
error = gfs2_meta_read ( gl , gl - > gl_name . ln_number , DIO_START ,
& bh ) ;
if ( ! error )
brelse ( bh ) ;
}
}
/**
* inode_go_drop_th - unlock a glock
* @ gl : the glock
*
* Invoked from rq_demote ( ) .
* Another node needs the lock in EXCLUSIVE mode , or lock ( unused for too long )
* is being purged from our node ' s glock cache ; we ' re dropping lock .
*/
static void inode_go_drop_th ( struct gfs2_glock * gl )
{
gfs2_pte_inval ( gl ) ;
gfs2_glock_drop_th ( gl ) ;
}
/**
* inode_go_sync - Sync the dirty data and / or metadata for an inode glock
* @ gl : the glock protecting the inode
* @ flags :
*
*/
static void inode_go_sync ( struct gfs2_glock * gl , int flags )
{
int meta = ( flags & DIO_METADATA ) ;
int data = ( flags & DIO_DATA ) ;
if ( test_bit ( GLF_DIRTY , & gl - > gl_flags ) ) {
if ( meta & & data ) {
gfs2_page_sync ( gl , flags | DIO_START ) ;
2006-04-07 19:17:32 +04:00
gfs2_log_flush ( gl - > gl_sbd , gl ) ;
2006-01-16 19:50:04 +03:00
gfs2_meta_sync ( gl , flags | DIO_START | DIO_WAIT ) ;
gfs2_page_sync ( gl , flags | DIO_WAIT ) ;
clear_bit ( GLF_DIRTY , & gl - > gl_flags ) ;
} else if ( meta ) {
2006-04-07 19:17:32 +04:00
gfs2_log_flush ( gl - > gl_sbd , gl ) ;
2006-01-16 19:50:04 +03:00
gfs2_meta_sync ( gl , flags | DIO_START | DIO_WAIT ) ;
} else if ( data )
gfs2_page_sync ( gl , flags | DIO_START | DIO_WAIT ) ;
if ( flags & DIO_RELEASE )
gfs2_ail_empty_gl ( gl ) ;
}
clear_bit ( GLF_SYNC , & gl - > gl_flags ) ;
}
/**
* inode_go_inval - prepare a inode glock to be released
* @ gl : the glock
* @ flags :
*
*/
static void inode_go_inval ( struct gfs2_glock * gl , int flags )
{
int meta = ( flags & DIO_METADATA ) ;
int data = ( flags & DIO_DATA ) ;
if ( meta ) {
gfs2_meta_inval ( gl ) ;
gl - > gl_vn + + ;
}
if ( data )
gfs2_page_inval ( gl ) ;
}
/**
* inode_go_demote_ok - Check to see if it ' s ok to unlock an inode glock
* @ gl : the glock
*
* Returns : 1 if it ' s ok
*/
static int inode_go_demote_ok ( struct gfs2_glock * gl )
{
struct gfs2_sbd * sdp = gl - > gl_sbd ;
int demote = 0 ;
2006-02-28 01:23:27 +03:00
if ( ! gl - > gl_object & & ! gl - > gl_aspace - > i_mapping - > nrpages )
2006-01-16 19:50:04 +03:00
demote = 1 ;
else if ( ! sdp - > sd_args . ar_localcaching & &
time_after_eq ( jiffies , gl - > gl_stamp +
gfs2_tune_get ( sdp , gt_demote_secs ) * HZ ) )
demote = 1 ;
return demote ;
}
/**
* inode_go_lock - operation done after an inode lock is locked by a process
* @ gl : the glock
* @ flags :
*
* Returns : errno
*/
static int inode_go_lock ( struct gfs2_holder * gh )
{
struct gfs2_glock * gl = gh - > gh_gl ;
2006-02-28 01:23:27 +03:00
struct gfs2_inode * ip = gl - > gl_object ;
2006-01-16 19:50:04 +03:00
int error = 0 ;
if ( ! ip )
return 0 ;
if ( ip - > i_vn ! = gl - > gl_vn ) {
error = gfs2_inode_refresh ( ip ) ;
if ( error )
return error ;
gfs2_inode_attr_in ( ip ) ;
}
if ( ( ip - > i_di . di_flags & GFS2_DIF_TRUNC_IN_PROG ) & &
( gl - > gl_state = = LM_ST_EXCLUSIVE ) & &
( gh - > gh_flags & GL_LOCAL_EXCL ) )
error = gfs2_truncatei_resume ( ip ) ;
return error ;
}
/**
* inode_go_unlock - operation done before an inode lock is unlocked by a
* process
* @ gl : the glock
* @ flags :
*
*/
static void inode_go_unlock ( struct gfs2_holder * gh )
{
struct gfs2_glock * gl = gh - > gh_gl ;
2006-02-28 01:23:27 +03:00
struct gfs2_inode * ip = gl - > gl_object ;
2006-01-16 19:50:04 +03:00
if ( ip & & test_bit ( GLF_DIRTY , & gl - > gl_flags ) )
gfs2_inode_attr_in ( ip ) ;
if ( ip )
gfs2_meta_cache_flush ( ip ) ;
}
/**
* inode_greedy -
* @ gl : the glock
*
*/
static void inode_greedy ( struct gfs2_glock * gl )
{
struct gfs2_sbd * sdp = gl - > gl_sbd ;
2006-02-28 01:23:27 +03:00
struct gfs2_inode * ip = gl - > gl_object ;
2006-01-16 19:50:04 +03:00
unsigned int quantum = gfs2_tune_get ( sdp , gt_greedy_quantum ) ;
unsigned int max = gfs2_tune_get ( sdp , gt_greedy_max ) ;
unsigned int new_time ;
spin_lock ( & ip - > i_spin ) ;
if ( time_after ( ip - > i_last_pfault + quantum , jiffies ) ) {
new_time = ip - > i_greedy + quantum ;
if ( new_time > max )
new_time = max ;
} else {
new_time = ip - > i_greedy - quantum ;
if ( ! new_time | | new_time > max )
new_time = 1 ;
}
ip - > i_greedy = new_time ;
spin_unlock ( & ip - > i_spin ) ;
2006-06-14 23:32:57 +04:00
iput ( & ip - > i_inode ) ;
2006-01-16 19:50:04 +03:00
}
/**
* rgrp_go_demote_ok - Check to see if it ' s ok to unlock a RG ' s glock
* @ gl : the glock
*
* Returns : 1 if it ' s ok
*/
static int rgrp_go_demote_ok ( struct gfs2_glock * gl )
{
return ! gl - > gl_aspace - > i_mapping - > nrpages ;
}
/**
* rgrp_go_lock - operation done after an rgrp lock is locked by
* a first holder on this node .
* @ gl : the glock
* @ flags :
*
* Returns : errno
*/
static int rgrp_go_lock ( struct gfs2_holder * gh )
{
2006-02-28 01:23:27 +03:00
return gfs2_rgrp_bh_get ( gh - > gh_gl - > gl_object ) ;
2006-01-16 19:50:04 +03:00
}
/**
* rgrp_go_unlock - operation done before an rgrp lock is unlocked by
* a last holder on this node .
* @ gl : the glock
* @ flags :
*
*/
static void rgrp_go_unlock ( struct gfs2_holder * gh )
{
2006-02-28 01:23:27 +03:00
gfs2_rgrp_bh_put ( gh - > gh_gl - > gl_object ) ;
2006-01-16 19:50:04 +03:00
}
/**
* trans_go_xmote_th - promote / demote the transaction glock
* @ gl : the glock
* @ state : the requested state
* @ flags :
*
*/
static void trans_go_xmote_th ( struct gfs2_glock * gl , unsigned int state ,
int flags )
{
struct gfs2_sbd * sdp = gl - > gl_sbd ;
if ( gl - > gl_state ! = LM_ST_UNLOCKED & &
test_bit ( SDF_JOURNAL_LIVE , & sdp - > sd_flags ) ) {
gfs2_meta_syncfs ( sdp ) ;
gfs2_log_shutdown ( sdp ) ;
}
gfs2_glock_xmote_th ( gl , state , flags ) ;
}
/**
* trans_go_xmote_bh - After promoting / demoting the transaction glock
* @ gl : the glock
*
*/
static void trans_go_xmote_bh ( struct gfs2_glock * gl )
{
struct gfs2_sbd * sdp = gl - > gl_sbd ;
2006-06-14 23:32:57 +04:00
struct gfs2_inode * ip = GFS2_I ( sdp - > sd_jdesc - > jd_inode ) ;
2006-02-28 01:23:27 +03:00
struct gfs2_glock * j_gl = ip - > i_gl ;
2006-01-16 19:50:04 +03:00
struct gfs2_log_header head ;
int error ;
if ( gl - > gl_state ! = LM_ST_UNLOCKED & &
test_bit ( SDF_JOURNAL_LIVE , & sdp - > sd_flags ) ) {
2006-06-14 23:32:57 +04:00
gfs2_meta_cache_flush ( GFS2_I ( sdp - > sd_jdesc - > jd_inode ) ) ;
2006-01-16 19:50:04 +03:00
j_gl - > gl_ops - > go_inval ( j_gl , DIO_METADATA | DIO_DATA ) ;
error = gfs2_find_jhead ( sdp - > sd_jdesc , & head ) ;
if ( error )
gfs2_consist ( sdp ) ;
if ( ! ( head . lh_flags & GFS2_LOG_HEAD_UNMOUNT ) )
gfs2_consist ( sdp ) ;
/* Initialize some head of the log stuff */
if ( ! test_bit ( SDF_SHUTDOWN , & sdp - > sd_flags ) ) {
sdp - > sd_log_sequence = head . lh_sequence + 1 ;
gfs2_log_pointers_init ( sdp , head . lh_blkno ) ;
}
}
}
/**
* trans_go_drop_th - unlock the transaction glock
* @ gl : the glock
*
* We want to sync the device even with localcaching . Remember
* that localcaching journal replay only marks buffers dirty .
*/
static void trans_go_drop_th ( struct gfs2_glock * gl )
{
struct gfs2_sbd * sdp = gl - > gl_sbd ;
if ( test_bit ( SDF_JOURNAL_LIVE , & sdp - > sd_flags ) ) {
gfs2_meta_syncfs ( sdp ) ;
gfs2_log_shutdown ( sdp ) ;
}
gfs2_glock_drop_th ( gl ) ;
}
/**
* quota_go_demote_ok - Check to see if it ' s ok to unlock a quota glock
* @ gl : the glock
*
* Returns : 1 if it ' s ok
*/
static int quota_go_demote_ok ( struct gfs2_glock * gl )
{
return ! atomic_read ( & gl - > gl_lvb_count ) ;
}
struct gfs2_glock_operations gfs2_meta_glops = {
. go_xmote_th = gfs2_glock_xmote_th ,
. go_drop_th = gfs2_glock_drop_th ,
. go_sync = meta_go_sync ,
. go_inval = meta_go_inval ,
. go_demote_ok = meta_go_demote_ok ,
. go_type = LM_TYPE_META
} ;
struct gfs2_glock_operations gfs2_inode_glops = {
. go_xmote_th = inode_go_xmote_th ,
. go_xmote_bh = inode_go_xmote_bh ,
. go_drop_th = inode_go_drop_th ,
. go_sync = inode_go_sync ,
. go_inval = inode_go_inval ,
. go_demote_ok = inode_go_demote_ok ,
. go_lock = inode_go_lock ,
. go_unlock = inode_go_unlock ,
. go_greedy = inode_greedy ,
. go_type = LM_TYPE_INODE
} ;
struct gfs2_glock_operations gfs2_rgrp_glops = {
. go_xmote_th = gfs2_glock_xmote_th ,
. go_drop_th = gfs2_glock_drop_th ,
. go_sync = meta_go_sync ,
. go_inval = meta_go_inval ,
. go_demote_ok = rgrp_go_demote_ok ,
. go_lock = rgrp_go_lock ,
. go_unlock = rgrp_go_unlock ,
. go_type = LM_TYPE_RGRP
} ;
struct gfs2_glock_operations gfs2_trans_glops = {
. go_xmote_th = trans_go_xmote_th ,
. go_xmote_bh = trans_go_xmote_bh ,
. go_drop_th = trans_go_drop_th ,
. go_type = LM_TYPE_NONDISK
} ;
struct gfs2_glock_operations gfs2_iopen_glops = {
. go_xmote_th = gfs2_glock_xmote_th ,
. go_drop_th = gfs2_glock_drop_th ,
. go_callback = gfs2_iopen_go_callback ,
. go_type = LM_TYPE_IOPEN
} ;
struct gfs2_glock_operations gfs2_flock_glops = {
. go_xmote_th = gfs2_glock_xmote_th ,
. go_drop_th = gfs2_glock_drop_th ,
. go_type = LM_TYPE_FLOCK
} ;
struct gfs2_glock_operations gfs2_nondisk_glops = {
. go_xmote_th = gfs2_glock_xmote_th ,
. go_drop_th = gfs2_glock_drop_th ,
. go_type = LM_TYPE_NONDISK
} ;
struct gfs2_glock_operations gfs2_quota_glops = {
. go_xmote_th = gfs2_glock_xmote_th ,
. go_drop_th = gfs2_glock_drop_th ,
. go_demote_ok = quota_go_demote_ok ,
. go_type = LM_TYPE_QUOTA
} ;
struct gfs2_glock_operations gfs2_journal_glops = {
. go_xmote_th = gfs2_glock_xmote_th ,
. go_drop_th = gfs2_glock_drop_th ,
. go_type = LM_TYPE_JOURNAL
} ;