2006-01-16 19:50:04 +03:00
/*
* Copyright ( C ) Sistina Software , Inc . 1997 - 2003 All rights reserved .
2006-05-18 23:09:15 +04:00
* Copyright ( C ) 2004 - 2006 Red Hat , Inc . All rights reserved .
2006-01-16 19:50:04 +03:00
*
* This copyrighted material is made available to anyone wishing to use ,
* modify , copy , or redistribute it subject to the terms and conditions
2006-09-01 19:05:15 +04:00
* of the GNU General Public License version 2.
2006-01-16 19:50:04 +03:00
*/
# ifndef __GLOCK_DOT_H__
# define __GLOCK_DOT_H__
Detach sched.h from mm.h
First thing mm.h does is including sched.h solely for can_do_mlock() inline
function which has "current" dereference inside. By dealing with can_do_mlock()
mm.h can be detached from sched.h which is good. See below, why.
This patch
a) removes unconditional inclusion of sched.h from mm.h
b) makes can_do_mlock() normal function in mm/mlock.c
c) exports can_do_mlock() to not break compilation
d) adds sched.h inclusions back to files that were getting it indirectly.
e) adds less bloated headers to some files (asm/signal.h, jiffies.h) that were
getting them indirectly
Net result is:
a) mm.h users would get less code to open, read, preprocess, parse, ... if
they don't need sched.h
b) sched.h stops being dependency for significant number of files:
on x86_64 allmodconfig touching sched.h results in recompile of 4083 files,
after patch it's only 3744 (-8.3%).
Cross-compile tested on
all arm defconfigs, all mips defconfigs, all powerpc defconfigs,
alpha alpha-up
arm
i386 i386-up i386-defconfig i386-allnoconfig
ia64 ia64-up
m68k
mips
parisc parisc-up
powerpc powerpc-up
s390 s390-up
sparc sparc-up
sparc64 sparc64-up
um-x86_64
x86_64 x86_64-up x86_64-defconfig x86_64-allnoconfig
as well as my two usual configs.
Signed-off-by: Alexey Dobriyan <adobriyan@gmail.com>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2007-05-21 01:22:52 +04:00
# include <linux/sched.h>
2009-01-12 13:43:39 +03:00
# include <linux/parser.h>
2006-09-05 18:39:21 +04:00
# include "incore.h"
2017-07-18 19:35:04 +03:00
# include "util.h"
2006-09-05 18:39:21 +04:00
2009-01-12 13:43:39 +03:00
/* Options for hostdata parser */
enum {
Opt_jid ,
Opt_id ,
Opt_first ,
Opt_nodir ,
Opt_err ,
} ;
/*
* lm_lockname types
*/
# define LM_TYPE_RESERVED 0x00
# define LM_TYPE_NONDISK 0x01
# define LM_TYPE_INODE 0x02
# define LM_TYPE_RGRP 0x03
# define LM_TYPE_META 0x04
# define LM_TYPE_IOPEN 0x05
# define LM_TYPE_FLOCK 0x06
# define LM_TYPE_PLOCK 0x07
# define LM_TYPE_QUOTA 0x08
# define LM_TYPE_JOURNAL 0x09
/*
* lm_lock ( ) states
*
* SHARED is compatible with SHARED , not with DEFERRED or EX .
* DEFERRED is compatible with DEFERRED , not with SHARED or EX .
*/
# define LM_ST_UNLOCKED 0
# define LM_ST_EXCLUSIVE 1
# define LM_ST_DEFERRED 2
# define LM_ST_SHARED 3
/*
* lm_lock ( ) flags
*
* LM_FLAG_TRY
* Don ' t wait to acquire the lock if it can ' t be granted immediately .
*
* LM_FLAG_TRY_1CB
* Send one blocking callback if TRY is set and the lock is not granted .
*
* LM_FLAG_NOEXP
* GFS sets this flag on lock requests it makes while doing journal recovery .
* These special requests should not be blocked due to the recovery like
* ordinary locks would be .
*
* LM_FLAG_ANY
* A SHARED request may also be granted in DEFERRED , or a DEFERRED request may
* also be granted in SHARED . The preferred state is whichever is compatible
* with other granted locks , or the specified state if no other locks exist .
*
* LM_FLAG_PRIORITY
* Override fairness considerations . Suppose a lock is held in a shared state
* and there is a pending request for the deferred state . A shared lock
* request with the priority flag would be allowed to bypass the deferred
* request and directly join the other shared lock . A shared lock request
* without the priority flag might be forced to wait until the deferred
* requested had acquired and released the lock .
*/
2015-07-24 17:45:43 +03:00
# define LM_FLAG_TRY 0x0001
# define LM_FLAG_TRY_1CB 0x0002
# define LM_FLAG_NOEXP 0x0004
# define LM_FLAG_ANY 0x0008
# define LM_FLAG_PRIORITY 0x0010
# define GL_ASYNC 0x0040
# define GL_EXACT 0x0080
# define GL_SKIP 0x0100
# define GL_NOCACHE 0x0400
2009-01-12 13:43:39 +03:00
/*
2010-11-29 15:50:38 +03:00
* lm_async_cb return flags
2009-01-12 13:43:39 +03:00
*
* LM_OUT_ST_MASK
* Masks the lower two bits of lock state in the returned value .
*
* LM_OUT_CANCELED
* The lock request was canceled .
*
*/
# define LM_OUT_ST_MASK 0x00000003
# define LM_OUT_CANCELED 0x00000008
2010-11-29 15:50:38 +03:00
# define LM_OUT_ERROR 0x00000004
2009-01-12 13:43:39 +03:00
/*
* lm_recovery_done ( ) messages
*/
# define LM_RD_GAVEUP 308
# define LM_RD_SUCCESS 309
# define GLR_TRYFAILED 13
2011-06-15 19:41:48 +04:00
# define GL_GLOCK_MAX_HOLD (long)(HZ / 5)
# define GL_GLOCK_DFT_HOLD (long)(HZ / 5)
# define GL_GLOCK_MIN_HOLD (long)(10)
# define GL_GLOCK_HOLD_INCR (long)(HZ / 20)
# define GL_GLOCK_HOLD_DECR (long)(HZ / 40)
2009-01-12 13:43:39 +03:00
struct lm_lockops {
const char * lm_proto_name ;
2012-01-10 02:18:05 +04:00
int ( * lm_mount ) ( struct gfs2_sbd * sdp , const char * table ) ;
void ( * lm_first_done ) ( struct gfs2_sbd * sdp ) ;
void ( * lm_recovery_result ) ( struct gfs2_sbd * sdp , unsigned int jid ,
unsigned int result ) ;
void ( * lm_unmount ) ( struct gfs2_sbd * sdp ) ;
2009-01-12 13:43:39 +03:00
void ( * lm_withdraw ) ( struct gfs2_sbd * sdp ) ;
2011-01-19 12:30:01 +03:00
void ( * lm_put_lock ) ( struct gfs2_glock * gl ) ;
2010-11-29 15:50:38 +03:00
int ( * lm_lock ) ( struct gfs2_glock * gl , unsigned int req_state ,
unsigned int flags ) ;
2009-01-12 13:43:39 +03:00
void ( * lm_cancel ) ( struct gfs2_glock * gl ) ;
const match_table_t * lm_tokens ;
} ;
2009-07-24 03:52:34 +04:00
extern struct workqueue_struct * gfs2_delete_workqueue ;
2008-02-22 19:07:18 +03:00
static inline struct gfs2_holder * gfs2_glock_is_locked_by_me ( struct gfs2_glock * gl )
2006-01-16 19:50:04 +03:00
{
struct gfs2_holder * gh ;
2008-02-07 11:13:19 +03:00
struct pid * pid ;
2006-01-16 19:50:04 +03:00
/* Look in glock's list of holders for one with current task as owner */
2015-10-29 18:58:09 +03:00
spin_lock ( & gl - > gl_lockref . lock ) ;
2008-02-07 11:13:19 +03:00
pid = task_pid ( current ) ;
2006-01-16 19:50:04 +03:00
list_for_each_entry ( gh , & gl - > gl_holders , gh_list ) {
2008-05-21 20:03:22 +04:00
if ( ! test_bit ( HIF_HOLDER , & gh - > gh_iflags ) )
break ;
2008-02-22 19:07:18 +03:00
if ( gh - > gh_owner_pid = = pid )
goto out ;
2006-01-16 19:50:04 +03:00
}
2008-02-22 19:07:18 +03:00
gh = NULL ;
out :
2015-10-29 18:58:09 +03:00
spin_unlock ( & gl - > gl_lockref . lock ) ;
2006-01-16 19:50:04 +03:00
2008-02-22 19:07:18 +03:00
return gh ;
2006-01-16 19:50:04 +03:00
}
static inline int gfs2_glock_is_held_excl ( struct gfs2_glock * gl )
{
2006-09-04 17:49:55 +04:00
return gl - > gl_state = = LM_ST_EXCLUSIVE ;
2006-01-16 19:50:04 +03:00
}
static inline int gfs2_glock_is_held_dfrd ( struct gfs2_glock * gl )
{
2006-09-04 17:49:55 +04:00
return gl - > gl_state = = LM_ST_DEFERRED ;
2006-01-16 19:50:04 +03:00
}
static inline int gfs2_glock_is_held_shrd ( struct gfs2_glock * gl )
{
2006-09-04 17:49:55 +04:00
return gl - > gl_state = = LM_ST_SHARED ;
2006-01-16 19:50:04 +03:00
}
2009-12-08 15:12:13 +03:00
static inline struct address_space * gfs2_glock2aspace ( struct gfs2_glock * gl )
{
if ( gl - > gl_ops - > go_flags & GLOF_ASPACE )
return ( struct address_space * ) ( gl + 1 ) ;
return NULL ;
}
2012-10-15 13:57:02 +04:00
extern int gfs2_glock_get ( struct gfs2_sbd * sdp , u64 number ,
const struct gfs2_glock_operations * glops ,
int create , struct gfs2_glock * * glp ) ;
2017-08-01 19:45:23 +03:00
extern void gfs2_glock_hold ( struct gfs2_glock * gl ) ;
2012-10-15 13:57:02 +04:00
extern void gfs2_glock_put ( struct gfs2_glock * gl ) ;
2017-08-01 19:45:23 +03:00
extern void gfs2_glock_queue_put ( struct gfs2_glock * gl ) ;
2012-10-15 13:57:02 +04:00
extern void gfs2_holder_init ( struct gfs2_glock * gl , unsigned int state ,
2015-07-24 17:45:43 +03:00
u16 flags , struct gfs2_holder * gh ) ;
extern void gfs2_holder_reinit ( unsigned int state , u16 flags ,
2012-10-15 13:57:02 +04:00
struct gfs2_holder * gh ) ;
extern void gfs2_holder_uninit ( struct gfs2_holder * gh ) ;
extern int gfs2_glock_nq ( struct gfs2_holder * gh ) ;
extern int gfs2_glock_poll ( struct gfs2_holder * gh ) ;
extern int gfs2_glock_wait ( struct gfs2_holder * gh ) ;
extern void gfs2_glock_dq ( struct gfs2_holder * gh ) ;
extern void gfs2_glock_dq_wait ( struct gfs2_holder * gh ) ;
extern void gfs2_glock_dq_uninit ( struct gfs2_holder * gh ) ;
extern int gfs2_glock_nq_num ( struct gfs2_sbd * sdp , u64 number ,
const struct gfs2_glock_operations * glops ,
2015-07-24 17:45:43 +03:00
unsigned int state , u16 flags ,
2012-10-15 13:57:02 +04:00
struct gfs2_holder * gh ) ;
extern int gfs2_glock_nq_m ( unsigned int num_gh , struct gfs2_holder * ghs ) ;
extern void gfs2_glock_dq_m ( unsigned int num_gh , struct gfs2_holder * ghs ) ;
2014-01-16 14:31:13 +04:00
extern void gfs2_dump_glock ( struct seq_file * seq , const struct gfs2_glock * gl ) ;
2012-10-15 13:57:02 +04:00
# define GLOCK_BUG_ON(gl,x) do { if (unlikely(x)) { gfs2_dump_glock(NULL, gl); BUG(); } } while(0)
extern __printf ( 2 , 3 )
2008-05-21 20:03:22 +04:00
void gfs2_print_dbg ( struct seq_file * seq , const char * fmt , . . . ) ;
2006-01-16 19:50:04 +03:00
2006-03-29 23:36:49 +04:00
/**
2010-10-16 17:19:22 +04:00
* gfs2_glock_nq_init - initialize a holder and enqueue it on a glock
2006-03-29 23:36:49 +04:00
* @ gl : the glock
* @ state : the state we ' re requesting
* @ flags : the modifier flags
* @ gh : the holder structure
*
* Returns : 0 , GLR_ * , or errno
*/
static inline int gfs2_glock_nq_init ( struct gfs2_glock * gl ,
2015-07-24 17:45:43 +03:00
unsigned int state , u16 flags ,
2006-03-29 23:36:49 +04:00
struct gfs2_holder * gh )
{
int error ;
gfs2_holder_init ( gl , state , flags , gh ) ;
error = gfs2_glock_nq ( gh ) ;
if ( error )
gfs2_holder_uninit ( gh ) ;
return error ;
}
2011-01-19 12:30:01 +03:00
extern void gfs2_glock_cb ( struct gfs2_glock * gl , unsigned int state ) ;
extern void gfs2_glock_complete ( struct gfs2_glock * gl , int ret ) ;
extern void gfs2_gl_hash_clear ( struct gfs2_sbd * sdp ) ;
extern void gfs2_glock_finish_truncate ( struct gfs2_inode * ip ) ;
extern void gfs2_glock_thaw ( struct gfs2_sbd * sdp ) ;
2011-03-30 19:33:25 +04:00
extern void gfs2_glock_add_to_lru ( struct gfs2_glock * gl ) ;
2011-03-09 13:58:04 +03:00
extern void gfs2_glock_free ( struct gfs2_glock * gl ) ;
2011-01-19 12:30:01 +03:00
extern int __init gfs2_glock_init ( void ) ;
extern void gfs2_glock_exit ( void ) ;
extern int gfs2_create_debugfs_file ( struct gfs2_sbd * sdp ) ;
extern void gfs2_delete_debugfs_file ( struct gfs2_sbd * sdp ) ;
extern int gfs2_register_debugfs ( void ) ;
extern void gfs2_unregister_debugfs ( void ) ;
2006-09-07 22:40:21 +04:00
2009-01-12 13:43:39 +03:00
extern const struct lm_lockops gfs2_dlm_ops ;
2016-06-17 15:31:27 +03:00
static inline void gfs2_holder_mark_uninitialized ( struct gfs2_holder * gh )
{
gh - > gh_gl = NULL ;
}
static inline bool gfs2_holder_initialized ( struct gfs2_holder * gh )
{
return gh - > gh_gl ;
}
2017-07-18 19:35:04 +03:00
/**
* glock_set_object - set the gl_object field of a glock
* @ gl : the glock
* @ object : the object
*/
2017-06-30 15:47:15 +03:00
static inline void glock_set_object ( struct gfs2_glock * gl , void * object )
{
spin_lock ( & gl - > gl_lockref . lock ) ;
2017-07-18 19:35:04 +03:00
if ( gfs2_assert_warn ( gl - > gl_name . ln_sbd , gl - > gl_object = = NULL ) )
gfs2_dump_glock ( NULL , gl ) ;
2017-06-30 15:47:15 +03:00
gl - > gl_object = object ;
spin_unlock ( & gl - > gl_lockref . lock ) ;
}
2017-07-18 19:35:04 +03:00
/**
* glock_clear_object - clear the gl_object field of a glock
* @ gl : the glock
* @ object : the object
*
* I ' d love to similarly add this :
* else if ( gfs2_assert_warn ( gl - > gl_sbd , gl - > gl_object = = object ) )
* gfs2_dump_glock ( NULL , gl ) ;
* Unfortunately , that ' s not possible because as soon as gfs2_delete_inode
* frees the block in the rgrp , another process can reassign it for an I_NEW
* inode in gfs2_create_inode because that calls new_inode , not gfs2_iget .
* That means gfs2_delete_inode may subsequently try to call this function
* for a glock that ' s already pointing to a brand new inode . If we clear the
* new inode ' s gl_object , we ' ll introduce metadata corruption . Function
* gfs2_delete_inode calls clear_inode which calls gfs2_clear_inode which also
* tries to clear gl_object , so it ' s more than just gfs2_delete_inode .
*
*/
static inline void glock_clear_object ( struct gfs2_glock * gl , void * object )
{
spin_lock ( & gl - > gl_lockref . lock ) ;
if ( gl - > gl_object = = object )
gl - > gl_object = NULL ;
spin_unlock ( & gl - > gl_lockref . lock ) ;
}
2006-01-16 19:50:04 +03:00
# endif /* __GLOCK_DOT_H__ */