2019-07-31 18:57:31 +03:00
/* SPDX-License-Identifier: GPL-2.0-only */
/*
2018-07-26 15:21:45 +03:00
* Copyright ( C ) 2017 - 2018 HUAWEI , Inc .
2020-07-13 16:09:44 +03:00
* https : //www.huawei.com/
2018-07-26 15:21:45 +03:00
* Created by Gao Xiang < gaoxiang25 @ huawei . com >
*/
2019-07-31 18:57:36 +03:00
# ifndef __EROFS_INTERNAL_H
# define __EROFS_INTERNAL_H
2018-07-26 15:21:45 +03:00
# include <linux/fs.h>
# include <linux/dcache.h>
# include <linux/mm.h>
# include <linux/pagemap.h>
# include <linux/bio.h>
# include <linux/buffer_head.h>
2019-08-23 00:36:59 +03:00
# include <linux/magic.h>
2018-07-26 15:21:45 +03:00
# include <linux/slab.h>
# include <linux/vmalloc.h>
# include "erofs_fs.h"
/* redefine pr_fmt "erofs: " */
# undef pr_fmt
# define pr_fmt(fmt) "erofs: " fmt
2019-09-04 05:09:09 +03:00
__printf ( 3 , 4 ) void _erofs_err ( struct super_block * sb ,
const char * function , const char * fmt , . . . ) ;
# define erofs_err(sb, fmt, ...) \
_erofs_err ( sb , __func__ , fmt " \n " , # # __VA_ARGS__ )
__printf ( 3 , 4 ) void _erofs_info ( struct super_block * sb ,
const char * function , const char * fmt , . . . ) ;
# define erofs_info(sb, fmt, ...) \
_erofs_info ( sb , __func__ , fmt " \n " , # # __VA_ARGS__ )
2018-07-26 15:21:45 +03:00
# ifdef CONFIG_EROFS_FS_DEBUG
2019-09-04 05:09:09 +03:00
# define erofs_dbg(x, ...) pr_debug(x "\n", ##__VA_ARGS__)
2018-07-26 15:21:45 +03:00
# define DBG_BUGON BUG_ON
# else
2019-09-04 05:09:09 +03:00
# define erofs_dbg(x, ...) ((void)0)
2018-11-22 20:15:59 +03:00
# define DBG_BUGON(x) ((void)(x))
2019-07-31 18:57:36 +03:00
# endif /* !CONFIG_EROFS_FS_DEBUG */
2018-07-26 15:21:45 +03:00
/* EROFS_SUPER_MAGIC_V1 to represent the whole file system */
# define EROFS_SUPER_MAGIC EROFS_SUPER_MAGIC_V1
typedef u64 erofs_nid_t ;
2019-07-31 18:57:36 +03:00
typedef u64 erofs_off_t ;
/* data type for filesystem-wide blocks number */
typedef u32 erofs_blk_t ;
2018-07-26 15:21:45 +03:00
2020-05-29 13:48:36 +03:00
struct erofs_fs_context {
# ifdef CONFIG_EROFS_FS_ZIP
/* current strategy of how to use managed cache */
unsigned char cache_strategy ;
2021-03-17 06:54:48 +03:00
/* strategy of sync decompression (false - auto, true - force on) */
bool readahead_sync_decompress ;
2020-05-29 13:48:36 +03:00
/* threshold for decompression synchronously */
unsigned int max_sync_decompress_pages ;
# endif
unsigned int mount_opt ;
} ;
2018-07-26 15:21:45 +03:00
struct erofs_sb_info {
2019-07-31 18:57:39 +03:00
# ifdef CONFIG_EROFS_FS_ZIP
2018-07-26 15:22:03 +03:00
/* list for all registered superblocks, mainly for shrinker */
struct list_head list ;
2018-07-26 15:22:04 +03:00
struct mutex umount_mutex ;
2018-07-26 15:22:03 +03:00
2020-02-20 05:46:42 +03:00
/* managed XArray arranged in physical block number */
struct xarray managed_pslots ;
2018-07-26 15:22:07 +03:00
2019-07-31 18:57:39 +03:00
unsigned int shrinker_run_no ;
2019-07-31 18:57:49 +03:00
/* pseudo inode to manage cached pages */
struct inode * managed_cache ;
2019-07-31 18:57:39 +03:00
# endif /* CONFIG_EROFS_FS_ZIP */
u32 blocks ;
u32 meta_blkaddr ;
# ifdef CONFIG_EROFS_FS_XATTR
u32 xattr_blkaddr ;
2018-07-26 15:21:58 +03:00
# endif
2018-07-26 15:21:45 +03:00
2019-07-31 18:57:39 +03:00
/* inode slot unit size in bit shift */
unsigned char islotbits ;
2018-07-26 15:21:45 +03:00
u32 build_time_nsec ;
u64 build_time ;
/* what we really care is nid, rather than ino.. */
erofs_nid_t root_nid ;
/* used for statfs, f_files - f_favail */
u64 inos ;
u8 uuid [ 16 ] ; /* 128-bit uuid for volume */
u8 volume_name [ 16 ] ; /* volume name */
2019-11-04 05:49:37 +03:00
u32 feature_compat ;
2019-09-04 05:08:53 +03:00
u32 feature_incompat ;
2019-06-13 11:35:41 +03:00
2020-05-29 13:48:36 +03:00
struct erofs_fs_context ctx ; /* options */
2018-07-26 15:21:45 +03:00
} ;
# define EROFS_SB(sb) ((struct erofs_sb_info *)(sb)->s_fs_info)
# define EROFS_I_SB(inode) ((struct erofs_sb_info *)(inode)->i_sb->s_fs_info)
2018-07-26 15:21:52 +03:00
/* Mount flags set via mount options or defaults */
# define EROFS_MOUNT_XATTR_USER 0x00000010
# define EROFS_MOUNT_POSIX_ACL 0x00000020
2020-05-29 13:48:36 +03:00
# define clear_opt(ctx, option) ((ctx)->mount_opt &= ~EROFS_MOUNT_##option)
# define set_opt(ctx, option) ((ctx)->mount_opt |= EROFS_MOUNT_##option)
# define test_opt(ctx, option) ((ctx)->mount_opt & EROFS_MOUNT_##option)
2018-07-26 15:21:45 +03:00
2019-07-31 18:57:49 +03:00
enum {
EROFS_ZIP_CACHE_DISABLED ,
EROFS_ZIP_CACHE_READAHEAD ,
EROFS_ZIP_CACHE_READAROUND
} ;
2020-05-29 13:48:36 +03:00
# ifdef CONFIG_EROFS_FS_ZIP
2019-07-31 18:57:36 +03:00
# define EROFS_LOCKED_MAGIC (INT_MIN | 0xE0F510CCL)
2018-07-26 15:22:05 +03:00
/* basic unit of the workstation of a super_block */
struct erofs_workgroup {
/* the workgroup index in the workstation */
pgoff_t index ;
/* overall workgroup reference count */
atomic_t refcount ;
} ;
2018-11-22 20:16:02 +03:00
# if defined(CONFIG_SMP)
static inline bool erofs_workgroup_try_to_freeze ( struct erofs_workgroup * grp ,
int val )
2018-07-26 15:22:05 +03:00
{
preempt_disable ( ) ;
2018-11-22 20:16:02 +03:00
if ( val ! = atomic_cmpxchg ( & grp - > refcount , val , EROFS_LOCKED_MAGIC ) ) {
2018-07-26 15:22:05 +03:00
preempt_enable ( ) ;
return false ;
}
return true ;
}
2018-11-22 20:16:02 +03:00
static inline void erofs_workgroup_unfreeze ( struct erofs_workgroup * grp ,
int orig_val )
2018-07-26 15:22:05 +03:00
{
2018-11-22 20:16:03 +03:00
/*
* other observers should notice all modifications
* in the freezing period .
*/
smp_mb ( ) ;
2018-11-22 20:16:02 +03:00
atomic_set ( & grp - > refcount , orig_val ) ;
2018-07-26 15:22:05 +03:00
preempt_enable ( ) ;
}
2018-11-22 20:16:01 +03:00
static inline int erofs_wait_on_workgroup_freezed ( struct erofs_workgroup * grp )
{
return atomic_cond_read_relaxed ( & grp - > refcount ,
VAL ! = EROFS_LOCKED_MAGIC ) ;
}
# else
2018-11-22 20:16:02 +03:00
static inline bool erofs_workgroup_try_to_freeze ( struct erofs_workgroup * grp ,
int val )
{
preempt_disable ( ) ;
/* no need to spin on UP platforms, let's just disable preemption. */
if ( val ! = atomic_read ( & grp - > refcount ) ) {
preempt_enable ( ) ;
return false ;
}
return true ;
}
static inline void erofs_workgroup_unfreeze ( struct erofs_workgroup * grp ,
int orig_val )
{
preempt_enable ( ) ;
}
2018-11-22 20:16:01 +03:00
static inline int erofs_wait_on_workgroup_freezed ( struct erofs_workgroup * grp )
{
int v = atomic_read ( & grp - > refcount ) ;
/* workgroup is never freezed on uniprocessor systems */
DBG_BUGON ( v = = EROFS_LOCKED_MAGIC ) ;
return v ;
}
2019-07-31 18:57:36 +03:00
# endif /* !CONFIG_SMP */
2018-07-26 15:22:07 +03:00
2019-07-31 18:57:36 +03:00
/* hard limit of pages per compressed cluster */
# define Z_EROFS_CLUSTER_MAX_PAGES (CONFIG_EROFS_FS_CLUSTER_PAGE_LIMIT)
# define EROFS_PCPUBUF_NR_PAGES Z_EROFS_CLUSTER_MAX_PAGES
2018-10-09 16:43:53 +03:00
# else
2019-07-31 18:57:36 +03:00
# define EROFS_PCPUBUF_NR_PAGES 0
# endif /* !CONFIG_EROFS_FS_ZIP */
2018-07-26 15:22:05 +03:00
2018-07-26 15:21:45 +03:00
/* we strictly follow PAGE_SIZE and no buffer head yet */
# define LOG_BLOCK_SIZE PAGE_SHIFT
# undef LOG_SECTORS_PER_BLOCK
# define LOG_SECTORS_PER_BLOCK (PAGE_SHIFT - 9)
# undef SECTORS_PER_BLOCK
# define SECTORS_PER_BLOCK (1 << SECTORS_PER_BLOCK)
# define EROFS_BLKSIZ (1 << LOG_BLOCK_SIZE)
# if (EROFS_BLKSIZ % 4096 || !EROFS_BLKSIZ)
# error erofs cannot be used in this platform
# endif
# define ROOT_NID(sb) ((sb)->root_nid)
# define erofs_blknr(addr) ((addr) / EROFS_BLKSIZ)
# define erofs_blkoff(addr) ((addr) % EROFS_BLKSIZ)
# define blknr_to_addr(nr) ((erofs_off_t)(nr) * EROFS_BLKSIZ)
static inline erofs_off_t iloc ( struct erofs_sb_info * sbi , erofs_nid_t nid )
{
return blknr_to_addr ( sbi - > meta_blkaddr ) + ( nid < < sbi - > islotbits ) ;
}
2021-03-29 04:23:05 +03:00
# define EROFS_FEATURE_FUNCS(name, compat, feature) \
static inline bool erofs_sb_has_ # # name ( struct erofs_sb_info * sbi ) \
{ \
return sbi - > feature_ # # compat & EROFS_FEATURE_ # # feature ; \
}
EROFS_FEATURE_FUNCS ( lz4_0padding , incompat , INCOMPAT_LZ4_0PADDING )
EROFS_FEATURE_FUNCS ( sb_chksum , compat , COMPAT_SB_CHKSUM )
2019-02-18 10:19:04 +03:00
/* atomic flag definitions */
2019-09-04 05:08:56 +03:00
# define EROFS_I_EA_INITED_BIT 0
# define EROFS_I_Z_INITED_BIT 1
2019-02-18 10:19:04 +03:00
/* bitlock definitions (arranged in reverse order) */
2019-09-04 05:08:56 +03:00
# define EROFS_I_BL_XATTR_BIT (BITS_PER_LONG - 1)
# define EROFS_I_BL_Z_BIT (BITS_PER_LONG - 2)
2018-07-26 15:21:45 +03:00
2019-09-04 05:08:56 +03:00
struct erofs_inode {
2018-07-26 15:21:45 +03:00
erofs_nid_t nid ;
2019-02-18 10:19:04 +03:00
/* atomic flags (including bitlocks) */
unsigned long flags ;
2018-07-26 15:21:45 +03:00
2019-09-04 05:08:54 +03:00
unsigned char datalayout ;
2018-07-26 15:21:45 +03:00
unsigned char inode_isize ;
unsigned short xattr_isize ;
2019-07-15 15:21:27 +03:00
unsigned int xattr_shared_count ;
unsigned int * xattr_shared_xattrs ;
2018-07-26 15:21:45 +03:00
2019-06-24 10:22:52 +03:00
union {
erofs_blk_t raw_blkaddr ;
# ifdef CONFIG_EROFS_FS_ZIP
struct {
unsigned short z_advise ;
unsigned char z_algorithmtype [ 2 ] ;
unsigned char z_logical_clusterbits ;
unsigned char z_physical_clusterbits [ 2 ] ;
} ;
2019-07-31 18:57:36 +03:00
# endif /* CONFIG_EROFS_FS_ZIP */
2019-06-24 10:22:52 +03:00
} ;
2018-07-26 15:21:45 +03:00
/* the corresponding vfs inode */
struct inode vfs_inode ;
} ;
2019-09-04 05:08:56 +03:00
# define EROFS_I(ptr) \
container_of ( ptr , struct erofs_inode , vfs_inode )
2018-07-26 15:21:45 +03:00
2019-09-04 05:09:05 +03:00
static inline unsigned long erofs_inode_datablocks ( struct inode * inode )
2018-07-26 15:21:45 +03:00
{
/* since i_size cannot be changed */
return DIV_ROUND_UP ( inode - > i_size , EROFS_BLKSIZ ) ;
}
2019-09-04 05:08:54 +03:00
static inline unsigned int erofs_bitrange ( unsigned int value , unsigned int bit ,
unsigned int bits )
{
return ( value > > bit ) & ( ( 1 < < bits ) - 1 ) ;
}
static inline unsigned int erofs_inode_version ( unsigned int value )
2018-07-26 15:21:45 +03:00
{
2019-09-04 05:08:54 +03:00
return erofs_bitrange ( value , EROFS_I_VERSION_BIT ,
EROFS_I_VERSION_BITS ) ;
2018-07-26 15:21:45 +03:00
}
2019-09-04 05:08:54 +03:00
static inline unsigned int erofs_inode_datalayout ( unsigned int value )
2018-07-26 15:21:45 +03:00
{
2019-09-04 05:08:54 +03:00
return erofs_bitrange ( value , EROFS_I_DATALAYOUT_BIT ,
EROFS_I_DATALAYOUT_BITS ) ;
2018-07-26 15:21:45 +03:00
}
extern const struct super_operations erofs_sops ;
extern const struct address_space_operations erofs_raw_access_aops ;
2019-11-08 06:37:33 +03:00
extern const struct address_space_operations z_erofs_aops ;
2018-07-26 15:21:45 +03:00
/*
2021-03-25 10:10:08 +03:00
* Logical to physical block mapping
2018-07-26 15:21:45 +03:00
*
* Different with other file systems , it is used for 2 access modes :
*
* 1 ) RAW access mode :
*
* Users pass a valid ( m_lblk , m_lofs - - usually 0 ) pair ,
* and get the valid m_pblk , m_pofs and the longest m_len ( in bytes ) .
*
* Note that m_lblk in the RAW access mode refers to the number of
* the compressed ondisk block rather than the uncompressed
* in - memory block for the compressed file .
*
* m_pofs equals to m_lofs except for the inline data page .
*
* 2 ) Normal access mode :
*
* If the inode is not compressed , it has no difference with
* the RAW access mode . However , if the inode is compressed ,
* users should pass a valid ( m_lblk , m_lofs ) pair , and get
* the needed m_pblk , m_pofs , m_len to get the compressed data
* and the updated m_lblk , m_lofs which indicates the start
* of the corresponding uncompressed data in the file .
*/
enum {
BH_Zipped = BH_PrivateStart ,
2019-06-24 10:22:58 +03:00
BH_FullMapped ,
2018-07-26 15:21:45 +03:00
} ;
/* Has a disk mapping */
# define EROFS_MAP_MAPPED (1 << BH_Mapped)
/* Located in metadata (could be copied from bd_inode) */
# define EROFS_MAP_META (1 << BH_Meta)
/* The extent has been compressed */
# define EROFS_MAP_ZIPPED (1 << BH_Zipped)
2019-06-24 10:22:58 +03:00
/* The length of extent is full */
# define EROFS_MAP_FULL_MAPPED (1 << BH_FullMapped)
2018-07-26 15:21:45 +03:00
struct erofs_map_blocks {
erofs_off_t m_pa , m_la ;
u64 m_plen , m_llen ;
unsigned int m_flags ;
2019-01-15 04:42:21 +03:00
struct page * mpage ;
2018-07-26 15:21:45 +03:00
} ;
2021-03-25 10:10:08 +03:00
/* Flags used by erofs_map_blocks_flatmode() */
2018-07-26 15:21:45 +03:00
# define EROFS_GET_BLOCKS_RAW 0x0001
2019-06-24 10:22:52 +03:00
/* zmap.c */
2019-01-15 04:42:21 +03:00
# ifdef CONFIG_EROFS_FS_ZIP
2019-06-24 10:22:52 +03:00
int z_erofs_fill_inode ( struct inode * inode ) ;
2019-01-15 04:42:21 +03:00
int z_erofs_map_blocks_iter ( struct inode * inode ,
struct erofs_map_blocks * map ,
int flags ) ;
# else
2019-08-14 13:37:05 +03:00
static inline int z_erofs_fill_inode ( struct inode * inode ) { return - EOPNOTSUPP ; }
2019-01-15 04:42:21 +03:00
static inline int z_erofs_map_blocks_iter ( struct inode * inode ,
struct erofs_map_blocks * map ,
int flags )
{
2019-08-14 13:37:05 +03:00
return - EOPNOTSUPP ;
2019-01-15 04:42:21 +03:00
}
2019-07-31 18:57:36 +03:00
# endif /* !CONFIG_EROFS_FS_ZIP */
2019-01-15 04:42:21 +03:00
2018-07-26 15:21:45 +03:00
/* data.c */
2019-09-04 05:09:03 +03:00
struct page * erofs_get_meta_page ( struct super_block * sb , erofs_blk_t blkaddr ) ;
2018-08-21 17:49:30 +03:00
2018-07-26 15:21:45 +03:00
/* inode.c */
2018-10-09 17:07:13 +03:00
static inline unsigned long erofs_inode_hash ( erofs_nid_t nid )
{
# if BITS_PER_LONG == 32
return ( nid > > 32 ) ^ ( nid & 0xffffffff ) ;
# else
return nid ;
# endif
}
2019-01-14 14:40:24 +03:00
extern const struct inode_operations erofs_generic_iops ;
extern const struct inode_operations erofs_symlink_iops ;
extern const struct inode_operations erofs_fast_symlink_iops ;
2018-07-26 15:21:52 +03:00
2019-01-16 11:59:56 +03:00
struct inode * erofs_iget ( struct super_block * sb , erofs_nid_t nid , bool dir ) ;
2021-01-21 16:19:43 +03:00
int erofs_getattr ( struct user_namespace * mnt_userns , const struct path * path ,
struct kstat * stat , u32 request_mask ,
unsigned int query_flags ) ;
2019-01-16 11:59:56 +03:00
2019-01-14 14:40:24 +03:00
/* namei.c */
extern const struct inode_operations erofs_dir_iops ;
int erofs_namei ( struct inode * dir , struct qstr * name ,
erofs_nid_t * nid , unsigned int * d_type ) ;
/* dir.c */
extern const struct file_operations erofs_dir_fops ;
2019-07-31 18:57:36 +03:00
/* utils.c / zdata.c */
2019-11-21 16:59:54 +03:00
struct page * erofs_allocpage ( struct list_head * pool , gfp_t gfp ) ;
2019-06-24 10:22:53 +03:00
# if (EROFS_PCPUBUF_NR_PAGES > 0)
void * erofs_get_pcpubuf ( unsigned int pagenr ) ;
# define erofs_put_pcpubuf(buf) do { \
( void ) & ( buf ) ; \
preempt_enable ( ) ; \
} while ( 0 )
# else
static inline void * erofs_get_pcpubuf ( unsigned int pagenr )
{
2019-08-14 13:37:05 +03:00
return ERR_PTR ( - EOPNOTSUPP ) ;
2019-06-24 10:22:53 +03:00
}
# define erofs_put_pcpubuf(buf) do {} while (0)
# endif
2019-07-31 18:57:39 +03:00
# ifdef CONFIG_EROFS_FS_ZIP
2019-07-31 18:57:36 +03:00
int erofs_workgroup_put ( struct erofs_workgroup * grp ) ;
struct erofs_workgroup * erofs_find_workgroup ( struct super_block * sb ,
2020-01-02 15:01:16 +03:00
pgoff_t index ) ;
2020-02-20 05:46:42 +03:00
struct erofs_workgroup * erofs_insert_workgroup ( struct super_block * sb ,
struct erofs_workgroup * grp ) ;
2019-07-31 18:57:36 +03:00
void erofs_workgroup_free_rcu ( struct erofs_workgroup * grp ) ;
2019-07-31 18:57:39 +03:00
void erofs_shrinker_register ( struct super_block * sb ) ;
void erofs_shrinker_unregister ( struct super_block * sb ) ;
int __init erofs_init_shrinker ( void ) ;
void erofs_exit_shrinker ( void ) ;
int __init z_erofs_init_zip_subsystem ( void ) ;
void z_erofs_exit_zip_subsystem ( void ) ;
2019-07-31 18:57:36 +03:00
int erofs_try_to_free_all_cached_pages ( struct erofs_sb_info * sbi ,
struct erofs_workgroup * egrp ) ;
int erofs_try_to_free_cached_page ( struct address_space * mapping ,
struct page * page ) ;
2019-07-31 18:57:39 +03:00
# else
static inline void erofs_shrinker_register ( struct super_block * sb ) { }
static inline void erofs_shrinker_unregister ( struct super_block * sb ) { }
static inline int erofs_init_shrinker ( void ) { return 0 ; }
static inline void erofs_exit_shrinker ( void ) { }
static inline int z_erofs_init_zip_subsystem ( void ) { return 0 ; }
static inline void z_erofs_exit_zip_subsystem ( void ) { }
# endif /* !CONFIG_EROFS_FS_ZIP */
2019-01-16 11:59:56 +03:00
2019-08-14 13:37:03 +03:00
# define EFSCORRUPTED EUCLEAN /* Filesystem is corrupted */
2019-07-31 18:57:36 +03:00
# endif /* __EROFS_INTERNAL_H */
2018-07-26 15:21:45 +03:00