2022-04-25 15:21:33 +03:00
// SPDX-License-Identifier: GPL-2.0-or-later
/*
* Copyright ( C ) 2022 , Alibaba Cloud
2022-09-18 07:34:53 +03:00
* Copyright ( C ) 2022 , Bytedance Inc . All rights reserved .
2022-04-25 15:21:33 +03:00
*/
# include <linux/fscache.h>
# include "internal.h"
2022-09-18 07:34:53 +03:00
static DEFINE_MUTEX ( erofs_domain_list_lock ) ;
2022-09-18 14:01:50 +03:00
static DEFINE_MUTEX ( erofs_domain_cookies_lock ) ;
2022-09-18 07:34:53 +03:00
static LIST_HEAD ( erofs_domain_list ) ;
2023-02-09 09:39:11 +03:00
static LIST_HEAD ( erofs_domain_cookies_list ) ;
2022-09-18 07:34:54 +03:00
static struct vfsmount * erofs_pseudo_mnt ;
2022-09-18 07:34:53 +03:00
2022-11-24 06:42:12 +03:00
struct erofs_fscache_request {
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
struct erofs_fscache_request * primary ;
2022-11-24 06:42:12 +03:00
struct netfs_cache_resources cache_resources ;
struct address_space * mapping ; /* The mapping being accessed */
loff_t start ; /* Start position */
size_t len ; /* Length of the request */
size_t submitted ; /* Length of submitted */
short error ; /* 0 or error that occurred */
refcount_t ref ;
} ;
static struct erofs_fscache_request * erofs_fscache_req_alloc ( struct address_space * mapping ,
2022-05-09 10:40:28 +03:00
loff_t start , size_t len )
{
2022-11-24 06:42:12 +03:00
struct erofs_fscache_request * req ;
2022-05-09 10:40:28 +03:00
2022-11-24 06:42:12 +03:00
req = kzalloc ( sizeof ( struct erofs_fscache_request ) , GFP_KERNEL ) ;
if ( ! req )
2022-05-09 10:40:28 +03:00
return ERR_PTR ( - ENOMEM ) ;
2022-11-24 06:42:12 +03:00
req - > mapping = mapping ;
req - > start = start ;
req - > len = len ;
refcount_set ( & req - > ref , 1 ) ;
2022-05-09 10:40:28 +03:00
2022-11-24 06:42:12 +03:00
return req ;
2022-05-09 10:40:28 +03:00
}
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
static struct erofs_fscache_request * erofs_fscache_req_chain ( struct erofs_fscache_request * primary ,
size_t len )
2022-05-09 10:40:28 +03:00
{
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
struct erofs_fscache_request * req ;
2022-05-09 10:40:28 +03:00
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
/* use primary request for the first submission */
if ( ! primary - > submitted ) {
refcount_inc ( & primary - > ref ) ;
return primary ;
}
2022-05-09 10:40:28 +03:00
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
req = erofs_fscache_req_alloc ( primary - > mapping ,
primary - > start + primary - > submitted , len ) ;
if ( ! IS_ERR ( req ) ) {
req - > primary = primary ;
refcount_inc ( & primary - > ref ) ;
2022-05-09 10:40:28 +03:00
}
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
return req ;
2022-05-09 10:40:28 +03:00
}
2022-11-24 06:42:12 +03:00
static void erofs_fscache_req_complete ( struct erofs_fscache_request * req )
2022-05-09 10:40:28 +03:00
{
struct folio * folio ;
2022-11-24 06:42:12 +03:00
bool failed = req - > error ;
pgoff_t start_page = req - > start / PAGE_SIZE ;
pgoff_t last_page = ( ( req - > start + req - > len ) / PAGE_SIZE ) - 1 ;
2022-05-09 10:40:28 +03:00
2022-11-24 06:42:12 +03:00
XA_STATE ( xas , & req - > mapping - > i_pages , start_page ) ;
2022-05-09 10:40:28 +03:00
rcu_read_lock ( ) ;
xas_for_each ( & xas , folio , last_page ) {
2022-11-14 15:19:43 +03:00
if ( xas_retry ( & xas , folio ) )
continue ;
2022-11-24 06:42:12 +03:00
if ( ! failed )
2022-05-09 10:40:28 +03:00
folio_mark_uptodate ( folio ) ;
folio_unlock ( folio ) ;
}
rcu_read_unlock ( ) ;
}
2022-11-24 06:42:12 +03:00
static void erofs_fscache_req_put ( struct erofs_fscache_request * req )
2022-05-09 10:40:28 +03:00
{
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
if ( refcount_dec_and_test ( & req - > ref ) ) {
if ( req - > cache_resources . ops )
req - > cache_resources . ops - > end_operation ( & req - > cache_resources ) ;
if ( ! req - > primary )
erofs_fscache_req_complete ( req ) ;
else
erofs_fscache_req_put ( req - > primary ) ;
kfree ( req ) ;
}
2022-05-09 10:40:28 +03:00
}
2022-11-24 06:42:12 +03:00
static void erofs_fscache_subreq_complete ( void * priv ,
2022-05-09 10:40:28 +03:00
ssize_t transferred_or_error , bool was_async )
{
2022-11-24 06:42:12 +03:00
struct erofs_fscache_request * req = priv ;
2022-05-09 10:40:28 +03:00
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
if ( IS_ERR_VALUE ( transferred_or_error ) ) {
if ( req - > primary )
req - > primary - > error = transferred_or_error ;
else
req - > error = transferred_or_error ;
}
2022-11-24 06:42:12 +03:00
erofs_fscache_req_put ( req ) ;
2022-05-09 10:40:28 +03:00
}
2022-04-25 15:21:36 +03:00
/*
2022-11-24 06:42:12 +03:00
* Read data from fscache ( cookie , pstart , len ) , and fill the read data into
* page cache described by ( req - > mapping , lstart , len ) . @ pstart describeis the
* start physical address in the cache file .
2022-04-25 15:21:36 +03:00
*/
2022-05-09 10:40:28 +03:00
static int erofs_fscache_read_folios_async ( struct fscache_cookie * cookie ,
2022-11-24 06:42:12 +03:00
struct erofs_fscache_request * req , loff_t pstart , size_t len )
2022-04-25 15:21:36 +03:00
{
enum netfs_io_source source ;
2022-11-24 06:42:12 +03:00
struct super_block * sb = req - > mapping - > host - > i_sb ;
struct netfs_cache_resources * cres = & req - > cache_resources ;
2022-04-25 15:21:36 +03:00
struct iov_iter iter ;
2022-11-24 06:42:12 +03:00
loff_t lstart = req - > start + req - > submitted ;
2022-04-25 15:21:36 +03:00
size_t done = 0 ;
int ret ;
2022-11-24 06:42:12 +03:00
DBG_BUGON ( len > req - > len - req - > submitted ) ;
2022-05-09 10:40:28 +03:00
2022-04-25 15:21:36 +03:00
ret = fscache_begin_read_operation ( cres , cookie ) ;
if ( ret )
2022-11-24 06:42:12 +03:00
return ret ;
2022-04-25 15:21:36 +03:00
while ( done < len ) {
2022-11-24 06:42:12 +03:00
loff_t sstart = pstart + done ;
size_t slen = len - done ;
unsigned long flags = 1 < < NETFS_SREQ_ONDEMAND ;
2022-05-09 10:40:28 +03:00
2022-11-24 06:42:12 +03:00
source = cres - > ops - > prepare_ondemand_read ( cres ,
sstart , & slen , LLONG_MAX , & flags , 0 ) ;
if ( WARN_ON ( slen = = 0 ) )
2022-04-25 15:21:36 +03:00
source = NETFS_INVALID_READ ;
if ( source ! = NETFS_READ_FROM_CACHE ) {
2022-11-24 06:42:12 +03:00
erofs_err ( sb , " failed to fscache prepare_read (source %d) " , source ) ;
return - EIO ;
2022-04-25 15:21:36 +03:00
}
2022-11-24 06:42:12 +03:00
refcount_inc ( & req - > ref ) ;
Changes since the last update:
- Enable large folios for iomap/fscache mode;
- Avoid sysfs warning due to mounting twice with the same fsid and
domain_id in fscache mode;
- Refine fscache interface among erofs, fscache, and cachefiles;
- Use kmap_local_page() only for metabuf;
- Fixes around crafted images found by syzbot;
- Minor cleanups and documentation updates.
-----BEGIN PGP SIGNATURE-----
iIcEABYIAC8WIQThPAmQN9sSA0DVxtI5NzHcH7XmBAUCY5S3khEceGlhbmdAa2Vy
bmVsLm9yZwAKCRA5NzHcH7XmBLr3AQDA5xpztSsxfe0Gp+bwf12ySuntimJxXmAj
83EHCfSC+AEAu4fcWkIF38MBBVJvFVjFaXCZKmFossbI5Rp8TuqPpgk=
=HDsJ
-----END PGP SIGNATURE-----
Merge tag 'erofs-for-6.2-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/xiang/erofs
Pull erofs updates from Gao Xiang:
"In this cycle, large folios are now enabled in the iomap/fscache mode
for uncompressed files first. In order to do that, we've also cleaned
up better interfaces between erofs and fscache, which are acked by
fscache/netfs folks and included in this pull request.
Other than that, there are random fixes around erofs over fscache and
crafted images by syzbot, minor cleanups and documentation updates.
Summary:
- Enable large folios for iomap/fscache mode
- Avoid sysfs warning due to mounting twice with the same fsid and
domain_id in fscache mode
- Refine fscache interface among erofs, fscache, and cachefiles
- Use kmap_local_page() only for metabuf
- Fixes around crafted images found by syzbot
- Minor cleanups and documentation updates"
* tag 'erofs-for-6.2-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/xiang/erofs:
erofs: validate the extent length for uncompressed pclusters
erofs: fix missing unmap if z_erofs_get_extent_compressedlen() fails
erofs: Fix pcluster memleak when its block address is zero
erofs: use kmap_local_page() only for erofs_bread()
erofs: enable large folios for fscache mode
erofs: support large folios for fscache mode
erofs: switch to prepare_ondemand_read() in fscache mode
fscache,cachefiles: add prepare_ondemand_read() callback
erofs: clean up cached I/O strategies
erofs: update documentation
erofs: check the uniqueness of fsid in shared domain in advance
erofs: enable large folios for iomap mode
2022-12-13 07:14:04 +03:00
iov_iter_xarray ( & iter , ITER_DEST , & req - > mapping - > i_pages ,
2022-11-24 06:42:12 +03:00
lstart + done , slen ) ;
2022-05-09 10:40:28 +03:00
2022-11-24 06:42:12 +03:00
ret = fscache_read ( cres , sstart , & iter , NETFS_READ_HOLE_FAIL ,
erofs_fscache_subreq_complete , req ) ;
2022-05-09 10:40:28 +03:00
if ( ret = = - EIOCBQUEUED )
ret = 0 ;
2022-04-25 15:21:36 +03:00
if ( ret ) {
erofs_err ( sb , " failed to fscache_read (ret %d) " , ret ) ;
2022-11-24 06:42:12 +03:00
return ret ;
2022-04-25 15:21:36 +03:00
}
2022-11-24 06:42:12 +03:00
done + = slen ;
2022-04-25 15:21:36 +03:00
}
2022-11-24 06:42:12 +03:00
DBG_BUGON ( done ! = len ) ;
return 0 ;
2022-04-25 15:21:36 +03:00
}
2022-05-25 05:55:07 +03:00
static int erofs_fscache_meta_read_folio ( struct file * data , struct folio * folio )
2022-04-25 15:21:39 +03:00
{
int ret ;
2023-02-09 09:39:10 +03:00
struct erofs_fscache * ctx = folio_mapping ( folio ) - > host - > i_private ;
2022-11-24 06:42:12 +03:00
struct erofs_fscache_request * req ;
2022-04-25 15:21:39 +03:00
2022-11-24 06:42:12 +03:00
req = erofs_fscache_req_alloc ( folio_mapping ( folio ) ,
2022-05-09 10:40:28 +03:00
folio_pos ( folio ) , folio_size ( folio ) ) ;
2022-11-24 06:42:12 +03:00
if ( IS_ERR ( req ) ) {
folio_unlock ( folio ) ;
return PTR_ERR ( req ) ;
2022-08-15 06:48:29 +03:00
}
2022-05-09 10:40:28 +03:00
2023-02-09 09:39:10 +03:00
ret = erofs_fscache_read_folios_async ( ctx - > cookie , req ,
folio_pos ( folio ) , folio_size ( folio ) ) ;
2022-11-24 06:42:12 +03:00
if ( ret )
req - > error = ret ;
erofs_fscache_req_put ( req ) ;
2022-04-25 15:21:39 +03:00
return ret ;
}
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
static int erofs_fscache_data_read_slice ( struct erofs_fscache_request * primary )
2022-04-25 15:21:40 +03:00
{
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
struct address_space * mapping = primary - > mapping ;
2022-09-22 09:24:14 +03:00
struct inode * inode = mapping - > host ;
2022-04-25 15:21:40 +03:00
struct super_block * sb = inode - > i_sb ;
2022-11-24 06:42:12 +03:00
struct erofs_fscache_request * req ;
2022-04-25 15:21:40 +03:00
struct erofs_map_blocks map ;
struct erofs_map_dev mdev ;
2022-09-22 09:24:14 +03:00
struct iov_iter iter ;
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
loff_t pos = primary - > start + primary - > submitted ;
2022-09-22 09:24:14 +03:00
size_t count ;
2022-04-25 15:21:40 +03:00
int ret ;
map . m_la = pos ;
2023-02-09 05:48:25 +03:00
ret = erofs_map_blocks ( inode , & map ) ;
2022-04-25 15:21:40 +03:00
if ( ret )
2022-09-22 09:24:14 +03:00
return ret ;
2022-04-25 15:21:40 +03:00
2022-09-22 09:24:14 +03:00
if ( map . m_flags & EROFS_MAP_META ) {
struct erofs_buf buf = __EROFS_BUF_INITIALIZER ;
erofs_blk_t blknr ;
size_t offset , size ;
void * src ;
/* For tail packing layout, the offset may be non-zero. */
offset = erofs_blkoff ( map . m_pa ) ;
blknr = erofs_blknr ( map . m_pa ) ;
size = map . m_llen ;
src = erofs_read_metabuf ( & buf , sb , blknr , EROFS_KMAP ) ;
if ( IS_ERR ( src ) )
return PTR_ERR ( src ) ;
2022-09-16 03:25:47 +03:00
iov_iter_xarray ( & iter , ITER_DEST , & mapping - > i_pages , pos , PAGE_SIZE ) ;
2022-11-04 08:40:27 +03:00
if ( copy_to_iter ( src + offset , size , & iter ) ! = size ) {
erofs_put_metabuf ( & buf ) ;
2022-09-22 09:24:14 +03:00
return - EFAULT ;
2022-11-04 08:40:27 +03:00
}
2022-09-22 09:24:14 +03:00
iov_iter_zero ( PAGE_SIZE - size , & iter ) ;
erofs_put_metabuf ( & buf ) ;
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
primary - > submitted + = PAGE_SIZE ;
return 0 ;
2022-04-25 15:21:40 +03:00
}
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
count = primary - > len - primary - > submitted ;
2022-09-22 09:24:14 +03:00
if ( ! ( map . m_flags & EROFS_MAP_MAPPED ) ) {
2022-09-16 03:25:47 +03:00
iov_iter_xarray ( & iter , ITER_DEST , & mapping - > i_pages , pos , count ) ;
2022-09-22 09:24:14 +03:00
iov_iter_zero ( count , & iter ) ;
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
primary - > submitted + = count ;
return 0 ;
2022-04-25 15:21:41 +03:00
}
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
count = min_t ( size_t , map . m_llen - ( pos - map . m_la ) , count ) ;
2022-11-04 08:40:28 +03:00
DBG_BUGON ( ! count | | count % PAGE_SIZE ) ;
2022-04-25 15:21:40 +03:00
mdev = ( struct erofs_map_dev ) {
. m_deviceid = map . m_deviceid ,
. m_pa = map . m_pa ,
} ;
ret = erofs_map_dev ( sb , & mdev ) ;
if ( ret )
2022-09-22 09:24:14 +03:00
return ret ;
2022-05-09 10:40:28 +03:00
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
req = erofs_fscache_req_chain ( primary , count ) ;
2022-11-24 06:42:12 +03:00
if ( IS_ERR ( req ) )
return PTR_ERR ( req ) ;
2022-04-25 15:21:40 +03:00
2022-11-24 06:42:12 +03:00
ret = erofs_fscache_read_folios_async ( mdev . m_fscache - > cookie ,
req , mdev . m_pa + ( pos - map . m_la ) , count ) ;
erofs_fscache_req_put ( req ) ;
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
primary - > submitted + = count ;
return ret ;
2022-04-25 15:21:40 +03:00
}
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
static int erofs_fscache_data_read ( struct erofs_fscache_request * req )
2022-04-25 15:21:42 +03:00
{
2022-09-22 09:24:14 +03:00
int ret ;
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
do {
ret = erofs_fscache_data_read_slice ( req ) ;
if ( ret )
req - > error = ret ;
} while ( ! ret & & req - > submitted < req - > len ) ;
2022-09-22 09:24:14 +03:00
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
return ret ;
}
static int erofs_fscache_read_folio ( struct file * file , struct folio * folio )
{
struct erofs_fscache_request * req ;
int ret ;
req = erofs_fscache_req_alloc ( folio_mapping ( folio ) ,
folio_pos ( folio ) , folio_size ( folio ) ) ;
if ( IS_ERR ( req ) ) {
2022-09-22 09:24:14 +03:00
folio_unlock ( folio ) ;
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
return PTR_ERR ( req ) ;
2022-04-25 15:21:42 +03:00
}
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
ret = erofs_fscache_data_read ( req ) ;
erofs_fscache_req_put ( req ) ;
return ret ;
2022-04-25 15:21:42 +03:00
}
static void erofs_fscache_readahead ( struct readahead_control * rac )
{
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
struct erofs_fscache_request * req ;
2022-04-25 15:21:42 +03:00
if ( ! readahead_count ( rac ) )
return ;
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
req = erofs_fscache_req_alloc ( rac - > mapping ,
readahead_pos ( rac ) , readahead_length ( rac ) ) ;
if ( IS_ERR ( req ) )
return ;
2022-04-25 15:21:42 +03:00
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
/* The request completion will drop refs on the folios. */
while ( readahead_folio ( rac ) )
;
2022-04-25 15:21:42 +03:00
erofs: support large folios for fscache mode
When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.
Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.
As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources. Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests. Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
2022-12-01 10:42:55 +03:00
erofs_fscache_data_read ( req ) ;
erofs_fscache_req_put ( req ) ;
2022-04-25 15:21:42 +03:00
}
2022-04-25 15:21:35 +03:00
static const struct address_space_operations erofs_fscache_meta_aops = {
2022-05-25 05:55:07 +03:00
. read_folio = erofs_fscache_meta_read_folio ,
2022-04-25 15:21:35 +03:00
} ;
2022-04-25 15:21:40 +03:00
const struct address_space_operations erofs_fscache_access_aops = {
2022-05-25 05:55:07 +03:00
. read_folio = erofs_fscache_read_folio ,
2022-04-25 15:21:42 +03:00
. readahead = erofs_fscache_readahead ,
2022-04-25 15:21:40 +03:00
} ;
2022-09-18 07:34:53 +03:00
static void erofs_fscache_domain_put ( struct erofs_domain * domain )
{
mutex_lock ( & erofs_domain_list_lock ) ;
if ( refcount_dec_and_test ( & domain - > ref ) ) {
list_del ( & domain - > list ) ;
2022-09-18 07:34:54 +03:00
if ( list_empty ( & erofs_domain_list ) ) {
kern_unmount ( erofs_pseudo_mnt ) ;
erofs_pseudo_mnt = NULL ;
}
2022-09-18 07:34:53 +03:00
mutex_unlock ( & erofs_domain_list_lock ) ;
fscache_relinquish_volume ( domain - > volume , NULL , false ) ;
kfree ( domain - > domain_id ) ;
kfree ( domain ) ;
return ;
}
mutex_unlock ( & erofs_domain_list_lock ) ;
}
static int erofs_fscache_register_volume ( struct super_block * sb )
{
struct erofs_sb_info * sbi = EROFS_SB ( sb ) ;
2022-10-21 05:31:53 +03:00
char * domain_id = sbi - > domain_id ;
2022-09-18 07:34:53 +03:00
struct fscache_volume * volume ;
char * name ;
int ret = 0 ;
name = kasprintf ( GFP_KERNEL , " erofs,%s " ,
2022-10-21 05:31:53 +03:00
domain_id ? domain_id : sbi - > fsid ) ;
2022-09-18 07:34:53 +03:00
if ( ! name )
return - ENOMEM ;
volume = fscache_acquire_volume ( name , NULL , NULL , 0 ) ;
if ( IS_ERR_OR_NULL ( volume ) ) {
erofs_err ( sb , " failed to register volume for %s " , name ) ;
ret = volume ? PTR_ERR ( volume ) : - EOPNOTSUPP ;
volume = NULL ;
}
sbi - > volume = volume ;
kfree ( name ) ;
return ret ;
}
static int erofs_fscache_init_domain ( struct super_block * sb )
{
int err ;
struct erofs_domain * domain ;
struct erofs_sb_info * sbi = EROFS_SB ( sb ) ;
domain = kzalloc ( sizeof ( struct erofs_domain ) , GFP_KERNEL ) ;
if ( ! domain )
return - ENOMEM ;
2022-10-21 05:31:53 +03:00
domain - > domain_id = kstrdup ( sbi - > domain_id , GFP_KERNEL ) ;
2022-09-18 07:34:53 +03:00
if ( ! domain - > domain_id ) {
kfree ( domain ) ;
return - ENOMEM ;
}
err = erofs_fscache_register_volume ( sb ) ;
if ( err )
goto out ;
2022-09-18 07:34:54 +03:00
if ( ! erofs_pseudo_mnt ) {
erofs_pseudo_mnt = kern_mount ( & erofs_fs_type ) ;
if ( IS_ERR ( erofs_pseudo_mnt ) ) {
err = PTR_ERR ( erofs_pseudo_mnt ) ;
goto out ;
}
}
2022-09-18 07:34:53 +03:00
domain - > volume = sbi - > volume ;
refcount_set ( & domain - > ref , 1 ) ;
list_add ( & domain - > list , & erofs_domain_list ) ;
sbi - > domain = domain ;
return 0 ;
out :
kfree ( domain - > domain_id ) ;
kfree ( domain ) ;
return err ;
}
static int erofs_fscache_register_domain ( struct super_block * sb )
{
int err ;
struct erofs_domain * domain ;
struct erofs_sb_info * sbi = EROFS_SB ( sb ) ;
mutex_lock ( & erofs_domain_list_lock ) ;
list_for_each_entry ( domain , & erofs_domain_list , list ) {
2022-10-21 05:31:53 +03:00
if ( ! strcmp ( domain - > domain_id , sbi - > domain_id ) ) {
2022-09-18 07:34:53 +03:00
sbi - > domain = domain ;
sbi - > volume = domain - > volume ;
refcount_inc ( & domain - > ref ) ;
mutex_unlock ( & erofs_domain_list_lock ) ;
return 0 ;
}
}
err = erofs_fscache_init_domain ( sb ) ;
mutex_unlock ( & erofs_domain_list_lock ) ;
return err ;
}
2022-09-18 14:01:50 +03:00
static
struct erofs_fscache * erofs_fscache_acquire_cookie ( struct super_block * sb ,
2022-11-25 14:08:22 +03:00
char * name ,
unsigned int flags )
2022-04-25 15:21:34 +03:00
{
struct fscache_volume * volume = EROFS_SB ( sb ) - > volume ;
struct erofs_fscache * ctx ;
struct fscache_cookie * cookie ;
2022-04-25 15:21:35 +03:00
int ret ;
2022-04-25 15:21:34 +03:00
ctx = kzalloc ( sizeof ( * ctx ) , GFP_KERNEL ) ;
if ( ! ctx )
2022-09-18 07:34:52 +03:00
return ERR_PTR ( - ENOMEM ) ;
2023-02-09 09:39:11 +03:00
INIT_LIST_HEAD ( & ctx - > node ) ;
refcount_set ( & ctx - > ref , 1 ) ;
2022-04-25 15:21:34 +03:00
cookie = fscache_acquire_cookie ( volume , FSCACHE_ADV_WANT_CACHE_SIZE ,
name , strlen ( name ) , NULL , 0 , 0 ) ;
if ( ! cookie ) {
erofs_err ( sb , " failed to get cookie for %s " , name ) ;
2022-04-25 15:21:35 +03:00
ret = - EINVAL ;
goto err ;
2022-04-25 15:21:34 +03:00
}
fscache_use_cookie ( cookie , false ) ;
ctx - > cookie = cookie ;
2022-11-25 14:08:22 +03:00
if ( flags & EROFS_REG_COOKIE_NEED_INODE ) {
2022-04-25 15:21:35 +03:00
struct inode * const inode = new_inode ( sb ) ;
if ( ! inode ) {
erofs_err ( sb , " failed to get anon inode for %s " , name ) ;
ret = - ENOMEM ;
goto err_cookie ;
}
set_nlink ( inode , 1 ) ;
inode - > i_size = OFFSET_MAX ;
inode - > i_mapping - > a_ops = & erofs_fscache_meta_aops ;
mapping_set_gfp_mask ( inode - > i_mapping , GFP_NOFS ) ;
2023-02-09 09:39:10 +03:00
inode - > i_private = ctx ;
2022-04-25 15:21:35 +03:00
ctx - > inode = inode ;
}
2022-09-18 07:34:52 +03:00
return ctx ;
2022-04-25 15:21:35 +03:00
err_cookie :
fscache_unuse_cookie ( ctx - > cookie , NULL , NULL ) ;
fscache_relinquish_cookie ( ctx - > cookie , false ) ;
err :
kfree ( ctx ) ;
2022-09-18 07:34:52 +03:00
return ERR_PTR ( ret ) ;
2022-04-25 15:21:34 +03:00
}
2022-09-18 14:01:50 +03:00
static void erofs_fscache_relinquish_cookie ( struct erofs_fscache * ctx )
2022-04-25 15:21:34 +03:00
{
fscache_unuse_cookie ( ctx - > cookie , NULL , NULL ) ;
fscache_relinquish_cookie ( ctx - > cookie , false ) ;
2022-04-25 15:21:35 +03:00
iput ( ctx - > inode ) ;
2023-02-09 09:39:11 +03:00
iput ( ctx - > anon_inode ) ;
2022-09-18 14:01:50 +03:00
kfree ( ctx - > name ) ;
2022-04-25 15:21:34 +03:00
kfree ( ctx ) ;
}
2022-09-18 14:01:50 +03:00
static
struct erofs_fscache * erofs_fscache_domain_init_cookie ( struct super_block * sb ,
2022-11-25 14:08:22 +03:00
char * name ,
unsigned int flags )
2022-09-18 14:01:50 +03:00
{
int err ;
struct inode * inode ;
struct erofs_fscache * ctx ;
struct erofs_domain * domain = EROFS_SB ( sb ) - > domain ;
2022-11-25 14:08:22 +03:00
ctx = erofs_fscache_acquire_cookie ( sb , name , flags ) ;
2022-09-18 14:01:50 +03:00
if ( IS_ERR ( ctx ) )
return ctx ;
ctx - > name = kstrdup ( name , GFP_KERNEL ) ;
if ( ! ctx - > name ) {
err = - ENOMEM ;
goto out ;
}
inode = new_inode ( erofs_pseudo_mnt - > mnt_sb ) ;
if ( ! inode ) {
err = - ENOMEM ;
goto out ;
}
ctx - > domain = domain ;
ctx - > anon_inode = inode ;
2023-02-09 09:39:11 +03:00
list_add ( & ctx - > node , & erofs_domain_cookies_list ) ;
2022-09-18 14:01:50 +03:00
inode - > i_private = ctx ;
refcount_inc ( & domain - > ref ) ;
return ctx ;
out :
erofs_fscache_relinquish_cookie ( ctx ) ;
return ERR_PTR ( err ) ;
}
static
struct erofs_fscache * erofs_domain_register_cookie ( struct super_block * sb ,
2022-11-25 14:08:22 +03:00
char * name ,
unsigned int flags )
2022-09-18 14:01:50 +03:00
{
struct erofs_fscache * ctx ;
struct erofs_domain * domain = EROFS_SB ( sb ) - > domain ;
mutex_lock ( & erofs_domain_cookies_lock ) ;
2023-02-09 09:39:11 +03:00
list_for_each_entry ( ctx , & erofs_domain_cookies_list , node ) {
if ( ctx - > domain ! = domain | | strcmp ( ctx - > name , name ) )
2022-09-18 14:01:50 +03:00
continue ;
2022-11-25 14:08:22 +03:00
if ( ! ( flags & EROFS_REG_COOKIE_NEED_NOEXIST ) ) {
2023-02-09 09:39:11 +03:00
refcount_inc ( & ctx - > ref ) ;
2022-11-25 14:08:22 +03:00
} else {
erofs_err ( sb , " %s already exists in domain %s " , name ,
domain - > domain_id ) ;
ctx = ERR_PTR ( - EEXIST ) ;
}
2022-09-18 14:01:50 +03:00
mutex_unlock ( & erofs_domain_cookies_lock ) ;
return ctx ;
}
2022-11-25 14:08:22 +03:00
ctx = erofs_fscache_domain_init_cookie ( sb , name , flags ) ;
2022-09-18 14:01:50 +03:00
mutex_unlock ( & erofs_domain_cookies_lock ) ;
return ctx ;
}
struct erofs_fscache * erofs_fscache_register_cookie ( struct super_block * sb ,
2022-11-25 14:08:22 +03:00
char * name ,
unsigned int flags )
2022-09-18 14:01:50 +03:00
{
2022-10-21 05:31:53 +03:00
if ( EROFS_SB ( sb ) - > domain_id )
2022-11-25 14:08:22 +03:00
return erofs_domain_register_cookie ( sb , name , flags ) ;
return erofs_fscache_acquire_cookie ( sb , name , flags ) ;
2022-09-18 14:01:50 +03:00
}
void erofs_fscache_unregister_cookie ( struct erofs_fscache * ctx )
{
2023-02-09 09:39:11 +03:00
struct erofs_domain * domain = NULL ;
2022-09-18 14:01:50 +03:00
if ( ! ctx )
return ;
2023-02-09 09:39:11 +03:00
if ( ! ctx - > domain )
return erofs_fscache_relinquish_cookie ( ctx ) ;
2022-09-18 14:01:50 +03:00
2023-02-09 09:39:11 +03:00
mutex_lock ( & erofs_domain_cookies_lock ) ;
if ( refcount_dec_and_test ( & ctx - > ref ) ) {
domain = ctx - > domain ;
list_del ( & ctx - > node ) ;
erofs_fscache_relinquish_cookie ( ctx ) ;
}
mutex_unlock ( & erofs_domain_cookies_lock ) ;
if ( domain )
erofs_fscache_domain_put ( domain ) ;
2022-09-18 14:01:50 +03:00
}
2022-04-25 15:21:33 +03:00
int erofs_fscache_register_fs ( struct super_block * sb )
{
2022-09-18 07:34:53 +03:00
int ret ;
2022-04-25 15:21:33 +03:00
struct erofs_sb_info * sbi = EROFS_SB ( sb ) ;
2022-09-18 07:34:52 +03:00
struct erofs_fscache * fscache ;
2022-11-25 14:08:22 +03:00
unsigned int flags ;
2022-04-25 15:21:33 +03:00
2022-10-21 05:31:53 +03:00
if ( sbi - > domain_id )
2022-09-18 07:34:53 +03:00
ret = erofs_fscache_register_domain ( sb ) ;
else
ret = erofs_fscache_register_volume ( sb ) ;
if ( ret )
return ret ;
2022-09-18 07:34:52 +03:00
2022-11-25 14:08:22 +03:00
/*
* When shared domain is enabled , using NEED_NOEXIST to guarantee
* the primary data blob ( aka fsid ) is unique in the shared domain .
*
* For non - shared - domain case , fscache_acquire_volume ( ) invoked by
* erofs_fscache_register_volume ( ) has already guaranteed
* the uniqueness of primary data blob .
*
* Acquired domain / volume will be relinquished in kill_sb ( ) on error .
*/
flags = EROFS_REG_COOKIE_NEED_INODE ;
if ( sbi - > domain_id )
flags | = EROFS_REG_COOKIE_NEED_NOEXIST ;
fscache = erofs_fscache_register_cookie ( sb , sbi - > fsid , flags ) ;
2022-09-18 07:34:52 +03:00
if ( IS_ERR ( fscache ) )
return PTR_ERR ( fscache ) ;
sbi - > s_fscache = fscache ;
return 0 ;
2022-04-25 15:21:33 +03:00
}
void erofs_fscache_unregister_fs ( struct super_block * sb )
{
struct erofs_sb_info * sbi = EROFS_SB ( sb ) ;
2022-09-18 07:34:52 +03:00
erofs_fscache_unregister_cookie ( sbi - > s_fscache ) ;
2022-09-18 07:34:53 +03:00
if ( sbi - > domain )
erofs_fscache_domain_put ( sbi - > domain ) ;
else
fscache_relinquish_volume ( sbi - > volume , NULL , false ) ;
2022-09-18 07:34:52 +03:00
sbi - > s_fscache = NULL ;
2022-04-25 15:21:33 +03:00
sbi - > volume = NULL ;
2022-09-18 07:34:53 +03:00
sbi - > domain = NULL ;
2022-04-25 15:21:33 +03:00
}