for-5.18/alloc-cleanups-2022-03-18
-----BEGIN PGP SIGNATURE----- iQJEBAABCAAuFiEEwPw5LcreJtl1+l5K99NY+ylx4KYFAmI0/7IQHGF4Ym9lQGtl cm5lbC5kawAKCRD301j7KXHgpm+FD/9wmpVQE30Y4Atgw4VygOXwlS6mhNHIVGUV Khd34mwVSh21hjBnyenWJX7KlcVT5YsfEmEz/CNvhxG7QwqjjFuu085n5eImJkHZ yP9N2jLzR5mECs9nZ9FGZO7h6hl8gEX5agR/QndoVBMyA3ho4SsXyHUoDoobgQ+R SuwAB0gvxA72wMGgIPtJ/y3EJxQyja+6kpOQiO+avQgccGXloi44YRLhqUJ5RkCE iJxeiTB4xJt46f2RnU6cS8yy4WqsTY4PLCnM2mpgkQKDfesqA+4wZh1K7TcNKVN1 MntawBwHEtp0JqZxQQqr6PFqqOLoIs22+JJagFxzk00jHy+s10dTL1O8Yj/If9Y2 l7ivguJog9yXPTdJ7THR2x1RTTxYN7p4DfU6L6cw7VBmG51mSpKwePlK15dqUIC9 mpYUgqrR4v1LYXLSpDHqPFRtARKvyzNmmk61qmA3SyWLOXbOuyMYWHYW5ta62S9B MnqaFjrKHrRzI2co8/Kqyv9t5ffb7eUvu9OUAmM5GnEpl/yg8iEUN+Rk55f0AnNa 1ABRUpiB6AOxRZZ/pqUBCB7wtkoyEk/O/Jqax0FrZWTYtdTQebMC8fbfPXCIdc3A HH7e7Kc8mr4QLhTS9oJ/pvNrm4kdkunTZe8CGQmocGiJjU5sWrMFRNbk2N7/fuAi kAYOByEjDA== =PzC6 -----END PGP SIGNATURE----- Merge tag 'for-5.18/alloc-cleanups-2022-03-18' of git://git.kernel.dk/linux-block Pull bio_alloc() cleanups from Jens Axboe: "Filesystem cleanups to pass the bio op to bio_alloc() instead of setting it just before bio submission". * tag 'for-5.18/alloc-cleanups-2022-03-18' of git://git.kernel.dk/linux-block: f2fs: pass the bio operation to bio_alloc_bioset f2fs: don't pass a bio to f2fs_target_device nilfs2: pass the operation to bio_alloc ext4: pass the operation to bio_alloc mpage: pass the operation to bio_alloc
This commit is contained in:
commit
d347ee54a7
@ -371,10 +371,9 @@ void ext4_io_submit(struct ext4_io_submit *io)
|
||||
struct bio *bio = io->io_bio;
|
||||
|
||||
if (bio) {
|
||||
int io_op_flags = io->io_wbc->sync_mode == WB_SYNC_ALL ?
|
||||
REQ_SYNC : 0;
|
||||
if (io->io_wbc->sync_mode == WB_SYNC_ALL)
|
||||
io->io_bio->bi_opf |= REQ_SYNC;
|
||||
io->io_bio->bi_write_hint = io->io_end->inode->i_write_hint;
|
||||
bio_set_op_attrs(io->io_bio, REQ_OP_WRITE, io_op_flags);
|
||||
submit_bio(io->io_bio);
|
||||
}
|
||||
io->io_bio = NULL;
|
||||
@ -397,7 +396,7 @@ static void io_submit_init_bio(struct ext4_io_submit *io,
|
||||
* bio_alloc will _always_ be able to allocate a bio if
|
||||
* __GFP_DIRECT_RECLAIM is set, see comments for bio_alloc_bioset().
|
||||
*/
|
||||
bio = bio_alloc(bh->b_bdev, BIO_MAX_VECS, 0, GFP_NOIO);
|
||||
bio = bio_alloc(bh->b_bdev, BIO_MAX_VECS, REQ_OP_WRITE, GFP_NOIO);
|
||||
fscrypt_set_bio_crypt_ctx_bh(bio, bh, GFP_NOIO);
|
||||
bio->bi_iter.bi_sector = bh->b_blocknr * (bh->b_size >> 9);
|
||||
bio->bi_end_io = ext4_end_bio;
|
||||
|
@ -354,7 +354,7 @@ static void f2fs_write_end_io(struct bio *bio)
|
||||
}
|
||||
|
||||
struct block_device *f2fs_target_device(struct f2fs_sb_info *sbi,
|
||||
block_t blk_addr, struct bio *bio)
|
||||
block_t blk_addr, sector_t *sector)
|
||||
{
|
||||
struct block_device *bdev = sbi->sb->s_bdev;
|
||||
int i;
|
||||
@ -369,10 +369,9 @@ struct block_device *f2fs_target_device(struct f2fs_sb_info *sbi,
|
||||
}
|
||||
}
|
||||
}
|
||||
if (bio) {
|
||||
bio_set_dev(bio, bdev);
|
||||
bio->bi_iter.bi_sector = SECTOR_FROM_BLOCK(blk_addr);
|
||||
}
|
||||
|
||||
if (sector)
|
||||
*sector = SECTOR_FROM_BLOCK(blk_addr);
|
||||
return bdev;
|
||||
}
|
||||
|
||||
@ -389,14 +388,40 @@ int f2fs_target_device_index(struct f2fs_sb_info *sbi, block_t blkaddr)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void __attach_io_flag(struct f2fs_io_info *fio, unsigned int io_flag)
|
||||
{
|
||||
unsigned int temp_mask = (1 << NR_TEMP_TYPE) - 1;
|
||||
unsigned int fua_flag = io_flag & temp_mask;
|
||||
unsigned int meta_flag = (io_flag >> NR_TEMP_TYPE) & temp_mask;
|
||||
|
||||
/*
|
||||
* data/node io flag bits per temp:
|
||||
* REQ_META | REQ_FUA |
|
||||
* 5 | 4 | 3 | 2 | 1 | 0 |
|
||||
* Cold | Warm | Hot | Cold | Warm | Hot |
|
||||
*/
|
||||
if ((1 << fio->temp) & meta_flag)
|
||||
fio->op_flags |= REQ_META;
|
||||
if ((1 << fio->temp) & fua_flag)
|
||||
fio->op_flags |= REQ_FUA;
|
||||
}
|
||||
|
||||
static struct bio *__bio_alloc(struct f2fs_io_info *fio, int npages)
|
||||
{
|
||||
struct f2fs_sb_info *sbi = fio->sbi;
|
||||
struct block_device *bdev;
|
||||
sector_t sector;
|
||||
struct bio *bio;
|
||||
|
||||
bio = bio_alloc_bioset(NULL, npages, 0, GFP_NOIO, &f2fs_bioset);
|
||||
if (fio->type == DATA)
|
||||
__attach_io_flag(fio, sbi->data_io_flag);
|
||||
else if (fio->type == NODE)
|
||||
__attach_io_flag(fio, sbi->node_io_flag);
|
||||
|
||||
f2fs_target_device(sbi, fio->new_blkaddr, bio);
|
||||
bdev = f2fs_target_device(sbi, fio->new_blkaddr, §or);
|
||||
bio = bio_alloc_bioset(bdev, npages, fio->op | fio->op_flags, GFP_NOIO,
|
||||
&f2fs_bioset);
|
||||
bio->bi_iter.bi_sector = sector;
|
||||
if (is_read_io(fio->op)) {
|
||||
bio->bi_end_io = f2fs_read_end_io;
|
||||
bio->bi_private = NULL;
|
||||
@ -500,34 +525,6 @@ void f2fs_submit_bio(struct f2fs_sb_info *sbi,
|
||||
__submit_bio(sbi, bio, type);
|
||||
}
|
||||
|
||||
static void __attach_io_flag(struct f2fs_io_info *fio)
|
||||
{
|
||||
struct f2fs_sb_info *sbi = fio->sbi;
|
||||
unsigned int temp_mask = (1 << NR_TEMP_TYPE) - 1;
|
||||
unsigned int io_flag, fua_flag, meta_flag;
|
||||
|
||||
if (fio->type == DATA)
|
||||
io_flag = sbi->data_io_flag;
|
||||
else if (fio->type == NODE)
|
||||
io_flag = sbi->node_io_flag;
|
||||
else
|
||||
return;
|
||||
|
||||
fua_flag = io_flag & temp_mask;
|
||||
meta_flag = (io_flag >> NR_TEMP_TYPE) & temp_mask;
|
||||
|
||||
/*
|
||||
* data/node io flag bits per temp:
|
||||
* REQ_META | REQ_FUA |
|
||||
* 5 | 4 | 3 | 2 | 1 | 0 |
|
||||
* Cold | Warm | Hot | Cold | Warm | Hot |
|
||||
*/
|
||||
if ((1 << fio->temp) & meta_flag)
|
||||
fio->op_flags |= REQ_META;
|
||||
if ((1 << fio->temp) & fua_flag)
|
||||
fio->op_flags |= REQ_FUA;
|
||||
}
|
||||
|
||||
static void __submit_merged_bio(struct f2fs_bio_info *io)
|
||||
{
|
||||
struct f2fs_io_info *fio = &io->fio;
|
||||
@ -535,9 +532,6 @@ static void __submit_merged_bio(struct f2fs_bio_info *io)
|
||||
if (!io->bio)
|
||||
return;
|
||||
|
||||
__attach_io_flag(fio);
|
||||
bio_set_op_attrs(io->bio, fio->op, fio->op_flags);
|
||||
|
||||
if (is_read_io(fio->op))
|
||||
trace_f2fs_prepare_read_bio(io->sbi->sb, fio->type, io->bio);
|
||||
else
|
||||
@ -595,10 +589,9 @@ static void __f2fs_submit_merged_write(struct f2fs_sb_info *sbi,
|
||||
/* change META to META_FLUSH in the checkpoint procedure */
|
||||
if (type >= META_FLUSH) {
|
||||
io->fio.type = META_FLUSH;
|
||||
io->fio.op = REQ_OP_WRITE;
|
||||
io->fio.op_flags = REQ_META | REQ_PRIO | REQ_SYNC;
|
||||
io->bio->bi_opf |= REQ_META | REQ_PRIO | REQ_SYNC;
|
||||
if (!test_opt(sbi, NOBARRIER))
|
||||
io->fio.op_flags |= REQ_PREFLUSH | REQ_FUA;
|
||||
io->bio->bi_opf |= REQ_PREFLUSH | REQ_FUA;
|
||||
}
|
||||
__submit_merged_bio(io);
|
||||
up_write(&io->io_rwsem);
|
||||
@ -679,9 +672,6 @@ int f2fs_submit_page_bio(struct f2fs_io_info *fio)
|
||||
if (fio->io_wbc && !is_read_io(fio->op))
|
||||
wbc_account_cgroup_owner(fio->io_wbc, page, PAGE_SIZE);
|
||||
|
||||
__attach_io_flag(fio);
|
||||
bio_set_op_attrs(bio, fio->op, fio->op_flags);
|
||||
|
||||
inc_page_count(fio->sbi, is_read_io(fio->op) ?
|
||||
__read_io_type(page): WB_DATA_TYPE(fio->page));
|
||||
|
||||
@ -875,10 +865,8 @@ int f2fs_merge_page_bio(struct f2fs_io_info *fio)
|
||||
alloc_new:
|
||||
if (!bio) {
|
||||
bio = __bio_alloc(fio, BIO_MAX_VECS);
|
||||
__attach_io_flag(fio);
|
||||
f2fs_set_bio_crypt_ctx(bio, fio->page->mapping->host,
|
||||
fio->page->index, fio, GFP_NOIO);
|
||||
bio_set_op_attrs(bio, fio->op, fio->op_flags);
|
||||
|
||||
add_bio_entry(fio->sbi, bio, page, fio->temp);
|
||||
} else {
|
||||
@ -984,15 +972,16 @@ static struct bio *f2fs_grab_read_bio(struct inode *inode, block_t blkaddr,
|
||||
struct bio *bio;
|
||||
struct bio_post_read_ctx *ctx = NULL;
|
||||
unsigned int post_read_steps = 0;
|
||||
sector_t sector;
|
||||
struct block_device *bdev = f2fs_target_device(sbi, blkaddr, §or);
|
||||
|
||||
bio = bio_alloc_bioset(NULL, bio_max_segs(nr_pages), REQ_OP_READ,
|
||||
bio = bio_alloc_bioset(bdev, bio_max_segs(nr_pages),
|
||||
REQ_OP_READ | op_flag,
|
||||
for_write ? GFP_NOIO : GFP_KERNEL, &f2fs_bioset);
|
||||
if (!bio)
|
||||
return ERR_PTR(-ENOMEM);
|
||||
|
||||
bio->bi_iter.bi_sector = sector;
|
||||
f2fs_set_bio_crypt_ctx(bio, inode, first_idx, NULL, GFP_NOFS);
|
||||
|
||||
f2fs_target_device(sbi, blkaddr, bio);
|
||||
bio->bi_end_io = f2fs_read_end_io;
|
||||
|
||||
if (fscrypt_inode_uses_fs_layer_crypto(inode))
|
||||
|
@ -3631,7 +3631,7 @@ int f2fs_submit_page_bio(struct f2fs_io_info *fio);
|
||||
int f2fs_merge_page_bio(struct f2fs_io_info *fio);
|
||||
void f2fs_submit_page_write(struct f2fs_io_info *fio);
|
||||
struct block_device *f2fs_target_device(struct f2fs_sb_info *sbi,
|
||||
block_t blk_addr, struct bio *bio);
|
||||
block_t blk_addr, sector_t *sector);
|
||||
int f2fs_target_device_index(struct f2fs_sb_info *sbi, block_t blkaddr);
|
||||
void f2fs_set_data_blkaddr(struct dnode_of_data *dn);
|
||||
void f2fs_update_data_blkaddr(struct dnode_of_data *dn, block_t blkaddr);
|
||||
|
50
fs/mpage.c
50
fs/mpage.c
@ -57,10 +57,9 @@ static void mpage_end_io(struct bio *bio)
|
||||
bio_put(bio);
|
||||
}
|
||||
|
||||
static struct bio *mpage_bio_submit(int op, int op_flags, struct bio *bio)
|
||||
static struct bio *mpage_bio_submit(struct bio *bio)
|
||||
{
|
||||
bio->bi_end_io = mpage_end_io;
|
||||
bio_set_op_attrs(bio, op, op_flags);
|
||||
guard_bio_eod(bio);
|
||||
submit_bio(bio);
|
||||
return NULL;
|
||||
@ -146,16 +145,15 @@ static struct bio *do_mpage_readpage(struct mpage_readpage_args *args)
|
||||
struct block_device *bdev = NULL;
|
||||
int length;
|
||||
int fully_mapped = 1;
|
||||
int op_flags;
|
||||
int op = REQ_OP_READ;
|
||||
unsigned nblocks;
|
||||
unsigned relative_block;
|
||||
gfp_t gfp;
|
||||
|
||||
if (args->is_readahead) {
|
||||
op_flags = REQ_RAHEAD;
|
||||
op |= REQ_RAHEAD;
|
||||
gfp = readahead_gfp_mask(page->mapping);
|
||||
} else {
|
||||
op_flags = 0;
|
||||
gfp = mapping_gfp_constraint(page->mapping, GFP_KERNEL);
|
||||
}
|
||||
|
||||
@ -264,7 +262,7 @@ static struct bio *do_mpage_readpage(struct mpage_readpage_args *args)
|
||||
* This page will go to BIO. Do we need to send this BIO off first?
|
||||
*/
|
||||
if (args->bio && (args->last_block_in_bio != blocks[0] - 1))
|
||||
args->bio = mpage_bio_submit(REQ_OP_READ, op_flags, args->bio);
|
||||
args->bio = mpage_bio_submit(args->bio);
|
||||
|
||||
alloc_new:
|
||||
if (args->bio == NULL) {
|
||||
@ -273,7 +271,7 @@ alloc_new:
|
||||
page))
|
||||
goto out;
|
||||
}
|
||||
args->bio = bio_alloc(bdev, bio_max_segs(args->nr_pages), 0,
|
||||
args->bio = bio_alloc(bdev, bio_max_segs(args->nr_pages), op,
|
||||
gfp);
|
||||
if (args->bio == NULL)
|
||||
goto confused;
|
||||
@ -282,7 +280,7 @@ alloc_new:
|
||||
|
||||
length = first_hole << blkbits;
|
||||
if (bio_add_page(args->bio, page, length, 0) < length) {
|
||||
args->bio = mpage_bio_submit(REQ_OP_READ, op_flags, args->bio);
|
||||
args->bio = mpage_bio_submit(args->bio);
|
||||
goto alloc_new;
|
||||
}
|
||||
|
||||
@ -290,7 +288,7 @@ alloc_new:
|
||||
nblocks = map_bh->b_size >> blkbits;
|
||||
if ((buffer_boundary(map_bh) && relative_block == nblocks) ||
|
||||
(first_hole != blocks_per_page))
|
||||
args->bio = mpage_bio_submit(REQ_OP_READ, op_flags, args->bio);
|
||||
args->bio = mpage_bio_submit(args->bio);
|
||||
else
|
||||
args->last_block_in_bio = blocks[blocks_per_page - 1];
|
||||
out:
|
||||
@ -298,7 +296,7 @@ out:
|
||||
|
||||
confused:
|
||||
if (args->bio)
|
||||
args->bio = mpage_bio_submit(REQ_OP_READ, op_flags, args->bio);
|
||||
args->bio = mpage_bio_submit(args->bio);
|
||||
if (!PageUptodate(page))
|
||||
block_read_full_page(page, args->get_block);
|
||||
else
|
||||
@ -361,7 +359,7 @@ void mpage_readahead(struct readahead_control *rac, get_block_t get_block)
|
||||
put_page(page);
|
||||
}
|
||||
if (args.bio)
|
||||
mpage_bio_submit(REQ_OP_READ, REQ_RAHEAD, args.bio);
|
||||
mpage_bio_submit(args.bio);
|
||||
}
|
||||
EXPORT_SYMBOL(mpage_readahead);
|
||||
|
||||
@ -378,7 +376,7 @@ int mpage_readpage(struct page *page, get_block_t get_block)
|
||||
|
||||
args.bio = do_mpage_readpage(&args);
|
||||
if (args.bio)
|
||||
mpage_bio_submit(REQ_OP_READ, 0, args.bio);
|
||||
mpage_bio_submit(args.bio);
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL(mpage_readpage);
|
||||
@ -469,7 +467,6 @@ static int __mpage_writepage(struct page *page, struct writeback_control *wbc,
|
||||
struct buffer_head map_bh;
|
||||
loff_t i_size = i_size_read(inode);
|
||||
int ret = 0;
|
||||
int op_flags = wbc_to_write_flags(wbc);
|
||||
|
||||
if (page_has_buffers(page)) {
|
||||
struct buffer_head *head = page_buffers(page);
|
||||
@ -577,7 +574,7 @@ page_is_mapped:
|
||||
* This page will go to BIO. Do we need to send this BIO off first?
|
||||
*/
|
||||
if (bio && mpd->last_block_in_bio != blocks[0] - 1)
|
||||
bio = mpage_bio_submit(REQ_OP_WRITE, op_flags, bio);
|
||||
bio = mpage_bio_submit(bio);
|
||||
|
||||
alloc_new:
|
||||
if (bio == NULL) {
|
||||
@ -586,9 +583,10 @@ alloc_new:
|
||||
page, wbc))
|
||||
goto out;
|
||||
}
|
||||
bio = bio_alloc(bdev, BIO_MAX_VECS, 0, GFP_NOFS);
|
||||
bio = bio_alloc(bdev, BIO_MAX_VECS,
|
||||
REQ_OP_WRITE | wbc_to_write_flags(wbc),
|
||||
GFP_NOFS);
|
||||
bio->bi_iter.bi_sector = blocks[0] << (blkbits - 9);
|
||||
|
||||
wbc_init_bio(wbc, bio);
|
||||
bio->bi_write_hint = inode->i_write_hint;
|
||||
}
|
||||
@ -601,7 +599,7 @@ alloc_new:
|
||||
wbc_account_cgroup_owner(wbc, page, PAGE_SIZE);
|
||||
length = first_unmapped << blkbits;
|
||||
if (bio_add_page(bio, page, length, 0) < length) {
|
||||
bio = mpage_bio_submit(REQ_OP_WRITE, op_flags, bio);
|
||||
bio = mpage_bio_submit(bio);
|
||||
goto alloc_new;
|
||||
}
|
||||
|
||||
@ -611,7 +609,7 @@ alloc_new:
|
||||
set_page_writeback(page);
|
||||
unlock_page(page);
|
||||
if (boundary || (first_unmapped != blocks_per_page)) {
|
||||
bio = mpage_bio_submit(REQ_OP_WRITE, op_flags, bio);
|
||||
bio = mpage_bio_submit(bio);
|
||||
if (boundary_block) {
|
||||
write_boundary_block(boundary_bdev,
|
||||
boundary_block, 1 << blkbits);
|
||||
@ -623,7 +621,7 @@ alloc_new:
|
||||
|
||||
confused:
|
||||
if (bio)
|
||||
bio = mpage_bio_submit(REQ_OP_WRITE, op_flags, bio);
|
||||
bio = mpage_bio_submit(bio);
|
||||
|
||||
if (mpd->use_writepage) {
|
||||
ret = mapping->a_ops->writepage(page, wbc);
|
||||
@ -679,11 +677,8 @@ mpage_writepages(struct address_space *mapping,
|
||||
};
|
||||
|
||||
ret = write_cache_pages(mapping, wbc, __mpage_writepage, &mpd);
|
||||
if (mpd.bio) {
|
||||
int op_flags = (wbc->sync_mode == WB_SYNC_ALL ?
|
||||
REQ_SYNC : 0);
|
||||
mpage_bio_submit(REQ_OP_WRITE, op_flags, mpd.bio);
|
||||
}
|
||||
if (mpd.bio)
|
||||
mpage_bio_submit(mpd.bio);
|
||||
}
|
||||
blk_finish_plug(&plug);
|
||||
return ret;
|
||||
@ -700,11 +695,8 @@ int mpage_writepage(struct page *page, get_block_t get_block,
|
||||
.use_writepage = 0,
|
||||
};
|
||||
int ret = __mpage_writepage(page, wbc, &mpd);
|
||||
if (mpd.bio) {
|
||||
int op_flags = (wbc->sync_mode == WB_SYNC_ALL ?
|
||||
REQ_SYNC : 0);
|
||||
mpage_bio_submit(REQ_OP_WRITE, op_flags, mpd.bio);
|
||||
}
|
||||
if (mpd.bio)
|
||||
mpage_bio_submit(mpd.bio);
|
||||
return ret;
|
||||
}
|
||||
EXPORT_SYMBOL(mpage_writepage);
|
||||
|
@ -337,8 +337,7 @@ static void nilfs_end_bio_write(struct bio *bio)
|
||||
}
|
||||
|
||||
static int nilfs_segbuf_submit_bio(struct nilfs_segment_buffer *segbuf,
|
||||
struct nilfs_write_info *wi, int mode,
|
||||
int mode_flags)
|
||||
struct nilfs_write_info *wi)
|
||||
{
|
||||
struct bio *bio = wi->bio;
|
||||
int err;
|
||||
@ -356,7 +355,6 @@ static int nilfs_segbuf_submit_bio(struct nilfs_segment_buffer *segbuf,
|
||||
|
||||
bio->bi_end_io = nilfs_end_bio_write;
|
||||
bio->bi_private = segbuf;
|
||||
bio_set_op_attrs(bio, mode, mode_flags);
|
||||
submit_bio(bio);
|
||||
segbuf->sb_nbio++;
|
||||
|
||||
@ -384,15 +382,15 @@ static void nilfs_segbuf_prepare_write(struct nilfs_segment_buffer *segbuf,
|
||||
|
||||
static int nilfs_segbuf_submit_bh(struct nilfs_segment_buffer *segbuf,
|
||||
struct nilfs_write_info *wi,
|
||||
struct buffer_head *bh, int mode)
|
||||
struct buffer_head *bh)
|
||||
{
|
||||
int len, err;
|
||||
|
||||
BUG_ON(wi->nr_vecs <= 0);
|
||||
repeat:
|
||||
if (!wi->bio) {
|
||||
wi->bio = bio_alloc(wi->nilfs->ns_bdev, wi->nr_vecs, 0,
|
||||
GFP_NOIO);
|
||||
wi->bio = bio_alloc(wi->nilfs->ns_bdev, wi->nr_vecs,
|
||||
REQ_OP_WRITE, GFP_NOIO);
|
||||
wi->bio->bi_iter.bi_sector = (wi->blocknr + wi->end) <<
|
||||
(wi->nilfs->ns_blocksize_bits - 9);
|
||||
}
|
||||
@ -403,7 +401,7 @@ static int nilfs_segbuf_submit_bh(struct nilfs_segment_buffer *segbuf,
|
||||
return 0;
|
||||
}
|
||||
/* bio is FULL */
|
||||
err = nilfs_segbuf_submit_bio(segbuf, wi, mode, 0);
|
||||
err = nilfs_segbuf_submit_bio(segbuf, wi);
|
||||
/* never submit current bh */
|
||||
if (likely(!err))
|
||||
goto repeat;
|
||||
@ -433,13 +431,13 @@ static int nilfs_segbuf_write(struct nilfs_segment_buffer *segbuf,
|
||||
nilfs_segbuf_prepare_write(segbuf, &wi);
|
||||
|
||||
list_for_each_entry(bh, &segbuf->sb_segsum_buffers, b_assoc_buffers) {
|
||||
res = nilfs_segbuf_submit_bh(segbuf, &wi, bh, REQ_OP_WRITE);
|
||||
res = nilfs_segbuf_submit_bh(segbuf, &wi, bh);
|
||||
if (unlikely(res))
|
||||
goto failed_bio;
|
||||
}
|
||||
|
||||
list_for_each_entry(bh, &segbuf->sb_payload_buffers, b_assoc_buffers) {
|
||||
res = nilfs_segbuf_submit_bh(segbuf, &wi, bh, REQ_OP_WRITE);
|
||||
res = nilfs_segbuf_submit_bh(segbuf, &wi, bh);
|
||||
if (unlikely(res))
|
||||
goto failed_bio;
|
||||
}
|
||||
@ -449,8 +447,8 @@ static int nilfs_segbuf_write(struct nilfs_segment_buffer *segbuf,
|
||||
* Last BIO is always sent through the following
|
||||
* submission.
|
||||
*/
|
||||
res = nilfs_segbuf_submit_bio(segbuf, &wi, REQ_OP_WRITE,
|
||||
REQ_SYNC);
|
||||
wi.bio->bi_opf |= REQ_SYNC;
|
||||
res = nilfs_segbuf_submit_bio(segbuf, &wi);
|
||||
}
|
||||
|
||||
failed_bio:
|
||||
|
Loading…
x
Reference in New Issue
Block a user