986e9842fb
This allows including a compression level when specifying a compression type, e.g. compression=zstd:15 Values from 1 through 15 indicate compression levels, 0 or unspecified indicates the default. For LZ4, values 3-15 specify that the HC algorithm should be used. Note that for compatibility, extents themselves only include the compression type, not the compression level. This means that specifying the same compression algorithm but different compression levels for the compression and background_compression options will have no effect. XXX: perhaps we could add a warning for this Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
203 lines
5.4 KiB
C
203 lines
5.4 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef _BCACHEFS_IO_H
|
|
#define _BCACHEFS_IO_H
|
|
|
|
#include "checksum.h"
|
|
#include "bkey_buf.h"
|
|
#include "io_types.h"
|
|
|
|
#define to_wbio(_bio) \
|
|
container_of((_bio), struct bch_write_bio, bio)
|
|
|
|
#define to_rbio(_bio) \
|
|
container_of((_bio), struct bch_read_bio, bio)
|
|
|
|
void bch2_bio_free_pages_pool(struct bch_fs *, struct bio *);
|
|
void bch2_bio_alloc_pages_pool(struct bch_fs *, struct bio *, size_t);
|
|
|
|
#ifndef CONFIG_BCACHEFS_NO_LATENCY_ACCT
|
|
void bch2_latency_acct(struct bch_dev *, u64, int);
|
|
#else
|
|
static inline void bch2_latency_acct(struct bch_dev *ca, u64 submit_time, int rw) {}
|
|
#endif
|
|
|
|
void bch2_submit_wbio_replicas(struct bch_write_bio *, struct bch_fs *,
|
|
enum bch_data_type, const struct bkey_i *, bool);
|
|
|
|
#define BLK_STS_REMOVED ((__force blk_status_t)128)
|
|
|
|
const char *bch2_blk_status_to_str(blk_status_t);
|
|
|
|
#define BCH_WRITE_FLAGS() \
|
|
x(ALLOC_NOWAIT) \
|
|
x(CACHED) \
|
|
x(DATA_ENCODED) \
|
|
x(PAGES_STABLE) \
|
|
x(PAGES_OWNED) \
|
|
x(ONLY_SPECIFIED_DEVS) \
|
|
x(WROTE_DATA_INLINE) \
|
|
x(FROM_INTERNAL) \
|
|
x(CHECK_ENOSPC) \
|
|
x(SYNC) \
|
|
x(MOVE) \
|
|
x(IN_WORKER) \
|
|
x(DONE) \
|
|
x(IO_ERROR) \
|
|
x(CONVERT_UNWRITTEN)
|
|
|
|
enum __bch_write_flags {
|
|
#define x(f) __BCH_WRITE_##f,
|
|
BCH_WRITE_FLAGS()
|
|
#undef x
|
|
};
|
|
|
|
enum bch_write_flags {
|
|
#define x(f) BCH_WRITE_##f = 1U << __BCH_WRITE_##f,
|
|
BCH_WRITE_FLAGS()
|
|
#undef x
|
|
};
|
|
|
|
static inline struct workqueue_struct *index_update_wq(struct bch_write_op *op)
|
|
{
|
|
return op->watermark == BCH_WATERMARK_copygc
|
|
? op->c->copygc_wq
|
|
: op->c->btree_update_wq;
|
|
}
|
|
|
|
int bch2_sum_sector_overwrites(struct btree_trans *, struct btree_iter *,
|
|
struct bkey_i *, bool *, s64 *, s64 *);
|
|
int bch2_extent_update(struct btree_trans *, subvol_inum,
|
|
struct btree_iter *, struct bkey_i *,
|
|
struct disk_reservation *, u64, s64 *, bool);
|
|
int bch2_extent_fallocate(struct btree_trans *, subvol_inum, struct btree_iter *,
|
|
unsigned, struct bch_io_opts, s64 *,
|
|
struct write_point_specifier);
|
|
|
|
int bch2_fpunch_at(struct btree_trans *, struct btree_iter *,
|
|
subvol_inum, u64, s64 *);
|
|
int bch2_fpunch(struct bch_fs *c, subvol_inum, u64, u64, s64 *);
|
|
|
|
static inline void bch2_write_op_init(struct bch_write_op *op, struct bch_fs *c,
|
|
struct bch_io_opts opts)
|
|
{
|
|
op->c = c;
|
|
op->end_io = NULL;
|
|
op->flags = 0;
|
|
op->written = 0;
|
|
op->error = 0;
|
|
op->csum_type = bch2_data_checksum_type(c, opts);
|
|
op->compression_opt = opts.compression;
|
|
op->nr_replicas = 0;
|
|
op->nr_replicas_required = c->opts.data_replicas_required;
|
|
op->watermark = BCH_WATERMARK_normal;
|
|
op->incompressible = 0;
|
|
op->open_buckets.nr = 0;
|
|
op->devs_have.nr = 0;
|
|
op->target = 0;
|
|
op->opts = opts;
|
|
op->subvol = 0;
|
|
op->pos = POS_MAX;
|
|
op->version = ZERO_VERSION;
|
|
op->write_point = (struct write_point_specifier) { 0 };
|
|
op->res = (struct disk_reservation) { 0 };
|
|
op->new_i_size = U64_MAX;
|
|
op->i_sectors_delta = 0;
|
|
op->devs_need_flush = NULL;
|
|
}
|
|
|
|
void bch2_write(struct closure *);
|
|
|
|
void bch2_write_point_do_index_updates(struct work_struct *);
|
|
|
|
static inline struct bch_write_bio *wbio_init(struct bio *bio)
|
|
{
|
|
struct bch_write_bio *wbio = to_wbio(bio);
|
|
|
|
memset(&wbio->wbio, 0, sizeof(wbio->wbio));
|
|
return wbio;
|
|
}
|
|
|
|
void bch2_write_op_to_text(struct printbuf *, struct bch_write_op *);
|
|
|
|
struct bch_devs_mask;
|
|
struct cache_promote_op;
|
|
struct extent_ptr_decoded;
|
|
|
|
int __bch2_read_indirect_extent(struct btree_trans *, unsigned *,
|
|
struct bkey_buf *);
|
|
|
|
static inline int bch2_read_indirect_extent(struct btree_trans *trans,
|
|
enum btree_id *data_btree,
|
|
unsigned *offset_into_extent,
|
|
struct bkey_buf *k)
|
|
{
|
|
if (k->k->k.type != KEY_TYPE_reflink_p)
|
|
return 0;
|
|
|
|
*data_btree = BTREE_ID_reflink;
|
|
return __bch2_read_indirect_extent(trans, offset_into_extent, k);
|
|
}
|
|
|
|
enum bch_read_flags {
|
|
BCH_READ_RETRY_IF_STALE = 1 << 0,
|
|
BCH_READ_MAY_PROMOTE = 1 << 1,
|
|
BCH_READ_USER_MAPPED = 1 << 2,
|
|
BCH_READ_NODECODE = 1 << 3,
|
|
BCH_READ_LAST_FRAGMENT = 1 << 4,
|
|
|
|
/* internal: */
|
|
BCH_READ_MUST_BOUNCE = 1 << 5,
|
|
BCH_READ_MUST_CLONE = 1 << 6,
|
|
BCH_READ_IN_RETRY = 1 << 7,
|
|
};
|
|
|
|
int __bch2_read_extent(struct btree_trans *, struct bch_read_bio *,
|
|
struct bvec_iter, struct bpos, enum btree_id,
|
|
struct bkey_s_c, unsigned,
|
|
struct bch_io_failures *, unsigned);
|
|
|
|
static inline void bch2_read_extent(struct btree_trans *trans,
|
|
struct bch_read_bio *rbio, struct bpos read_pos,
|
|
enum btree_id data_btree, struct bkey_s_c k,
|
|
unsigned offset_into_extent, unsigned flags)
|
|
{
|
|
__bch2_read_extent(trans, rbio, rbio->bio.bi_iter, read_pos,
|
|
data_btree, k, offset_into_extent, NULL, flags);
|
|
}
|
|
|
|
void __bch2_read(struct bch_fs *, struct bch_read_bio *, struct bvec_iter,
|
|
subvol_inum, struct bch_io_failures *, unsigned flags);
|
|
|
|
static inline void bch2_read(struct bch_fs *c, struct bch_read_bio *rbio,
|
|
subvol_inum inum)
|
|
{
|
|
struct bch_io_failures failed = { .nr = 0 };
|
|
|
|
BUG_ON(rbio->_state);
|
|
|
|
rbio->c = c;
|
|
rbio->start_time = local_clock();
|
|
rbio->subvol = inum.subvol;
|
|
|
|
__bch2_read(c, rbio, rbio->bio.bi_iter, inum, &failed,
|
|
BCH_READ_RETRY_IF_STALE|
|
|
BCH_READ_MAY_PROMOTE|
|
|
BCH_READ_USER_MAPPED);
|
|
}
|
|
|
|
static inline struct bch_read_bio *rbio_init(struct bio *bio,
|
|
struct bch_io_opts opts)
|
|
{
|
|
struct bch_read_bio *rbio = to_rbio(bio);
|
|
|
|
rbio->_state = 0;
|
|
rbio->promote = NULL;
|
|
rbio->opts = opts;
|
|
return rbio;
|
|
}
|
|
|
|
void bch2_fs_io_exit(struct bch_fs *);
|
|
int bch2_fs_io_init(struct bch_fs *);
|
|
|
|
#endif /* _BCACHEFS_IO_H */
|