mirror of
git://sourceware.org/git/lvm2.git
synced 2024-12-21 13:34:40 +03:00
1570e76233
Add a "device index" (di) for each device, and use this in the bcache api to the rest of lvm. This replaces the file descriptor (fd) in the api. The rest of lvm uses new functions bcache_set_fd(), bcache_clear_fd(), and bcache_change_fd() to control which fd bcache uses for io to a particular device. . lvm opens a dev and gets and fd. fd = open(dev); . lvm passes fd to the bcache layer and gets a di to use in the bcache api for the dev. di = bcache_set_fd(fd); . lvm uses bcache functions, passing di for the dev. bcache_write_bytes(di, ...), etc. . bcache translates di to fd to do io. . lvm closes the device and clears the di/fd bcache state. close(fd); bcache_clear_fd(di); In the bcache layer, a di-to-fd translation table (int *_fd_table) is added. When bcache needs to perform io on a di, it uses _fd_table[di]. In the following commit, lvm will make use of the new bcache_change_fd() function to change the fd that bcache uses for the dev, without dropping cached blocks.
288 lines
7.2 KiB
C
288 lines
7.2 KiB
C
/*
|
|
* Copyright (C) 2018 Red Hat, Inc. All rights reserved.
|
|
*
|
|
* This file is part of LVM2.
|
|
*
|
|
* This copyrighted material is made available to anyone wishing to use,
|
|
* modify, copy, or redistribute it subject to the terms and conditions
|
|
* of the GNU Lesser General Public License v.2.1.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public License
|
|
* along with this program; if not, write to the Free Software Foundation,
|
|
* Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "lib/device/bcache.h"
|
|
|
|
// FIXME: need to define this in a common place (that doesn't pull in deps)
|
|
#ifndef SECTOR_SHIFT
|
|
#define SECTOR_SHIFT 9
|
|
#endif
|
|
|
|
//----------------------------------------------------------------
|
|
|
|
static void byte_range_to_block_range(struct bcache *cache, uint64_t start, size_t len,
|
|
block_address *bb, block_address *be)
|
|
{
|
|
block_address block_size = bcache_block_sectors(cache) << SECTOR_SHIFT;
|
|
*bb = start / block_size;
|
|
*be = (start + len + block_size - 1) / block_size;
|
|
}
|
|
|
|
static uint64_t _min(uint64_t lhs, uint64_t rhs)
|
|
{
|
|
if (rhs < lhs)
|
|
return rhs;
|
|
|
|
return lhs;
|
|
}
|
|
|
|
//----------------------------------------------------------------
|
|
|
|
void bcache_prefetch_bytes(struct bcache *cache, int di, uint64_t start, size_t len)
|
|
{
|
|
block_address bb, be;
|
|
|
|
byte_range_to_block_range(cache, start, len, &bb, &be);
|
|
while (bb < be) {
|
|
bcache_prefetch(cache, di, bb);
|
|
bb++;
|
|
}
|
|
}
|
|
|
|
//----------------------------------------------------------------
|
|
|
|
bool bcache_read_bytes(struct bcache *cache, int di, uint64_t start, size_t len, void *data)
|
|
{
|
|
struct block *b;
|
|
block_address bb, be;
|
|
uint64_t block_size = bcache_block_sectors(cache) << SECTOR_SHIFT;
|
|
uint64_t block_offset = start % block_size;
|
|
|
|
bcache_prefetch_bytes(cache, di, start, len);
|
|
|
|
byte_range_to_block_range(cache, start, len, &bb, &be);
|
|
|
|
for (; bb != be; bb++) {
|
|
if (!bcache_get(cache, di, bb, 0, &b))
|
|
return false;
|
|
|
|
size_t blen = _min(block_size - block_offset, len);
|
|
memcpy(data, ((unsigned char *) b->data) + block_offset, blen);
|
|
bcache_put(b);
|
|
|
|
block_offset = 0;
|
|
len -= blen;
|
|
data = ((unsigned char *) data) + blen;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool bcache_invalidate_bytes(struct bcache *cache, int di, uint64_t start, size_t len)
|
|
{
|
|
block_address bb, be;
|
|
bool result = true;
|
|
|
|
byte_range_to_block_range(cache, start, len, &bb, &be);
|
|
|
|
for (; bb != be; bb++) {
|
|
if (!bcache_invalidate(cache, di, bb))
|
|
result = false;
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
//----------------------------------------------------------------
|
|
|
|
// Writing bytes and zeroing bytes are very similar, so we factor out
|
|
// this common code.
|
|
|
|
struct updater;
|
|
|
|
typedef bool (*partial_update_fn)(struct updater *u, int di, block_address bb, uint64_t offset, size_t len);
|
|
typedef bool (*whole_update_fn)(struct updater *u, int di, block_address bb, block_address be);
|
|
|
|
struct updater {
|
|
struct bcache *cache;
|
|
partial_update_fn partial_fn;
|
|
whole_update_fn whole_fn;
|
|
void *data;
|
|
};
|
|
|
|
static bool _update_bytes(struct updater *u, int di, uint64_t start, size_t len)
|
|
{
|
|
struct bcache *cache = u->cache;
|
|
block_address bb, be;
|
|
uint64_t block_size = bcache_block_sectors(cache) << SECTOR_SHIFT;
|
|
uint64_t block_offset = start % block_size;
|
|
uint64_t nr_whole;
|
|
|
|
byte_range_to_block_range(cache, start, len, &bb, &be);
|
|
|
|
// If the last block is partial, we will require a read, so let's
|
|
// prefetch it.
|
|
if ((start + len) % block_size)
|
|
bcache_prefetch(cache, di, (start + len) / block_size);
|
|
|
|
// First block may be partial
|
|
if (block_offset) {
|
|
size_t blen = _min(block_size - block_offset, len);
|
|
if (!u->partial_fn(u, di, bb, block_offset, blen))
|
|
return false;
|
|
|
|
len -= blen;
|
|
if (!len)
|
|
return true;
|
|
|
|
bb++;
|
|
}
|
|
|
|
// Now we write out a set of whole blocks
|
|
nr_whole = len / block_size;
|
|
if (!u->whole_fn(u, di, bb, bb + nr_whole))
|
|
return false;
|
|
|
|
bb += nr_whole;
|
|
len -= nr_whole * block_size;
|
|
|
|
if (!len)
|
|
return true;
|
|
|
|
// Finally we write a partial end block
|
|
return u->partial_fn(u, di, bb, 0, len);
|
|
}
|
|
|
|
//----------------------------------------------------------------
|
|
|
|
static bool _write_partial(struct updater *u, int di, block_address bb,
|
|
uint64_t offset, size_t len)
|
|
{
|
|
struct block *b;
|
|
|
|
if (!bcache_get(u->cache, di, bb, GF_DIRTY, &b))
|
|
return false;
|
|
|
|
memcpy(((unsigned char *) b->data) + offset, u->data, len);
|
|
u->data = ((unsigned char *) u->data) + len;
|
|
|
|
bcache_put(b);
|
|
return true;
|
|
}
|
|
|
|
static bool _write_whole(struct updater *u, int di, block_address bb, block_address be)
|
|
{
|
|
struct block *b;
|
|
uint64_t block_size = bcache_block_sectors(u->cache) << SECTOR_SHIFT;
|
|
|
|
for (; bb != be; bb++) {
|
|
// We don't need to read the block since we are overwriting
|
|
// it completely.
|
|
if (!bcache_get(u->cache, di, bb, GF_ZERO, &b))
|
|
return false;
|
|
memcpy(b->data, u->data, block_size);
|
|
u->data = ((unsigned char *) u->data) + block_size;
|
|
bcache_put(b);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool bcache_write_bytes(struct bcache *cache, int di, uint64_t start, size_t len, void *data)
|
|
{
|
|
struct updater u;
|
|
|
|
u.cache = cache;
|
|
u.partial_fn = _write_partial;
|
|
u.whole_fn = _write_whole;
|
|
u.data = data;
|
|
|
|
return _update_bytes(&u, di, start, len);
|
|
}
|
|
|
|
//----------------------------------------------------------------
|
|
|
|
static bool _zero_partial(struct updater *u, int di, block_address bb, uint64_t offset, size_t len)
|
|
{
|
|
struct block *b;
|
|
|
|
if (!bcache_get(u->cache, di, bb, GF_DIRTY, &b))
|
|
return false;
|
|
|
|
memset(((unsigned char *) b->data) + offset, 0, len);
|
|
bcache_put(b);
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool _zero_whole(struct updater *u, int di, block_address bb, block_address be)
|
|
{
|
|
struct block *b;
|
|
|
|
for (; bb != be; bb++) {
|
|
if (!bcache_get(u->cache, di, bb, GF_ZERO, &b))
|
|
return false;
|
|
bcache_put(b);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool bcache_zero_bytes(struct bcache *cache, int di, uint64_t start, size_t len)
|
|
{
|
|
struct updater u;
|
|
|
|
u.cache = cache;
|
|
u.partial_fn = _zero_partial;
|
|
u.whole_fn = _zero_whole;
|
|
u.data = NULL;
|
|
|
|
return _update_bytes(&u, di, start, len);
|
|
}
|
|
|
|
//----------------------------------------------------------------
|
|
|
|
static bool _set_partial(struct updater *u, int di, block_address bb, uint64_t offset, size_t len)
|
|
{
|
|
struct block *b;
|
|
uint8_t val = *((uint8_t *) u->data);
|
|
|
|
if (!bcache_get(u->cache, di, bb, GF_DIRTY, &b))
|
|
return false;
|
|
|
|
memset(((unsigned char *) b->data) + offset, val, len);
|
|
bcache_put(b);
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool _set_whole(struct updater *u, int di, block_address bb, block_address be)
|
|
{
|
|
struct block *b;
|
|
uint8_t val = *((uint8_t *) u->data);
|
|
uint64_t len = bcache_block_sectors(u->cache) * 512;
|
|
|
|
for (; bb != be; bb++) {
|
|
if (!bcache_get(u->cache, di, bb, GF_ZERO, &b))
|
|
return false;
|
|
memset((unsigned char *) b->data, val, len);
|
|
bcache_put(b);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool bcache_set_bytes(struct bcache *cache, int di, uint64_t start, size_t len, uint8_t val)
|
|
{
|
|
struct updater u;
|
|
|
|
u.cache = cache;
|
|
u.partial_fn = _set_partial;
|
|
u.whole_fn = _set_whole;
|
|
u.data = &val;
|
|
|
|
return _update_bytes(&u, di, start, len);
|
|
}
|
|
|