mirror of
git://sourceware.org/git/lvm2.git
synced 2024-12-21 13:34:40 +03:00
b493811968
Introduce 'hard limit' for max number of cache chunks. When cache target operates with too many chunks (>10e6). When user is aware of related possible troubles he may increase the limit in lvm.conf. Also verbosely inform user about possible solution. Code works for both lvcreate and lvconvert. Lvconvert fully supports change of chunk_size when caching LV (and validates for compatible settings).
790 lines
23 KiB
C
790 lines
23 KiB
C
/*
|
|
* Copyright (C) 2014-2015 Red Hat, Inc. All rights reserved.
|
|
*
|
|
* This file is part of LVM2.
|
|
*
|
|
* This copyrighted material is made available to anyone wishing to use,
|
|
* modify, copy, or redistribute it subject to the terms and conditions
|
|
* of the GNU Lesser General Public License v.2.1.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public License
|
|
* along with this program; if not, write to the Free Software Foundation,
|
|
* Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "lib.h"
|
|
#include "metadata.h"
|
|
#include "locking.h"
|
|
#include "lvm-string.h"
|
|
#include "toolcontext.h"
|
|
#include "display.h"
|
|
#include "segtype.h"
|
|
#include "activate.h"
|
|
#include "defaults.h"
|
|
#include "lv_alloc.h"
|
|
|
|
/* https://github.com/jthornber/thin-provisioning-tools/blob/master/caching/cache_metadata_size.cc */
|
|
#define DM_TRANSACTION_OVERHEAD 4096 /* KiB */
|
|
#define DM_BYTES_PER_BLOCK 16 /* bytes */
|
|
#define DM_HINT_OVERHEAD_PER_BLOCK 8 /* bytes */
|
|
#define DM_MAX_HINT_WIDTH (4+16) /* bytes. FIXME Configurable? */
|
|
|
|
const char *display_cache_mode(const struct lv_segment *seg)
|
|
{
|
|
if (seg_is_cache(seg))
|
|
seg = first_seg(seg->pool_lv);
|
|
|
|
if (!seg_is_cache_pool(seg) ||
|
|
(seg->cache_mode == CACHE_MODE_UNDEFINED))
|
|
return "";
|
|
|
|
return get_cache_mode_name(seg);
|
|
}
|
|
|
|
|
|
const char *get_cache_mode_name(const struct lv_segment *pool_seg)
|
|
{
|
|
switch (pool_seg->cache_mode) {
|
|
default:
|
|
log_error(INTERNAL_ERROR "Cache pool %s has undefined cache mode, using writethrough instead.",
|
|
display_lvname(pool_seg->lv));
|
|
/* Fall through */
|
|
case CACHE_MODE_WRITETHROUGH:
|
|
return "writethrough";
|
|
case CACHE_MODE_WRITEBACK:
|
|
return "writeback";
|
|
case CACHE_MODE_PASSTHROUGH:
|
|
return "passthrough";
|
|
}
|
|
}
|
|
|
|
int set_cache_mode(cache_mode_t *mode, const char *cache_mode)
|
|
{
|
|
if (!strcasecmp(cache_mode, "writethrough"))
|
|
*mode = CACHE_MODE_WRITETHROUGH;
|
|
else if (!strcasecmp(cache_mode, "writeback"))
|
|
*mode = CACHE_MODE_WRITEBACK;
|
|
else if (!strcasecmp(cache_mode, "passthrough"))
|
|
*mode = CACHE_MODE_PASSTHROUGH;
|
|
else {
|
|
log_error("Unknown cache mode: %s.", cache_mode);
|
|
return 0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
int cache_set_cache_mode(struct lv_segment *seg, cache_mode_t mode)
|
|
{
|
|
struct cmd_context *cmd = seg->lv->vg->cmd;
|
|
const char *str;
|
|
int id;
|
|
|
|
if (seg_is_cache(seg))
|
|
seg = first_seg(seg->pool_lv);
|
|
else if (seg_is_cache_pool(seg)) {
|
|
if (mode == CACHE_MODE_UNDEFINED)
|
|
return 1; /* Defaults only for cache */
|
|
} else {
|
|
log_error(INTERNAL_ERROR "Cannot set cache mode for non cache volume %s.",
|
|
display_lvname(seg->lv));
|
|
return 0;
|
|
}
|
|
|
|
if (mode != CACHE_MODE_UNDEFINED) {
|
|
seg->cache_mode = mode;
|
|
return 1;
|
|
}
|
|
|
|
if (seg->cache_mode != CACHE_MODE_UNDEFINED)
|
|
return 1; /* Default already set in cache pool */
|
|
|
|
/* Figure default settings from config/profiles */
|
|
id = allocation_cache_mode_CFG;
|
|
|
|
/* If present, check backward compatible settings */
|
|
if (!find_config_node(cmd, cmd->cft, id) &&
|
|
find_config_node(cmd, cmd->cft, allocation_cache_pool_cachemode_CFG))
|
|
id = allocation_cache_pool_cachemode_CFG;
|
|
|
|
if (!(str = find_config_tree_str(cmd, id, NULL))) {
|
|
log_error(INTERNAL_ERROR "Cache mode is not determined.");
|
|
return 0;
|
|
}
|
|
|
|
if (!(set_cache_mode(&seg->cache_mode, str)))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* At least warn a user if certain cache stacks may present some problems
|
|
*/
|
|
void cache_check_for_warns(const struct lv_segment *seg)
|
|
{
|
|
struct logical_volume *origin_lv = seg_lv(seg, 0);
|
|
|
|
if (lv_is_raid(origin_lv) &&
|
|
first_seg(seg->pool_lv)->cache_mode == CACHE_MODE_WRITEBACK)
|
|
log_warn("WARNING: Data redundancy is lost with writeback "
|
|
"caching of raid logical volume!");
|
|
|
|
if (lv_is_thin_pool_data(seg->lv))
|
|
log_warn("WARNING: Cached thin pool's data cannot be currently "
|
|
"resized and require manual uncache before resize!");
|
|
}
|
|
|
|
/*
|
|
* Returns minimum size of cache metadata volume for give data and chunk size
|
|
* (all values in sector)
|
|
* Default meta size is: (Overhead + mapping size + hint size)
|
|
*/
|
|
static uint64_t _cache_min_metadata_size(uint64_t data_size, uint32_t chunk_size)
|
|
{
|
|
uint64_t min_meta_size;
|
|
|
|
min_meta_size = data_size / chunk_size; /* nr_chunks */
|
|
min_meta_size *= (DM_BYTES_PER_BLOCK + DM_MAX_HINT_WIDTH + DM_HINT_OVERHEAD_PER_BLOCK);
|
|
min_meta_size = (min_meta_size + (SECTOR_SIZE - 1)) >> SECTOR_SHIFT; /* in sectors */
|
|
min_meta_size += DM_TRANSACTION_OVERHEAD * (1024 >> SECTOR_SHIFT);
|
|
|
|
return min_meta_size;
|
|
}
|
|
|
|
int update_cache_pool_params(const struct segment_type *segtype,
|
|
struct volume_group *vg, unsigned attr,
|
|
int passed_args, uint32_t pool_data_extents,
|
|
uint32_t *pool_metadata_extents,
|
|
int *chunk_size_calc_method, uint32_t *chunk_size)
|
|
{
|
|
uint64_t min_meta_size;
|
|
uint32_t extent_size = vg->extent_size;
|
|
uint64_t pool_metadata_size = (uint64_t) *pool_metadata_extents * extent_size;
|
|
uint64_t pool_data_size = (uint64_t) pool_data_extents * extent_size;
|
|
uint64_t max_chunks =
|
|
get_default_allocation_cache_pool_max_chunks_CFG(vg->cmd, vg->profile);
|
|
/* min chunk size in a multiple of DM_CACHE_MIN_DATA_BLOCK_SIZE */
|
|
uint64_t min_chunk_size = (((pool_data_size + max_chunks - 1) / max_chunks +
|
|
DM_CACHE_MIN_DATA_BLOCK_SIZE - 1) /
|
|
DM_CACHE_MIN_DATA_BLOCK_SIZE) * DM_CACHE_MIN_DATA_BLOCK_SIZE;
|
|
|
|
if (!(passed_args & PASS_ARG_CHUNK_SIZE)) {
|
|
*chunk_size = DEFAULT_CACHE_POOL_CHUNK_SIZE * 2;
|
|
if (*chunk_size < min_chunk_size) {
|
|
/*
|
|
* When using more then 'standard' default,
|
|
* keep user informed he might be using things in untintended direction
|
|
*/
|
|
log_print_unless_silent("Using %s chunk size instead of default %s, "
|
|
"so cache pool has less then " FMTu64 " chunks.",
|
|
display_size(vg->cmd, min_chunk_size),
|
|
display_size(vg->cmd, *chunk_size),
|
|
max_chunks);
|
|
*chunk_size = min_chunk_size;
|
|
} else
|
|
log_verbose("Setting chunk size to %s.",
|
|
display_size(vg->cmd, *chunk_size));
|
|
} else if (*chunk_size < min_chunk_size) {
|
|
log_error("Chunk size %s is less then required minimal chunk size %s "
|
|
"for a cache pool of %s size and limit " FMTu64 " chunks.",
|
|
display_size(vg->cmd, *chunk_size),
|
|
display_size(vg->cmd, min_chunk_size),
|
|
display_size(vg->cmd, pool_data_size),
|
|
max_chunks);
|
|
log_error("To allow use of more chunks, see setting allocation/cache_pool_max_chunks.");
|
|
return 0;
|
|
}
|
|
|
|
if (!validate_pool_chunk_size(vg->cmd, segtype, *chunk_size))
|
|
return_0;
|
|
|
|
min_meta_size = _cache_min_metadata_size((uint64_t) pool_data_extents * extent_size, *chunk_size);
|
|
|
|
/* Round up to extent size */
|
|
if (min_meta_size % extent_size)
|
|
min_meta_size += extent_size - min_meta_size % extent_size;
|
|
|
|
if (!pool_metadata_size)
|
|
pool_metadata_size = min_meta_size;
|
|
|
|
if (pool_metadata_size > (2 * DEFAULT_CACHE_POOL_MAX_METADATA_SIZE)) {
|
|
pool_metadata_size = 2 * DEFAULT_CACHE_POOL_MAX_METADATA_SIZE;
|
|
if (passed_args & PASS_ARG_POOL_METADATA_SIZE)
|
|
log_warn("WARNING: Maximum supported pool metadata size is %s.",
|
|
display_size(vg->cmd, pool_metadata_size));
|
|
} else if (pool_metadata_size < min_meta_size) {
|
|
if (passed_args & PASS_ARG_POOL_METADATA_SIZE)
|
|
log_warn("WARNING: Minimum required pool metadata size is %s "
|
|
"(needs extra %s).",
|
|
display_size(vg->cmd, min_meta_size),
|
|
display_size(vg->cmd, min_meta_size - pool_metadata_size));
|
|
pool_metadata_size = min_meta_size;
|
|
}
|
|
|
|
if (!(*pool_metadata_extents =
|
|
extents_from_size(vg->cmd, pool_metadata_size, extent_size)))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* Validate if existing cache-pool can be used with given chunk size
|
|
* i.e. cache-pool metadata size fits all info.
|
|
*/
|
|
int validate_lv_cache_chunk_size(struct logical_volume *pool_lv, uint32_t chunk_size)
|
|
{
|
|
struct volume_group *vg = pool_lv->vg;
|
|
uint64_t max_chunks = get_default_allocation_cache_pool_max_chunks_CFG(vg->cmd, vg->profile);
|
|
uint64_t min_size = _cache_min_metadata_size(pool_lv->size, chunk_size);
|
|
uint64_t chunks = pool_lv->size / chunk_size;
|
|
int r = 1;
|
|
|
|
if (min_size > first_seg(pool_lv)->metadata_lv->size) {
|
|
log_error("Cannot use chunk size %s with cache pool %s metadata size %s.",
|
|
display_size(vg->cmd, chunk_size),
|
|
display_lvname(pool_lv),
|
|
display_size(vg->cmd, first_seg(pool_lv)->metadata_lv->size));
|
|
log_error("Minimal size for cache pool %s metadata with chunk size %s would be %s.",
|
|
display_lvname(pool_lv),
|
|
display_size(vg->cmd, chunk_size),
|
|
display_size(vg->cmd, min_size));
|
|
r = 0;
|
|
}
|
|
|
|
if (chunks > max_chunks) {
|
|
log_error("Cannot use too small chunk size %s with cache pool %s data volume size %s.",
|
|
display_size(vg->cmd, chunk_size),
|
|
display_lvname(pool_lv),
|
|
display_size(pool_lv->vg->cmd, pool_lv->size));
|
|
log_error("Maximum configured chunks for a cache pool is " FMTu64 ".",
|
|
max_chunks);
|
|
log_error("Use smaller cache pool (<%s) or bigger cache chunk size (>=%s) or enable higher "
|
|
"values in 'allocation/cache_pool_max_chunks'.",
|
|
display_size(vg->cmd, chunk_size * max_chunks),
|
|
display_size(vg->cmd, pool_lv->size / max_chunks));
|
|
r = 0;
|
|
}
|
|
|
|
return r;
|
|
}
|
|
/*
|
|
* Validate arguments for converting origin into cached volume with given cache pool.
|
|
*
|
|
* Always validates origin_lv, and when it is known also cache pool_lv
|
|
*/
|
|
int validate_lv_cache_create_pool(const struct logical_volume *pool_lv)
|
|
{
|
|
struct lv_segment *seg;
|
|
|
|
if (!lv_is_cache_pool(pool_lv)) {
|
|
log_error("Logical volume %s is not a cache pool.",
|
|
display_lvname(pool_lv));
|
|
return 0;
|
|
}
|
|
|
|
if (lv_is_locked(pool_lv)) {
|
|
log_error("Cannot use locked cache pool %s.",
|
|
display_lvname(pool_lv));
|
|
return 0;
|
|
}
|
|
|
|
if (!dm_list_empty(&pool_lv->segs_using_this_lv)) {
|
|
seg = get_only_segment_using_this_lv(pool_lv);
|
|
log_error("Logical volume %s is already in use by %s",
|
|
display_lvname(pool_lv),
|
|
seg ? display_lvname(seg->lv) : "another LV");
|
|
return 0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
int validate_lv_cache_create_origin(const struct logical_volume *origin_lv)
|
|
{
|
|
if (lv_is_locked(origin_lv)) {
|
|
log_error("Cannot use locked origin volume %s.",
|
|
display_lvname(origin_lv));
|
|
return 0;
|
|
}
|
|
|
|
/* For now we only support conversion of thin pool data volume */
|
|
if (!lv_is_visible(origin_lv) && !lv_is_thin_pool_data(origin_lv)) {
|
|
log_error("Can't convert internal LV %s.", display_lvname(origin_lv));
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Only linear, striped or raid supported.
|
|
* FIXME Tidy up all these type restrictions.
|
|
*/
|
|
if (lv_is_cache_type(origin_lv) ||
|
|
lv_is_mirror_type(origin_lv) ||
|
|
lv_is_thin_volume(origin_lv) || lv_is_thin_pool_metadata(origin_lv) ||
|
|
lv_is_origin(origin_lv) || lv_is_merging_origin(origin_lv) ||
|
|
lv_is_cow(origin_lv) || lv_is_merging_cow(origin_lv) ||
|
|
lv_is_external_origin(origin_lv) ||
|
|
lv_is_virtual(origin_lv)) {
|
|
log_error("Cache is not supported with %s segment type of the original logical volume %s.",
|
|
first_seg(origin_lv)->segtype->name, display_lvname(origin_lv));
|
|
return 0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* lv_cache_create
|
|
* @pool
|
|
* @origin
|
|
*
|
|
* Given a cache_pool and an origin, link the two and create a
|
|
* cached LV.
|
|
*
|
|
* Returns: cache LV on success, NULL on failure
|
|
*/
|
|
struct logical_volume *lv_cache_create(struct logical_volume *pool_lv,
|
|
struct logical_volume *origin_lv)
|
|
{
|
|
const struct segment_type *segtype;
|
|
struct cmd_context *cmd = pool_lv->vg->cmd;
|
|
struct logical_volume *cache_lv = origin_lv;
|
|
struct lv_segment *seg;
|
|
|
|
if (!validate_lv_cache_create_pool(pool_lv) ||
|
|
!validate_lv_cache_create_origin(cache_lv))
|
|
return_NULL;
|
|
|
|
if (lv_is_thin_pool(cache_lv))
|
|
cache_lv = seg_lv(first_seg(cache_lv), 0); /* cache _tdata */
|
|
|
|
if (!(segtype = get_segtype_from_string(cmd, SEG_TYPE_NAME_CACHE)))
|
|
return_NULL;
|
|
|
|
if (!insert_layer_for_lv(cmd, cache_lv, CACHE, "_corig"))
|
|
return_NULL;
|
|
|
|
seg = first_seg(cache_lv);
|
|
seg->segtype = segtype;
|
|
|
|
if (!attach_pool_lv(seg, pool_lv, NULL, NULL, NULL))
|
|
return_NULL;
|
|
|
|
return cache_lv;
|
|
}
|
|
|
|
/*
|
|
* Checks cache status and loops until there are not dirty blocks
|
|
* Set 1 to *is_clean when there are no dirty blocks on return.
|
|
*/
|
|
int lv_cache_wait_for_clean(struct logical_volume *cache_lv, int *is_clean)
|
|
{
|
|
const struct logical_volume *lock_lv = lv_lock_holder(cache_lv);
|
|
struct lv_segment *cache_seg = first_seg(cache_lv);
|
|
struct lv_status_cache *status;
|
|
int cleaner_policy;
|
|
uint64_t dirty_blocks;
|
|
|
|
*is_clean = 0;
|
|
|
|
//FIXME: use polling to do this...
|
|
for (;;) {
|
|
if (!lv_cache_status(cache_lv, &status))
|
|
return_0;
|
|
|
|
if (status->cache->fail) {
|
|
dm_pool_destroy(status->mem);
|
|
log_warn("WARNING: Skippping flush for failed cache %s.",
|
|
display_lvname(cache_lv));
|
|
return 1;
|
|
}
|
|
|
|
cleaner_policy = !strcmp(status->cache->policy_name, "cleaner");
|
|
dirty_blocks = status->cache->dirty_blocks;
|
|
|
|
/* No clear policy and writeback mode means dirty */
|
|
if (!cleaner_policy &&
|
|
(status->cache->feature_flags & DM_CACHE_FEATURE_WRITEBACK))
|
|
dirty_blocks++;
|
|
dm_pool_destroy(status->mem);
|
|
|
|
if (!dirty_blocks)
|
|
break;
|
|
|
|
log_print_unless_silent("Flushing " FMTu64 " blocks for cache %s.",
|
|
dirty_blocks, display_lvname(cache_lv));
|
|
if (cleaner_policy) {
|
|
/* TODO: Use centralized place */
|
|
usleep(500000);
|
|
continue;
|
|
}
|
|
|
|
/* Switch to cleaner policy to flush the cache */
|
|
cache_seg->cleaner_policy = 1;
|
|
/* Reaload kernel with "cleaner" policy */
|
|
if (!lv_update_and_reload_origin(cache_lv))
|
|
return_0;
|
|
}
|
|
|
|
/*
|
|
* TODO: add check if extra suspend resume is necessary
|
|
* ATM this is workaround for missing cache sync when cache gets clean
|
|
*/
|
|
if (1) {
|
|
if (!lv_refresh_suspend_resume(lock_lv))
|
|
return_0;
|
|
}
|
|
|
|
cache_seg->cleaner_policy = 0;
|
|
*is_clean = 1;
|
|
|
|
return 1;
|
|
}
|
|
/*
|
|
* lv_cache_remove
|
|
* @cache_lv
|
|
*
|
|
* Given a cache LV, remove the cache layer. This will unlink
|
|
* the origin and cache_pool, remove the cache LV layer, and promote
|
|
* the origin to a usable non-cached LV of the same name as the
|
|
* given cache_lv.
|
|
*
|
|
* Returns: 1 on success, 0 on failure
|
|
*/
|
|
int lv_cache_remove(struct logical_volume *cache_lv)
|
|
{
|
|
struct lv_segment *cache_seg = first_seg(cache_lv);
|
|
struct logical_volume *corigin_lv;
|
|
struct logical_volume *cache_pool_lv;
|
|
int is_clear;
|
|
|
|
if (!lv_is_cache(cache_lv)) {
|
|
log_error(INTERNAL_ERROR "LV %s is not cache volume.",
|
|
display_lvname(cache_lv));
|
|
return 0;
|
|
}
|
|
|
|
if (lv_is_pending_delete(cache_lv)) {
|
|
log_error(INTERNAL_ERROR "LV %s is already dropped cache volume.",
|
|
display_lvname(cache_lv));
|
|
goto remove; /* Already dropped */
|
|
}
|
|
|
|
/* Localy active volume is needed for writeback */
|
|
if (!lv_is_active_locally(cache_lv)) {
|
|
/* Give up any remote locks */
|
|
if (!deactivate_lv(cache_lv->vg->cmd, cache_lv)) {
|
|
log_error("Cannot deactivate remotely active cache volume %s.",
|
|
display_lvname(cache_lv));
|
|
return 0;
|
|
}
|
|
|
|
switch (first_seg(cache_seg->pool_lv)->cache_mode) {
|
|
case CACHE_MODE_WRITETHROUGH:
|
|
case CACHE_MODE_PASSTHROUGH:
|
|
/* For inactive pass/writethrough just drop cache layer */
|
|
corigin_lv = seg_lv(cache_seg, 0);
|
|
if (!detach_pool_lv(cache_seg))
|
|
return_0;
|
|
if (!remove_layer_from_lv(cache_lv, corigin_lv))
|
|
return_0;
|
|
if (!lv_remove(corigin_lv))
|
|
return_0;
|
|
return 1;
|
|
default:
|
|
/* Otherwise localy activate volume to sync dirty blocks */
|
|
cache_lv->status |= LV_TEMPORARY;
|
|
if (!activate_lv_excl_local(cache_lv->vg->cmd, cache_lv) ||
|
|
!lv_is_active_locally(cache_lv)) {
|
|
log_error("Failed to active cache locally %s.",
|
|
display_lvname(cache_lv));
|
|
return 0;
|
|
}
|
|
cache_lv->status &= ~LV_TEMPORARY;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* FIXME:
|
|
* Before the link can be broken, we must ensure that the
|
|
* cache has been flushed. This may already be the case
|
|
* if the cache mode is writethrough (or the cleaner
|
|
* policy is in place from a previous half-finished attempt
|
|
* to remove the cache_pool). It could take a long time to
|
|
* flush the cache - it should probably be done in the background.
|
|
*
|
|
* Also, if we do perform the flush in the background and we
|
|
* happen to also be removing the cache/origin LV, then we
|
|
* could check if the cleaner policy is in place and simply
|
|
* remove the cache_pool then without waiting for the flush to
|
|
* complete.
|
|
*/
|
|
if (!lv_cache_wait_for_clean(cache_lv, &is_clear))
|
|
return_0;
|
|
|
|
cache_pool_lv = cache_seg->pool_lv;
|
|
if (!detach_pool_lv(cache_seg))
|
|
return_0;
|
|
|
|
/*
|
|
* Drop layer from cache LV and make _corigin to appear again as regular LV
|
|
* And use 'existing' _corigin volume to keep reference on cache-pool
|
|
* This way we still have a way to reference _corigin in dm table and we
|
|
* know it's been 'cache' LV and we can drop all needed table entries via
|
|
* activation and deactivation of it.
|
|
*
|
|
* This 'cache' LV without origin is temporary LV, which still could be
|
|
* easily operated by lvm2 commands - it could be activate/deactivated/removed.
|
|
* However in the dm-table it will use 'error' target for _corigin volume.
|
|
*/
|
|
corigin_lv = seg_lv(cache_seg, 0);
|
|
lv_set_visible(corigin_lv);
|
|
|
|
if (!remove_layer_from_lv(cache_lv, corigin_lv))
|
|
return_0;
|
|
|
|
/* Replace 'error' with 'cache' segtype */
|
|
cache_seg = first_seg(corigin_lv);
|
|
if (!(cache_seg->segtype = get_segtype_from_string(corigin_lv->vg->cmd, SEG_TYPE_NAME_CACHE)))
|
|
return_0;
|
|
|
|
if (!(cache_seg->areas = dm_pool_zalloc(cache_lv->vg->vgmem, sizeof(*cache_seg->areas))))
|
|
return_0;
|
|
|
|
if (!set_lv_segment_area_lv(cache_seg, 0, cache_lv, 0, 0))
|
|
return_0;
|
|
|
|
cache_seg->area_count = 1;
|
|
corigin_lv->le_count = cache_lv->le_count;
|
|
corigin_lv->size = cache_lv->size;
|
|
corigin_lv->status |= LV_PENDING_DELETE;
|
|
|
|
/* Reattach cache pool */
|
|
if (!attach_pool_lv(cache_seg, cache_pool_lv, NULL, NULL, NULL))
|
|
return_0;
|
|
|
|
/* Suspend/resume also deactivates deleted LV via support of LV_PENDING_DELETE */
|
|
if (!lv_update_and_reload(cache_lv))
|
|
return_0;
|
|
cache_lv = corigin_lv;
|
|
remove:
|
|
if (!detach_pool_lv(cache_seg))
|
|
return_0;
|
|
|
|
if (!lv_remove(cache_lv)) /* Will use LV_PENDING_DELETE */
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
int lv_is_cache_origin(const struct logical_volume *lv)
|
|
{
|
|
struct lv_segment *seg;
|
|
|
|
/* Make sure there's exactly one segment in segs_using_this_lv! */
|
|
if (dm_list_empty(&lv->segs_using_this_lv) ||
|
|
(dm_list_size(&lv->segs_using_this_lv) > 1))
|
|
return 0;
|
|
|
|
seg = get_only_segment_using_this_lv(lv);
|
|
return seg && lv_is_cache(seg->lv) && !lv_is_pending_delete(seg->lv) && (seg_lv(seg, 0) == lv);
|
|
}
|
|
|
|
static const char *_get_default_cache_policy(struct cmd_context *cmd)
|
|
{
|
|
const struct segment_type *segtype = get_segtype_from_string(cmd, SEG_TYPE_NAME_CACHE);
|
|
unsigned attr = ~0;
|
|
const char *def = NULL;
|
|
|
|
if (!segtype ||
|
|
!segtype->ops->target_present ||
|
|
!segtype->ops->target_present(cmd, NULL, &attr)) {
|
|
log_warn("WARNING: Cannot detect default cache policy, using \""
|
|
DEFAULT_CACHE_POLICY "\".");
|
|
return DEFAULT_CACHE_POLICY;
|
|
}
|
|
|
|
if (attr & CACHE_FEATURE_POLICY_SMQ)
|
|
def = "smq";
|
|
else if (attr & CACHE_FEATURE_POLICY_MQ)
|
|
def = "mq";
|
|
else {
|
|
log_error("Default cache policy is not available.");
|
|
return NULL;
|
|
}
|
|
|
|
log_debug_metadata("Detected default cache_policy \"%s\".", def);
|
|
|
|
return def;
|
|
}
|
|
|
|
int cache_set_policy(struct lv_segment *seg, const char *name,
|
|
const struct dm_config_tree *settings)
|
|
{
|
|
struct dm_config_node *cn;
|
|
const struct dm_config_node *cns;
|
|
struct dm_config_tree *old = NULL, *new = NULL, *tmp = NULL;
|
|
int r = 0;
|
|
const int passed_seg_is_cache = seg_is_cache(seg);
|
|
|
|
if (passed_seg_is_cache)
|
|
seg = first_seg(seg->pool_lv);
|
|
|
|
if (name) {
|
|
if (!(seg->policy_name = dm_pool_strdup(seg->lv->vg->vgmem, name))) {
|
|
log_error("Failed to duplicate policy name.");
|
|
return 0;
|
|
}
|
|
} else if (!seg->policy_name && passed_seg_is_cache) {
|
|
if (!(seg->policy_name = find_config_tree_str(seg->lv->vg->cmd, allocation_cache_policy_CFG, NULL)) &&
|
|
!(seg->policy_name = _get_default_cache_policy(seg->lv->vg->cmd)))
|
|
return_0;
|
|
}
|
|
|
|
if (settings) {
|
|
if (!seg->policy_name) {
|
|
log_error(INTERNAL_ERROR "Can't set policy settings without policy name.");
|
|
return 0;
|
|
}
|
|
|
|
if (seg->policy_settings) {
|
|
if (!(old = dm_config_create()))
|
|
goto_out;
|
|
if (!(new = dm_config_create()))
|
|
goto_out;
|
|
new->root = settings->root;
|
|
old->root = seg->policy_settings;
|
|
new->cascade = old;
|
|
if (!(tmp = dm_config_flatten(new)))
|
|
goto_out;
|
|
}
|
|
|
|
if ((cn = dm_config_find_node((tmp) ? tmp->root : settings->root, "policy_settings")) &&
|
|
!(seg->policy_settings = dm_config_clone_node_with_mem(seg->lv->vg->vgmem, cn, 0)))
|
|
goto_out;
|
|
} else if (passed_seg_is_cache && /* Look for command's profile cache_policies */
|
|
(cns = find_config_tree_node(seg->lv->vg->cmd, allocation_cache_settings_CFG_SECTION, NULL))) {
|
|
/* Try to find our section for given policy */
|
|
for (cn = cns->child; cn; cn = cn->sib) {
|
|
/* Only matching section names */
|
|
if (cn->v || strcmp(cn->key, seg->policy_name) != 0)
|
|
continue;
|
|
|
|
if (!cn->child)
|
|
break;
|
|
|
|
if (!(new = dm_config_create()))
|
|
goto_out;
|
|
|
|
if (!(new->root = dm_config_clone_node_with_mem(new->mem,
|
|
cn->child, 1)))
|
|
goto_out;
|
|
|
|
if (!(seg->policy_settings = dm_config_create_node(new, "policy_settings")))
|
|
goto_out;
|
|
|
|
seg->policy_settings->child = new->root;
|
|
|
|
break; /* Only first match counts */
|
|
}
|
|
}
|
|
|
|
restart: /* remove any 'default" nodes */
|
|
cn = seg->policy_settings ? seg->policy_settings->child : NULL;
|
|
while (cn) {
|
|
if (cn->v->type == DM_CFG_STRING && !strcmp(cn->v->v.str, "default")) {
|
|
dm_config_remove_node(seg->policy_settings, cn);
|
|
goto restart;
|
|
}
|
|
cn = cn->sib;
|
|
}
|
|
|
|
r = 1;
|
|
|
|
out:
|
|
if (tmp)
|
|
dm_config_destroy(tmp);
|
|
if (new)
|
|
dm_config_destroy(new);
|
|
if (old)
|
|
dm_config_destroy(old);
|
|
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* Universal 'wrapper' function do-it-all
|
|
* to update all commonly specified cache parameters
|
|
*/
|
|
int cache_set_params(struct lv_segment *seg,
|
|
cache_mode_t mode,
|
|
const char *policy_name,
|
|
const struct dm_config_tree *policy_settings,
|
|
uint32_t chunk_size)
|
|
{
|
|
struct lv_segment *pool_seg;
|
|
|
|
if (!cache_set_cache_mode(seg, mode))
|
|
return_0;
|
|
|
|
if (!cache_set_policy(seg, policy_name, policy_settings))
|
|
return_0;
|
|
|
|
pool_seg = seg_is_cache(seg) ? first_seg(seg->pool_lv) : seg;
|
|
|
|
if (chunk_size) {
|
|
if (!validate_lv_cache_chunk_size(pool_seg->lv, chunk_size))
|
|
return_0;
|
|
pool_seg->chunk_size = chunk_size;
|
|
} else {
|
|
/* TODO: some calc_policy solution for cache ? */
|
|
if (!recalculate_pool_chunk_size_with_dev_hints(pool_seg->lv, 0,
|
|
THIN_CHUNK_SIZE_CALC_METHOD_GENERIC))
|
|
return_0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* Wipe cache pool metadata area before use.
|
|
*
|
|
* Activates metadata volume as 'cache-pool' so regular wiping
|
|
* of existing visible volume may proceed.
|
|
*/
|
|
int wipe_cache_pool(struct logical_volume *cache_pool_lv)
|
|
{
|
|
int r;
|
|
|
|
/* Only unused cache-pool could be activated and wiped */
|
|
if (!lv_is_cache_pool(cache_pool_lv) ||
|
|
!dm_list_empty(&cache_pool_lv->segs_using_this_lv)) {
|
|
log_error(INTERNAL_ERROR "Failed to wipe cache pool for volume %s.",
|
|
display_lvname(cache_pool_lv));
|
|
return 0;
|
|
}
|
|
|
|
cache_pool_lv->status |= LV_TEMPORARY;
|
|
if (!activate_lv_local(cache_pool_lv->vg->cmd, cache_pool_lv)) {
|
|
log_error("Aborting. Failed to activate cache pool %s.",
|
|
display_lvname(cache_pool_lv));
|
|
return 0;
|
|
}
|
|
cache_pool_lv->status &= ~LV_TEMPORARY;
|
|
if (!(r = wipe_lv(cache_pool_lv, (struct wipe_params) { .do_zero = 1 }))) {
|
|
log_error("Aborting. Failed to wipe cache pool %s.",
|
|
display_lvname(cache_pool_lv));
|
|
/* Delay return of error after deactivation */
|
|
}
|
|
|
|
/* Deactivate cleared cache-pool metadata */
|
|
if (!deactivate_lv(cache_pool_lv->vg->cmd, cache_pool_lv)) {
|
|
log_error("Aborting. Could not deactivate cache pool %s.",
|
|
display_lvname(cache_pool_lv));
|
|
r = 0;
|
|
}
|
|
|
|
return r;
|
|
}
|