mirror of
git://sourceware.org/git/lvm2.git
synced 2024-12-22 17:35:59 +03:00
7a28b243fa
Just add some dots to messages and remove unneeded stack trace from return after log_error.
1013 lines
25 KiB
C
1013 lines
25 KiB
C
/*
|
|
* Copyright (C) 2003-2004 Sistina Software, Inc. All rights reserved.
|
|
* Copyright (C) 2004-2015 Red Hat, Inc. All rights reserved.
|
|
*
|
|
* This file is part of LVM2.
|
|
*
|
|
* This copyrighted material is made available to anyone wishing to use,
|
|
* modify, copy, or redistribute it subject to the terms and conditions
|
|
* of the GNU Lesser General Public License v.2.1.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public License
|
|
* along with this program; if not, write to the Free Software Foundation,
|
|
* Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "tools.h"
|
|
|
|
#include "polldaemon.h"
|
|
#include "display.h"
|
|
#include "pvmove_poll.h"
|
|
#include "lvmpolld-client.h"
|
|
|
|
#define PVMOVE_FIRST_TIME 0x00000001 /* Called for first time */
|
|
|
|
struct pvmove_params {
|
|
char *pv_name_arg; /* original unmodified arg */
|
|
char *lv_name_arg; /* original unmodified arg */
|
|
alloc_policy_t alloc;
|
|
int pv_count;
|
|
char **pv_names;
|
|
|
|
union lvid *lvid;
|
|
char *id_vg_name;
|
|
char *id_lv_name;
|
|
unsigned in_progress;
|
|
int setup_result;
|
|
int found_pv;
|
|
};
|
|
|
|
static int _pvmove_target_present(struct cmd_context *cmd, int clustered)
|
|
{
|
|
const struct segment_type *segtype;
|
|
unsigned attr = 0;
|
|
int found = 1;
|
|
static int _clustered_found = -1;
|
|
|
|
if (clustered && _clustered_found >= 0)
|
|
return _clustered_found;
|
|
|
|
if (!(segtype = get_segtype_from_string(cmd, SEG_TYPE_NAME_MIRROR)))
|
|
return_0;
|
|
|
|
if (activation() && segtype->ops->target_present &&
|
|
!segtype->ops->target_present(cmd, NULL, clustered ? &attr : NULL))
|
|
found = 0;
|
|
|
|
if (activation() && clustered) {
|
|
if (found && (attr & MIRROR_LOG_CLUSTERED))
|
|
_clustered_found = found = 1;
|
|
else
|
|
_clustered_found = found = 0;
|
|
}
|
|
|
|
return found;
|
|
}
|
|
|
|
static unsigned _pvmove_is_exclusive(struct cmd_context *cmd,
|
|
struct volume_group *vg)
|
|
{
|
|
if (vg_is_clustered(vg))
|
|
if (!_pvmove_target_present(cmd, 1))
|
|
return 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Allow /dev/vgname/lvname, vgname/lvname or lvname */
|
|
static const char *_extract_lvname(struct cmd_context *cmd, const char *vgname,
|
|
const char *arg)
|
|
{
|
|
const char *lvname;
|
|
|
|
/* Is an lvname supplied directly? */
|
|
if (!strchr(arg, '/'))
|
|
return arg;
|
|
|
|
lvname = skip_dev_dir(cmd, arg, NULL);
|
|
while (*lvname == '/')
|
|
lvname++;
|
|
if (!strchr(lvname, '/')) {
|
|
log_error("--name takes a logical volume name.");
|
|
return NULL;
|
|
}
|
|
if (strncmp(vgname, lvname, strlen(vgname)) ||
|
|
(lvname += strlen(vgname), *lvname != '/')) {
|
|
log_error("Named LV and old PV must be in the same VG.");
|
|
return NULL;
|
|
}
|
|
while (*lvname == '/')
|
|
lvname++;
|
|
if (!*lvname) {
|
|
log_error("Incomplete LV name supplied with --name.");
|
|
return NULL;
|
|
}
|
|
return lvname;
|
|
}
|
|
|
|
/* Create list of PVs for allocation of replacement extents */
|
|
static struct dm_list *_get_allocatable_pvs(struct cmd_context *cmd, int argc,
|
|
char **argv, struct volume_group *vg,
|
|
struct physical_volume *pv,
|
|
alloc_policy_t alloc)
|
|
{
|
|
struct dm_list *allocatable_pvs, *pvht, *pvh;
|
|
struct pv_list *pvl;
|
|
|
|
if (argc)
|
|
allocatable_pvs = create_pv_list(cmd->mem, vg, argc, argv, 1);
|
|
else
|
|
allocatable_pvs = clone_pv_list(cmd->mem, &vg->pvs);
|
|
|
|
if (!allocatable_pvs)
|
|
return_NULL;
|
|
|
|
dm_list_iterate_safe(pvh, pvht, allocatable_pvs) {
|
|
pvl = dm_list_item(pvh, struct pv_list);
|
|
|
|
/* Don't allocate onto the PV we're clearing! */
|
|
if ((alloc != ALLOC_ANYWHERE) && (pvl->pv->dev == pv_dev(pv))) {
|
|
dm_list_del(&pvl->list);
|
|
continue;
|
|
}
|
|
|
|
/* Remove PV if full */
|
|
if (pvl->pv->pe_count == pvl->pv->pe_alloc_count)
|
|
dm_list_del(&pvl->list);
|
|
}
|
|
|
|
if (dm_list_empty(allocatable_pvs)) {
|
|
log_error("No extents available for allocation.");
|
|
return NULL;
|
|
}
|
|
|
|
return allocatable_pvs;
|
|
}
|
|
|
|
/*
|
|
* If @lv_name's a RAID SubLV, check for any PVs
|
|
* on @trim_list holding it's sibling (rimage/rmeta)
|
|
* and remove it from the @trim_list in order to allow
|
|
* for pvmove collocation of DataLV/MetaLV pairs.
|
|
*/
|
|
static int _remove_sibling_pvs_from_trim_list(struct logical_volume *lv,
|
|
const char *lv_name,
|
|
struct dm_list *trim_list)
|
|
{
|
|
char *idx, *suffix;
|
|
char sublv_name[NAME_LEN];
|
|
struct logical_volume *sublv;
|
|
struct dm_list untrim_list, *pvh1, *pvh2;
|
|
struct pv_list *pvl1, *pvl2;
|
|
|
|
/* Give up with success unless @lv_name _and_ valid raid segment type */
|
|
if (!lv_name || !*lv_name ||
|
|
!seg_is_raid(first_seg(lv)) ||
|
|
seg_is_raid0(first_seg(lv)) ||
|
|
!strcmp(lv->name, lv_name))
|
|
return 1;
|
|
|
|
dm_list_init(&untrim_list);
|
|
|
|
if (!(suffix = first_substring(lv_name, "_rimage_", "_rmeta_", NULL))) {
|
|
log_error("Can't find rimage or rmeta suffix.");
|
|
return 0;
|
|
}
|
|
|
|
if (!(idx = strchr(suffix + 1, '_'))) {
|
|
log_error("Can't find '_' after suffix %s.", suffix);
|
|
return 0;
|
|
}
|
|
idx++;
|
|
|
|
/* Create the siblings name (e.g. "raidlv_rmeta_N" -> "raidlv_rimage_N" */
|
|
if (dm_snprintf(sublv_name, sizeof(sublv_name), "%s_r%s_%s", lv_name,
|
|
strstr(suffix, "_rimage_") ? "meta" : "image", idx) < 0) {
|
|
log_error("Raid sublv for name %s too long.", lv_name);
|
|
return 0;
|
|
}
|
|
|
|
if (!(sublv = find_lv(lv->vg, sublv_name))) {
|
|
log_error("Can't find sub LV %s.", sublv_name);
|
|
return 0;
|
|
}
|
|
|
|
if (!get_pv_list_for_lv(lv->vg->cmd->mem, sublv, &untrim_list)) {
|
|
log_error("Can't find PVs for sub LV %s.", sublv_name);
|
|
return 0;
|
|
}
|
|
|
|
dm_list_iterate(pvh1, &untrim_list) {
|
|
pvl1 = dm_list_item(pvh1, struct pv_list);
|
|
|
|
dm_list_iterate(pvh2, trim_list) {
|
|
pvl2 = dm_list_item(pvh2, struct pv_list);
|
|
|
|
if (pvl1->pv == pvl2->pv) {
|
|
log_debug("Removing PV %s from trim list.",
|
|
pvl2->pv->dev->pvid);
|
|
dm_list_del(&pvl2->list);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* _trim_allocatable_pvs
|
|
* @alloc_list
|
|
* @trim_list
|
|
*
|
|
* Remove PVs in 'trim_list' from 'alloc_list'.
|
|
*
|
|
* Returns: 1 on success, 0 on error
|
|
*/
|
|
static int _trim_allocatable_pvs(struct dm_list *alloc_list,
|
|
struct dm_list *trim_list,
|
|
alloc_policy_t alloc)
|
|
{
|
|
struct dm_list *pvht, *pvh, *trim_pvh;
|
|
struct pv_list *pvl, *trim_pvl;
|
|
|
|
if (!alloc_list) {
|
|
log_error(INTERNAL_ERROR "alloc_list is NULL.");
|
|
return 0;
|
|
}
|
|
|
|
if (!trim_list || dm_list_empty(trim_list))
|
|
return 1; /* alloc_list stays the same */
|
|
|
|
dm_list_iterate_safe(pvh, pvht, alloc_list) {
|
|
pvl = dm_list_item(pvh, struct pv_list);
|
|
|
|
dm_list_iterate(trim_pvh, trim_list) {
|
|
trim_pvl = dm_list_item(trim_pvh, struct pv_list);
|
|
|
|
/* Don't allocate onto a trim PV */
|
|
if ((alloc != ALLOC_ANYWHERE) &&
|
|
(pvl->pv == trim_pvl->pv)) {
|
|
dm_list_del(&pvl->list);
|
|
break; /* goto next in alloc_list */
|
|
}
|
|
}
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* Replace any LV segments on given PV with temporary mirror.
|
|
* Returns list of LVs changed.
|
|
*/
|
|
static int _insert_pvmove_mirrors(struct cmd_context *cmd,
|
|
struct logical_volume *lv_mirr,
|
|
struct dm_list *source_pvl,
|
|
struct logical_volume *lv,
|
|
struct dm_list *lvs_changed)
|
|
|
|
{
|
|
struct pv_list *pvl;
|
|
uint32_t prev_le_count;
|
|
|
|
/* Only 1 PV may feature in source_pvl */
|
|
pvl = dm_list_item(source_pvl->n, struct pv_list);
|
|
|
|
prev_le_count = lv_mirr->le_count;
|
|
if (!insert_layer_for_segments_on_pv(cmd, lv, lv_mirr, PVMOVE,
|
|
pvl, lvs_changed))
|
|
return_0;
|
|
|
|
/* check if layer was inserted */
|
|
if (lv_mirr->le_count - prev_le_count) {
|
|
lv->status |= LOCKED;
|
|
|
|
log_verbose("Moving %u extents of logical volume %s.",
|
|
lv_mirr->le_count - prev_le_count,
|
|
display_lvname(lv));
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* Is 'lv' a sub_lv of the LV by the name of 'lv_name'?
|
|
*
|
|
* Returns: 1 if true, 0 otherwise
|
|
*/
|
|
static int _sub_lv_of(struct logical_volume *lv, const char *lv_name)
|
|
{
|
|
struct lv_segment *seg;
|
|
|
|
/* Sub-LVs only ever have one segment using them */
|
|
if (dm_list_size(&lv->segs_using_this_lv) != 1)
|
|
return 0;
|
|
|
|
if (!(seg = get_only_segment_using_this_lv(lv)))
|
|
return_0;
|
|
|
|
if (!strcmp(seg->lv->name, lv_name))
|
|
return 1;
|
|
|
|
/* Continue up the tree */
|
|
return _sub_lv_of(seg->lv, lv_name);
|
|
}
|
|
|
|
/*
|
|
* parent_lv_is_cache_type
|
|
*
|
|
* FIXME: This function can be removed when 'pvmove' is supported for
|
|
* cache types.
|
|
*
|
|
* If this LV is below a cache LV (at any depth), return 1.
|
|
*/
|
|
static int _parent_lv_is_cache_type(struct logical_volume *lv)
|
|
{
|
|
struct lv_segment *seg;
|
|
|
|
/* Sub-LVs only ever have one segment using them */
|
|
if (dm_list_size(&lv->segs_using_this_lv) != 1)
|
|
return 0;
|
|
|
|
if (!(seg = get_only_segment_using_this_lv(lv)))
|
|
return_0;
|
|
|
|
if (lv_is_cache_type(seg->lv))
|
|
return 1;
|
|
|
|
/* Continue up the tree */
|
|
return _parent_lv_is_cache_type(seg->lv);
|
|
}
|
|
|
|
/* Create new LV with mirror segments for the required copies */
|
|
static struct logical_volume *_set_up_pvmove_lv(struct cmd_context *cmd,
|
|
struct volume_group *vg,
|
|
struct dm_list *source_pvl,
|
|
const char *lv_name,
|
|
struct dm_list *allocatable_pvs,
|
|
alloc_policy_t alloc,
|
|
struct dm_list **lvs_changed,
|
|
unsigned *exclusive)
|
|
{
|
|
struct logical_volume *lv_mirr, *lv;
|
|
struct lv_segment *seg;
|
|
struct lv_list *lvl;
|
|
struct dm_list trim_list;
|
|
uint32_t log_count = 0;
|
|
int lv_found = 0;
|
|
int lv_skipped = 0;
|
|
int lv_active_count = 0;
|
|
int lv_exclusive_count = 0;
|
|
|
|
/* FIXME Cope with non-contiguous => splitting existing segments */
|
|
if (!(lv_mirr = lv_create_empty("pvmove%d", NULL,
|
|
LVM_READ | LVM_WRITE,
|
|
ALLOC_CONTIGUOUS, vg))) {
|
|
log_error("Creation of temporary pvmove LV failed.");
|
|
return NULL;
|
|
}
|
|
|
|
lv_mirr->status |= (PVMOVE | LOCKED);
|
|
|
|
if (!(*lvs_changed = dm_pool_alloc(cmd->mem, sizeof(**lvs_changed)))) {
|
|
log_error("lvs_changed list struct allocation failed.");
|
|
return NULL;
|
|
}
|
|
|
|
dm_list_init(*lvs_changed);
|
|
|
|
/*
|
|
* First,
|
|
* use top-level RAID and mirror LVs to build a list of PVs
|
|
* that must be avoided during allocation. This is necessary
|
|
* to maintain redundancy of those targets, but it is also
|
|
* sub-optimal. Avoiding entire PVs in this way limits our
|
|
* ability to find space for other segment types. In the
|
|
* majority of cases, however, this method will suffice and
|
|
* in the cases where it does not, the user can issue the
|
|
* pvmove on a per-LV basis.
|
|
*
|
|
* FIXME: Eliminating entire PVs places too many restrictions
|
|
* on allocation.
|
|
*/
|
|
dm_list_iterate_items(lvl, &vg->lvs) {
|
|
lv = lvl->lv;
|
|
if (lv == lv_mirr)
|
|
continue;
|
|
|
|
if (lv_name && strcmp(lv->name, top_level_lv_name(vg, lv_name)))
|
|
continue;
|
|
|
|
/*
|
|
* RAID, thin and snapshot-related LVs are not
|
|
* processed in a cluster, so we don't have to
|
|
* worry about avoiding certain PVs in that context.
|
|
*
|
|
* Allow clustered mirror, but not raid mirror.
|
|
*/
|
|
if (vg_is_clustered(vg) && !lv_is_mirror_type(lv))
|
|
continue;
|
|
|
|
if (!lv_is_on_pvs(lv, source_pvl))
|
|
continue;
|
|
|
|
if (lv_is_converting(lv) || lv_is_merging(lv)) {
|
|
log_error("Unable to pvmove when %s volumes are present.",
|
|
lv_is_converting(lv) ?
|
|
"converting" : "merging");
|
|
return NULL;
|
|
}
|
|
|
|
if (seg_is_raid(first_seg(lv)) ||
|
|
seg_is_mirrored(first_seg(lv))) {
|
|
dm_list_init(&trim_list);
|
|
|
|
if (!get_pv_list_for_lv(vg->cmd->mem, lv, &trim_list))
|
|
return_NULL;
|
|
|
|
/*
|
|
* Remove any PVs holding SubLV siblings to allow
|
|
* for collocation (e.g. *rmeta_0 -> *rimage_0).
|
|
*
|
|
* Callee checks for lv_name and valid raid segment type.
|
|
*
|
|
* FIXME: don't rely on namespace
|
|
*/
|
|
if (!_remove_sibling_pvs_from_trim_list(lv, lv_name, &trim_list))
|
|
return_NULL;
|
|
|
|
if (!_trim_allocatable_pvs(allocatable_pvs,
|
|
&trim_list, alloc))
|
|
return_NULL;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Second,
|
|
* use bottom-level LVs (like *_mimage_*, *_mlog, *_rmeta_*, etc)
|
|
* to find segments to be moved and then set up mirrors.
|
|
*/
|
|
dm_list_iterate_items(lvl, &vg->lvs) {
|
|
lv = lvl->lv;
|
|
if (lv == lv_mirr)
|
|
continue;
|
|
|
|
if (lv_name) {
|
|
if (strcmp(lv->name, lv_name) && !_sub_lv_of(lv, lv_name))
|
|
continue;
|
|
lv_found = 1;
|
|
}
|
|
|
|
if (!lv_is_on_pvs(lv, source_pvl))
|
|
continue;
|
|
|
|
if (lv_is_cache_type(lv)) {
|
|
log_print_unless_silent("Skipping %s LV, %s.",
|
|
lv_is_cache(lv) ? "cache" :
|
|
lv_is_cache_pool(lv) ?
|
|
"cache-pool" : "cache-related",
|
|
lv->name);
|
|
lv_skipped = 1;
|
|
continue;
|
|
}
|
|
|
|
if (_parent_lv_is_cache_type(lv)) {
|
|
log_print_unless_silent("Skipping %s because a parent"
|
|
" is of cache type.", lv->name);
|
|
lv_skipped = 1;
|
|
continue;
|
|
}
|
|
|
|
/*
|
|
* If the VG is clustered, we are unable to handle
|
|
* snapshots, origins, thin types, RAID or mirror
|
|
*/
|
|
if ((vg_is_clustered(vg) || is_lockd_type(vg->lock_type)) &&
|
|
(lv_is_origin(lv) || lv_is_cow(lv) ||
|
|
lv_is_thin_type(lv) || lv_is_raid_type(lv))) {
|
|
log_print_unless_silent("Skipping %s LV %s.",
|
|
lv_is_origin(lv) ? "origin" :
|
|
lv_is_cow(lv) ?
|
|
"snapshot-related" :
|
|
lv_is_thin_volume(lv) ? "thin" :
|
|
lv_is_thin_pool(lv) ?
|
|
"thin-pool" :
|
|
lv_is_thin_type(lv) ?
|
|
"thin-related" :
|
|
seg_is_raid(first_seg(lv)) ?
|
|
"RAID" :
|
|
lv_is_raid_type(lv) ?
|
|
"RAID-related" : "",
|
|
lv->name);
|
|
lv_skipped = 1;
|
|
continue;
|
|
}
|
|
|
|
seg = first_seg(lv);
|
|
if (seg_is_raid(seg) || seg_is_mirrored(seg) ||
|
|
lv_is_thin_volume(lv) || lv_is_thin_pool(lv)) {
|
|
/*
|
|
* Pass over top-level LVs - they were handled.
|
|
* Allow sub-LVs to proceed.
|
|
*/
|
|
continue;
|
|
}
|
|
|
|
if (lv_is_locked(lv)) {
|
|
lv_skipped = 1;
|
|
log_print_unless_silent("Skipping locked LV %s.", lv->name);
|
|
continue;
|
|
}
|
|
|
|
if (vg_is_clustered(vg) &&
|
|
lv_is_active_exclusive_remotely(lv)) {
|
|
lv_skipped = 1;
|
|
log_print_unless_silent("Skipping LV %s which is activated "
|
|
"exclusively on remote node.", lv->name);
|
|
continue;
|
|
}
|
|
|
|
if (vg_is_clustered(vg)) {
|
|
if (lv_is_active_exclusive_locally(lv))
|
|
lv_exclusive_count++;
|
|
else if (lv_is_active(lv))
|
|
lv_active_count++;
|
|
}
|
|
|
|
if (!_insert_pvmove_mirrors(cmd, lv_mirr, source_pvl, lv,
|
|
*lvs_changed))
|
|
return_NULL;
|
|
}
|
|
|
|
if (lv_name && !lv_found) {
|
|
/* NOTE: Is this now an internal error? It is already checked in _pvmove_setup_single */
|
|
log_error("Logical volume %s not found.", lv_name);
|
|
return NULL;
|
|
}
|
|
|
|
/* Is temporary mirror empty? */
|
|
if (!lv_mirr->le_count) {
|
|
if (lv_skipped)
|
|
log_error("All data on source PV skipped. "
|
|
"It contains locked, hidden or "
|
|
"non-top level LVs only.");
|
|
log_error("No data to move for %s.", vg->name);
|
|
return NULL;
|
|
}
|
|
|
|
if (vg_is_clustered(vg) && lv_active_count && *exclusive) {
|
|
log_error("Cannot move in clustered VG %s, "
|
|
"clustered mirror (cmirror) not detected "
|
|
"and LVs are activated non-exclusively.",
|
|
vg->name);
|
|
return NULL;
|
|
}
|
|
|
|
if (vg_is_clustered(vg) && lv_exclusive_count) {
|
|
if (lv_active_count) {
|
|
log_error("Cannot move in clustered VG %s "
|
|
"if some LVs are activated "
|
|
"exclusively while others don't.",
|
|
vg->name);
|
|
return NULL;
|
|
}
|
|
*exclusive = 1;
|
|
}
|
|
|
|
if (!lv_add_mirrors(cmd, lv_mirr, 1, 1, 0, 0, log_count,
|
|
allocatable_pvs, alloc,
|
|
(arg_is_set(cmd, atomic_ARG)) ?
|
|
MIRROR_BY_SEGMENTED_LV : MIRROR_BY_SEG)) {
|
|
log_error("Failed to convert pvmove LV to mirrored.");
|
|
return NULL;
|
|
}
|
|
|
|
if (!split_parent_segments_for_layer(cmd, lv_mirr)) {
|
|
log_error("Failed to split segments being moved.");
|
|
return NULL;
|
|
}
|
|
|
|
return lv_mirr;
|
|
}
|
|
|
|
static int _activate_lv(struct cmd_context *cmd, struct logical_volume *lv_mirr,
|
|
unsigned exclusive)
|
|
{
|
|
int r = 0;
|
|
|
|
if (exclusive || lv_is_active_exclusive(lv_mirr))
|
|
r = activate_lv_excl(cmd, lv_mirr);
|
|
else
|
|
r = activate_lv(cmd, lv_mirr);
|
|
|
|
if (!r)
|
|
stack;
|
|
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* Called to set up initial pvmove LV only.
|
|
* (Not called after first or any other section completes.)
|
|
*/
|
|
static int _update_metadata(struct cmd_context *cmd, struct volume_group *vg,
|
|
struct logical_volume *lv_mirr,
|
|
struct dm_list *lvs_changed, unsigned exclusive)
|
|
{
|
|
int r = 0;
|
|
|
|
log_verbose("Setting up pvmove in on-disk volume group metadata.");
|
|
if (!vg_write(vg)) {
|
|
log_error("ABORTING: Volume group metadata update failed.");
|
|
return 0;
|
|
}
|
|
|
|
if (!suspend_lvs(cmd, lvs_changed, vg)) {
|
|
log_error("ABORTING: Temporary pvmove mirror activation failed.");
|
|
/* FIXME Add a recovery path for first time too. */
|
|
return 0;
|
|
}
|
|
|
|
/* Commit on-disk metadata */
|
|
if (!vg_commit(vg)) {
|
|
log_error("ABORTING: Volume group metadata update failed.");
|
|
if (!resume_lvs(cmd, lvs_changed))
|
|
log_error("Unable to resume logical volumes.");
|
|
return 0;
|
|
}
|
|
|
|
/* Activate the temporary mirror LV */
|
|
/* Only the first mirror segment gets activated as a mirror */
|
|
/* FIXME: Add option to use a log */
|
|
if (!exclusive && _pvmove_is_exclusive(cmd, vg))
|
|
exclusive = 1;
|
|
|
|
if (!_activate_lv(cmd, lv_mirr, exclusive)) {
|
|
if (test_mode()) {
|
|
r = 1;
|
|
goto out;
|
|
}
|
|
|
|
/*
|
|
* FIXME Run --abort internally here.
|
|
*/
|
|
log_error("ABORTING: Temporary pvmove mirror activation failed. Run pvmove --abort.");
|
|
goto out;
|
|
}
|
|
|
|
r = 1;
|
|
|
|
out:
|
|
if (!resume_lvs(cmd, lvs_changed)) {
|
|
log_error("Unable to resume logical volumes.");
|
|
r = 0;
|
|
}
|
|
|
|
if (r)
|
|
backup(vg);
|
|
|
|
return r;
|
|
}
|
|
|
|
static int _copy_id_components(struct cmd_context *cmd,
|
|
const struct logical_volume *lv, char **vg_name,
|
|
char **lv_name, union lvid *lvid)
|
|
{
|
|
if (!(*vg_name = dm_pool_strdup(cmd->mem, lv->vg->name)) ||
|
|
!(*lv_name = dm_pool_strdup(cmd->mem, lv->name))) {
|
|
log_error("Failed to clone VG or LV name.");
|
|
return 0;
|
|
}
|
|
|
|
*lvid = lv->lvid;
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int _pvmove_setup_single(struct cmd_context *cmd,
|
|
struct volume_group *vg,
|
|
struct physical_volume *pv,
|
|
struct processing_handle *handle)
|
|
{
|
|
struct pvmove_params *pp = (struct pvmove_params *) handle->custom_handle;
|
|
const char *lv_name = NULL;
|
|
struct dm_list *source_pvl;
|
|
struct dm_list *allocatable_pvs;
|
|
struct dm_list *lvs_changed;
|
|
struct logical_volume *lv_mirr;
|
|
struct logical_volume *lv = NULL;
|
|
const char *pv_name = pv_dev_name(pv);
|
|
unsigned flags = PVMOVE_FIRST_TIME;
|
|
unsigned exclusive;
|
|
int r = ECMD_FAILED;
|
|
|
|
pp->found_pv = 1;
|
|
pp->setup_result = ECMD_FAILED;
|
|
|
|
if (pp->lv_name_arg) {
|
|
if (!(lv_name = _extract_lvname(cmd, vg->name, pp->lv_name_arg))) {
|
|
log_error("Failed to find an LV name.");
|
|
pp->setup_result = EINVALID_CMD_LINE;
|
|
return ECMD_FAILED;
|
|
}
|
|
|
|
if (!validate_name(lv_name)) {
|
|
log_error("Logical volume name %s is invalid.", lv_name);
|
|
pp->setup_result = EINVALID_CMD_LINE;
|
|
return ECMD_FAILED;
|
|
}
|
|
|
|
if (!(lv = find_lv(vg, lv_name))) {
|
|
log_error("Logical volume %s not found.", lv_name);
|
|
return ECMD_FAILED;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* We would need to avoid any PEs used by LVs that are active (ex) on
|
|
* other hosts. For LVs that are active on multiple hosts (sh), we
|
|
* would need to used cluster mirrors.
|
|
*/
|
|
if (is_lockd_type(vg->lock_type)) {
|
|
if (!lv) {
|
|
log_error("pvmove in a shared VG requires a named LV.");
|
|
return ECMD_FAILED;
|
|
}
|
|
|
|
if (lv_is_lockd_sanlock_lv(lv)) {
|
|
log_error("pvmove not allowed on internal sanlock LV.");
|
|
return ECMD_FAILED;
|
|
}
|
|
|
|
if (!lockd_lv(cmd, lv, "ex", LDLV_PERSISTENT)) {
|
|
log_error("pvmove in a shared VG requires exclusive lock on named LV.");
|
|
return ECMD_FAILED;
|
|
}
|
|
}
|
|
|
|
exclusive = _pvmove_is_exclusive(cmd, vg);
|
|
|
|
if ((lv_mirr = find_pvmove_lv(vg, pv_dev(pv), PVMOVE))) {
|
|
log_print_unless_silent("Detected pvmove in progress for %s.", pv_name);
|
|
if (pp->pv_count || lv_name)
|
|
log_warn("WARNING: Ignoring remaining command line arguments.");
|
|
|
|
if (!(lvs_changed = lvs_using_lv(cmd, vg, lv_mirr))) {
|
|
log_error("ABORTING: Failed to generate list of moving LVs.");
|
|
goto out;
|
|
}
|
|
|
|
/* Ensure mirror LV is active */
|
|
if (!_activate_lv(cmd, lv_mirr, exclusive)) {
|
|
log_error("ABORTING: Temporary mirror activation failed.");
|
|
goto out;
|
|
}
|
|
|
|
flags &= ~PVMOVE_FIRST_TIME;
|
|
} else {
|
|
/* Determine PE ranges to be moved */
|
|
if (!(source_pvl = create_pv_list(cmd->mem, vg, 1,
|
|
&pp->pv_name_arg, 0)))
|
|
goto_out;
|
|
|
|
if (pp->alloc == ALLOC_INHERIT)
|
|
pp->alloc = vg->alloc;
|
|
|
|
/* Get PVs we can use for allocation */
|
|
if (!(allocatable_pvs = _get_allocatable_pvs(cmd, pp->pv_count, pp->pv_names,
|
|
vg, pv, pp->alloc)))
|
|
goto_out;
|
|
|
|
if (!archive(vg))
|
|
goto_out;
|
|
|
|
if (!(lv_mirr = _set_up_pvmove_lv(cmd, vg, source_pvl, lv_name,
|
|
allocatable_pvs, pp->alloc,
|
|
&lvs_changed, &exclusive)))
|
|
goto_out;
|
|
}
|
|
|
|
/* Lock lvs_changed and activate (with old metadata) */
|
|
if (!activate_lvs(cmd, lvs_changed, exclusive))
|
|
goto_out;
|
|
|
|
/* FIXME Presence of a mirror once set PVMOVE - now remove associated logic */
|
|
/* init_pvmove(1); */
|
|
/* vg->status |= PVMOVE; */
|
|
|
|
if (!_copy_id_components(cmd, lv_mirr, &pp->id_vg_name, &pp->id_lv_name, pp->lvid))
|
|
goto out;
|
|
|
|
if (flags & PVMOVE_FIRST_TIME)
|
|
if (!_update_metadata(cmd, vg, lv_mirr, lvs_changed, exclusive))
|
|
goto_out;
|
|
|
|
/* LVs are all in status LOCKED */
|
|
pp->setup_result = ECMD_PROCESSED;
|
|
r = ECMD_PROCESSED;
|
|
out:
|
|
return r;
|
|
}
|
|
|
|
static int _pvmove_read_single(struct cmd_context *cmd,
|
|
struct volume_group *vg,
|
|
struct physical_volume *pv,
|
|
struct processing_handle *handle)
|
|
{
|
|
struct pvmove_params *pp = (struct pvmove_params *) handle->custom_handle;
|
|
struct logical_volume *lv;
|
|
int ret = ECMD_FAILED;
|
|
|
|
pp->found_pv = 1;
|
|
|
|
if (!(lv = find_pvmove_lv(vg, pv_dev(pv), PVMOVE))) {
|
|
log_print_unless_silent("%s: No pvmove in progress - already finished or aborted.",
|
|
pv_dev_name(pv));
|
|
ret = ECMD_PROCESSED;
|
|
pp->in_progress = 0;
|
|
} else if (_copy_id_components(cmd, lv, &pp->id_vg_name, &pp->id_lv_name, pp->lvid)) {
|
|
ret = ECMD_PROCESSED;
|
|
pp->in_progress = 1;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static struct poll_functions _pvmove_fns = {
|
|
.get_copy_name_from_lv = get_pvmove_pvname_from_lv_mirr,
|
|
.poll_progress = poll_mirror_progress,
|
|
.update_metadata = pvmove_update_metadata,
|
|
.finish_copy = pvmove_finish,
|
|
};
|
|
|
|
static struct poll_operation_id *_pvmove_create_id(struct cmd_context *cmd,
|
|
const char *pv_name,
|
|
const char *vg_name,
|
|
const char *lv_name,
|
|
const char *uuid)
|
|
{
|
|
struct poll_operation_id *id;
|
|
|
|
if (!vg_name || !lv_name || !pv_name || !uuid) {
|
|
log_error(INTERNAL_ERROR "Wrong params for _pvmove_create_id.");
|
|
return NULL;
|
|
}
|
|
|
|
if (!(id = dm_pool_alloc(cmd->mem, sizeof(*id)))) {
|
|
log_error("Poll operation ID allocation failed.");
|
|
return NULL;
|
|
}
|
|
|
|
if (!(id->vg_name = dm_pool_strdup(cmd->mem, vg_name)) ||
|
|
!(id->lv_name = dm_pool_strdup(cmd->mem, lv_name)) ||
|
|
!(id->display_name = dm_pool_strdup(cmd->mem, pv_name)) ||
|
|
!(id->uuid = dm_pool_strdup(cmd->mem, uuid))) {
|
|
log_error("Failed to copy one or more poll operation ID members.");
|
|
dm_pool_free(cmd->mem, id);
|
|
return NULL;
|
|
}
|
|
|
|
return id;
|
|
}
|
|
|
|
int pvmove_poll(struct cmd_context *cmd, const char *pv_name,
|
|
const char *uuid, const char *vg_name,
|
|
const char *lv_name, unsigned background)
|
|
{
|
|
struct poll_operation_id *id = NULL;
|
|
|
|
if (uuid &&
|
|
!(id = _pvmove_create_id(cmd, pv_name, vg_name, lv_name, uuid))) {
|
|
log_error("Failed to allocate poll identifier for pvmove.");
|
|
return ECMD_FAILED;
|
|
}
|
|
|
|
if (test_mode())
|
|
return ECMD_PROCESSED;
|
|
|
|
return poll_daemon(cmd, background, PVMOVE, &_pvmove_fns, "Moved", id);
|
|
}
|
|
|
|
int pvmove(struct cmd_context *cmd, int argc, char **argv)
|
|
{
|
|
struct pvmove_params pp = { 0 };
|
|
struct processing_handle *handle = NULL;
|
|
union lvid *lvid = NULL;
|
|
char *pv_name = NULL;
|
|
char *colon;
|
|
unsigned is_abort = arg_is_set(cmd, abort_ARG);
|
|
|
|
/* dm raid1 target must be present in every case */
|
|
if (!_pvmove_target_present(cmd, 0)) {
|
|
log_error("Required device-mapper target(s) not "
|
|
"detected in your kernel.");
|
|
return ECMD_FAILED;
|
|
}
|
|
|
|
if (lvmlockd_use() && !lvmpolld_use()) {
|
|
/*
|
|
* Don't want to spend the time making lvmlockd
|
|
* work without lvmpolld.
|
|
*/
|
|
log_error("Enable lvmpolld when using lvmlockd.");
|
|
return ECMD_FAILED;
|
|
}
|
|
|
|
if (lvmlockd_use() && !argc) {
|
|
/*
|
|
* FIXME: move process_each_vg from polldaemon up to here,
|
|
* then we can remove this limitation.
|
|
*/
|
|
log_error("Specify pvmove args when using lvmlockd.");
|
|
return ECMD_FAILED;
|
|
}
|
|
|
|
if (argc) {
|
|
if (!(lvid = dm_pool_alloc(cmd->mem, sizeof(*lvid)))) {
|
|
log_error("Failed to allocate lvid.");
|
|
return ECMD_FAILED;
|
|
}
|
|
pp.lvid = lvid;
|
|
|
|
if (!(pp.pv_name_arg = dm_pool_strdup(cmd->mem, argv[0]))) {
|
|
log_error("Failed to clone PV name.");
|
|
return ECMD_FAILED;
|
|
}
|
|
|
|
if (!(pv_name = dm_pool_strdup(cmd->mem, argv[0]))) {
|
|
log_error("Failed to clone PV name.");
|
|
return ECMD_FAILED;
|
|
}
|
|
|
|
dm_unescape_colons_and_at_signs(pv_name, &colon, NULL);
|
|
|
|
/* Drop any PE lists from PV name */
|
|
if (colon)
|
|
*colon = '\0';
|
|
|
|
argc--;
|
|
argv++;
|
|
|
|
pp.pv_count = argc;
|
|
pp.pv_names = argv;
|
|
|
|
if (arg_is_set(cmd, name_ARG)) {
|
|
if (!(pp.lv_name_arg = dm_pool_strdup(cmd->mem, arg_value(cmd, name_ARG)))) {
|
|
log_error("Failed to clone LV name.");
|
|
return ECMD_FAILED;
|
|
}
|
|
}
|
|
|
|
pp.alloc = (alloc_policy_t) arg_uint_value(cmd, alloc_ARG, ALLOC_INHERIT);
|
|
|
|
pp.in_progress = 1;
|
|
|
|
/* Normal pvmove setup requires ex lock from lvmlockd. */
|
|
if (is_abort)
|
|
cmd->lockd_vg_default_sh = 1;
|
|
|
|
if (!(handle = init_processing_handle(cmd, NULL))) {
|
|
log_error("Failed to initialize processing handle.");
|
|
return ECMD_FAILED;
|
|
}
|
|
|
|
handle->custom_handle = &pp;
|
|
|
|
process_each_pv(cmd, 1, &pv_name, NULL, 0,
|
|
is_abort ? 0 : READ_FOR_UPDATE,
|
|
handle,
|
|
is_abort ? &_pvmove_read_single : &_pvmove_setup_single);
|
|
|
|
destroy_processing_handle(cmd, handle);
|
|
|
|
if (!is_abort) {
|
|
if (!pp.found_pv) {
|
|
stack;
|
|
return EINVALID_CMD_LINE;
|
|
}
|
|
|
|
if (pp.setup_result != ECMD_PROCESSED) {
|
|
stack;
|
|
return pp.setup_result;
|
|
}
|
|
} else {
|
|
if (!pp.found_pv)
|
|
return_ECMD_FAILED;
|
|
|
|
if (!pp.in_progress)
|
|
return ECMD_PROCESSED;
|
|
}
|
|
|
|
/*
|
|
* The command may sit and report progress for some time,
|
|
* and we do not want or need the lockd locks held during
|
|
* that time.
|
|
*/
|
|
lockd_gl(cmd, "un", 0);
|
|
}
|
|
|
|
return pvmove_poll(cmd, pv_name, lvid ? lvid->s : NULL,
|
|
pp.id_vg_name, pp.id_lv_name,
|
|
arg_is_set(cmd, background_ARG));
|
|
}
|