1
0
mirror of git://sourceware.org/git/lvm2.git synced 2024-10-27 10:25:13 +03:00
lvm2/tools/vgsplit.c
David Teigland b4402bd821 exported vg handling
The exported VG checking/enforcement was scattered and
inconsistent.  This centralizes it and makes it consistent,
following the existing approach for foreign and shared
VGs/PVs, which are very similar to exported VGs/PVs.

The access policy that now applies to foreign/shared/exported
VGs/PVs, is that if a foreign/shared/exported VG/PV is named
on the command line (i.e. explicitly requested by the user),
and the command is not permitted to operate on it because it
is foreign/shared/exported, then an access error is reported
and the command exits with an error.  But, if the command is
processing all VGs/PVs, and happens to come across a
foreign/shared/exported VG/PV (that is not explicitly named on
the command line), then the command silently skips it and does
not produce an error.

A command using tags or --select handles inaccessible VGs/PVs
the same way as a command processing all VGs/PVs, and will
not report/return errors if these inaccessible VGs/PVs exist.

The new policy fixes the exit codes on a somewhat random set of
commands that previously exited with an error if they were
looking at all VGs/PVs and an exported VG existed on the system.

There should be no change to which commands are allowed/disallowed
on exported VGs/PVs.

Certain LV commands (lvs/lvdisplay/lvscan) would previously not
display LVs from an exported VG (for unknown reasons).  This has
not changed.  The lvm fullreport command would previously report
info about an exported VG but not about the LVs in it.  This
has changed to include all info from the exported VG.
2019-06-25 15:39:08 -05:00

708 lines
18 KiB
C

/*
* Copyright (C) 2001-2004 Sistina Software, Inc. All rights reserved.
* Copyright (C) 2004-2019 Red Hat, Inc. All rights reserved.
*
* This file is part of LVM2.
*
* This copyrighted material is made available to anyone wishing to use,
* modify, copy, or redistribute it subject to the terms and conditions
* of the GNU Lesser General Public License v.2.1.
*
* You should have received a copy of the GNU Lesser General Public License
* along with this program; if not, write to the Free Software Foundation,
* Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "tools.h"
static int _lv_is_in_vg(struct volume_group *vg, struct logical_volume *lv)
{
if (!lv || lv->vg != vg)
return 0;
return 1;
}
static struct dm_list *_lvh_in_vg(struct logical_volume *lv, struct volume_group *vg)
{
struct dm_list *lvh;
dm_list_iterate(lvh, &vg->lvs)
if (lv == dm_list_item(lvh, struct lv_list)->lv)
return lvh;
return NULL;
}
static int _lv_tree_move(struct dm_list *lvh,
struct dm_list **lvht,
struct volume_group *vg_from,
struct volume_group *vg_to)
{
uint32_t s;
struct logical_volume *lv = dm_list_item(lvh, struct lv_list)->lv;
struct lv_segment *seg = first_seg(lv);
struct dm_list *lvh1;
/* Update the list pointer refering to the item moving to @vg_to. */
if (lvh == *lvht)
*lvht = dm_list_next(lvh, lvh);
dm_list_move(&vg_to->lvs, lvh);
lv->vg = vg_to;
lv->lvid.id[0] = lv->vg->id;
if (seg)
for (s = 0; s < seg->area_count; s++)
if (seg_type(seg, s) == AREA_LV && seg_lv(seg, s)) {
if ((lvh1 = _lvh_in_vg(seg_lv(seg, s), vg_from))) {
if (!_lv_tree_move(lvh1, lvht, vg_from, vg_to))
return 0;
} else if (!_lvh_in_vg(seg_lv(seg, s), vg_to))
return 0;
}
return 1;
}
static int _move_one_lv(struct volume_group *vg_from,
struct volume_group *vg_to,
struct dm_list *lvh,
struct dm_list **lvht)
{
struct logical_volume *lv = dm_list_item(lvh, struct lv_list)->lv;
struct logical_volume *parent_lv;
if (lv_is_active(lv)) {
if ((parent_lv = lv_parent(lv)))
log_error("Logical volume %s (part of %s) must be inactive.", display_lvname(lv), parent_lv->name);
else
log_error("Logical volume %s must be inactive.", display_lvname(lv));
return 0;
}
/* Bail out, if any allocations of @lv are still on PVs of @vg_from */
if (lv_is_on_pvs(lv, &vg_from->pvs)) {
log_error("Can't split LV %s between "
"two Volume Groups", lv->name);
return 0;
}
if (!_lv_tree_move(lvh, lvht, vg_from, vg_to))
return 0;
/* Moved pool metadata spare LV */
if (vg_from->pool_metadata_spare_lv == lv) {
vg_to->pool_metadata_spare_lv = lv;
vg_from->pool_metadata_spare_lv = NULL;
}
return 1;
}
static int _move_lvs(struct volume_group *vg_from, struct volume_group *vg_to)
{
struct dm_list *lvh, *lvht;
struct logical_volume *lv;
struct lv_segment *seg;
struct physical_volume *pv;
struct volume_group *vg_with;
unsigned s;
dm_list_iterate_safe(lvh, lvht, &vg_from->lvs) {
lv = dm_list_item(lvh, struct lv_list)->lv;
if (lv_is_snapshot(lv))
continue;
if (lv_is_raid(lv))
continue;
if (lv_is_mirrored(lv))
continue;
if (lv_is_thin_pool(lv) ||
lv_is_thin_volume(lv))
continue;
if (lv_is_cache(lv) || lv_is_cache_pool(lv))
/* further checks by _move_cache() */
continue;
/* Ensure all the PVs used by this LV remain in the same */
/* VG as each other */
vg_with = NULL;
dm_list_iterate_items(seg, &lv->segments) {
for (s = 0; s < seg->area_count; s++) {
/* FIXME Check AREA_LV too */
if (seg_type(seg, s) != AREA_PV)
continue;
pv = seg_pv(seg, s);
if (vg_with) {
if (!pv_is_in_vg(vg_with, pv)) {
log_error("Can't split Logical "
"Volume %s between "
"two Volume Groups",
lv->name);
return 0;
}
continue;
}
if (pv_is_in_vg(vg_from, pv)) {
vg_with = vg_from;
continue;
}
if (pv_is_in_vg(vg_to, pv)) {
vg_with = vg_to;
continue;
}
log_error("Physical Volume %s not found",
pv_dev_name(pv));
return 0;
}
}
if (vg_with == vg_from)
continue;
/* Move this LV */
if (!_move_one_lv(vg_from, vg_to, lvh, &lvht))
return_0;
}
/* FIXME Ensure no LVs contain segs pointing at LVs in the other VG */
return 1;
}
/*
* Move the hidden / internal "snapshotN" LVs.from 'vg_from' to 'vg_to'.
*/
static int _move_snapshots(struct volume_group *vg_from,
struct volume_group *vg_to)
{
struct dm_list *lvh, *lvht;
struct logical_volume *lv;
struct lv_segment *seg;
int cow_from = 0;
int origin_from = 0;
dm_list_iterate_safe(lvh, lvht, &vg_from->lvs) {
lv = dm_list_item(lvh, struct lv_list)->lv;
if (!lv_is_snapshot(lv))
continue;
dm_list_iterate_items(seg, &lv->segments) {
cow_from = _lv_is_in_vg(vg_from, seg->cow);
origin_from = _lv_is_in_vg(vg_from, seg->origin);
if (cow_from && origin_from)
continue;
if ((!cow_from && origin_from) ||
(cow_from && !origin_from)) {
log_error("Can't split snapshot %s between"
" two Volume Groups", seg->cow->name);
return 0;
}
/*
* At this point, the cow and origin should already be
* in vg_to.
*/
if (_lv_is_in_vg(vg_to, seg->cow) &&
_lv_is_in_vg(vg_to, seg->origin)) {
if (!_move_one_lv(vg_from, vg_to, lvh, &lvht))
return_0;
}
}
}
return 1;
}
static int _move_mirrors(struct volume_group *vg_from,
struct volume_group *vg_to)
{
struct dm_list *lvh, *lvht;
struct logical_volume *lv;
struct lv_segment *seg, *log_seg;
unsigned s, seg_in, log_in;
dm_list_iterate_safe(lvh, lvht, &vg_from->lvs) {
lv = dm_list_item(lvh, struct lv_list)->lv;
if (lv_is_raid(lv))
continue;
if (!lv_is_mirrored(lv))
continue;
/* Ignore, if no allocations on PVs of @vg_to */
if (!lv_is_on_pvs(lv, &vg_to->pvs))
continue;
seg = first_seg(lv);
seg_in = 0;
for (s = 0; s < seg->area_count; s++)
if (_lv_is_in_vg(vg_to, seg_lv(seg, s)))
seg_in++;
log_in = !seg->log_lv;
if (seg->log_lv) {
log_seg = first_seg(seg->log_lv);
if (seg_is_mirrored(log_seg)) {
log_in = 1;
/* Ensure each log dev is in vg_to */
for (s = 0; s < log_seg->area_count; s++)
log_in = log_in &&
_lv_is_in_vg(vg_to,
seg_lv(log_seg, s));
} else
log_in = _lv_is_in_vg(vg_to, seg->log_lv);
}
if ((seg_in && seg_in < seg->area_count) ||
(seg_in && seg->log_lv && !log_in) ||
(!seg_in && seg->log_lv && log_in)) {
log_error("Can't split mirror %s between "
"two Volume Groups", lv->name);
return 0;
}
if (seg_in == seg->area_count && log_in) {
if (!_move_one_lv(vg_from, vg_to, lvh, &lvht))
return_0;
}
}
return 1;
}
/*
* Check for any RAID LVs with allocations on PVs of @vg_to.
*
* If these don't have any allocations on PVs of @vg_from,
* move their whole lv stack across to @vg_to including the
* top-level RAID LV.
*/
static int _move_raids(struct volume_group *vg_from,
struct volume_group *vg_to)
{
struct dm_list *lvh, *lvht;
struct logical_volume *lv;
dm_list_iterate_safe(lvh, lvht, &vg_from->lvs) {
lv = dm_list_item(lvh, struct lv_list)->lv;
if (!lv_is_raid(lv))
continue;
/* Ignore, if no allocations on PVs of @vg_to */
if (!lv_is_on_pvs(lv, &vg_to->pvs))
continue;
/* If allocations are on PVs of @vg_to -> move RAID LV stack across */
if (!_move_one_lv(vg_from, vg_to, lvh, &lvht))
return_0;
}
return 1;
}
static int _move_thins(struct volume_group *vg_from,
struct volume_group *vg_to)
{
struct dm_list *lvh, *lvht;
struct logical_volume *lv, *data_lv;
struct lv_segment *seg;
dm_list_iterate_safe(lvh, lvht, &vg_from->lvs) {
lv = dm_list_item(lvh, struct lv_list)->lv;
if (lv_is_thin_volume(lv)) {
seg = first_seg(lv);
data_lv = seg_lv(first_seg(seg->pool_lv), 0);
/* Ignore, if no allocations on PVs of @vg_to */
if (!lv_is_on_pvs(data_lv, &vg_to->pvs) &&
(seg->external_lv && !lv_is_on_pvs(seg->external_lv, &vg_to->pvs)))
continue;
if ((_lv_is_in_vg(vg_to, data_lv) ||
_lv_is_in_vg(vg_to, seg->external_lv))) {
if (_lv_is_in_vg(vg_from, seg->external_lv) ||
_lv_is_in_vg(vg_from, data_lv)) {
log_error("Can't split external origin %s "
"and pool %s between two Volume Groups.",
display_lvname(seg->external_lv),
display_lvname(seg->pool_lv));
return 0;
}
if (!_move_one_lv(vg_from, vg_to, lvh, &lvht))
return_0;
}
} else if (lv_is_thin_pool(lv)) {
seg = first_seg(lv);
data_lv = seg_lv(seg, 0);
/* Ignore, if no allocations on PVs of @vg_to */
if (!lv_is_on_pvs(data_lv, &vg_to->pvs))
continue;
if (_lv_is_in_vg(vg_to, data_lv) ||
_lv_is_in_vg(vg_to, seg->metadata_lv)) {
if (_lv_is_in_vg(vg_from, seg->metadata_lv) ||
_lv_is_in_vg(vg_from, data_lv)) {
log_error("Can't split pool data and metadata %s "
"between two Volume Groups.",
lv->name);
return 0;
}
if (!_move_one_lv(vg_from, vg_to, lvh, &lvht))
return_0;
}
}
}
return 1;
}
static int _move_cache(struct volume_group *vg_from,
struct volume_group *vg_to)
{
int is_moving;
struct dm_list *lvh, *lvht;
struct logical_volume *lv, *data, *meta, *orig;
struct lv_segment *seg, *cache_seg;
dm_list_iterate_safe(lvh, lvht, &vg_from->lvs) {
lv = dm_list_item(lvh, struct lv_list)->lv;
seg = first_seg(lv);
if (!lv_is_cache(lv) && !lv_is_cache_pool(lv))
continue;
if (lv_is_cache(lv) && lv_is_cache_vol(seg->pool_lv)) {
log_error("Cannot split while LV %s has cache attached.", display_lvname(lv));
return 0;
}
if (lv_is_cache(lv)) {
orig = seg_lv(seg, 0);
seg = first_seg(seg->pool_lv);
} else { /* lv_is_cache_pool */
orig = NULL;
if (!dm_list_empty(&seg->lv->segs_using_this_lv)) {
if (!(cache_seg = get_only_segment_using_this_lv(seg->lv)))
return_0;
orig = seg_lv(cache_seg, 0);
}
}
data = seg_lv(seg, 0);
meta = seg->metadata_lv;
if ((orig && !lv_is_on_pvs(orig, &vg_to->pvs)) &&
!lv_is_on_pvs(data, &vg_to->pvs) &&
!lv_is_on_pvs(meta, &vg_to->pvs))
continue;
/* Ensure all components are coming along */
if (orig) {
is_moving = _lv_is_in_vg(vg_to, orig);
if (_lv_is_in_vg(vg_to, data) != is_moving) {
log_error("Cannot split cache origin %s and its cache pool data %s "
"into separate VGs.",
display_lvname(orig), display_lvname(data));
return 0;
}
if (_lv_is_in_vg(vg_to, meta) != is_moving) {
log_error("Cannot split cache origin %s and its cache pool metadata %s "
"into separate VGs.",
display_lvname(orig), display_lvname(meta));
return 0;
}
} else if (_lv_is_in_vg(vg_to, data) != _lv_is_in_vg(vg_to, meta)) {
log_error("Cannot split cache pool data %s and its metadata %s "
"into separate VGs.",
display_lvname(data), display_lvname(meta));
return 0;
}
if (!_move_one_lv(vg_from, vg_to, lvh, &lvht))
return_0;
}
return 1;
}
/*
* Has the user given an option related to a new vg as the split destination?
*/
static int _new_vg_option_specified(struct cmd_context *cmd)
{
return(arg_is_set(cmd, clustered_ARG) ||
arg_is_set(cmd, alloc_ARG) ||
arg_is_set(cmd, maxphysicalvolumes_ARG) ||
arg_is_set(cmd, maxlogicalvolumes_ARG) ||
arg_is_set(cmd, vgmetadatacopies_ARG));
}
int vgsplit(struct cmd_context *cmd, int argc, char **argv)
{
struct vgcreate_params vp_new;
struct vgcreate_params vp_def;
const char *vg_name_from, *vg_name_to;
struct volume_group *vg_to = NULL, *vg_from = NULL;
struct lvmcache_vginfo *vginfo_to;
int opt;
int existing_vg = 0;
int r = ECMD_FAILED;
const char *lv_name;
if ((arg_is_set(cmd, name_ARG) + argc) < 3) {
log_error("Existing VG, new VG and either physical volumes "
"or logical volume required.");
return EINVALID_CMD_LINE;
}
if (arg_is_set(cmd, name_ARG) && (argc > 2)) {
log_error("A logical volume name cannot be given with "
"physical volumes.");
return ECMD_FAILED;
}
if (!lock_global(cmd, "ex"))
return_ECMD_FAILED;
clear_hint_file(cmd);
if (arg_is_set(cmd, name_ARG))
lv_name = arg_value(cmd, name_ARG);
else
lv_name = NULL;
vg_name_from = skip_dev_dir(cmd, argv[0], NULL);
vg_name_to = skip_dev_dir(cmd, argv[1], NULL);
argc -= 2;
argv += 2;
if (!strcmp(vg_name_to, vg_name_from)) {
log_error("Duplicate volume group name \"%s\"", vg_name_from);
return ECMD_FAILED;
}
lvmcache_label_scan(cmd);
if (!(vginfo_to = lvmcache_vginfo_from_vgname(vg_name_to, NULL))) {
if (!validate_name(vg_name_to)) {
log_error("Invalid vg name %s.", vg_name_to);
return ECMD_FAILED;
}
if (!lock_vol(cmd, vg_name_to, LCK_VG_WRITE, NULL)) {
log_error("Failed to lock new VG name %s.", vg_name_to);
return ECMD_FAILED;
}
if (!(vg_to = vg_create(cmd, vg_name_to))) {
log_error("Failed to create new VG %s.", vg_name_to);
unlock_vg(cmd, NULL, vg_name_to);
return ECMD_FAILED;
}
} else {
if (!(vg_to = vg_read_for_update(cmd, vg_name_to, NULL, 0, 0))) {
log_error("Failed to read VG %s.", vg_name_to);
return ECMD_FAILED;
}
existing_vg = 1;
}
if (!(vg_from = vg_read_for_update(cmd, vg_name_from, NULL, 0, 0))) {
log_error("Failed to read VG %s.", vg_name_to);
unlock_and_release_vg(cmd, vg_to, vg_name_to);
return ECMD_FAILED;
}
cmd->fmt = vg_from->fid->fmt;
if (existing_vg) {
if (_new_vg_option_specified(cmd)) {
log_error("Volume group \"%s\" exists, but new VG option specified", vg_name_to);
goto bad;
}
if (!vgs_are_compatible(cmd, vg_from, vg_to))
goto_bad;
} else {
if (!vgcreate_params_set_defaults(cmd, &vp_def, vg_from)) {
r = EINVALID_CMD_LINE;
goto_bad;
}
vp_def.vg_name = vg_name_to;
if (!vgcreate_params_set_from_args(cmd, &vp_new, &vp_def)) {
r = EINVALID_CMD_LINE;
goto_bad;
}
if (!vgcreate_params_validate(cmd, &vp_new)) {
r = EINVALID_CMD_LINE;
goto_bad;
}
if (!vg_set_extent_size(vg_to, vp_new.extent_size) ||
!vg_set_max_lv(vg_to, vp_new.max_lv) ||
!vg_set_max_pv(vg_to, vp_new.max_pv) ||
!vg_set_alloc_policy(vg_to, vp_new.alloc) ||
!vg_set_system_id(vg_to, vp_new.system_id) ||
!vg_set_mda_copies(vg_to, vp_new.vgmetadatacopies))
goto_bad;
}
/* Archive vg_from before changing it */
if (!archive(vg_from))
goto_bad;
/* Move PVs across to new structure */
for (opt = 0; opt < argc; opt++) {
dm_unescape_colons_and_at_signs(argv[opt], NULL, NULL);
if (!move_pv(vg_from, vg_to, argv[opt]))
goto_bad;
}
/* If an LV given on the cmdline, move used_by PVs */
if (lv_name && !move_pvs_used_by_lv(vg_from, vg_to, lv_name))
goto_bad;
/*
* First move any required RAID LVs across recursively.
* Reject if they get split between VGs.
*
* This moves the whole LV stack across, thus _move_lvs() below
* ain't hit any of their MetaLVs/DataLVs any more but'll still
* work for all other type specific moves following it.
*/
if (!(_move_raids(vg_from, vg_to)))
goto_bad;
/* Move required sub LVs across, checking consistency */
if (!(_move_lvs(vg_from, vg_to)))
goto_bad;
/* Move required mirrors across */
if (!(_move_mirrors(vg_from, vg_to)))
goto_bad;
/* Move required pools across */
if (!(_move_thins(vg_from, vg_to)))
goto_bad;
/* Move required cache LVs across */
if (!(_move_cache(vg_from, vg_to)))
goto_bad;
/* Move required snapshots across */
if (!(_move_snapshots(vg_from, vg_to)))
goto_bad;
/* Split metadata areas and check if both vgs have at least one area */
if (!(vg_split_mdas(cmd, vg_from, vg_to)) && vg_from->pv_count) {
log_error("Cannot split: Nowhere to store metadata for new Volume Group");
goto bad;
}
/* Set proper name for all PVs in new VG */
if (!vg_rename(cmd, vg_to, vg_name_to))
goto_bad;
/* Set old VG name so the metadata operations recognise that the PVs are in an existing VG */
vg_to->old_name = vg_from->name;
/* store it on disks */
log_verbose("Writing out updated volume groups");
/*
* First, write out the new VG as EXPORTED. We do this first in case
* there is a crash - we will still have the new VG information, in an
* exported state. Recovery after this point would importing and removal
* of the new VG and redoing the vgsplit.
* FIXME: recover automatically or instruct the user?
*/
vg_to->status |= EXPORTED_VG;
if (!archive(vg_to))
goto_bad;
if (!vg_write(vg_to) || !vg_commit(vg_to))
goto_bad;
backup(vg_to);
/*
* Next, write out the updated old VG. If we crash after this point,
* recovery is a vgimport on the new VG.
* FIXME: recover automatically or instruct the user?
*/
if (vg_from->pv_count) {
if (!vg_write(vg_from) || !vg_commit(vg_from))
goto_bad;
backup(vg_from);
}
/*
* Finally, remove the EXPORTED flag from the new VG and write it out.
* We need to unlock vg_to because vg_read_for_update wants to lock it.
*/
if (!test_mode()) {
unlock_vg(cmd, NULL, vg_name_to);
release_vg(vg_to);
/*
* This command uses the exported vg flag internally, but
* exported VGs are not allowed to be split from the command
* level, so ALLOW_EXPORTED is not set in commands.h.
*/
cmd->include_exported_vgs = 1;
vg_to = vg_read_for_update(cmd, vg_name_to, NULL, 0, 0);
if (vg_read_error(vg_to)) {
log_error("Volume group \"%s\" became inconsistent: "
"please fix manually", vg_name_to);
goto bad;
}
}
vg_to->status &= ~EXPORTED_VG;
if (!vg_write(vg_to) || !vg_commit(vg_to))
goto_bad;
backup(vg_to);
log_print_unless_silent("%s volume group \"%s\" successfully split from \"%s\"",
existing_vg ? "Existing" : "New",
vg_to->name, vg_from->name);
r = ECMD_PROCESSED;
bad:
/*
* vg_to references elements moved from vg_from
* so vg_to has to be freed first.
*/
unlock_and_release_vg(cmd, vg_to, vg_name_to);
unlock_and_release_vg(cmd, vg_from, vg_name_from);
return r;
}