1
0
mirror of git://sourceware.org/git/lvm2.git synced 2024-10-28 03:27:58 +03:00
lvm2/tools/pvchange.c
Peter Rajnoha 7f2eebf519 select: initialize selection handle for process_each_* fns with initial report type
This is a followup patch for previous patchset that enables selection in
process_each_* fns to fix an issue where field prefixes are not
automatically used for fields in selection criteria.

Use initial report type that matches the intention of each process_each_* functions:
- _process_pvs_in_vg - PVS
- process_each_vg - VGS
- process_each_lv and process_each_lv_in_vg - LVS

This is not normally needed for the selection handle init, BUT we would
miss the field prefix matching, e.g.

lvchange -ay -S 'name=lvol0'

The "name" above would not work if we didn't initialize reporting with
the LVS type at its start. If we pass proper init type, reporting code
can deduce the prefix automatically ("lv_name" in this case).

This report type is then changed further based on what selection criteria we
have. When doing pure selection, not report output, the final report type
is purely based on combination of this initial report type and report types
of the fields used in selection criteria.
2015-02-10 16:10:17 +01:00

277 lines
7.5 KiB
C

/*
* Copyright (C) 2001-2004 Sistina Software, Inc. All rights reserved.
* Copyright (C) 2004-2007 Red Hat, Inc. All rights reserved.
*
* This file is part of LVM2.
*
* This copyrighted material is made available to anyone wishing to use,
* modify, copy, or redistribute it subject to the terms and conditions
* of the GNU Lesser General Public License v.2.1.
*
* You should have received a copy of the GNU Lesser General Public License
* along with this program; if not, write to the Free Software Foundation,
* Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
*/
#include "tools.h"
static int _pvchange_single(struct cmd_context *cmd, struct volume_group *vg,
struct physical_volume *pv,
void *handle __attribute__((unused)))
{
const char *pv_name = pv_dev_name(pv);
char uuid[64] __attribute__((aligned(8)));
int allocatable = arg_int_value(cmd, allocatable_ARG, 0);
int mda_ignore = arg_int_value(cmd, metadataignore_ARG, 0);
int tagargs = arg_count(cmd, addtag_ARG) + arg_count(cmd, deltag_ARG);
/* If in a VG, must change using volume group. */
if (!is_orphan(pv)) {
if (tagargs && !(vg->fid->fmt->features & FMT_TAGS)) {
log_error("Volume group containing %s does not "
"support tags", pv_name);
return 0;
}
if (arg_count(cmd, uuid_ARG) && lvs_in_vg_activated(vg)) {
log_error("Volume group containing %s has active "
"logical volumes", pv_name);
return 0;
}
if (!archive(vg))
return 0;
} else {
if (tagargs) {
log_error("Can't change tag on Physical Volume %s not "
"in volume group", pv_name);
return 0;
}
}
if (arg_count(cmd, allocatable_ARG)) {
if (is_orphan(pv) &&
!(pv->fmt->features & FMT_ORPHAN_ALLOCATABLE)) {
log_error("Allocatability not supported by orphan "
"%s format PV %s", pv->fmt->name, pv_name);
return 0;
}
/* change allocatability for a PV */
if (allocatable && (pv_status(pv) & ALLOCATABLE_PV)) {
log_warn("Physical volume \"%s\" is already "
"allocatable.", pv_name);
return 1;
}
if (!allocatable && !(pv_status(pv) & ALLOCATABLE_PV)) {
log_warn("Physical volume \"%s\" is already "
"unallocatable.", pv_name);
return 1;
}
if (allocatable) {
log_verbose("Setting physical volume \"%s\" "
"allocatable", pv_name);
pv->status |= ALLOCATABLE_PV;
} else {
log_verbose("Setting physical volume \"%s\" NOT "
"allocatable", pv_name);
pv->status &= ~ALLOCATABLE_PV;
}
}
if (tagargs) {
/* tag or deltag */
if (arg_count(cmd, addtag_ARG) && !change_tag(cmd, NULL, NULL, pv, addtag_ARG))
return_0;
if (arg_count(cmd, deltag_ARG) && !change_tag(cmd, NULL, NULL, pv, deltag_ARG))
return_0;
}
if (arg_count(cmd, metadataignore_ARG)) {
if ((vg_mda_copies(vg) != VGMETADATACOPIES_UNMANAGED) &&
(arg_count(cmd, force_ARG) == PROMPT) &&
yes_no_prompt("Override preferred number of copies "
"of VG %s metadata? [y/n]: ",
pv_vg_name(pv)) == 'n') {
log_error("Physical volume %s not changed", pv_name);
return 0;
}
if (!pv_change_metadataignore(pv, mda_ignore))
return_0;
}
if (arg_count(cmd, uuid_ARG)) {
/* --uuid: Change PV ID randomly */
memcpy(&pv->old_id, &pv->id, sizeof(pv->id));
if (!id_create(&pv->id)) {
log_error("Failed to generate new random UUID for %s.",
pv_name);
return 0;
}
if (!id_write_format(&pv->id, uuid, sizeof(uuid)))
return 0;
log_verbose("Changing uuid of %s to %s.", pv_name, uuid);
if (!is_orphan(pv) && (!pv_write(cmd, pv, 1))) {
log_error("pv_write with new uuid failed "
"for %s.", pv_name);
return 0;
}
}
log_verbose("Updating physical volume \"%s\"", pv_name);
if (!is_orphan(pv)) {
if (!vg_write(vg) || !vg_commit(vg)) {
log_error("Failed to store physical volume \"%s\" in "
"volume group \"%s\"", pv_name, vg->name);
return 0;
}
backup(vg);
} else if (!(pv_write(cmd, pv, 0))) {
log_error("Failed to store physical volume \"%s\"",
pv_name);
return 0;
}
log_print_unless_silent("Physical volume \"%s\" changed", pv_name);
return 1;
}
int pvchange(struct cmd_context *cmd, int argc, char **argv)
{
int opt = 0;
int done = 0;
int total = 0;
int selected;
struct processing_handle *handle = NULL;
struct volume_group *vg;
const char *vg_name;
char *pv_name;
struct pv_list *pvl;
struct dm_list *vgnames;
struct dm_str_list *sll;
int r = ECMD_PROCESSED;
if (!(arg_count(cmd, allocatable_ARG) + arg_is_set(cmd, addtag_ARG) +
arg_is_set(cmd, deltag_ARG) + arg_count(cmd, uuid_ARG) +
arg_count(cmd, metadataignore_ARG))) {
log_error("Please give one or more of -x, -uuid, "
"--addtag, --deltag or --metadataignore");
r = EINVALID_CMD_LINE;
goto out;
}
/* FIXME: use process_each_pv for pvchange. */
if (!(handle = init_processing_handle(cmd)) ||
(handle->internal_report_for_select && !init_selection_handle(cmd, handle, PVS))) {
log_error("Failed to initialize processing handle.");
r = ECMD_FAILED;
goto out;
}
if (!(arg_count(cmd, all_ARG)) && !argc && !handle->internal_report_for_select) {
log_error("Please give a physical volume path "
"or use -S for selection.");
r = EINVALID_CMD_LINE;
goto out;
}
if (arg_count(cmd, all_ARG) && argc) {
log_error("Option --all and PhysicalVolumePath are exclusive.");
r = EINVALID_CMD_LINE;
goto out;
}
if (argc) {
log_verbose("Using physical volume(s) on command line");
for (; opt < argc; opt++) {
total++;
pv_name = argv[opt];
dm_unescape_colons_and_at_signs(pv_name, NULL, NULL);
vg_name = find_vgname_from_pvname(cmd, pv_name);
if (!vg_name) {
log_error("Failed to read physical volume %s",
pv_name);
continue;
}
vg = vg_read_for_update(cmd, vg_name, NULL, 0);
if (vg_read_error(vg)) {
release_vg(vg);
stack;
continue;
}
pvl = find_pv_in_vg(vg, pv_name);
if (!pvl || !pvl->pv) {
unlock_and_release_vg(cmd, vg, vg_name);
log_error("Unable to find %s in %s",
pv_name, vg_name);
continue;
}
done += _pvchange_single(cmd, vg,
pvl->pv, NULL);
unlock_and_release_vg(cmd, vg, vg_name);
}
} else {
log_verbose("Scanning for physical volume names");
/* FIXME: share code with toollib */
/*
* Take the global lock here so the lvmcache remains
* consistent across orphan/non-orphan vg locks. If we don't
* take the lock here, pvs with 0 mdas in a non-orphan VG will
* be processed twice.
*/
if (!lock_vol(cmd, VG_GLOBAL, LCK_VG_WRITE, NULL)) {
log_error("Unable to obtain global lock.");
r = ECMD_FAILED;
goto out;
}
/* populate lvmcache */
if (!lvmetad_vg_list_to_lvmcache(cmd))
stack;
if ((vgnames = get_vgnames(cmd, 1)) &&
!dm_list_empty(vgnames)) {
dm_list_iterate_items(sll, vgnames) {
vg = vg_read_for_update(cmd, sll->str, NULL, 0);
if (vg_read_error(vg)) {
release_vg(vg);
stack;
continue;
}
dm_list_iterate_items(pvl, &vg->pvs) {
if (select_match_pv(cmd, handle, vg, pvl->pv,
&selected) && selected) {
total++;
done += _pvchange_single(cmd, vg,
pvl->pv,
NULL);
}
}
unlock_and_release_vg(cmd, vg, sll->str);
}
}
unlock_vg(cmd, VG_GLOBAL);
}
log_print_unless_silent("%d physical volume%s changed / %d physical volume%s "
"not changed",
done, done == 1 ? "" : "s",
total - done, (total - done) == 1 ? "" : "s");
out:
destroy_processing_handle(cmd, handle, 1);
if (r == ECMD_PROCESSED)
return (total == done) ? ECMD_PROCESSED : ECMD_FAILED;
else
return r;
}