mirror of
git://sourceware.org/git/lvm2.git
synced 2024-12-21 13:34:40 +03:00
a77ded3001
"pvcreate_each_params" was a temporary name used to transition from the old "pvcreate_params". Remove the old pvcreate_params struct and rename the new pvcreate_each_params struct to pvcreate_params. Rename various pvcreate_each_params terms to simply pvcreate_params.
231 lines
6.4 KiB
C
231 lines
6.4 KiB
C
/*
|
|
* Copyright (C) 2001-2004 Sistina Software, Inc. All rights reserved.
|
|
* Copyright (C) 2004-2009 Red Hat, Inc. All rights reserved.
|
|
*
|
|
* This file is part of LVM2.
|
|
*
|
|
* This copyrighted material is made available to anyone wishing to use,
|
|
* modify, copy, or redistribute it subject to the terms and conditions
|
|
* of the GNU Lesser General Public License v.2.1.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public License
|
|
* along with this program; if not, write to the Free Software Foundation,
|
|
* Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "tools.h"
|
|
|
|
int vgcreate(struct cmd_context *cmd, int argc, char **argv)
|
|
{
|
|
struct processing_handle *handle;
|
|
struct pvcreate_params pp;
|
|
struct vgcreate_params vp_new;
|
|
struct vgcreate_params vp_def;
|
|
struct volume_group *vg;
|
|
const char *tag;
|
|
const char *clustered_message = "";
|
|
char *vg_name;
|
|
struct arg_value_group_list *current_group;
|
|
uint32_t rc;
|
|
|
|
if (!argc) {
|
|
log_error("Please provide volume group name and "
|
|
"physical volumes");
|
|
return EINVALID_CMD_LINE;
|
|
}
|
|
|
|
vg_name = argv[0];
|
|
argc--;
|
|
argv++;
|
|
|
|
pvcreate_params_set_defaults(&pp);
|
|
|
|
if (!pvcreate_params_from_args(cmd, &pp))
|
|
return EINVALID_CMD_LINE;
|
|
|
|
pp.pv_count = argc;
|
|
pp.pv_names = argv;
|
|
|
|
/* Don't create a new PV on top of an existing PV like pvcreate does. */
|
|
pp.preserve_existing = 1;
|
|
|
|
/* pvcreate within vgcreate cannot be forced. */
|
|
pp.force = 0;
|
|
|
|
if (!vgcreate_params_set_defaults(cmd, &vp_def, NULL))
|
|
return EINVALID_CMD_LINE;
|
|
vp_def.vg_name = vg_name;
|
|
if (!vgcreate_params_set_from_args(cmd, &vp_new, &vp_def))
|
|
return EINVALID_CMD_LINE;
|
|
|
|
if (!vgcreate_params_validate(cmd, &vp_new))
|
|
return EINVALID_CMD_LINE;
|
|
|
|
/*
|
|
* Needed to change the global VG namespace,
|
|
* and to change the set of orphan PVs.
|
|
*/
|
|
if (!lockd_gl_create(cmd, "ex", vp_new.lock_type))
|
|
return_ECMD_FAILED;
|
|
cmd->lockd_gl_disable = 1;
|
|
|
|
lvmcache_seed_infos_from_lvmetad(cmd);
|
|
|
|
/*
|
|
* Check if the VG name already exists. This should be done before
|
|
* creating PVs on any of the devices.
|
|
*/
|
|
if ((rc = vg_lock_newname(cmd, vp_new.vg_name)) != SUCCESS) {
|
|
if (rc == FAILED_EXIST)
|
|
log_error("A volume group called %s already exists.", vp_new.vg_name);
|
|
else
|
|
log_error("Can't get lock for %s.", vp_new.vg_name);
|
|
return ECMD_FAILED;
|
|
}
|
|
|
|
/*
|
|
* FIXME: we have to unlock/relock the new VG name around the pvcreate
|
|
* step because pvcreate needs to destroy lvmcache, which doesn't allow
|
|
* any locks to be held. There shouldn't be any reason to require this
|
|
* VG lock to be released, so the lvmcache destroy rule about locks
|
|
* seems to be unwarranted here.
|
|
*/
|
|
unlock_vg(cmd, vp_new.vg_name);
|
|
|
|
if (!(handle = init_processing_handle(cmd))) {
|
|
log_error("Failed to initialize processing handle.");
|
|
return ECMD_FAILED;
|
|
}
|
|
|
|
if (!pvcreate_each_device(cmd, handle, &pp)) {
|
|
destroy_processing_handle(cmd, handle);
|
|
return_ECMD_FAILED;
|
|
}
|
|
|
|
/* Relock the new VG name, see comment above. */
|
|
if (!lock_vol(cmd, vp_new.vg_name, LCK_VG_WRITE, NULL)) {
|
|
destroy_processing_handle(cmd, handle);
|
|
return_ECMD_FAILED;
|
|
}
|
|
|
|
/*
|
|
* pvcreate_each_device returns with the VG_ORPHANS write lock held,
|
|
* which was used to do pvcreate. Now to create the VG using those
|
|
* PVs, the VG lock will be taken (with the orphan lock already held.)
|
|
*/
|
|
|
|
if (!(vg = vg_create(cmd, vp_new.vg_name)))
|
|
goto_bad;
|
|
|
|
if (vg->fid->fmt->features & FMT_CONFIG_PROFILE)
|
|
vg->profile = vg->cmd->profile_params->global_metadata_profile;
|
|
|
|
if (!vg_set_extent_size(vg, vp_new.extent_size) ||
|
|
!vg_set_max_lv(vg, vp_new.max_lv) ||
|
|
!vg_set_max_pv(vg, vp_new.max_pv) ||
|
|
!vg_set_alloc_policy(vg, vp_new.alloc) ||
|
|
!vg_set_clustered(vg, vp_new.clustered) ||
|
|
!vg_set_system_id(vg, vp_new.system_id) ||
|
|
!vg_set_mda_copies(vg, vp_new.vgmetadatacopies))
|
|
goto_bad;
|
|
|
|
/* attach the pv's */
|
|
if (!vg_extend_each_pv(vg, &pp))
|
|
goto_bad;
|
|
|
|
if (vp_new.max_lv != vg->max_lv)
|
|
log_warn("WARNING: Setting maxlogicalvolumes to %d "
|
|
"(0 means unlimited)", vg->max_lv);
|
|
|
|
if (vp_new.max_pv != vg->max_pv)
|
|
log_warn("WARNING: Setting maxphysicalvolumes to %d "
|
|
"(0 means unlimited)", vg->max_pv);
|
|
|
|
if (arg_count(cmd, addtag_ARG)) {
|
|
dm_list_iterate_items(current_group, &cmd->arg_value_groups) {
|
|
if (!grouped_arg_is_set(current_group->arg_values, addtag_ARG))
|
|
continue;
|
|
|
|
if (!(tag = grouped_arg_str_value(current_group->arg_values, addtag_ARG, NULL))) {
|
|
log_error("Failed to get tag");
|
|
goto bad;
|
|
}
|
|
|
|
if (!vg_change_tag(vg, tag, 1))
|
|
goto_bad;
|
|
}
|
|
}
|
|
|
|
if (vg_is_clustered(vg))
|
|
clustered_message = "Clustered ";
|
|
else if (locking_is_clustered())
|
|
clustered_message = "Non-clustered ";
|
|
|
|
if (!archive(vg))
|
|
goto_bad;
|
|
|
|
/* Store VG on disk(s) */
|
|
if (!vg_write(vg) || !vg_commit(vg))
|
|
goto_bad;
|
|
|
|
/*
|
|
* The VG is initially written without lock_type set, i.e. it starts as
|
|
* a local VG. lockd_init_vg() then writes the VG a second time with
|
|
* both lock_type and lock_args set.
|
|
*/
|
|
if (!lockd_init_vg(cmd, vg, vp_new.lock_type, 0)) {
|
|
log_error("Failed to initialize lock args for lock type %s",
|
|
vp_new.lock_type);
|
|
vg_remove_pvs(vg);
|
|
vg_remove_direct(vg);
|
|
goto_bad;
|
|
}
|
|
|
|
unlock_vg(cmd, VG_ORPHANS);
|
|
unlock_vg(cmd, vp_new.vg_name);
|
|
|
|
backup(vg);
|
|
|
|
log_print_unless_silent("%s%colume group \"%s\" successfully created%s%s",
|
|
clustered_message, *clustered_message ? 'v' : 'V', vg->name,
|
|
vg->system_id ? " with system ID " : "", vg->system_id ? : "");
|
|
|
|
/*
|
|
* Start the VG lockspace because it will likely be used right away.
|
|
* Optionally wait for the start to complete so the VG can be fully
|
|
* used after this command completes (otherwise, the VG can only be
|
|
* read without locks until the lockspace is done starting.)
|
|
*/
|
|
if (is_lockd_type(vg->lock_type)) {
|
|
const char *start_opt = arg_str_value(cmd, lockopt_ARG, NULL);
|
|
|
|
if (!lockd_start_vg(cmd, vg, 1)) {
|
|
log_error("Failed to start locking");
|
|
goto out;
|
|
}
|
|
|
|
lockd_gl(cmd, "un", 0);
|
|
|
|
if (!start_opt || !strcmp(start_opt, "wait")) {
|
|
/* It is OK if the user does Ctrl-C to cancel the wait. */
|
|
log_print_unless_silent("Starting locking. Waiting until locks are ready...");
|
|
lockd_start_wait(cmd);
|
|
|
|
} else if (!strcmp(start_opt, "nowait")) {
|
|
log_print_unless_silent("Starting locking. VG is read-only until locks are ready.");
|
|
}
|
|
|
|
}
|
|
out:
|
|
release_vg(vg);
|
|
destroy_processing_handle(cmd, handle);
|
|
return ECMD_PROCESSED;
|
|
|
|
bad:
|
|
unlock_vg(cmd, vp_new.vg_name);
|
|
unlock_vg(cmd, VG_ORPHANS);
|
|
release_vg(vg);
|
|
destroy_processing_handle(cmd, handle);
|
|
return ECMD_FAILED;
|
|
}
|