/* * Copyright (C) 2005-2014 Red Hat, Inc. All rights reserved. * * This file is part of LVM2. * * This copyrighted material is made available to anyone wishing to use, * modify, copy, or redistribute it subject to the terms and conditions * of the GNU Lesser General Public License v.2.1. * * You should have received a copy of the GNU Lesser General Public License * along with this program; if not, write to the Free Software Foundation, * Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA */ #include "tools.h" #include "polldaemon.h" #include "lv_alloc.h" struct lvconvert_params { int cache; int force; int snapshot; int split; int splitcache; int splitsnapshot; int merge; int merge_mirror; int poolmetadataspare; int thin; int uncache; int yes; int zero; const char *lv_name; const char *lv_split_name; const char *lv_name_full; const char *vg_name; int wait_completion; int need_polling; int thin_chunk_size_calc_policy; uint32_t chunk_size; uint32_t region_size; uint32_t mirrors; sign_t mirrors_sign; uint32_t keep_mimages; uint32_t stripes; uint32_t stripe_size; uint32_t read_ahead; uint64_t feature_flags; /* cache_pool */ const struct segment_type *segtype; unsigned target_attr; alloc_policy_t alloc; int pv_count; char **pvs; struct dm_list *pvh; int replace_pv_count; char **replace_pvs; struct dm_list *replace_pvh; struct logical_volume *lv_to_poll; uint32_t pool_metadata_extents; int passed_args; uint64_t pool_metadata_size; const char *origin_name; const char *pool_data_name; struct logical_volume *pool_data_lv; const char *pool_metadata_name; struct logical_volume *pool_metadata_lv; thin_discards_t discards; }; static int _lvconvert_validate_names(struct lvconvert_params *lp) { int i, j; const char *names[] = { (lp->lv_name == lp->pool_data_name) ? NULL : lp->lv_name, "converted", lp->pool_data_name, "pool", lp->pool_metadata_name, "pool metadata", lp->origin_name, "origin", }; for (i = 0; i < DM_ARRAY_SIZE(names); i += 2) if (names[i]) for (j = i + 2; j < DM_ARRAY_SIZE(names); j += 2) if (names[j] && !strcmp(names[i], names[j])) { log_error("Can't use same name %s for %s and %s volume.", names[i], names[i + 1], names[j + 1]); return 0; } return 1; } static int _lvconvert_name_params(struct lvconvert_params *lp, struct cmd_context *cmd, int *pargc, char ***pargv) { char *ptr; const char *vg_name = NULL; if (lp->merge) { if (!*pargc) { log_error("Please specify a logical volume path."); return 0; } return 1; } if (!*pargc) { if (lp->cache) { log_error("Logical volume name for caching is missing."); return 0; } if (lp->thin) { log_error("Please specify a logical volume to act as " "the external origin."); return 0; } if (lp->snapshot) { log_error("Please specify a logical volume to act as " "the snapshot exception store."); return 0; } if (lp->split) { log_error("Logical volume for split is missing."); return 0; } if (lp->splitcache) { log_error("Cache logical volume for split is missing."); return 0; } if (lp->uncache) { log_error("Cache logical volume for uncache is missing."); return 0; } if (!lp->lv_name_full) { log_error("Please provide logical volume path."); return 0; } } else if (!lp->lv_name_full) { lp->lv_name_full = (*pargv)[0]; (*pargv)++, (*pargc)--; } if (!validate_restricted_lvname_param(cmd, &lp->vg_name, &lp->pool_metadata_name)) return_0; if (!validate_restricted_lvname_param(cmd, &lp->vg_name, &lp->pool_data_name)) return_0; if (!validate_restricted_lvname_param(cmd, &lp->vg_name, &lp->origin_name)) return_0; if (!validate_restricted_lvname_param(cmd, &lp->vg_name, &lp->lv_split_name)) return_0; if (strchr(lp->lv_name_full, '/') && (vg_name = extract_vgname(cmd, lp->lv_name_full)) && lp->vg_name && strcmp(vg_name, lp->vg_name)) { log_error("Please use a single volume group name " "(\"%s\" or \"%s\")", vg_name, lp->vg_name); return 0; } if (!lp->vg_name) lp->vg_name = vg_name; if (!validate_name(lp->vg_name)) { log_error("Please provide a valid volume group name"); return 0; } if ((ptr = strrchr(lp->lv_name_full, '/'))) lp->lv_name = ptr + 1; else lp->lv_name = lp->lv_name_full; if (!lp->merge_mirror && !arg_count(cmd, repair_ARG) && !arg_count(cmd, splitmirrors_ARG) && !strstr(lp->lv_name, "_tdata") && !strstr(lp->lv_name, "_tmeta") && !apply_lvname_restrictions(lp->lv_name)) return_0; if (*pargc) { if (lp->snapshot) { log_error("Too many arguments provided for snapshots."); return 0; } if (lp->splitsnapshot) { log_error("Too many arguments provided with --splitsnapshot."); return 0; } if (lp->splitcache) { log_error("Too many arguments provided with --splitcache."); return 0; } if (lp->split) { log_error("Too many arguments provided with --split."); return 0; } if (lp->uncache) { log_error("Too many arguments provided with --uncache."); return 0; } if (lp->pool_data_name && lp->pool_metadata_name) { log_error("Too many arguments provided for pool."); return 0; } } if (!_lvconvert_validate_names(lp)) return_0; return 1; } static int _check_conversion_type(struct cmd_context *cmd, const char *type_str) { if (!type_str || !*type_str) return 1; if (!strcmp(type_str, "mirror")) { if (!arg_count(cmd, mirrors_ARG)) { log_error("--type mirror requires -m/--mirrors"); return 0; } return 1; } /* FIXME: Check thin-pool and thin more thoroughly! */ if (!strcmp(type_str, "snapshot") || !strncmp(type_str, "raid", 4) || !strcmp(type_str, "cache-pool") || !strcmp(type_str, "cache") || !strcmp(type_str, "thin-pool") || !strcmp(type_str, "thin")) return 1; log_error("Conversion using --type %s is not supported.", type_str); return 0; } /* -s/--snapshot and --type snapshot are synonyms */ static int _snapshot_type_requested(struct cmd_context *cmd, const char *type_str) { return (arg_count(cmd, snapshot_ARG) || !strcmp(type_str, "snapshot")); } /* mirror/raid* (1,10,4,5,6 and their variants) reshape */ static int _mirror_or_raid_type_requested(struct cmd_context *cmd, const char *type_str) { return (arg_count(cmd, mirrors_ARG) || !strncmp(type_str, "raid", 4) || !strcmp(type_str, "mirror")); } static int _read_pool_params(struct cmd_context *cmd, int *pargc, char ***pargv, const char *type_str, struct lvconvert_params *lp) { int cachepool = 0; int thinpool = 0; if ((lp->pool_data_name = arg_str_value(cmd, cachepool_ARG, NULL))) { if (type_str[0] && strcmp(type_str, "cache") && strcmp(type_str, "cache-pool")) { log_error("--cachepool argument is only valid with " " the cache or cache-pool segment type."); return 0; } cachepool = 1; type_str = "cache-pool"; } else if (!strcmp(type_str, "cache-pool")) cachepool = 1; else if ((lp->pool_data_name = arg_str_value(cmd, thinpool_ARG, NULL))) { if (type_str[0] && strcmp(type_str, "thin") && strcmp(type_str, "thin-pool")) { log_error("--thinpool argument is only valid with " " the thin or thin-pool segment type."); return 0; } thinpool = 1; type_str = "thin-pool"; } else if (!strcmp(type_str, "thin-pool")) thinpool = 1; if (cachepool) { const char *cachemode = arg_str_value(cmd, cachemode_ARG, NULL); if (!cachemode) cachemode = find_config_tree_str(cmd, allocation_cache_pool_cachemode_CFG, NULL); if (!set_cache_pool_feature(&lp->feature_flags, cachemode)) return_0; } else { if (arg_from_list_is_set(cmd, "is valid only with cache pools", cachepool_ARG, cachemode_ARG, -1)) return_0; if (lp->cache) { log_error("--cache requires --cachepool."); return 0; } } if (thinpool) { lp->discards = (thin_discards_t) arg_uint_value(cmd, discards_ARG, THIN_DISCARDS_PASSDOWN); lp->origin_name = arg_str_value(cmd, originname_ARG, NULL); } else { if (arg_from_list_is_set(cmd, "is valid only with thin pools", discards_ARG, originname_ARG, thinpool_ARG, zero_ARG, -1)) return_0; if (lp->thin) { log_error("--thin requires --thinpool."); return 0; } } if (thinpool || cachepool) { if (arg_from_list_is_set(cmd, "is invalid with pools", merge_ARG, mirrors_ARG, repair_ARG, snapshot_ARG, splitmirrors_ARG, splitsnapshot_ARG, -1)) return_0; if (!(lp->segtype = get_segtype_from_string(cmd, type_str))) return_0; if (!get_pool_params(cmd, lp->segtype, &lp->passed_args, &lp->pool_metadata_size, &lp->poolmetadataspare, &lp->chunk_size, &lp->discards, &lp->zero)) return_0; if ((lp->pool_metadata_name = arg_str_value(cmd, poolmetadata_ARG, NULL)) && arg_from_list_is_set(cmd, "is invalid with --poolmetadata", stripesize_ARG, stripes_long_ARG, readahead_ARG, -1)) return_0; if (!lp->pool_data_name) { if (!*pargc) { log_error("Please specify the pool data LV."); return 0; } lp->pool_data_name = (*pargv)[0]; (*pargv)++, (*pargc)--; } if (!lp->thin && !lp->cache) lp->lv_name_full = lp->pool_data_name; /* Hmm _read_activation_params */ lp->read_ahead = arg_uint_value(cmd, readahead_ARG, cmd->default_settings.read_ahead); } else if (arg_from_list_is_set(cmd, "is valid only with pools", poolmetadatasize_ARG, poolmetadataspare_ARG, -1)) return_0; return 1; } static int _read_params(struct cmd_context *cmd, int argc, char **argv, struct lvconvert_params *lp) { int i; const char *tmp_str; struct arg_value_group_list *group; int region_size; int pagesize = lvm_getpagesize(); const char *type_str = arg_str_value(cmd, type_ARG, ""); if (!_check_conversion_type(cmd, type_str)) return_0; if (arg_count(cmd, repair_ARG) && arg_outside_list_is_set(cmd, "cannot be used with --repair", repair_ARG, alloc_ARG, use_policies_ARG, stripes_long_ARG, stripesize_ARG, force_ARG, noudevsync_ARG, test_ARG, -1)) return_0; if (arg_is_set(cmd, mirrorlog_ARG) && arg_is_set(cmd, corelog_ARG)) { log_error("--mirrorlog and --corelog are incompatible."); return 0; } if (arg_is_set(cmd, splitsnapshot_ARG)) { if (arg_outside_list_is_set(cmd, "cannot be used with --splitsnapshot", splitsnapshot_ARG, force_ARG, noudevsync_ARG, test_ARG, -1)) return_0; lp->splitsnapshot = 1; } if (arg_is_set(cmd, split_ARG)) { if (arg_outside_list_is_set(cmd, "cannot be used with --split", split_ARG, name_ARG, force_ARG, noudevsync_ARG, test_ARG, -1)) return_0; lp->split = 1; } if (arg_is_set(cmd, splitcache_ARG)) { if (arg_outside_list_is_set(cmd, "cannot be used with --splitcache", splitcache_ARG, force_ARG, noudevsync_ARG, test_ARG, -1)) return_0; lp->splitcache = 1; } if (arg_is_set(cmd, uncache_ARG)) { if (arg_outside_list_is_set(cmd, "cannot be used with --uncache", uncache_ARG, force_ARG, noudevsync_ARG, test_ARG, -1)) return_0; lp->uncache = 1; } if ((_snapshot_type_requested(cmd, type_str) || arg_count(cmd, merge_ARG)) && (arg_count(cmd, mirrorlog_ARG) || _mirror_or_raid_type_requested(cmd, type_str) || arg_count(cmd, repair_ARG) || arg_count(cmd, thinpool_ARG))) { log_error("--snapshot/--type snapshot or --merge argument " "cannot be mixed with --mirrors/--type mirror/--type raid*, " "--mirrorlog, --repair or --thinpool."); return 0; } if ((arg_count(cmd, stripes_long_ARG) || arg_count(cmd, stripesize_ARG)) && !(_mirror_or_raid_type_requested(cmd, type_str) || arg_count(cmd, repair_ARG) || arg_count(cmd, thinpool_ARG))) { log_error("--stripes or --stripesize argument is only valid " "with --mirrors/--type mirror/--type raid*, --repair and --thinpool"); return 0; } if (arg_count(cmd, cache_ARG)) lp->cache = 1; if (!strcmp(type_str, "cache")) lp->cache = 1; else if (lp->cache) { if (type_str[0]) { log_error("--cache is incompatible with --type %s", type_str); return 0; } type_str = "cache"; } if (arg_count(cmd, thin_ARG)) lp->thin = 1; if (!strcmp(type_str, "thin")) lp->thin = 1; else if (lp->thin) { if (type_str[0]) { log_error("--thin is incompatible with --type %s", type_str); return 0; } type_str = "thin"; } if (!_read_pool_params(cmd, &argc, &argv, type_str, lp)) return_0; if (!arg_count(cmd, background_ARG)) lp->wait_completion = 1; if (_snapshot_type_requested(cmd, type_str)) { if (arg_count(cmd, merge_ARG)) { log_error("--snapshot and --merge are mutually exclusive."); return 0; } lp->snapshot = 1; } if (lp->split) { lp->lv_split_name = arg_str_value(cmd, name_ARG, NULL); /* * The '--splitmirrors n' argument is equivalent to '--mirrors -n' * (note the minus sign), except that it signifies the additional * intent to keep the mimage that is detached, rather than * discarding it. */ } else if (arg_count(cmd, splitmirrors_ARG)) { if (_mirror_or_raid_type_requested(cmd, type_str)) { log_error("--mirrors/--type mirror/--type raid* and --splitmirrors are " "mutually exclusive."); return 0; } if (!arg_count(cmd, name_ARG) && !arg_count(cmd, trackchanges_ARG)) { log_error("Please name the new logical volume using '--name'"); return 0; } lp->lv_split_name = arg_str_value(cmd, name_ARG, NULL); lp->keep_mimages = 1; lp->mirrors = arg_uint_value(cmd, splitmirrors_ARG, 0); lp->mirrors_sign = SIGN_MINUS; } else if (arg_count(cmd, name_ARG)) { log_error("The 'name' argument is only valid" " with --splitmirrors"); return 0; } if (arg_count(cmd, merge_ARG)) { if ((argc == 1) && strstr(argv[0], "_rimage_")) lp->merge_mirror = 1; else lp->merge = 1; } if (arg_count(cmd, mirrors_ARG)) { /* * --splitmirrors has been chosen as the mechanism for * specifying the intent of detaching and keeping a mimage * versus an additional qualifying argument being added here. */ lp->mirrors = arg_uint_value(cmd, mirrors_ARG, 0); lp->mirrors_sign = arg_sign_value(cmd, mirrors_ARG, SIGN_NONE); } lp->alloc = (alloc_policy_t) arg_uint_value(cmd, alloc_ARG, ALLOC_INHERIT); /* There are six types of lvconvert. */ if (lp->merge) { /* Snapshot merge */ if (arg_outside_list_is_set(cmd, "cannot be used with --merge", merge_ARG, background_ARG, interval_ARG, force_ARG, noudevsync_ARG, test_ARG, -1)) return_0; if (!(lp->segtype = get_segtype_from_string(cmd, "snapshot"))) return_0; } else if (lp->splitsnapshot) /* Destroy snapshot retaining cow as separate LV */ ; else if (lp->snapshot) { /* Snapshot creation from pre-existing cow */ if (!argc) { log_error("Please provide logical volume path for snapshot origin."); return 0; } lp->origin_name = argv[0]; argv++, argc--; if (arg_count(cmd, regionsize_ARG)) { log_error("--regionsize is only available with mirrors"); return 0; } if (arg_count(cmd, stripesize_ARG) || arg_count(cmd, stripes_long_ARG)) { log_error("--stripes and --stripesize are only available with striped mirrors"); return 0; } if (arg_count(cmd, chunksize_ARG) && (arg_sign_value(cmd, chunksize_ARG, SIGN_NONE) == SIGN_MINUS)) { log_error("Negative chunk size is invalid."); return 0; } lp->chunk_size = arg_uint_value(cmd, chunksize_ARG, 8); if (lp->chunk_size < 8 || lp->chunk_size > 1024 || (lp->chunk_size & (lp->chunk_size - 1))) { log_error("Chunk size must be a power of 2 in the " "range 4K to 512K"); return 0; } log_verbose("Setting chunk size to %s.", display_size(cmd, lp->chunk_size)); if (!(lp->segtype = get_segtype_from_string(cmd, "snapshot"))) return_0; lp->zero = (lp->segtype->flags & SEG_CANNOT_BE_ZEROED) ? 0 : arg_int_value(cmd, zero_ARG, 1); } else if (arg_count(cmd, replace_ARG)) { /* RAID device replacement */ lp->replace_pv_count = arg_count(cmd, replace_ARG); lp->replace_pvs = dm_pool_alloc(cmd->mem, sizeof(char *) * lp->replace_pv_count); if (!lp->replace_pvs) return_0; i = 0; dm_list_iterate_items(group, &cmd->arg_value_groups) { if (!grouped_arg_is_set(group->arg_values, replace_ARG)) continue; if (!(tmp_str = grouped_arg_str_value(group->arg_values, replace_ARG, NULL))) { log_error("Failed to get '--replace' argument"); return 0; } if (!(lp->replace_pvs[i++] = dm_pool_strdup(cmd->mem, tmp_str))) return_0; } } else if (_mirror_or_raid_type_requested(cmd, type_str) || arg_is_set(cmd, repair_ARG) || arg_is_set(cmd, mirrorlog_ARG) || arg_is_set(cmd, corelog_ARG)) { /* Mirrors (and some RAID functions) */ if (arg_count(cmd, chunksize_ARG)) { log_error("--chunksize is only available with snapshots or pools."); return 0; } if (arg_count(cmd, zero_ARG)) { log_error("--zero is only available with snapshots or thin pools."); return 0; } /* * --regionsize is only valid if converting an LV into a mirror. * Checked when we know the state of the LV being converted. */ if (arg_count(cmd, regionsize_ARG)) { if (arg_sign_value(cmd, regionsize_ARG, SIGN_NONE) == SIGN_MINUS) { log_error("Negative regionsize is invalid"); return 0; } lp->region_size = arg_uint_value(cmd, regionsize_ARG, 0); } else { region_size = get_default_region_size(cmd); if (region_size < 0) { log_error("Negative regionsize in " "configuration file is invalid"); return 0; } lp->region_size = region_size; } if (lp->region_size % (pagesize >> SECTOR_SHIFT)) { log_error("Region size (%" PRIu32 ") must be " "a multiple of machine memory " "page size (%d)", lp->region_size, pagesize >> SECTOR_SHIFT); return 0; } if (lp->region_size & (lp->region_size - 1)) { log_error("Region size (%" PRIu32 ") must be a power of 2", lp->region_size); return 0; } if (!lp->region_size) { log_error("Non-zero region size must be supplied."); return 0; } /* Default is never striped, regardless of existing LV configuration. */ if (!get_stripe_params(cmd, &lp->stripes, &lp->stripe_size)) return_0; if (arg_count(cmd, mirrors_ARG) && !lp->mirrors) { /* down-converting to linear/stripe? */ if (!(lp->segtype = get_segtype_from_string(cmd, "striped"))) return_0; } else if (arg_count(cmd, type_ARG)) { /* changing mirror type? */ if (!(lp->segtype = get_segtype_from_string(cmd, arg_str_value(cmd, type_ARG, find_config_tree_str(cmd, global_mirror_segtype_default_CFG, NULL))))) return_0; } /* else segtype will default to current type */ } lp->force = arg_count(cmd, force_ARG); lp->yes = arg_count(cmd, yes_ARG); if (activation() && lp->segtype && lp->segtype->ops->target_present && !lp->segtype->ops->target_present(cmd, NULL, &lp->target_attr)) { log_error("%s: Required device-mapper target(s) not " "detected in your kernel", lp->segtype->name); return 0; } if (!_lvconvert_name_params(lp, cmd, &argc, &argv)) return_0; lp->pv_count = argc; lp->pvs = argv; return 1; } static struct volume_group *_get_lvconvert_vg(struct cmd_context *cmd, const char *name, const char *uuid __attribute__((unused))) { dev_close_all(); if (name && !strchr(name, '/')) return vg_read_for_update(cmd, name, NULL, 0); /* 'name' is the full LV name; must extract_vgname() */ return vg_read_for_update(cmd, extract_vgname(cmd, name), NULL, 0); } static struct logical_volume *_get_lvconvert_lv(struct cmd_context *cmd __attribute__((unused)), struct volume_group *vg, const char *name, const char *uuid, uint64_t lv_type __attribute__((unused))) { struct logical_volume *lv = find_lv(vg, name); if (!lv || (uuid && strcmp(uuid, (char *)&lv->lvid))) return NULL; return lv; } static int _finish_lvconvert_mirror(struct cmd_context *cmd, struct volume_group *vg, struct logical_volume *lv, struct dm_list *lvs_changed __attribute__((unused))) { if (!lv_is_converting(lv)) return 1; if (!collapse_mirrored_lv(lv)) { log_error("Failed to remove temporary sync layer."); return 0; } lv->status &= ~CONVERTING; if (!lv_update_and_reload(lv)) return_0; log_print_unless_silent("Logical volume %s converted.", lv->name); return 1; } /* Swap lvid and LV names */ static int _swap_lv_identifiers(struct cmd_context *cmd, struct logical_volume *a, struct logical_volume *b) { union lvid lvid; const char *name; lvid = a->lvid; a->lvid = b->lvid; b->lvid = lvid; name = a->name; a->name = b->name; if (!lv_rename_update(cmd, b, name, 0)) return_0; return 1; } static void _move_lv_attributes(struct logical_volume *to, struct logical_volume *from) { /* Maybe move this code into _finish_thin_merge() */ to->status = from->status; // FIXME maybe some masking ? to->alloc = from->alloc; to->profile = from->profile; to->read_ahead = from->read_ahead; to->major = from->major; to->minor = from->minor; to->timestamp = from->timestamp; to->hostname = from->hostname; /* Move tags */ dm_list_init(&to->tags); dm_list_splice(&to->tags, &from->tags); /* Anything else to preserve? */ } /* Finalise merging of lv into merge_lv */ static int _finish_thin_merge(struct cmd_context *cmd, struct logical_volume *merge_lv, struct logical_volume *lv) { if (!_swap_lv_identifiers(cmd, merge_lv, lv)) { log_error("Failed to swap %s with merging %s.", lv->name, merge_lv->name); return 0; } /* Preserve origins' attributes */ _move_lv_attributes(lv, merge_lv); /* Removed LV has to be visible */ if (!lv_remove_single(cmd, merge_lv, DONT_PROMPT, 1)) return_0; return 1; } static int _finish_lvconvert_merge(struct cmd_context *cmd, struct volume_group *vg, struct logical_volume *lv, struct dm_list *lvs_changed __attribute__((unused))) { struct lv_segment *snap_seg = find_snapshot(lv); if (!lv_is_merging_origin(lv)) { log_error("Logical volume %s has no merging snapshot.", lv->name); return 0; } log_print_unless_silent("Merge of snapshot into logical volume %s has finished.", lv->name); if (seg_is_thin_volume(snap_seg)) { clear_snapshot_merge(lv); if (!_finish_thin_merge(cmd, lv, snap_seg->lv)) return_0; } else if (!lv_remove_single(cmd, snap_seg->cow, DONT_PROMPT, 0)) { log_error("Could not remove snapshot %s merged into %s.", snap_seg->cow->name, lv->name); return 0; } return 1; } static progress_t _poll_merge_progress(struct cmd_context *cmd, struct logical_volume *lv, const char *name __attribute__((unused)), struct daemon_parms *parms) { dm_percent_t percent = DM_PERCENT_0; if (!lv_is_merging_origin(lv) || !lv_snapshot_percent(lv, &percent)) { log_error("%s: Failed query for merging percentage. Aborting merge.", lv->name); return PROGRESS_CHECK_FAILED; } else if (percent == DM_PERCENT_INVALID) { log_error("%s: Merging snapshot invalidated. Aborting merge.", lv->name); return PROGRESS_CHECK_FAILED; } else if (percent == LVM_PERCENT_MERGE_FAILED) { log_error("%s: Merge failed. Retry merge or inspect manually.", lv->name); return PROGRESS_CHECK_FAILED; } if (parms->progress_display) log_print_unless_silent("%s: %s: %.1f%%", lv->name, parms->progress_title, 100.0 - dm_percent_to_float(percent)); else log_verbose("%s: %s: %.1f%%", lv->name, parms->progress_title, 100.0 - dm_percent_to_float(percent)); if (percent == DM_PERCENT_0) return PROGRESS_FINISHED_ALL; return PROGRESS_UNFINISHED; } static progress_t _poll_thin_merge_progress(struct cmd_context *cmd, struct logical_volume *lv, const char *name __attribute__((unused)), struct daemon_parms *parms) { uint32_t device_id; if (!lv_thin_device_id(lv, &device_id)) { stack; return PROGRESS_CHECK_FAILED; } /* * There is no need to poll more than once, * a thin snapshot merge is immediate. */ if (device_id != find_snapshot(lv)->device_id) { log_error("LV %s is not merged.", lv->name); return PROGRESS_CHECK_FAILED; } return PROGRESS_FINISHED_ALL; /* Merging happend */ } static struct poll_functions _lvconvert_mirror_fns = { .get_copy_vg = _get_lvconvert_vg, .get_copy_lv = _get_lvconvert_lv, .poll_progress = poll_mirror_progress, .finish_copy = _finish_lvconvert_mirror, }; static struct poll_functions _lvconvert_merge_fns = { .get_copy_vg = _get_lvconvert_vg, .get_copy_lv = _get_lvconvert_lv, .poll_progress = _poll_merge_progress, .finish_copy = _finish_lvconvert_merge, }; static struct poll_functions _lvconvert_thin_merge_fns = { .get_copy_vg = _get_lvconvert_vg, .get_copy_lv = _get_lvconvert_lv, .poll_progress = _poll_thin_merge_progress, .finish_copy = _finish_lvconvert_merge, }; int lvconvert_poll(struct cmd_context *cmd, struct logical_volume *lv, unsigned background) { /* * FIXME allocate an "object key" structure with split * out members (vg_name, lv_name, uuid, etc) and pass that * around the lvconvert and polldaemon code * - will avoid needless work, e.g. extract_vgname() * - unfortunately there are enough overloaded "name" dragons in * the polldaemon, lvconvert, pvmove code that a comprehensive * audit/rework is needed */ char uuid[sizeof(lv->lvid)]; char lv_full_name[NAME_LEN]; if (dm_snprintf(lv_full_name, sizeof(lv_full_name), "%s/%s", lv->vg->name, lv->name) < 0) { log_error(INTERNAL_ERROR "Name \"%s/%s\" is too long.", lv->vg->name, lv->name); return ECMD_FAILED; } memcpy(uuid, &lv->lvid, sizeof(lv->lvid)); if (lv_is_merging_origin(lv)) return poll_daemon(cmd, lv_full_name, uuid, background, 0, seg_is_thin_volume(find_snapshot(lv)) ? &_lvconvert_thin_merge_fns : &_lvconvert_merge_fns, "Merged"); return poll_daemon(cmd, lv_full_name, uuid, background, 0, &_lvconvert_mirror_fns, "Converted"); } static int _insert_lvconvert_layer(struct cmd_context *cmd, struct logical_volume *lv) { char *format, *layer_name; size_t len; int i; /* * We would like to give the same number for this layer * and the newly added mimage. * However, LV name of newly added mimage is determined *after* * the LV name of this layer is determined. * * So, use generate_lv_name() to generate mimage name first * and take the number from it. */ len = strlen(lv->name) + 32; if (!(format = alloca(len)) || !(layer_name = alloca(len)) || dm_snprintf(format, len, "%s_mimage_%%d", lv->name) < 0) { log_error("lvconvert: layer name allocation failed."); return 0; } if (!generate_lv_name(lv->vg, format, layer_name, len) || sscanf(layer_name, format, &i) != 1) { log_error("lvconvert: layer name generation failed."); return 0; } if (dm_snprintf(layer_name, len, MIRROR_SYNC_LAYER "_%d", i) < 0) { log_error("layer name allocation failed."); return 0; } if (!insert_layer_for_lv(cmd, lv, 0, layer_name)) { log_error("Failed to insert resync layer"); return 0; } return 1; } static int _failed_mirrors_count(struct logical_volume *lv) { struct lv_segment *lvseg; int ret = 0; unsigned s; dm_list_iterate_items(lvseg, &lv->segments) { if (!seg_is_mirrored(lvseg)) return -1; for (s = 0; s < lvseg->area_count; s++) { if (seg_type(lvseg, s) == AREA_LV) { if (is_temporary_mirror_layer(seg_lv(lvseg, s))) ret += _failed_mirrors_count(seg_lv(lvseg, s)); else if (seg_lv(lvseg, s)->status & PARTIAL_LV) ++ ret; else if (seg_type(lvseg, s) == AREA_PV && is_missing_pv(seg_pv(lvseg, s))) ++ret; } } } return ret; } static int _failed_logs_count(struct logical_volume *lv) { int ret = 0; unsigned s; struct logical_volume *log_lv = first_seg(lv)->log_lv; if (log_lv && (log_lv->status & PARTIAL_LV)) { if (lv_is_mirrored(log_lv)) ret += _failed_mirrors_count(log_lv); else ret += 1; } for (s = 0; s < first_seg(lv)->area_count; s++) { if (seg_type(first_seg(lv), s) == AREA_LV && is_temporary_mirror_layer(seg_lv(first_seg(lv), s))) ret += _failed_logs_count(seg_lv(first_seg(lv), s)); } return ret; } static struct dm_list *_failed_pv_list(struct volume_group *vg) { struct dm_list *failed_pvs; struct pv_list *pvl, *new_pvl; if (!(failed_pvs = dm_pool_alloc(vg->vgmem, sizeof(*failed_pvs)))) { log_error("Allocation of list of failed_pvs failed."); return NULL; } dm_list_init(failed_pvs); dm_list_iterate_items(pvl, &vg->pvs) { if (!is_missing_pv(pvl->pv)) continue; /* * Finally, --repair will remove empty PVs. * But we only want remove these which are output of repair, * Do not count these which are already empty here. * FIXME: code should traverse PV in LV not in whole VG. * FIXME: layer violation? should it depend on vgreduce --removemising? */ if (pvl->pv->pe_alloc_count == 0) continue; if (!(new_pvl = dm_pool_alloc(vg->vgmem, sizeof(*new_pvl)))) { log_error("Allocation of failed_pvs list entry failed."); return NULL; } new_pvl->pv = pvl->pv; dm_list_add(failed_pvs, &new_pvl->list); } return failed_pvs; } static int _is_partial_lv(struct logical_volume *lv, void *baton __attribute__((unused))) { return lv->status & PARTIAL_LV; } /* * Walk down the stacked mirror LV to the original mirror LV. */ static struct logical_volume *_original_lv(struct logical_volume *lv) { struct logical_volume *next_lv = lv, *tmp_lv; while ((tmp_lv = find_temporary_mirror(next_lv))) next_lv = tmp_lv; return next_lv; } static void _lvconvert_mirrors_repair_ask(struct cmd_context *cmd, int failed_log, int failed_mirrors, int *replace_log, int *replace_mirrors) { const char *leg_policy, *log_policy; int force = arg_count(cmd, force_ARG); int yes = arg_count(cmd, yes_ARG); if (arg_count(cmd, use_policies_ARG)) { leg_policy = find_config_tree_str(cmd, activation_mirror_image_fault_policy_CFG, NULL); log_policy = find_config_tree_str(cmd, activation_mirror_log_fault_policy_CFG, NULL); *replace_mirrors = strcmp(leg_policy, "remove"); *replace_log = strcmp(log_policy, "remove"); return; } if (force != PROMPT) { *replace_log = *replace_mirrors = 0; return; } *replace_log = *replace_mirrors = 1; if (yes) return; if (failed_log && yes_no_prompt("Attempt to replace failed mirror log? [y/n]: ") == 'n') *replace_log = 0; if (failed_mirrors && yes_no_prompt("Attempt to replace failed mirror images " "(requires full device resync)? [y/n]: ") == 'n') *replace_mirrors = 0; } /* * _get_log_count * @lv: the mirror LV * * Get the number of on-disk copies of the log. * 0 = 'core' * 1 = 'disk' * 2+ = 'mirrored' */ static uint32_t _get_log_count(struct logical_volume *lv) { struct logical_volume *log_lv; log_lv = first_seg(_original_lv(lv))->log_lv; if (log_lv) return lv_mirror_count(log_lv); return 0; } static int _lv_update_mirrored_log(struct logical_volume *lv, struct dm_list *operable_pvs, int log_count) { int old_log_count; struct logical_volume *log_lv; /* * When log_count is 0, mirrored log doesn't need to be * updated here but it will be removed later. */ if (!log_count) return 1; log_lv = first_seg(_original_lv(lv))->log_lv; if (!log_lv || !lv_is_mirrored(log_lv)) return 1; old_log_count = _get_log_count(lv); if (old_log_count == log_count) return 1; /* Reducing redundancy of the log */ return remove_mirror_images(log_lv, log_count, is_mirror_image_removable, operable_pvs, 0U); } static int _lv_update_log_type(struct cmd_context *cmd, struct lvconvert_params *lp, struct logical_volume *lv, struct dm_list *operable_pvs, int log_count) { int old_log_count; uint32_t region_size = (lp) ? lp->region_size : first_seg(lv)->region_size; alloc_policy_t alloc = (lp) ? lp->alloc : lv->alloc; struct logical_volume *original_lv; struct logical_volume *log_lv; old_log_count = _get_log_count(lv); if (old_log_count == log_count) return 1; original_lv = _original_lv(lv); /* Remove an existing log completely */ if (!log_count) { if (!remove_mirror_log(cmd, original_lv, operable_pvs, arg_count(cmd, yes_ARG) || arg_count(cmd, force_ARG))) return_0; return 1; } log_lv = first_seg(original_lv)->log_lv; /* Adding redundancy to the log */ if (old_log_count < log_count) { region_size = adjusted_mirror_region_size(lv->vg->extent_size, lv->le_count, region_size, 0, vg_is_clustered(lv->vg)); if (!add_mirror_log(cmd, original_lv, log_count, region_size, operable_pvs, alloc)) return_0; /* * FIXME: This simple approach won't work in cluster mirrors, * but it doesn't matter because we don't support * mirrored logs in cluster mirrors. */ if (old_log_count && !lv_update_and_reload(log_lv)) return_0; return 1; } /* Reducing redundancy of the log */ return remove_mirror_images(log_lv, log_count, is_mirror_image_removable, operable_pvs, 1U); } /* * Reomove missing and empty PVs from VG, if are also in provided list */ static void _remove_missing_empty_pv(struct volume_group *vg, struct dm_list *remove_pvs) { struct pv_list *pvl, *pvl_vg, *pvlt; int removed = 0; if (!remove_pvs) return; dm_list_iterate_items(pvl, remove_pvs) { dm_list_iterate_items_safe(pvl_vg, pvlt, &vg->pvs) { if (!id_equal(&pvl->pv->id, &pvl_vg->pv->id) || !is_missing_pv(pvl_vg->pv) || pvl_vg->pv->pe_alloc_count != 0) continue; /* FIXME: duplication of vgreduce code, move this to library */ vg->free_count -= pvl_vg->pv->pe_count; vg->extent_count -= pvl_vg->pv->pe_count; del_pvl_from_vgs(vg, pvl_vg); free_pv_fid(pvl_vg->pv); removed++; } } if (removed) { if (!vg_write(vg) || !vg_commit(vg)) { stack; return; } log_warn("%d missing and now unallocated Physical Volumes removed from VG.", removed); } } /* * _lvconvert_mirrors_parse_params * * This function performs the following: * 1) Gets the old values of mimage and log counts * 2) Parses the CLI args to find the new desired values * 3) Adjusts 'lp->mirrors' to the appropriate absolute value. * (Remember, 'lp->mirrors' is specified in terms of the number of "copies" * vs. the number of mimages. It can also be a relative value.) * 4) Sets 'lp->need_polling' if collapsing * 5) Validates other mirror params * * Returns: 1 on success, 0 on error */ static int _lvconvert_mirrors_parse_params(struct cmd_context *cmd, struct logical_volume *lv, struct lvconvert_params *lp, uint32_t *old_mimage_count, uint32_t *old_log_count, uint32_t *new_mimage_count, uint32_t *new_log_count) { int repair = arg_count(cmd, repair_ARG); *old_mimage_count = lv_mirror_count(lv); *old_log_count = _get_log_count(lv); /* * Collapsing a stack of mirrors: * * If called with no argument, try collapsing the resync layers */ if (!arg_count(cmd, mirrors_ARG) && !arg_count(cmd, mirrorlog_ARG) && !arg_count(cmd, corelog_ARG) && !arg_count(cmd, regionsize_ARG) && !arg_count(cmd, splitmirrors_ARG) && !repair) { *new_mimage_count = *old_mimage_count; *new_log_count = *old_log_count; if (find_temporary_mirror(lv) || lv_is_converting(lv)) lp->need_polling = 1; return 1; } /* * Adjusting mimage count? */ if (!arg_count(cmd, mirrors_ARG) && !arg_count(cmd, splitmirrors_ARG)) lp->mirrors = *old_mimage_count; else if (lp->mirrors_sign == SIGN_PLUS) lp->mirrors = *old_mimage_count + lp->mirrors; else if (lp->mirrors_sign == SIGN_MINUS) lp->mirrors = (*old_mimage_count > lp->mirrors) ? *old_mimage_count - lp->mirrors: 0; else lp->mirrors += 1; *new_mimage_count = lp->mirrors; /* Too many mimages? */ if (lp->mirrors > DEFAULT_MIRROR_MAX_IMAGES) { log_error("Only up to %d images in mirror supported currently.", DEFAULT_MIRROR_MAX_IMAGES); return 0; } /* Did the user try to subtract more legs than available? */ if (lp->mirrors < 1) { log_error("Unable to reduce images by specified amount - only %d in %s", *old_mimage_count, lv->name); return 0; } /* * FIXME: It would be nice to say what we are adjusting to, but * I really don't know whether to specify the # of copies or mimages. */ if (*old_mimage_count != *new_mimage_count) log_verbose("Adjusting mirror image count of %s", lv->name); /* * Adjust log type * * If we are converting from a mirror to another mirror or simply * changing the log type, we start by assuming they want the log * type the same and then parse the given args. OTOH, If we are * converting from linear to mirror, then we start from the default * position that the user would like a 'disk' log. */ *new_log_count = (*old_mimage_count > 1) ? *old_log_count : 1; if (!arg_count(cmd, corelog_ARG) && !arg_count(cmd, mirrorlog_ARG)) return 1; *new_log_count = arg_int_value(cmd, mirrorlog_ARG, arg_is_set(cmd, corelog_ARG) ? MIRROR_LOG_CORE : DEFAULT_MIRRORLOG); /* * No mirrored logs for cluster mirrors until * log daemon is multi-threaded. */ if ((*new_log_count == MIRROR_LOG_MIRRORED) && vg_is_clustered(lv->vg)) { log_error("Log type, \"mirrored\", is unavailable to cluster mirrors"); return 0; } log_verbose("Setting logging type to %s", get_mirror_log_name(*new_log_count)); /* * Region size must not change on existing mirrors */ if (arg_count(cmd, regionsize_ARG) && lv_is_mirrored(lv) && (lp->region_size != first_seg(lv)->region_size)) { log_error("Mirror log region size cannot be changed on " "an existing mirror."); return 0; } /* * For the most part, we cannot handle multi-segment mirrors. Bail out * early if we have encountered one. */ if (lv_is_mirrored(lv) && dm_list_size(&lv->segments) != 1) { log_error("Logical volume %s has multiple " "mirror segments.", lv->name); return 0; } return 1; } /* * _lvconvert_mirrors_aux * * Add/remove mirror images and adjust log type. 'operable_pvs' * are the set of PVs open to removal or allocation - depending * on the operation being performed. */ static int _lvconvert_mirrors_aux(struct cmd_context *cmd, struct logical_volume *lv, struct lvconvert_params *lp, struct dm_list *operable_pvs, uint32_t new_mimage_count, uint32_t new_log_count) { uint32_t region_size; struct lv_segment *seg; struct logical_volume *layer_lv; uint32_t old_mimage_count = lv_mirror_count(lv); uint32_t old_log_count = _get_log_count(lv); if ((lp->mirrors == 1) && !lv_is_mirrored(lv)) { log_warn("Logical volume %s is already not mirrored.", lv->name); return 1; } region_size = adjusted_mirror_region_size(lv->vg->extent_size, lv->le_count, lp->region_size, 0, vg_is_clustered(lv->vg)); if (!operable_pvs) operable_pvs = lp->pvh; seg = first_seg(lv); /* * Up-convert from linear to mirror */ if (!lv_is_mirrored(lv)) { /* FIXME Share code with lvcreate */ /* * FIXME should we give not only lp->pvh, but also all PVs * currently taken by the mirror? Would make more sense from * user perspective. */ if (!lv_add_mirrors(cmd, lv, new_mimage_count - 1, lp->stripes, lp->stripe_size, region_size, new_log_count, operable_pvs, lp->alloc, MIRROR_BY_LV)) return_0; if (lp->wait_completion) lp->need_polling = 1; goto out; } /* * Up-convert m-way mirror to n-way mirror */ if (new_mimage_count > old_mimage_count) { if (lv->status & LV_NOTSYNCED) { log_error("Can't add mirror to out-of-sync mirrored " "LV: use lvchange --resync first."); return 0; } /* * We allow snapshots of mirrors, but for now, we * do not allow up converting mirrors that are under * snapshots. The layering logic is somewhat complex, * and preliminary test show that the conversion can't * seem to get the correct %'age of completion. */ if (lv_is_origin(lv)) { log_error("Can't add additional mirror images to " "mirrors that are under snapshots"); return 0; } /* * Is there already a convert in progress? We do not * currently allow more than one. */ if (find_temporary_mirror(lv) || lv_is_converting(lv)) { log_error("%s is already being converted. Unable to start another conversion.", lv->name); return 0; } /* * Log addition/removal should be done before the layer * insertion to make the end result consistent with * linear-to-mirror conversion. */ if (!_lv_update_log_type(cmd, lp, lv, operable_pvs, new_log_count)) return_0; /* Insert a temporary layer for syncing, * only if the original lv is using disk log. */ if (seg->log_lv && !_insert_lvconvert_layer(cmd, lv)) { log_error("Failed to insert resync layer"); return 0; } /* FIXME: can't have multiple mlogs. force corelog. */ if (!lv_add_mirrors(cmd, lv, new_mimage_count - old_mimage_count, lp->stripes, lp->stripe_size, region_size, 0U, operable_pvs, lp->alloc, MIRROR_BY_LV)) { layer_lv = seg_lv(first_seg(lv), 0); if (!remove_layer_from_lv(lv, layer_lv) || (lv_is_active(layer_lv) && !deactivate_lv(cmd, layer_lv)) || !lv_remove(layer_lv) || !vg_write(lv->vg) || !vg_commit(lv->vg)) { log_error("ABORTING: Failed to remove " "temporary mirror layer %s.", layer_lv->name); log_error("Manual cleanup with vgcfgrestore " "and dmsetup may be required."); return 0; } return_0; } if (seg->log_lv) lv->status |= CONVERTING; lp->need_polling = 1; goto out_skip_log_convert; } /* * Down-convert (reduce # of mimages). */ if (new_mimage_count < old_mimage_count) { uint32_t nmc = old_mimage_count - new_mimage_count; uint32_t nlc = (!new_log_count || lp->mirrors == 1) ? 1U : 0U; /* FIXME: Why did nlc used to be calculated that way? */ /* Reduce number of mirrors */ if (lp->keep_mimages) { if (arg_count(cmd, trackchanges_ARG)) { log_error("--trackchanges is not available " "to 'mirror' segment type"); return 0; } if (!lv_split_mirror_images(lv, lp->lv_split_name, nmc, operable_pvs)) return_0; } else if (!lv_remove_mirrors(cmd, lv, nmc, nlc, is_mirror_image_removable, operable_pvs, 0)) return_0; goto out; /* Just in case someone puts code between */ } out: /* * Converting the log type */ if (lv_is_mirrored(lv) && (old_log_count != new_log_count)) { if (!_lv_update_log_type(cmd, lp, lv, operable_pvs, new_log_count)) return_0; } out_skip_log_convert: if (!lv_update_and_reload(lv)) return_0; return 1; } int mirror_remove_missing(struct cmd_context *cmd, struct logical_volume *lv, int force) { struct dm_list *failed_pvs; int log_count = _get_log_count(lv) - _failed_logs_count(lv); if (!(failed_pvs = _failed_pv_list(lv->vg))) return_0; if (force && _failed_mirrors_count(lv) == (int)lv_mirror_count(lv)) { log_error("No usable images left in %s.", lv->name); return lv_remove_with_dependencies(cmd, lv, DONT_PROMPT, 0); } /* * We must adjust the log first, or the entire mirror * will get stuck during a suspend. */ if (!_lv_update_mirrored_log(lv, failed_pvs, log_count)) return_0; if (_failed_mirrors_count(lv) > 0 && !lv_remove_mirrors(cmd, lv, _failed_mirrors_count(lv), log_count ? 0U : 1U, _is_partial_lv, NULL, 0)) return_0; if (lv_is_mirrored(lv) && !_lv_update_log_type(cmd, NULL, lv, failed_pvs, log_count)) return_0; if (!lv_update_and_reload(lv)) return_0; return 1; } /* * _lvconvert_mirrors_repair * * This function operates in two phases. First, all of the bad * devices are removed from the mirror. Then, if desired by the * user, the devices are replaced. * * 'old_mimage_count' and 'old_log_count' are there so we know * what to convert to after the removal of devices. */ static int _lvconvert_mirrors_repair(struct cmd_context *cmd, struct logical_volume *lv, struct lvconvert_params *lp) { int failed_logs; int failed_mimages; int replace_logs = 0; int replace_mimages = 0; uint32_t log_count; uint32_t original_mimages = lv_mirror_count(lv); uint32_t original_logs = _get_log_count(lv); cmd->handles_missing_pvs = 1; cmd->partial_activation = 1; lp->need_polling = 0; lv_check_transient(lv); /* TODO check this in lib for all commands? */ if (!(lv->status & PARTIAL_LV)) { log_print_unless_silent("%s is consistent. Nothing to repair.", lv->name); return 1; } failed_mimages = _failed_mirrors_count(lv); failed_logs = _failed_logs_count(lv); if (!mirror_remove_missing(cmd, lv, 0)) return_0; if (failed_mimages) log_print_unless_silent("Mirror status: %d of %d images failed.", failed_mimages, original_mimages); /* * Count the failed log devices */ if (failed_logs) log_print_unless_silent("Mirror log status: %d of %d images failed.", failed_logs, original_logs); /* * Find out our policies */ _lvconvert_mirrors_repair_ask(cmd, failed_logs, failed_mimages, &replace_logs, &replace_mimages); /* * Second phase - replace faulty devices */ lp->mirrors = replace_mimages ? original_mimages : (original_mimages - failed_mimages); /* * It does not make sense to replace the log if the volume is no longer * a mirror. */ if (lp->mirrors == 1) replace_logs = 0; log_count = replace_logs ? original_logs : (original_logs - failed_logs); while (replace_mimages || replace_logs) { log_warn("Trying to up-convert to %d images, %d logs.", lp->mirrors, log_count); if (_lvconvert_mirrors_aux(cmd, lv, lp, NULL, lp->mirrors, log_count)) break; if (lp->mirrors > 2) --lp->mirrors; else if (log_count > 0) --log_count; else break; /* nowhere to go, anymore... */ } if (replace_mimages && lv_mirror_count(lv) != original_mimages) log_warn("WARNING: Failed to replace %d of %d images in volume %s", original_mimages - lv_mirror_count(lv), original_mimages, lv->name); if (replace_logs && _get_log_count(lv) != original_logs) log_warn("WARNING: Failed to replace %d of %d logs in volume %s", original_logs - _get_log_count(lv), original_logs, lv->name); /* if (!arg_count(cmd, use_policies_ARG) && (lp->mirrors != old_mimage_count || log_count != old_log_count)) return 0; */ return 1; } static int _lvconvert_validate_thin(struct logical_volume *lv, struct lvconvert_params *lp) { if (!lv_is_thin_pool(lv) && !lv_is_thin_volume(lv)) return 1; log_error("Converting thin%s segment type for \"%s/%s\" to %s is not supported.", lv_is_thin_pool(lv) ? " pool" : "", lv->vg->name, lv->name, lp->segtype->name); if (lv_is_thin_volume(lv)) return 0; /* Give advice for thin pool conversion */ log_error("For pool data volume conversion use \"%s/%s\".", lv->vg->name, seg_lv(first_seg(lv), 0)->name); log_error("For pool metadata volume conversion use \"%s/%s\".", lv->vg->name, first_seg(lv)->metadata_lv->name); return 0; } /* * _lvconvert_mirrors * * Determine what is being done. Are we doing a conversion, repair, or * collapsing a stack? Once determined, call helper functions. */ static int _lvconvert_mirrors(struct cmd_context *cmd, struct logical_volume *lv, struct lvconvert_params *lp) { int repair = arg_count(cmd, repair_ARG); uint32_t old_mimage_count; uint32_t old_log_count; uint32_t new_mimage_count; uint32_t new_log_count; if (lp->merge_mirror) { log_error("Unable to merge mirror images" "of segment type 'mirror'"); return 0; } if (!_lvconvert_validate_thin(lv, lp)) return_0; if (lv_is_thin_type(lv)) { log_error("Mirror segment type cannot be used for thinpool%s.\n" "Try \"%s\" segment type instead.", lv_is_thin_pool_data(lv) ? "s" : " metadata", SEG_TYPE_NAME_RAID1); return 0; } if (lv_is_cache_type(lv)) { log_error("Mirrors are not yet supported on cache LVs %s.", display_lvname(lv)); return 0; } /* Adjust mimage and/or log count */ if (!_lvconvert_mirrors_parse_params(cmd, lv, lp, &old_mimage_count, &old_log_count, &new_mimage_count, &new_log_count)) return 0; if (((old_mimage_count < new_mimage_count && old_log_count > new_log_count) || (old_mimage_count > new_mimage_count && old_log_count < new_log_count)) && lp->pv_count) { log_error("Cannot both allocate and free extents when " "specifying physical volumes to use."); log_error("Please specify the operation in two steps."); return 0; } /* Nothing to do? (Probably finishing collapse.) */ if ((old_mimage_count == new_mimage_count) && (old_log_count == new_log_count) && !repair) return 1; if (repair) return _lvconvert_mirrors_repair(cmd, lv, lp); if (!_lvconvert_mirrors_aux(cmd, lv, lp, NULL, new_mimage_count, new_log_count)) return 0; if (!lp->need_polling) log_print_unless_silent("Logical volume %s converted.", lv->name); backup(lv->vg); return 1; } static int _is_valid_raid_conversion(const struct segment_type *from_segtype, const struct segment_type *to_segtype) { if (from_segtype == to_segtype) return 1; if (!segtype_is_raid(from_segtype) && !segtype_is_raid(to_segtype)) return_0; /* Not converting to or from RAID? */ return 1; } static void _lvconvert_raid_repair_ask(struct cmd_context *cmd, struct lvconvert_params *lp, int *replace_dev) { const char *dev_policy; *replace_dev = 1; if (arg_count(cmd, use_policies_ARG)) { dev_policy = find_config_tree_str(cmd, activation_raid_fault_policy_CFG, NULL); if (!strcmp(dev_policy, "allocate") || !strcmp(dev_policy, "replace")) return; /* else if (!strcmp(dev_policy, "anything_else")) -- no replace */ *replace_dev = 0; return; } if (!lp->yes && yes_no_prompt("Attempt to replace failed RAID images " "(requires full device resync)? [y/n]: ") == 'n') { *replace_dev = 0; } } static int _lvconvert_raid(struct logical_volume *lv, struct lvconvert_params *lp) { int replace = 0, image_count = 0; struct dm_list *failed_pvs; struct cmd_context *cmd = lv->vg->cmd; struct lv_segment *seg = first_seg(lv); dm_percent_t sync_percent; if (!arg_count(cmd, type_ARG)) lp->segtype = seg->segtype; /* Can only change image count for raid1 and linear */ if (arg_count(cmd, mirrors_ARG) && !seg_is_mirrored(seg) && !seg_is_linear(seg)) { log_error("'--mirrors/-m' is not compatible with %s", lvseg_name(seg)); return 0; } if (!_lvconvert_validate_thin(lv, lp)) return_0; if (!_is_valid_raid_conversion(seg->segtype, lp->segtype)) { log_error("Unable to convert %s/%s from %s to %s", lv->vg->name, lv->name, lvseg_name(seg), lp->segtype->name); return 0; } /* Change number of RAID1 images */ if (arg_count(cmd, mirrors_ARG) || arg_count(cmd, splitmirrors_ARG)) { image_count = lv_raid_image_count(lv); if (lp->mirrors_sign == SIGN_PLUS) image_count += lp->mirrors; else if (lp->mirrors_sign == SIGN_MINUS) image_count -= lp->mirrors; else image_count = lp->mirrors + 1; if (image_count < 1) { log_error("Unable to %s images by specified amount", arg_count(cmd, splitmirrors_ARG) ? "split" : "reduce"); return 0; } } if (lp->merge_mirror) return lv_raid_merge(lv); if (arg_count(cmd, trackchanges_ARG)) return lv_raid_split_and_track(lv, lp->pvh); if (arg_count(cmd, splitmirrors_ARG)) return lv_raid_split(lv, lp->lv_split_name, image_count, lp->pvh); if (arg_count(cmd, mirrors_ARG)) return lv_raid_change_image_count(lv, image_count, lp->pvh); if (arg_count(cmd, type_ARG)) return lv_raid_reshape(lv, lp->segtype); if (arg_count(cmd, replace_ARG)) return lv_raid_replace(lv, lp->replace_pvh, lp->pvh); if (arg_count(cmd, repair_ARG)) { if (!lv_is_active_exclusive_locally(lv_lock_holder(lv))) { log_error("%s/%s must be active %sto perform this" " operation.", lv->vg->name, lv->name, vg_is_clustered(lv->vg) ? "exclusive locally " : ""); return 0; } if (!lv_raid_percent(lv, &sync_percent)) { log_error("Unable to determine sync status of %s/%s.", lv->vg->name, lv->name); return 0; } if (sync_percent != DM_PERCENT_100) { log_warn("WARNING: %s/%s is not in-sync.", lv->vg->name, lv->name); log_warn("WARNING: Portions of the array may be unrecoverable."); /* * The kernel will not allow a device to be replaced * in an array that is not in-sync unless we override * by forcing the array to be considered "in-sync". */ init_mirror_in_sync(1); } _lvconvert_raid_repair_ask(cmd, lp, &replace); if (replace) { if (!(failed_pvs = _failed_pv_list(lv->vg))) return_0; if (!lv_raid_replace(lv, failed_pvs, lp->pvh)) { log_error("Failed to replace faulty devices in" " %s/%s.", lv->vg->name, lv->name); return 0; } log_print_unless_silent("Faulty devices in %s/%s successfully" " replaced.", lv->vg->name, lv->name); return 1; } /* "warn" if policy not set to replace */ if (arg_count(cmd, use_policies_ARG)) log_warn("Use 'lvconvert --repair %s/%s' to replace " "failed device.", lv->vg->name, lv->name); return 1; } log_error("Conversion operation not yet supported."); return 0; } static int _lvconvert_splitsnapshot(struct cmd_context *cmd, struct logical_volume *cow, struct lvconvert_params *lp) { struct volume_group *vg = cow->vg; if (!lv_is_cow(cow)) { log_error("%s/%s is not a snapshot.", vg->name, cow->name); return 0; } if (lv_is_origin(cow) || lv_is_external_origin(cow)) { log_error("Unable to split LV %s/%s that is a snapshot origin.", vg->name, cow->name); return 0; } if (lv_is_merging_cow(cow)) { log_error("Unable to split off snapshot %s/%s being merged into its origin.", vg->name, cow->name); return 0; } if (lv_is_virtual_origin(origin_from_cow(cow))) { log_error("Unable to split off snapshot %s/%s with virtual origin.", vg->name, cow->name); return 0; } if (lv_is_thin_pool(cow) || lv_is_pool_metadata_spare(cow)) { log_error("Unable to split off LV %s/%s needed by thin volume(s).", vg->name, cow->name); return 0; } if (!(vg->fid->fmt->features & FMT_MDAS)) { log_error("Unable to split off snapshot %s/%s using old LVM1-style metadata.", vg->name, cow->name); return 0; } if (!vg_check_status(vg, LVM_WRITE)) return_0; if (lv_is_pvmove(cow) || lv_is_mirror_type(cow) || lv_is_raid_type(cow) || lv_is_thin_type(cow)) { log_error("LV %s/%s type is unsupported with --splitsnapshot.", vg->name, cow->name); return 0; } if (lv_is_active_locally(cow)) { if (!lv_check_not_in_use(cow)) return_0; if ((lp->force == PROMPT) && !lp->yes && lv_is_visible(cow) && lv_is_active(cow)) { if (yes_no_prompt("Do you really want to split off active " "logical volume %s? [y/n]: ", cow->name) == 'n') { log_error("Logical volume %s not split.", cow->name); return 0; } } } if (!archive(vg)) return_0; log_verbose("Splitting snapshot %s/%s from its origin.", vg->name, cow->name); if (!vg_remove_snapshot(cow)) return_0; backup(vg); log_print_unless_silent("Logical Volume %s/%s split from its origin.", vg->name, cow->name); return 1; } static int _lvconvert_split_cached(struct cmd_context *cmd, struct logical_volume *lv) { struct logical_volume *cache_pool_lv = first_seg(lv)->pool_lv; log_debug("Detaching cache pool %s from cached LV %s.", display_lvname(cache_pool_lv), display_lvname(lv)); if (!archive(lv->vg)) return_0; if (!lv_cache_remove(lv)) return_0; if (!vg_write(lv->vg) || !vg_commit(lv->vg)) return_0; backup(lv->vg); log_print_unless_silent("Logical volume %s is not cached and cache pool %s is unused.", display_lvname(lv), display_lvname(cache_pool_lv)); return 1; } static int _lvconvert_splitcache(struct cmd_context *cmd, struct logical_volume *lv, struct lvconvert_params *lp) { struct lv_segment *seg; if (lv_is_thin_pool(lv)) lv = seg_lv(first_seg(lv), 0); /* cached _tdata ? */ /* When passed used cache-pool of used cached LV -> split cached LV */ if (lv_is_cache_pool(lv) && (dm_list_size(&lv->segs_using_this_lv) == 1) && (seg = get_only_segment_using_this_lv(lv)) && seg_is_cache(seg)) lv = seg->lv; /* Supported LV types for split */ if (!lv_is_cache(lv)) { log_error("Split of %s is not cache.", display_lvname(lv)); return 0; } if (!_lvconvert_split_cached(cmd, lv)) return_0; return 1; } static int _lvconvert_split(struct cmd_context *cmd, struct logical_volume *lv, struct lvconvert_params *lp) { struct lv_segment *seg; if (lv_is_thin_pool(lv) && lv_is_cache(seg_lv(first_seg(lv), 0))) lv = seg_lv(first_seg(lv), 0); /* cached _tdata ? */ /* When passed used cache-pool of used cached LV -> split cached LV */ if (lv_is_cache_pool(lv) && (dm_list_size(&lv->segs_using_this_lv) == 1) && (seg = get_only_segment_using_this_lv(lv)) && seg_is_cache(seg)) lv = seg->lv; /* Supported LV types for split */ if (lv_is_cache(lv)) { if (!_lvconvert_split_cached(cmd, lv)) return_0; /* Add more types here */ } else { log_error("Split of %s is unsupported.", display_lvname(lv)); return 0; } return 1; } static int _lvconvert_uncache(struct cmd_context *cmd, struct logical_volume *lv, struct lvconvert_params *lp) { if (lv_is_thin_pool(lv)) lv = seg_lv(first_seg(lv), 0); /* cached _tdata ? */ if (!lv_is_cache(lv)) { log_error("Cannot uncache non-cached logical volume %s.", display_lvname(lv)); return 0; } if (!lv_remove_single(cmd, first_seg(lv)->pool_lv, lp->force, 0)) return_0; log_print_unless_silent("Logical volume %s is not cached.", display_lvname(lv)); return 1; } static int _lvconvert_snapshot(struct cmd_context *cmd, struct logical_volume *lv, struct lvconvert_params *lp) { struct logical_volume *org; if (lv_is_cache_type(lv)) { log_error("Snapshots are not yet supported with cache type LVs %s.", display_lvname(lv)); return 0; } if (lv_is_mirrored(lv)) { log_error("Unable to convert mirrored LV \"%s\" into a snapshot.", lv->name); return 0; } if (lv_is_origin(lv)) { /* Unsupported stack */ log_error("Unable to convert origin \"%s\" into a snapshot.", lv->name); return 0; } if (lv_is_pool(lv)) { log_error("Unable to convert pool LVs %s into a snapshot.", display_lvname(lv)); return 0; } if (!(org = find_lv(lv->vg, lp->origin_name))) { log_error("Couldn't find origin volume %s in Volume group %s.", lp->origin_name, lv->vg->name); return 0; } if (org == lv) { log_error("Unable to use %s as both snapshot and origin.", display_lvname(lv)); return 0; } if (!cow_has_min_chunks(lv->vg, lv->le_count, lp->chunk_size)) return_0; if (lv_is_locked(org) || lv_is_cache_type(org) || lv_is_thin_type(org) || lv_is_pvmove(org) || lv_is_mirrored(org) || lv_is_cow(org)) { log_error("Unable to convert an LV into a snapshot of a %s LV.", lv_is_locked(org) ? "locked" : lv_is_cache_type(org) ? "cache type" : lv_is_thin_type(org) ? "thin type" : lv_is_pvmove(org) ? "pvmove" : lv_is_mirrored(org) ? "mirrored" : "snapshot"); return 0; } log_warn("WARNING: Converting logical volume %s to snapshot exception store.", display_lvname(lv)); log_warn("THIS WILL DESTROY CONTENT OF LOGICAL VOLUME (filesystem etc.)"); if (!lp->yes && yes_no_prompt("Do you really want to convert %s? [y/n]: ", display_lvname(lv)) == 'n') { log_error("Conversion aborted."); return 0; } if (!deactivate_lv(cmd, lv)) { log_error("Couldn't deactivate LV %s.", lv->name); return 0; } if (!lp->zero || !(lv->status & LVM_WRITE)) log_warn("WARNING: \"%s\" not zeroed", lv->name); else { lv->status |= LV_TEMPORARY; if (!activate_lv_local(cmd, lv) || !wipe_lv(lv, (struct wipe_params) { .do_zero = 1 })) { log_error("Aborting. Failed to wipe snapshot exception store."); return 0; } lv->status &= ~LV_TEMPORARY; /* Deactivates cleared metadata LV */ if (!deactivate_lv_local(lv->vg->cmd, lv)) { log_error("Failed to deactivate zeroed snapshot exception store."); return 0; } } if (!archive(lv->vg)) return_0; if (!vg_add_snapshot(org, lv, NULL, org->le_count, lp->chunk_size)) { log_error("Couldn't create snapshot."); return 0; } /* store vg on disk(s) */ if (!lv_update_and_reload(org)) return_0; log_print_unless_silent("Logical volume %s converted to snapshot.", lv->name); return 1; } static int _lvconvert_merge_old_snapshot(struct cmd_context *cmd, struct logical_volume *lv, struct lvconvert_params *lp) { int merge_on_activate = 0; struct logical_volume *origin = origin_from_cow(lv); struct lv_segment *snap_seg = find_snapshot(lv); struct lvinfo info; dm_percent_t snap_percent; /* Check if merge is possible */ if (!lv_is_cow(lv)) { log_error("\"%s\" is not a mergeable logical volume.", lv->name); return 0; } if (lv_is_merging_cow(lv)) { log_error("Snapshot %s is already merging.", lv->name); return 0; } if (lv_is_merging_origin(origin)) { log_error("Snapshot %s is already merging into the origin.", find_snapshot(origin)->cow->name); return 0; } if (lv_is_virtual_origin(origin)) { log_error("Snapshot %s has virtual origin.", lv->name); return 0; } if (lv_is_external_origin(origin_from_cow(lv))) { log_error("Cannot merge snapshot \"%s\" into " "the read-only external origin \"%s\".", lv->name, origin_from_cow(lv)->name); return 0; } /* FIXME: test when snapshot is remotely active */ if (lv_info(cmd, lv, 0, &info, 1, 0) && info.exists && info.live_table && (!lv_snapshot_percent(lv, &snap_percent) || snap_percent == DM_PERCENT_INVALID)) { log_error("Unable to merge invalidated snapshot LV \"%s\".", lv->name); return 0; } if (snap_seg->segtype->ops->target_present && !snap_seg->segtype->ops->target_present(cmd, snap_seg, NULL)) { log_error("Can't initialize snapshot merge. " "Missing support in kernel?"); return 0; } if (!archive(lv->vg)) return_0; /* * Prevent merge with open device(s) as it would likely lead * to application/filesystem failure. Merge on origin's next * activation if either the origin or snapshot LV are currently * open. * * FIXME testing open_count is racey; snapshot-merge target's * constructor and DM should prevent appropriate devices from * being open. */ if (lv_is_active_locally(origin)) { if (!lv_check_not_in_use(origin)) { log_print_unless_silent("Can't merge over open origin volume."); merge_on_activate = 1; } else if (!lv_check_not_in_use(lv)) { log_print_unless_silent("Can't merge when snapshot is open."); merge_on_activate = 1; } } else if (vg_is_clustered(origin->vg) && lv_is_active(origin)) { /* When it's active somewhere else */ log_print_unless_silent("Can't check whether remotely active snapshot is open."); merge_on_activate = 1; } init_snapshot_merge(snap_seg, origin); if (merge_on_activate) { /* Store and commit vg but skip starting the merge */ if (!vg_write(lv->vg) || !vg_commit(lv->vg)) return_0; backup(lv->vg); } else { /* Perform merge */ if (!lv_update_and_reload(origin)) return_0; lp->need_polling = 1; lp->lv_to_poll = origin; } if (merge_on_activate) log_print_unless_silent("Merging of snapshot %s will occur on " "next activation of %s.", display_lvname(lv), display_lvname(origin)); else log_print_unless_silent("Merging of volume %s started.", lv->name); return 1; } static int _lvconvert_merge_thin_snapshot(struct cmd_context *cmd, struct logical_volume *lv, struct lvconvert_params *lp) { int origin_is_active = 0, r = 0; struct lv_segment *snap_seg = first_seg(lv); struct logical_volume *origin = snap_seg->origin; if (!origin) { log_error("\"%s\" is not a mergeable logical volume.", lv->name); return 0; } /* Check if merge is possible */ if (lv_is_merging_origin(origin)) { log_error("Snapshot %s is already merging into the origin.", find_snapshot(origin)->lv->name); return 0; } if (lv_is_external_origin(origin)) { log_error("\"%s\" is read-only external origin \"%s\".", lv->name, origin_from_cow(lv)->name); return 0; } if (lv_is_origin(origin)) { log_error("Merging into the old snapshot origin %s is not supported.", origin->name); return 0; } if (!archive(lv->vg)) return_0; // FIXME: allow origin to be specified // FIXME: verify snapshot is descendant of specified origin /* * Prevent merge with open device(s) as it would likely lead * to application/filesystem failure. Merge on origin's next * activation if either the origin or snapshot LV can't be * deactivated. */ if (!deactivate_lv(cmd, lv)) log_print_unless_silent("Delaying merge since snapshot is open."); else if ((origin_is_active = lv_is_active(origin)) && !deactivate_lv(cmd, origin)) log_print_unless_silent("Delaying merge since origin volume is open."); else { /* * Both thin snapshot and origin are inactive, * replace the origin LV with its snapshot LV. */ if (!_finish_thin_merge(cmd, origin, lv)) goto_out; if (origin_is_active && !activate_lv(cmd, lv)) { log_error("Failed to reactivate origin %s.", lv->name); goto out; } r = 1; goto out; } init_snapshot_merge(snap_seg, origin); /* Commit vg, merge will start with next activation */ if (!vg_write(lv->vg) || !vg_commit(lv->vg)) return_0; r = 1; out: backup(lv->vg); if (r) log_print_unless_silent("Merging of thin snapshot %s will occur on " "next activation of %s.", display_lvname(lv), display_lvname(origin)); return r; } static int _lvconvert_pool_repair(struct cmd_context *cmd, struct logical_volume *pool_lv, struct lvconvert_params *lp) { const char *dmdir = dm_dir(); const char *thin_dump = find_config_tree_str_allow_empty(cmd, global_thin_dump_executable_CFG, NULL); const char *thin_repair = find_config_tree_str_allow_empty(cmd, global_thin_repair_executable_CFG, NULL); const struct dm_config_node *cn; const struct dm_config_value *cv; int ret = 0, status; int args = 0; const char *argv[19]; /* Max supported 10 args */ char *split, *dm_name, *trans_id_str; char meta_path[PATH_MAX]; char pms_path[PATH_MAX]; uint64_t trans_id; struct logical_volume *pmslv; struct logical_volume *mlv = first_seg(pool_lv)->metadata_lv; struct pipe_data pdata; FILE *f; if (!thin_repair[0]) { log_error("Thin repair commnand is not configured. Repair is disabled."); return 0; /* Checking disabled */ } pmslv = pool_lv->vg->pool_metadata_spare_lv; /* Check we have pool metadata spare LV */ if (!handle_pool_metadata_spare(pool_lv->vg, 0, NULL, 1)) return_0; if (pmslv != pool_lv->vg->pool_metadata_spare_lv) { if (!vg_write(pool_lv->vg) || !vg_commit(pool_lv->vg)) return_0; pmslv = pool_lv->vg->pool_metadata_spare_lv; } if (!(dm_name = dm_build_dm_name(cmd->mem, mlv->vg->name, mlv->name, NULL)) || (dm_snprintf(meta_path, sizeof(meta_path), "%s/%s", dmdir, dm_name) < 0)) { log_error("Failed to build thin metadata path."); return 0; } if (!(dm_name = dm_build_dm_name(cmd->mem, pmslv->vg->name, pmslv->name, NULL)) || (dm_snprintf(pms_path, sizeof(pms_path), "%s/%s", dmdir, dm_name) < 0)) { log_error("Failed to build pool metadata spare path."); return 0; } if ((cn = find_config_tree_node(cmd, global_thin_repair_options_CFG, NULL))) { for (cv = cn->v; cv && args < 16; cv = cv->next) { if (cv->type != DM_CFG_STRING) { log_error("Invalid string in config file: " "global/thin_repair_options"); return 0; } argv[++args] = cv->v.str; } } else { /* Use default options (no support for options with spaces) */ if (!(split = dm_pool_strdup(cmd->mem, DEFAULT_THIN_REPAIR_OPTIONS))) { log_error("Failed to duplicate thin repair string."); return 0; } args = dm_split_words(split, 16, 0, (char**) argv + 1); } if (args == 10) { log_error("Too many options for thin repair command."); return 0; } argv[0] = thin_repair; argv[++args] = "-i"; argv[++args] = meta_path; argv[++args] = "-o"; argv[++args] = pms_path; argv[++args] = NULL; if (pool_is_active(pool_lv)) { log_error("Only inactive pool can be repaired."); return 0; } if (!activate_lv_local(cmd, pmslv)) { log_error("Cannot activate pool metadata spare volume %s.", pmslv->name); return 0; } if (!activate_lv_local(cmd, mlv)) { log_error("Cannot activate thin pool metadata volume %s.", mlv->name); goto deactivate_pmslv; } if (!(ret = exec_cmd(cmd, (const char * const *)argv, &status, 1))) { log_error("Repair of thin metadata volume of thin pool %s/%s failed (status:%d). " "Manual repair required!", pool_lv->vg->name, pool_lv->name, status); goto deactivate_mlv; } if (thin_dump[0]) { argv[0] = thin_dump; argv[1] = pms_path; argv[2] = NULL; if (!(f = pipe_open(cmd, argv, 0, &pdata))) log_warn("WARNING: Cannot read output from %s %s.", thin_dump, pms_path); else { /* * Scan only the 1st. line for transation id. * Watch out, if the thin_dump format changes */ if ((fgets(meta_path, sizeof(meta_path), f) > 0) && (trans_id_str = strstr(meta_path, "transaction=\"")) && (sscanf(trans_id_str + 13, "%" PRIu64, &trans_id) == 1) && (trans_id != first_seg(pool_lv)->transaction_id) && ((trans_id - 1) != first_seg(pool_lv)->transaction_id)) log_error("Transaction id %" PRIu64 " from pool \"%s/%s\" " "does not match repaired transaction id " "%" PRIu64 " from %s.", first_seg(pool_lv)->transaction_id, pool_lv->vg->name, pool_lv->name, trans_id, pms_path); (void) pipe_close(&pdata); /* killing pipe */ } } deactivate_mlv: if (!deactivate_lv(cmd, mlv)) { log_error("Cannot deactivate thin pool metadata volume %s.", mlv->name); return 0; } deactivate_pmslv: if (!deactivate_lv(cmd, pmslv)) { log_error("Cannot deactivate thin pool metadata volume %s.", mlv->name); return 0; } if (!ret) return 0; if (pmslv == pool_lv->vg->pool_metadata_spare_lv) { pool_lv->vg->pool_metadata_spare_lv = NULL; pmslv->status &= ~POOL_METADATA_SPARE; lv_set_visible(pmslv); } /* Try to allocate new pool metadata spare LV */ if (!handle_pool_metadata_spare(pool_lv->vg, 0, NULL, 1)) stack; if (dm_snprintf(meta_path, sizeof(meta_path), "%s_meta%%d", pool_lv->name) < 0) { log_error("Can't prepare new metadata name for %s.", pool_lv->name); return 0; } if (!generate_lv_name(pool_lv->vg, meta_path, pms_path, sizeof(pms_path))) { log_error("Can't generate new name for %s.", meta_path); return 0; } if (!detach_pool_metadata_lv(first_seg(pool_lv), &mlv)) return_0; if (!_swap_lv_identifiers(cmd, mlv, pmslv)) return_0; /* Used _pmspare will become _tmeta */ if (!attach_pool_metadata_lv(first_seg(pool_lv), pmslv)) return_0; /* Used _tmeta will become visible _meta%d */ if (!lv_rename_update(cmd, mlv, pms_path, 0)) return_0; if (!vg_write(pool_lv->vg) || !vg_commit(pool_lv->vg)) return_0; log_warn("WARNING: If everything works, remove \"%s/%s\".", mlv->vg->name, mlv->name); log_warn("WARNING: Use pvmove command to move \"%s/%s\" on the best fitting PV.", pool_lv->vg->name, first_seg(pool_lv)->metadata_lv->name); return 1; } /* Currently converts only to thin volume with external origin */ static int _lvconvert_thin(struct cmd_context *cmd, struct logical_volume *lv, struct lvconvert_params *lp) { struct logical_volume *torigin_lv, *pool_lv = lp->pool_data_lv; struct volume_group *vg = lv->vg; struct lvcreate_params lvc = { .activate = CHANGE_AEY, .alloc = ALLOC_INHERIT, .lv_name = lp->origin_name, .major = -1, .minor = -1, .permission = LVM_READ, .pool_name = pool_lv->name, .pvh = &vg->pvs, .read_ahead = DM_READ_AHEAD_AUTO, .stripes = 1, .virtual_extents = lv->le_count, }; if (lv == pool_lv) { log_error("Can't use same LV %s for thin pool and thin volume.", display_lvname(pool_lv)); return 0; } if (lv_is_locked(lv) || !lv_is_visible(lv) || lv_is_cache_type(lv) || lv_is_cow(lv) || lv_is_pool(lv) || lv_is_pool_data(lv) || lv_is_pool_metadata(lv)) { log_error("Can't use%s%s %s %s as external origin.", lv_is_locked(lv) ? " locked" : "", lv_is_visible(lv) ? "" : " hidden", lvseg_name(first_seg(lv)), display_lvname(lv)); return 0; } dm_list_init(&lvc.tags); if (!pool_supports_external_origin(first_seg(pool_lv), lv)) return_0; if (!(lvc.segtype = get_segtype_from_string(cmd, "thin"))) return_0; if (!archive(vg)) return_0; /* New thin LV needs to be created (all messages sent to pool) */ if (!(torigin_lv = lv_create_single(vg, &lvc))) return_0; /* Deactivate prepared Thin LV */ if (!deactivate_lv(cmd, torigin_lv)) { log_error("Aborting. Unable to deactivate new LV. " "Manual intervention required."); return 0; } /* * Crashing till this point will leave plain thin volume * which could be easily removed by the user after i.e. power-off */ if (!_swap_lv_identifiers(cmd, torigin_lv, lv)) { stack; goto revert_new_lv; } /* Preserve read-write status of original LV here */ torigin_lv->status |= (lv->status & LVM_WRITE); if (!attach_thin_external_origin(first_seg(torigin_lv), lv)) { stack; goto revert_new_lv; } if (!lv_update_and_reload(torigin_lv)) { stack; goto deactivate_and_revert_new_lv; } log_print_unless_silent("Converted %s to thin volume with " "external origin %s.", display_lvname(torigin_lv), display_lvname(lv)); return 1; deactivate_and_revert_new_lv: if (!_swap_lv_identifiers(cmd, torigin_lv, lv)) stack; if (!deactivate_lv(cmd, torigin_lv)) { log_error("Unable to deactivate failed new LV. " "Manual intervention required."); return 0; } if (!detach_thin_external_origin(first_seg(torigin_lv))) return_0; revert_new_lv: /* FIXME Better to revert to backup of metadata? */ if (!lv_remove(torigin_lv) || !vg_write(vg) || !vg_commit(vg)) log_error("Manual intervention may be required to remove " "abandoned LV(s) before retrying."); else backup(vg); return 0; } static int _lvconvert_update_pool_params(struct logical_volume *pool_lv, struct lvconvert_params *lp) { if (lp->pool_metadata_size && !(lp->pool_metadata_extents = extents_from_size(pool_lv->vg->cmd, lp->pool_metadata_size, pool_lv->vg->extent_size))) return_0; return update_pool_params(lp->segtype, pool_lv->vg, lp->target_attr, lp->passed_args, pool_lv->le_count, &lp->pool_metadata_extents, &lp->thin_chunk_size_calc_policy, &lp->chunk_size, &lp->discards, &lp->zero); } /* * Converts a data lv and a metadata lv into a thin or cache pool lv. * * Thin lvconvert version which * rename metadata * convert/layers thinpool over data * attach metadata * * pool_lv might or might not already be a pool. */ static int _lvconvert_pool(struct cmd_context *cmd, struct logical_volume *pool_lv, struct lvconvert_params *lp) { int r = 0; const char *old_name; struct lv_segment *seg; struct volume_group *vg = pool_lv->vg; struct logical_volume *data_lv; struct logical_volume *metadata_lv = NULL; struct logical_volume *pool_metadata_lv; char metadata_name[NAME_LEN], data_name[NAME_LEN]; int activate_pool; if (lp->pool_data_name) { if ((lp->thin || lp->cache) && !strcmp(lp->pool_data_name, pool_lv->name)) { log_error("Converted volume %s and pool volume must differ.", display_lvname(pool_lv)); return 0; } if (!(pool_lv = find_lv(vg, lp->pool_data_name))) { log_error("Unknown pool data LV %s.", lp->pool_data_name); return 0; } } if (!lv_is_visible(pool_lv)) { log_error("Can't convert internal LV %s.", display_lvname(pool_lv)); return 0; } if (lv_is_locked(pool_lv)) { log_error("Can't convert locked LV %s.", display_lvname(pool_lv)); return 0; } if (lv_is_thin_pool(pool_lv) && (segtype_is_cache_pool(lp->segtype) || lp->cache)) { log_error("Can't convert thin pool LV %s.", display_lvname(pool_lv)); return 0; } if (lv_is_cache(pool_lv) && !segtype_is_thin_pool(lp->segtype)) { log_error("Cached LV %s could be only converted into a thin pool volume.", display_lvname(pool_lv)); return 0; } if (lv_is_cache_pool(pool_lv) && (segtype_is_thin_pool(lp->segtype) || lp->thin)) { log_error("Cannot convert cache pool %s as pool data volume.", display_lvname(pool_lv)); return 0; } if (lv_is_mirror(pool_lv)) { log_error("Mirror logical volumes cannot be used as pools."); log_print_unless_silent("Try \"%s\" segment type instead.", SEG_TYPE_NAME_RAID1); return 0; } /* * Only linear, striped and raid supported. * FIXME Tidy up all these type restrictions. */ if (!lv_is_pool(pool_lv) && (lv_is_thin_type(pool_lv) || lv_is_cow(pool_lv) || lv_is_merging_cow(pool_lv) || lv_is_origin(pool_lv) ||lv_is_merging_origin(pool_lv) || lv_is_external_origin(pool_lv) || lv_is_virtual(pool_lv))) { log_error("Pool data LV %s is of an unsupported type.", display_lvname(pool_lv)); return 0; } if (lp->pool_metadata_name) { if (!(lp->pool_metadata_lv = find_lv(vg, lp->pool_metadata_name))) { log_error("Unknown pool metadata LV %s.", lp->pool_metadata_name); return 0; } lp->pool_metadata_extents = lp->pool_metadata_lv->le_count; metadata_lv = lp->pool_metadata_lv; if (metadata_lv == pool_lv) { log_error("Can't use same LV for pool data and metadata LV %s.", display_lvname(metadata_lv)); return 0; } if (!lv_is_visible(metadata_lv)) { log_error("Can't convert internal LV %s.", display_lvname(metadata_lv)); return 0; } if (lv_is_locked(metadata_lv)) { log_error("Can't convert locked LV %s.", display_lvname(metadata_lv)); return 0; } if (lv_is_mirror(metadata_lv)) { log_error("Mirror logical volumes cannot be used for pool metadata."); log_print_unless_silent("Try \"%s\" segment type instead.", SEG_TYPE_NAME_RAID1); return 0; } /* FIXME Tidy up all these type restrictions. */ if (lv_is_cache_type(metadata_lv) || lv_is_thin_type(metadata_lv) || lv_is_cow(metadata_lv) || lv_is_merging_cow(metadata_lv) || lv_is_origin(metadata_lv) || lv_is_merging_origin(metadata_lv) || lv_is_external_origin(metadata_lv) || lv_is_virtual(metadata_lv)) { log_error("Pool metadata LV %s is of an unsupported type.", display_lvname(metadata_lv)); return 0; } if (!lv_is_pool(pool_lv)) { if (!_lvconvert_update_pool_params(pool_lv, lp)) return_0; if (lp->pool_metadata_extents > metadata_lv->le_count) { log_error("Logical volume %s is too small for metadata.", display_lvname(metadata_lv)); return 0; } } } if (lv_is_pool(pool_lv)) { lp->pool_data_lv = pool_lv; if (!metadata_lv) { if (arg_from_list_is_set(cmd, "is invalid with existing pool", cachemode_ARG, chunksize_ARG, discards_ARG, zero_ARG, poolmetadatasize_ARG, -1)) return_0; if (lp->thin || lp->cache) /* already pool, can continue converting volume */ return 1; log_error("LV %s is already pool.", display_lvname(pool_lv)); return 0; } if (lp->thin || lp->cache) { log_error("--%s and pool metadata swap is not supported.", lp->thin ? "thin" : "cache"); return 0; } /* FIXME cache pool */ if (lv_is_thin_pool(pool_lv) && pool_is_active(pool_lv)) { /* If any volume referencing pool active - abort here */ log_error("Cannot convert pool %s with active volumes.", display_lvname(pool_lv)); return 0; } lp->passed_args |= PASS_ARG_CHUNK_SIZE | PASS_ARG_DISCARDS | PASS_ARG_ZERO; seg = first_seg(pool_lv); /* Normally do NOT change chunk size when swapping */ if (arg_count(cmd, chunksize_ARG) && (lp->chunk_size != seg->chunk_size) && !dm_list_empty(&pool_lv->segs_using_this_lv)) { if (lp->force == PROMPT) { log_error("Chunk size can be only changed with --force. Conversion aborted."); return 0; } log_warn("WARNING: Changing chunk size %s to " "%s for %s pool volume.", display_size(cmd, seg->chunk_size), display_size(cmd, lp->chunk_size), display_lvname(pool_lv)); /* Ok, user has likely some serious reason for this */ if (!lp->yes && yes_no_prompt("Do you really want to change chunk size " "for %s pool volume? [y/n]: ", display_lvname(pool_lv)) == 'n') { log_error("Conversion aborted."); return 0; } } else lp->chunk_size = seg->chunk_size; if (!_lvconvert_update_pool_params(pool_lv, lp)) return_0; if (metadata_lv->le_count < lp->pool_metadata_extents) log_print_unless_silent("Continuing with swap..."); if (!arg_count(cmd, discards_ARG)) lp->discards = seg->discards; if (!arg_count(cmd, zero_ARG)) lp->zero = seg->zero_new_blocks; if (!lp->yes && yes_no_prompt("Do you want to swap metadata of %s " "pool with metadata volume %s? [y/n]: ", display_lvname(pool_lv), display_lvname(metadata_lv)) == 'n') { log_error("Conversion aborted."); return 0; } } else { log_warn("WARNING: Converting logical volume %s%s%s to pool's data%s.", display_lvname(pool_lv), metadata_lv ? " and " : "", metadata_lv ? display_lvname(metadata_lv) : "", metadata_lv ? " and metadata volumes" : " volume"); log_warn("THIS WILL DESTROY CONTENT OF LOGICAL VOLUME (filesystem etc.)"); if (!lp->yes && yes_no_prompt("Do you really want to convert %s%s%s? [y/n]: ", display_lvname(pool_lv), metadata_lv ? " and " : "", metadata_lv ? display_lvname(metadata_lv) : "") == 'n') { log_error("Conversion aborted."); return 0; } } if (segtype_is_cache_pool(lp->segtype)) activate_pool = 0; /* Cannot activate cache pool */ else /* Allow to have only thinpool active and restore it's active state */ activate_pool = lv_is_active(pool_lv); if ((dm_snprintf(metadata_name, sizeof(metadata_name), "%s%s", pool_lv->name, (segtype_is_cache_pool(lp->segtype)) ? "_cmeta" : "_tmeta") < 0) || (dm_snprintf(data_name, sizeof(data_name), "%s%s", pool_lv->name, (segtype_is_cache_pool(lp->segtype)) ? "_cdata" : "_tdata") < 0)) { log_error("Failed to create internal lv names, " "pool name is too long."); return 0; } if (!metadata_lv) { if (!_lvconvert_update_pool_params(pool_lv, lp)) return_0; if (!get_stripe_params(cmd, &lp->stripes, &lp->stripe_size)) return_0; if (!archive(vg)) return_0; if (!(metadata_lv = alloc_pool_metadata(pool_lv, metadata_name, lp->read_ahead, lp->stripes, lp->stripe_size, lp->pool_metadata_extents, lp->alloc, lp->pvh))) return_0; } else { if (!deactivate_lv(cmd, metadata_lv)) { log_error("Aborting. Failed to deactivate %s.", display_lvname(metadata_lv)); return 0; } if (!archive(vg)) return_0; /* Swap normal LV with pool's metadata LV ? */ if (lv_is_pool(pool_lv)) { /* Swap names between old and new metadata LV */ seg = first_seg(pool_lv); if (!detach_pool_metadata_lv(seg, &pool_metadata_lv)) return_0; old_name = metadata_lv->name; if (!lv_rename_update(cmd, metadata_lv, "pvmove_tmeta", 0)) return_0; if (!lv_rename_update(cmd, pool_metadata_lv, old_name, 0)) return_0; goto mda_write; } metadata_lv->status |= LV_TEMPORARY; if (!activate_lv_local(cmd, metadata_lv)) { log_error("Aborting. Failed to activate metadata lv."); return 0; } if (!wipe_lv(metadata_lv, (struct wipe_params) { .do_zero = 1 })) { log_error("Aborting. Failed to wipe metadata lv."); return 0; } } /* We are changing target type, so deactivate first */ if (!deactivate_lv(cmd, metadata_lv)) { log_error("Aborting. Failed to deactivate metadata lv. " "Manual intervention required."); return 0; } if (!deactivate_lv(cmd, pool_lv)) { log_error("Aborting. Failed to deactivate logical volume %s.", display_lvname(pool_lv)); return 0; } data_lv = pool_lv; old_name = data_lv->name; /* Use for pool name */ /* * Since we wish to have underlaying devs to match _[ct]data * rename data LV to match pool LV subtree first, * also checks for visible LV. */ /* FIXME: any more types prohibited here? */ if (!lv_rename_update(cmd, data_lv, data_name, 0)) return_0; if (!(pool_lv = lv_create_empty(old_name, NULL, ((segtype_is_cache_pool(lp->segtype)) ? CACHE_POOL : THIN_POOL) | VISIBLE_LV | LVM_READ | LVM_WRITE, ALLOC_INHERIT, vg))) { log_error("Creation of pool LV failed."); return 0; } /* Allocate a new pool segment */ if (!(seg = alloc_lv_segment(lp->segtype, pool_lv, 0, data_lv->le_count, pool_lv->status, 0, NULL, 1, data_lv->le_count, 0, 0, 0, NULL))) return_0; /* Add the new segment to the layer LV */ dm_list_add(&pool_lv->segments, &seg->list); pool_lv->le_count = data_lv->le_count; pool_lv->size = data_lv->size; if (!attach_pool_data_lv(seg, data_lv)) return_0; /* FIXME: revert renamed LVs in fail path? */ /* FIXME: any common code with metadata/thin_manip.c extend_pool() ? */ seg->low_water_mark = 0; seg->transaction_id = 0; mda_write: seg->chunk_size = lp->chunk_size; seg->discards = lp->discards; seg->zero_new_blocks = lp->zero ? 1 : 0; seg->feature_flags = lp->feature_flags; /* cache-pool */ /* Rename deactivated metadata LV to have _tmeta suffix */ /* Implicit checks if metadata_lv is visible */ if (lp->pool_metadata_name && !lv_rename_update(cmd, metadata_lv, metadata_name, 0)) return_0; if (!attach_pool_metadata_lv(seg, metadata_lv)) return_0; if (!handle_pool_metadata_spare(vg, metadata_lv->le_count, lp->pvh, lp->poolmetadataspare)) return_0; if (!vg_write(vg) || !vg_commit(vg)) return_0; if (seg->zero_new_blocks && seg->chunk_size >= DEFAULT_THIN_POOL_CHUNK_SIZE_PERFORMANCE * 2) log_warn("WARNING: Pool zeroing and large %s chunk size slows down " "provisioning.", display_size(cmd, seg->chunk_size)); if (activate_pool && !activate_lv_excl(cmd, pool_lv)) { log_error("Failed to activate pool logical volume %s.", display_lvname(pool_lv)); /* Deactivate subvolumes */ if (!deactivate_lv(cmd, seg_lv(seg, 0))) log_error("Failed to deactivate pool data logical volume."); if (!deactivate_lv(cmd, seg->metadata_lv)) log_error("Failed to deactivate pool metadata logical volume."); goto out; } r = 1; lp->pool_data_lv = pool_lv; out: backup(vg); if (r) log_print_unless_silent("Converted %s to %s pool.", display_lvname(pool_lv), (segtype_is_cache_pool(lp->segtype)) ? "cache" : "thin"); return r; #if 0 revert_new_lv: /* TBD */ if (!lp->pool_metadata_lv_name) { if (!deactivate_lv(cmd, metadata_lv)) { log_error("Failed to deactivate metadata lv."); return 0; } if (!lv_remove(metadata_lv) || !vg_write(vg) || !vg_commit(vg)) log_error("Manual intervention may be required to remove " "abandoned LV(s) before retrying."); else backup(vg); } return 0; #endif } /* * Convert origin into a cache LV by attaching a cache pool. */ static int _lvconvert_cache(struct cmd_context *cmd, struct logical_volume *origin_lv, struct lvconvert_params *lp) { struct logical_volume *pool_lv = lp->pool_data_lv; struct logical_volume *cache_lv; if (!validate_lv_cache_create_pool(pool_lv)) return_0; if (!archive(origin_lv->vg)) return_0; if (!(cache_lv = lv_cache_create(pool_lv, origin_lv))) return_0; if (!lv_update_and_reload(cache_lv)) return_0; log_print_unless_silent("Logical volume %s is now cached.", display_lvname(cache_lv)); return 1; } static int _lvconvert_single(struct cmd_context *cmd, struct logical_volume *lv, void *handle) { struct logical_volume *origin = NULL; struct lvconvert_params *lp = handle; struct dm_list *failed_pvs; if (lv_is_locked(lv)) { log_error("Cannot convert locked LV %s", lv->name); return ECMD_FAILED; } if (lv_is_cow(lv) && !lp->merge && !lp->splitsnapshot) { log_error("Can't convert snapshot logical volume \"%s\"", lv->name); return ECMD_FAILED; } if (lv_is_pvmove(lv)) { log_error("Unable to convert pvmove LV %s", lv->name); return ECMD_FAILED; } if (lp->splitsnapshot) { if (!_lvconvert_splitsnapshot(cmd, lv, lp)) return_ECMD_FAILED; return ECMD_PROCESSED; } if (lp->splitcache) { if (!_lvconvert_splitcache(cmd, lv, lp)) return_ECMD_FAILED; return ECMD_PROCESSED; } if (lp->split) { if (!_lvconvert_split(cmd, lv, lp)) return_ECMD_FAILED; return ECMD_PROCESSED; } if (lp->uncache) { if (!_lvconvert_uncache(cmd, lv, lp)) return_ECMD_FAILED; return ECMD_PROCESSED; } /* Validate origin prior we start conversion of pool */ if (lp->cache && !validate_lv_cache_create_origin(lv)) return_ECMD_FAILED; if (lp->thin) { if (lv_is_cache_type(lv) || lv_is_pool(lv) || lv_is_thin_pool_data(lv) || lv_is_thin_pool_metadata(lv)) { log_error("Can't convert %s %s to external origin.", first_seg(lv)->segtype->name, display_lvname(lv)); return ECMD_FAILED; } } if (arg_count(cmd, repair_ARG)) { if (lv_is_pool(lv)) { if (!_lvconvert_pool_repair(cmd, lv, lp)) return_ECMD_FAILED; return ECMD_PROCESSED; } if (!lv_is_mirrored(lv) && !lv_is_raid(lv)) { if (arg_count(cmd, use_policies_ARG)) return ECMD_PROCESSED; /* nothing to be done here */ log_error("Can't repair LV \"%s\" of segtype %s.", lv->name, lvseg_name(first_seg(lv))); return ECMD_FAILED; } } /* forward splitmirror operations to the cache origin, which may be raid * or old-style mirror */ if (arg_count(cmd, splitmirrors_ARG) && lv_is_cache_type(lv) && (origin = seg_lv(first_seg(lv), 0)) && lv_is_cache_origin(origin)) { log_warn("WARNING: Selected operation does not work with cache-type LVs."); log_warn("Proceeding using the cache origin LV %s instead", origin->name); lv = origin; } if (!lp->segtype) { /* segtype not explicitly set in _read_params */ lp->segtype = first_seg(lv)->segtype; /* * If we are converting to mirror/raid1 and * the segtype was not specified, then we need * to consult the default. */ if (arg_count(cmd, mirrors_ARG) && !lv_is_mirrored(lv)) { if (!(lp->segtype = get_segtype_from_string(cmd, find_config_tree_str(cmd, global_mirror_segtype_default_CFG, NULL)))) return_ECMD_FAILED; } } if (lp->merge) { if ((lv_is_thin_volume(lv) && !_lvconvert_merge_thin_snapshot(cmd, lv, lp)) || (!lv_is_thin_volume(lv) && !_lvconvert_merge_old_snapshot(cmd, lv, lp))) { log_print_unless_silent("Unable to merge LV \"%s\" into its origin.", lv->name); return ECMD_FAILED; } } else if (lp->snapshot) { if (!_lvconvert_snapshot(cmd, lv, lp)) return_ECMD_FAILED; } else if (segtype_is_pool(lp->segtype) || lp->thin || lp->cache) { if (!_lvconvert_pool(cmd, lv, lp)) return_ECMD_FAILED; if ((lp->thin && !_lvconvert_thin(cmd, lv, lp)) || (lp->cache && !_lvconvert_cache(cmd, lv, lp))) return_ECMD_FAILED; } else if (segtype_is_raid(lp->segtype) || (lv->status & RAID) || lp->merge_mirror) { if (!archive(lv->vg)) return_ECMD_FAILED; if (!_lvconvert_raid(lv, lp)) return_ECMD_FAILED; if (!(failed_pvs = _failed_pv_list(lv->vg))) return_ECMD_FAILED; /* If repairing and using policies, remove missing PVs from VG */ if (arg_count(cmd, repair_ARG) && arg_count(cmd, use_policies_ARG)) _remove_missing_empty_pv(lv->vg, failed_pvs); } else if (arg_count(cmd, mirrors_ARG) || arg_count(cmd, splitmirrors_ARG) || lv_is_mirrored(lv)) { if (!archive(lv->vg)) return_ECMD_FAILED; if (!_lvconvert_mirrors(cmd, lv, lp)) return_ECMD_FAILED; if (!(failed_pvs = _failed_pv_list(lv->vg))) return_ECMD_FAILED; /* If repairing and using policies, remove missing PVs from VG */ if (arg_count(cmd, repair_ARG) && arg_count(cmd, use_policies_ARG)) _remove_missing_empty_pv(lv->vg, failed_pvs); } return ECMD_PROCESSED; } /* * FIXME move to toollib along with the rest of the drop/reacquire * VG locking that is used by _lvconvert_merge_single() */ static struct logical_volume *get_vg_lock_and_logical_volume(struct cmd_context *cmd, const char *vg_name, const char *lv_name) { /* * Returns NULL if the requested LV doesn't exist; * otherwise the caller must release_vg(lv->vg) * - it is also up to the caller to unlock_vg() as needed */ struct volume_group *vg; struct logical_volume* lv = NULL; vg = _get_lvconvert_vg(cmd, vg_name, NULL); if (vg_read_error(vg)) { release_vg(vg); return_NULL; } if (!(lv = _get_lvconvert_lv(cmd, vg, lv_name, NULL, 0))) { log_error("Can't find LV %s in VG %s", lv_name, vg_name); unlock_and_release_vg(cmd, vg, vg_name); return NULL; } return lv; } static int _poll_logical_volume(struct cmd_context *cmd, struct logical_volume *lv, int wait_completion) { struct lvinfo info; if (!lv_info(cmd, lv, 0, &info, 0, 0) || !info.exists) { log_print_unless_silent("Conversion starts after activation."); return ECMD_PROCESSED; } return lvconvert_poll(cmd, lv, wait_completion ? 0 : 1U); } static int lvconvert_single(struct cmd_context *cmd, struct lvconvert_params *lp) { struct logical_volume *lv; int ret = ECMD_FAILED; int saved_ignore_suspended_devices = ignore_suspended_devices(); if (arg_count(cmd, repair_ARG)) { init_ignore_suspended_devices(1); cmd->handles_missing_pvs = 1; } if (!(lv = get_vg_lock_and_logical_volume(cmd, lp->vg_name, lp->lv_name))) goto_out; /* * lp->pvh holds the list of PVs available for allocation or removal */ if (lp->pv_count) { if (!(lp->pvh = create_pv_list(cmd->mem, lv->vg, lp->pv_count, lp->pvs, 0))) goto_bad; } else lp->pvh = &lv->vg->pvs; if (lp->replace_pv_count && !(lp->replace_pvh = create_pv_list(cmd->mem, lv->vg, lp->replace_pv_count, lp->replace_pvs, 0))) goto_bad; lp->lv_to_poll = lv; ret = _lvconvert_single(cmd, lv, lp); bad: unlock_vg(cmd, lp->vg_name); if (ret == ECMD_PROCESSED && lp->need_polling) ret = _poll_logical_volume(cmd, lp->lv_to_poll, lp->wait_completion); release_vg(lv->vg); out: init_ignore_suspended_devices(saved_ignore_suspended_devices); return ret; } static int _lvconvert_merge_single(struct cmd_context *cmd, struct logical_volume *lv, struct processing_handle *handle) { struct lvconvert_params *lp = (struct lvconvert_params *) handle->custom_handle; const char *vg_name; struct logical_volume *refreshed_lv; int ret; /* * FIXME can't trust lv's VG to be current given that caller * is process_each_lv() -- _poll_logical_volume() may have * already updated the VG's metadata in an earlier iteration. * - preemptively drop the VG lock, as is needed for * _poll_logical_volume(), refresh LV (and VG in the process). */ vg_name = lv->vg->name; unlock_vg(cmd, vg_name); refreshed_lv = get_vg_lock_and_logical_volume(cmd, vg_name, lv->name); if (!refreshed_lv) { log_error("ABORTING: Can't reread LV %s/%s", vg_name, lv->name); return ECMD_FAILED; } lp->lv_to_poll = refreshed_lv; if ((ret = _lvconvert_single(cmd, refreshed_lv, lp)) != ECMD_PROCESSED) stack; if (ret == ECMD_PROCESSED && lp->need_polling) { /* * Must drop VG lock, because lvconvert_poll() needs it, * then reacquire it after polling completes */ unlock_vg(cmd, vg_name); if ((ret = _poll_logical_volume(cmd, lp->lv_to_poll, lp->wait_completion)) != ECMD_PROCESSED) stack; /* use LCK_VG_WRITE to match lvconvert()'s READ_FOR_UPDATE */ if (!lock_vol(cmd, vg_name, LCK_VG_WRITE, NULL)) { log_error("ABORTING: Can't relock VG for %s " "after polling finished", vg_name); ret = ECMD_FAILED; } } release_vg(refreshed_lv->vg); return ret; } int lvconvert(struct cmd_context * cmd, int argc, char **argv) { int ret; struct lvconvert_params lp = { .target_attr = ~0, }; struct processing_handle *handle = NULL; if (!(handle = init_processing_handle(cmd))) { log_error("Failed to initialize processing handle."); ret = ECMD_FAILED; goto out; } handle->custom_handle = &lp; if (!_read_params(cmd, argc, argv, &lp)) { ret = EINVALID_CMD_LINE; goto_out; } if (lp.merge) ret = process_each_lv(cmd, argc, argv, READ_FOR_UPDATE, handle, &_lvconvert_merge_single); else ret = lvconvert_single(cmd, &lp); out: destroy_processing_handle(cmd, handle); return ret; }