/* * Copyright (C) 2003-2004 Sistina Software, Inc. All rights reserved. * Copyright (C) 2004-2015 Red Hat, Inc. All rights reserved. * * This file is part of LVM2. * * This copyrighted material is made available to anyone wishing to use, * modify, copy, or redistribute it subject to the terms and conditions * of the GNU Lesser General Public License v.2.1. * * You should have received a copy of the GNU Lesser General Public License * along with this program; if not, write to the Free Software Foundation, * Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include "tools.h" #include "lib/lvmpolld/polldaemon.h" #include "lib/display/display.h" #include "pvmove_poll.h" #include "lib/lvmpolld/lvmpolld-client.h" #define PVMOVE_FIRST_TIME 0x00000001 /* Called for first time */ struct pvmove_params { char *pv_name_arg; /* original unmodified arg */ char *lv_name_arg; /* original unmodified arg */ alloc_policy_t alloc; int pv_count; char **pv_names; union lvid *lvid; char *id_vg_name; char *id_lv_name; unsigned in_progress; int setup_result; int found_pv; }; static int _pvmove_target_present(struct cmd_context *cmd, int clustered) { const struct segment_type *segtype; int found = 1; if (!(segtype = get_segtype_from_string(cmd, SEG_TYPE_NAME_MIRROR))) return_0; if (activation() && segtype->ops->target_present && !segtype->ops->target_present(cmd, NULL, NULL)) found = 0; return found; } static unsigned _pvmove_is_exclusive(struct cmd_context *cmd, struct volume_group *vg) { return 0; } /* Allow /dev/vgname/lvname, vgname/lvname or lvname */ static const char *_extract_lvname(struct cmd_context *cmd, const char *vgname, const char *arg) { const char *lvname; /* Is an lvname supplied directly? */ if (!strchr(arg, '/')) return arg; lvname = skip_dev_dir(cmd, arg, NULL); while (*lvname == '/') lvname++; if (!strchr(lvname, '/')) { log_error("--name takes a logical volume name."); return NULL; } if (strncmp(vgname, lvname, strlen(vgname)) || (lvname += strlen(vgname), *lvname != '/')) { log_error("Named LV and old PV must be in the same VG."); return NULL; } while (*lvname == '/') lvname++; if (!*lvname) { log_error("Incomplete LV name supplied with --name."); return NULL; } return lvname; } /* Create list of PVs for allocation of replacement extents */ static struct dm_list *_get_allocatable_pvs(struct cmd_context *cmd, int argc, char **argv, struct volume_group *vg, struct physical_volume *pv, alloc_policy_t alloc) { struct dm_list *allocatable_pvs, *pvht, *pvh; struct pv_list *pvl; if (argc) allocatable_pvs = create_pv_list(cmd->mem, vg, argc, argv, 1); else allocatable_pvs = clone_pv_list(cmd->mem, &vg->pvs); if (!allocatable_pvs) return_NULL; dm_list_iterate_safe(pvh, pvht, allocatable_pvs) { pvl = dm_list_item(pvh, struct pv_list); /* Don't allocate onto the PV we're clearing! */ if ((alloc != ALLOC_ANYWHERE) && (pvl->pv->dev == pv_dev(pv))) { dm_list_del(&pvl->list); continue; } /* Remove PV if full */ if (pvl->pv->pe_count == pvl->pv->pe_alloc_count) dm_list_del(&pvl->list); } if (dm_list_empty(allocatable_pvs)) { log_error("No extents available for allocation."); return NULL; } return allocatable_pvs; } /* * If @lv_name's a RAID SubLV, check for any PVs * on @trim_list holding it's sibling (rimage/rmeta) * and remove it from the @trim_list in order to allow * for pvmove collocation of DataLV/MetaLV pairs. */ static int _remove_sibling_pvs_from_trim_list(struct logical_volume *lv, const char *lv_name, struct dm_list *trim_list) { char *idx, *suffix; const char *sibling; char sublv_name[NAME_LEN]; struct logical_volume *sublv; struct dm_list untrim_list, *pvh1, *pvh2; struct pv_list *pvl1, *pvl2; /* Give up with success unless @lv_name _and_ valid raid segment type */ if (!lv_name || !*lv_name || !seg_is_raid(first_seg(lv)) || seg_is_raid0(first_seg(lv)) || !strcmp(lv->name, lv_name)) return 1; dm_list_init(&untrim_list); if (!dm_strncpy(sublv_name, lv_name, sizeof(sublv_name))) { log_error(INTERNAL_ERROR "LV name %s is too long.", lv_name); return 0; } if ((suffix = strstr(sublv_name, "_rimage_"))) sibling = "meta"; else if ((suffix = strstr(sublv_name, "_rmeta_"))) sibling = "image"; else { log_error("Can't find rimage or rmeta suffix."); return 0; } if (!(idx = strchr(suffix + 1, '_'))) { log_error("Can't find '_' after suffix %s.", suffix); return 0; } idx++; /* Create the siblings name (e.g. "raidlv_rmeta_N" -> "raidlv_rimage_N" */ if (dm_snprintf(suffix + 2, sizeof(sublv_name) - 2 - (suffix - sublv_name), "%s_%s", sibling, idx) < 0) { log_error("Raid sublv for name %s too long.", lv_name); return 0; } if (!(sublv = find_lv(lv->vg, sublv_name))) { log_error("Can't find sub LV %s.", sublv_name); return 0; } if (!get_pv_list_for_lv(lv->vg->cmd->mem, sublv, &untrim_list)) { log_error("Can't find PVs for sub LV %s.", sublv_name); return 0; } dm_list_iterate(pvh1, &untrim_list) { pvl1 = dm_list_item(pvh1, struct pv_list); dm_list_iterate(pvh2, trim_list) { pvl2 = dm_list_item(pvh2, struct pv_list); if (pvl1->pv == pvl2->pv) { log_debug("Removing PV %s from trim list.", pvl2->pv->dev->pvid); dm_list_del(&pvl2->list); break; } } } return 1; } /* * _trim_allocatable_pvs * @alloc_list * @trim_list * * Remove PVs in 'trim_list' from 'alloc_list'. * * Returns: 1 on success, 0 on error */ static int _trim_allocatable_pvs(struct dm_list *alloc_list, struct dm_list *trim_list, alloc_policy_t alloc) { struct dm_list *pvht, *pvh, *trim_pvh; struct pv_list *pvl, *trim_pvl; if (!alloc_list) { log_error(INTERNAL_ERROR "alloc_list is NULL."); return 0; } if (!trim_list || dm_list_empty(trim_list)) return 1; /* alloc_list stays the same */ dm_list_iterate_safe(pvh, pvht, alloc_list) { pvl = dm_list_item(pvh, struct pv_list); dm_list_iterate(trim_pvh, trim_list) { trim_pvl = dm_list_item(trim_pvh, struct pv_list); /* Don't allocate onto a trim PV */ if ((alloc != ALLOC_ANYWHERE) && (pvl->pv == trim_pvl->pv)) { dm_list_del(&pvl->list); break; /* goto next in alloc_list */ } } } return 1; } /* * Replace any LV segments on given PV with temporary mirror. * Returns list of LVs changed. */ static int _insert_pvmove_mirrors(struct cmd_context *cmd, struct logical_volume *lv_mirr, struct dm_list *source_pvl, struct logical_volume *lv, struct dm_list *lvs_changed) { struct pv_list *pvl; uint32_t prev_le_count; /* Only 1 PV may feature in source_pvl */ pvl = dm_list_item(source_pvl->n, struct pv_list); prev_le_count = lv_mirr->le_count; if (!insert_layer_for_segments_on_pv(cmd, lv, lv_mirr, PVMOVE, pvl, lvs_changed)) return_0; /* check if layer was inserted */ if (lv_mirr->le_count - prev_le_count) { lv->status |= LOCKED; log_verbose("Moving %u extents of logical volume %s.", lv_mirr->le_count - prev_le_count, display_lvname(lv)); } return 1; } /* * Is 'lv' a sub_lv of the LV by the name of 'lv_name'? * * Returns: 1 if true, 0 otherwise */ static int _sub_lv_of(struct logical_volume *lv, const char *lv_name) { struct lv_segment *seg; /* Sub-LVs only ever have one segment using them */ if (dm_list_size(&lv->segs_using_this_lv) != 1) return 0; if (!(seg = get_only_segment_using_this_lv(lv))) return_0; if (!strcmp(seg->lv->name, lv_name)) return 1; /* Continue up the tree */ return _sub_lv_of(seg->lv, lv_name); } /* Create new LV with mirror segments for the required copies */ static struct logical_volume *_set_up_pvmove_lv(struct cmd_context *cmd, struct volume_group *vg, struct dm_list *source_pvl, const char *lv_name, struct dm_list *allocatable_pvs, alloc_policy_t alloc, struct dm_list **lvs_changed, unsigned *exclusive) { struct logical_volume *lv_mirr, *lv; struct lv_segment *seg; struct lv_list *lvl; struct dm_list trim_list; uint32_t log_count = 0; int lv_found = 0; int lv_skipped = 0; int needs_exclusive = *exclusive; const struct logical_volume *holder; const char *new_lv_name; /* FIXME Cope with non-contiguous => splitting existing segments */ if (!(lv_mirr = lv_create_empty("pvmove%d", NULL, LVM_READ | LVM_WRITE, ALLOC_CONTIGUOUS, vg))) { log_error("Creation of temporary pvmove LV failed."); return NULL; } lv_mirr->status |= (PVMOVE | LOCKED); if (!(*lvs_changed = dm_pool_alloc(cmd->mem, sizeof(**lvs_changed)))) { log_error("lvs_changed list struct allocation failed."); return NULL; } dm_list_init(*lvs_changed); /* * First, * use top-level RAID and mirror LVs to build a list of PVs * that must be avoided during allocation. This is necessary * to maintain redundancy of those targets, but it is also * sub-optimal. Avoiding entire PVs in this way limits our * ability to find space for other segment types. In the * majority of cases, however, this method will suffice and * in the cases where it does not, the user can issue the * pvmove on a per-LV basis. * * FIXME: Eliminating entire PVs places too many restrictions * on allocation. */ dm_list_iterate_items(lvl, &vg->lvs) { lv = lvl->lv; if (lv == lv_mirr) continue; if (lv_name) { if (!(new_lv_name = top_level_lv_name(vg, lv_name))) return_NULL; if (strcmp(lv->name, new_lv_name)) continue; } if (!lv_is_on_pvs(lv, source_pvl)) continue; if (lv_is_converting(lv) || lv_is_merging(lv)) { log_error("Unable to pvmove when %s volume %s is present.", lv_is_converting(lv) ? "converting" : "merging", display_lvname(lv)); return NULL; } if (lv_is_writecache_cachevol(lv)) { log_error("Unable to pvmove device used for writecache."); return NULL; } if (lv_is_writecache(lv)) { struct logical_volume *lv_cachevol = first_seg(lv)->writecache; if (lv_is_on_pvs(lv_cachevol, source_pvl)) { log_error("Unable to move device used for writecache cachevol %s.", display_lvname(lv_cachevol)); return NULL; } } if (lv_is_raid(lv) && lv_raid_has_integrity(lv)) { log_error("Unable to pvmove device used for raid with integrity."); return NULL; } seg = first_seg(lv); if (!needs_exclusive) { /* Presence of exclusive LV decides whether pvmove must be also exclusive */ if (!seg_only_exclusive(seg)) { holder = lv_lock_holder(lv); if (seg_only_exclusive(first_seg(holder)) || lv_is_origin(holder) || lv_is_cow(holder)) needs_exclusive = 1; } else needs_exclusive = 1; } if (seg_is_raid(seg) || seg_is_mirrored(seg)) { dm_list_init(&trim_list); if (!get_pv_list_for_lv(vg->cmd->mem, lv, &trim_list)) return_NULL; /* * Remove any PVs holding SubLV siblings to allow * for collocation (e.g. *rmeta_0 -> *rimage_0). * * Callee checks for lv_name and valid raid segment type. * * FIXME: don't rely on namespace */ if (!_remove_sibling_pvs_from_trim_list(lv, lv_name, &trim_list)) return_NULL; if (!_trim_allocatable_pvs(allocatable_pvs, &trim_list, alloc)) return_NULL; } } /* * Second, * use bottom-level LVs (like *_mimage_*, *_mlog, *_rmeta_*, etc) * to find segments to be moved and then set up mirrors. */ dm_list_iterate_items(lvl, &vg->lvs) { lv = lvl->lv; if (lv == lv_mirr) continue; if (lv_name) { if (strcmp(lv->name, lv_name) && !_sub_lv_of(lv, lv_name)) continue; lv_found = 1; } seg = first_seg(lv); if (seg_is_cache(seg) || seg_is_cache_pool(seg) || seg_is_mirrored(seg) || seg_is_raid(seg) || seg_is_snapshot(seg) || seg_is_thin(seg) || seg_is_thin_pool(seg)) continue; /* bottom-level LVs only... */ if (!lv_is_on_pvs(lv, source_pvl)) continue; if (lv_is_locked(lv)) { lv_skipped = 1; log_print_unless_silent("Skipping locked LV %s.", display_lvname(lv)); continue; } holder = lv_lock_holder(lv); if (needs_exclusive) { /* With exclusive pvmove skip LV when: * - is active remotely * - is not active locally and cannot be activated exclusively locally * * Note: lvm2 can proceed with exclusive pvmove for 'just' locally active LVs * in the case it's NOT active anywhere else, since LOCKED LVs cannot be * later activated by user. */ if ((!lv_is_active(holder) && !activate_lv(cmd, holder))) { lv_skipped = 1; log_print_unless_silent("Skipping LV %s which is not locally exclusive%s.", display_lvname(lv), /* Report missing cmirrord cases that matterd. * With exclusive LV types cmirrord would not help. */ (*exclusive && !lv_is_origin(holder) && !seg_only_exclusive(first_seg(holder))) ? " and clustered mirror (cmirror) not detected" : ""); continue; } } else if (!activate_lv(cmd, holder)) { lv_skipped = 1; log_print_unless_silent("Skipping LV %s which cannot be activated.", display_lvname(lv)); continue; } if (!_insert_pvmove_mirrors(cmd, lv_mirr, source_pvl, lv, *lvs_changed)) return_NULL; } if (lv_name && !lv_found) { /* NOTE: Is this now an internal error? It is already checked in _pvmove_setup_single */ log_error("Logical volume %s not found.", lv_name); return NULL; } /* Is temporary mirror empty? */ if (!lv_mirr->le_count) { if (lv_skipped) log_error("All data on source PV skipped. " "It contains locked, hidden or " "non-top level LVs only."); log_error("No data to move for %s.", vg->name); return NULL; } if (!lv_add_mirrors(cmd, lv_mirr, 1, 1, 0, get_default_region_size(cmd), log_count, allocatable_pvs, alloc, (arg_is_set(cmd, atomic_ARG)) ? MIRROR_BY_SEGMENTED_LV : MIRROR_BY_SEG)) { log_error("Failed to convert pvmove LV to mirrored."); return NULL; } if (!split_parent_segments_for_layer(cmd, lv_mirr)) { log_error("Failed to split segments being moved."); return NULL; } if (needs_exclusive) *exclusive = 1; return lv_mirr; } static int _activate_lv(struct cmd_context *cmd, struct logical_volume *lv_mirr, unsigned exclusive) { int r = 0; r = activate_lv(cmd, lv_mirr); if (!r) stack; return r; } /* * Called to set up initial pvmove LV only. * (Not called after first or any other section completes.) */ static int _update_metadata(struct logical_volume *lv_mirr, struct dm_list *lvs_changed, unsigned exclusive) { struct lv_list *lvl; struct logical_volume *lv = lv_mirr; dm_list_iterate_items(lvl, lvs_changed) { lv = lvl->lv; break; } if (!lv_update_and_reload(lv)) return_0; /* Ensure mirror LV is active */ if (!_activate_lv(lv_mirr->vg->cmd, lv_mirr, exclusive)) { if (test_mode()) return 1; /* * FIXME Run --abort internally here. */ log_error("ABORTING: Temporary pvmove mirror activation failed. Run pvmove --abort."); return 0; } return 1; } static int _copy_id_components(struct cmd_context *cmd, const struct logical_volume *lv, char **vg_name, char **lv_name, union lvid *lvid) { if (!(*vg_name = dm_pool_strdup(cmd->mem, lv->vg->name)) || !(*lv_name = dm_pool_strdup(cmd->mem, lv->name))) { log_error("Failed to clone VG or LV name."); return 0; } *lvid = lv->lvid; return 1; } static int _pvmove_setup_single(struct cmd_context *cmd, struct volume_group *vg, struct physical_volume *pv, struct processing_handle *handle) { struct pvmove_params *pp = (struct pvmove_params *) handle->custom_handle; const char *lv_name = NULL; struct dm_list *source_pvl; struct dm_list *allocatable_pvs; struct dm_list *lvs_changed; struct logical_volume *lv_mirr; struct logical_volume *lv = NULL; struct lv_list *lvl; const struct logical_volume *lvh; const char *pv_name = pv_dev_name(pv); unsigned flags = PVMOVE_FIRST_TIME; unsigned exclusive; int r = ECMD_FAILED; pp->found_pv = 1; pp->setup_result = ECMD_FAILED; if (pp->lv_name_arg) { if (!(lv_name = _extract_lvname(cmd, vg->name, pp->lv_name_arg))) { log_error("Failed to find an LV name."); pp->setup_result = EINVALID_CMD_LINE; return ECMD_FAILED; } if (!validate_name(lv_name)) { log_error("Logical volume name %s is invalid.", lv_name); pp->setup_result = EINVALID_CMD_LINE; return ECMD_FAILED; } if (!(lv = find_lv(vg, lv_name))) { log_error("Logical volume %s not found.", lv_name); return ECMD_FAILED; } if (lv_is_raid(lv) && lv_raid_has_integrity(lv)) { log_error("pvmove not allowed on raid LV with integrity."); return ECMD_FAILED; } } /* * We would need to avoid any PEs used by LVs that are active (ex) on * other hosts. For LVs that are active on multiple hosts (sh), we * would need to used cluster mirrors. */ if (vg_is_shared(vg)) { if (!lv) { log_error("pvmove in a shared VG requires a named LV."); return ECMD_FAILED; } if (lv_is_lockd_sanlock_lv(lv)) { log_error("pvmove not allowed on internal sanlock LV."); return ECMD_FAILED; } if (!lockd_lv(cmd, lv, "ex", LDLV_PERSISTENT)) { log_error("pvmove in a shared VG requires exclusive lock on named LV."); return ECMD_FAILED; } } exclusive = _pvmove_is_exclusive(cmd, vg); if ((lv_mirr = find_pvmove_lv(vg, pv_dev(pv), PVMOVE))) { log_print_unless_silent("Detected pvmove in progress for %s.", pv_name); if (pp->pv_count || lv_name) log_warn("WARNING: Ignoring remaining command line arguments."); if (!(lvs_changed = lvs_using_lv(cmd, vg, lv_mirr))) { log_error("ABORTING: Failed to generate list of moving LVs."); goto out; } dm_list_iterate_items(lvl, lvs_changed) { lvh = lv_lock_holder(lvl->lv); /* Exclusive LV decides whether pvmove must be also exclusive */ if (lv_is_origin(lvh) || seg_only_exclusive(first_seg(lvh))) exclusive = 1; } /* Ensure mirror LV is active */ if (!_activate_lv(cmd, lv_mirr, exclusive)) { log_error("ABORTING: Temporary mirror activation failed."); goto out; } flags &= ~PVMOVE_FIRST_TIME; } else { /* Determine PE ranges to be moved */ if (!(source_pvl = create_pv_list(cmd->mem, vg, 1, &pp->pv_name_arg, 0))) goto_out; if (pp->alloc == ALLOC_INHERIT) pp->alloc = vg->alloc; /* Get PVs we can use for allocation */ if (!(allocatable_pvs = _get_allocatable_pvs(cmd, pp->pv_count, pp->pv_names, vg, pv, pp->alloc))) goto_out; if (!(lv_mirr = _set_up_pvmove_lv(cmd, vg, source_pvl, lv_name, allocatable_pvs, pp->alloc, &lvs_changed, &exclusive))) goto_out; } /* Lock lvs_changed and activate (with old metadata) */ if (!activate_lvs(cmd, lvs_changed, exclusive)) goto_out; /* FIXME Presence of a mirror once set PVMOVE - now remove associated logic */ /* init_pvmove(1); */ /* vg->status |= PVMOVE; */ if (!_copy_id_components(cmd, lv_mirr, &pp->id_vg_name, &pp->id_lv_name, pp->lvid)) goto out; if (flags & PVMOVE_FIRST_TIME) if (!_update_metadata(lv_mirr, lvs_changed, exclusive)) goto_out; /* LVs are all in status LOCKED */ pp->setup_result = ECMD_PROCESSED; r = ECMD_PROCESSED; out: return r; } static int _pvmove_read_single(struct cmd_context *cmd, struct volume_group *vg, struct physical_volume *pv, struct processing_handle *handle) { struct pvmove_params *pp = (struct pvmove_params *) handle->custom_handle; struct logical_volume *lv; int ret = ECMD_FAILED; pp->found_pv = 1; if (!(lv = find_pvmove_lv(vg, pv_dev(pv), PVMOVE))) { log_print_unless_silent("%s: No pvmove in progress - already finished or aborted.", pv_dev_name(pv)); ret = ECMD_PROCESSED; pp->in_progress = 0; } else if (_copy_id_components(cmd, lv, &pp->id_vg_name, &pp->id_lv_name, pp->lvid)) { ret = ECMD_PROCESSED; pp->in_progress = 1; } return ret; } static struct poll_functions _pvmove_fns = { .get_copy_name_from_lv = get_pvmove_pvname_from_lv_mirr, .poll_progress = poll_mirror_progress, .update_metadata = pvmove_update_metadata, .finish_copy = pvmove_finish, }; static struct poll_operation_id *_pvmove_create_id(struct cmd_context *cmd, const char *pv_name, const char *vg_name, const char *lv_name, const char *uuid) { struct poll_operation_id *id; if (!vg_name || !lv_name || !pv_name || !uuid) { log_error(INTERNAL_ERROR "Wrong params for _pvmove_create_id."); return NULL; } if (!(id = dm_pool_alloc(cmd->mem, sizeof(*id)))) { log_error("Poll operation ID allocation failed."); return NULL; } if (!(id->vg_name = dm_pool_strdup(cmd->mem, vg_name)) || !(id->lv_name = dm_pool_strdup(cmd->mem, lv_name)) || !(id->display_name = dm_pool_strdup(cmd->mem, pv_name)) || !(id->uuid = dm_pool_strdup(cmd->mem, uuid))) { log_error("Failed to copy one or more poll operation ID members."); dm_pool_free(cmd->mem, id); return NULL; } return id; } int pvmove_poll(struct cmd_context *cmd, const char *pv_name, const char *uuid, const char *vg_name, const char *lv_name, unsigned background) { struct poll_operation_id *id = NULL; if (uuid && !(id = _pvmove_create_id(cmd, pv_name, vg_name, lv_name, uuid))) { log_error("Failed to allocate poll identifier for pvmove."); return ECMD_FAILED; } if (test_mode()) return ECMD_PROCESSED; return poll_daemon(cmd, background, PVMOVE, &_pvmove_fns, "Moved", id); } int pvmove(struct cmd_context *cmd, int argc, char **argv) { struct pvmove_params pp = { 0 }; struct processing_handle *handle = NULL; union lvid *lvid = NULL; char *pv_name = NULL; char *colon; unsigned is_abort = arg_is_set(cmd, abort_ARG); /* dm raid1 target must be present in every case */ if (!_pvmove_target_present(cmd, 0)) { log_error("Required device-mapper target(s) not " "detected in your kernel."); return ECMD_FAILED; } if (lvmlockd_use() && !lvmpolld_use()) { /* * Don't want to spend the time making lvmlockd * work without lvmpolld. */ log_error("Enable lvmpolld when using lvmlockd."); return ECMD_FAILED; } if (lvmlockd_use() && !argc) { /* * FIXME: move process_each_vg from polldaemon up to here, * then we can remove this limitation. */ log_error("Specify pvmove args when using lvmlockd."); return ECMD_FAILED; } if (argc) { if (!(lvid = dm_pool_alloc(cmd->mem, sizeof(*lvid)))) { log_error("Failed to allocate lvid."); return ECMD_FAILED; } pp.lvid = lvid; if (!(pp.pv_name_arg = dm_pool_strdup(cmd->mem, argv[0]))) { log_error("Failed to clone PV name."); return ECMD_FAILED; } if (!(pv_name = dm_pool_strdup(cmd->mem, argv[0]))) { log_error("Failed to clone PV name."); return ECMD_FAILED; } dm_unescape_colons_and_at_signs(pv_name, &colon, NULL); /* Drop any PE lists from PV name */ if (colon) *colon = '\0'; argc--; argv++; pp.pv_count = argc; pp.pv_names = argv; if (arg_is_set(cmd, name_ARG)) { if (!(pp.lv_name_arg = dm_pool_strdup(cmd->mem, arg_value(cmd, name_ARG)))) { log_error("Failed to clone LV name."); return ECMD_FAILED; } } pp.alloc = (alloc_policy_t) arg_uint_value(cmd, alloc_ARG, ALLOC_INHERIT); pp.in_progress = 1; /* Normal pvmove setup requires ex lock from lvmlockd. */ if (is_abort) cmd->lockd_vg_default_sh = 1; if (!(handle = init_processing_handle(cmd, NULL))) { log_error("Failed to initialize processing handle."); return ECMD_FAILED; } handle->custom_handle = &pp; process_each_pv(cmd, 1, &pv_name, NULL, 0, is_abort ? 0 : READ_FOR_UPDATE, handle, is_abort ? &_pvmove_read_single : &_pvmove_setup_single); destroy_processing_handle(cmd, handle); if (!is_abort) { if (!pp.found_pv) { stack; return EINVALID_CMD_LINE; } if (pp.setup_result != ECMD_PROCESSED) { stack; return pp.setup_result; } } else { if (!pp.found_pv) return_ECMD_FAILED; if (!pp.in_progress) return ECMD_PROCESSED; } /* * The command may sit and report progress for some time, * and we do not want or need the global lock held during * that time. */ lock_global(cmd, "un"); } return pvmove_poll(cmd, pv_name, lvid ? lvid->s : NULL, pp.id_vg_name, pp.id_lv_name, arg_is_set(cmd, background_ARG)); }