/* * Copyright (C) 2001-2004 Sistina Software, Inc. All rights reserved. * Copyright (C) 2004-2018 Red Hat, Inc. All rights reserved. * * This file is part of LVM2. * * This copyrighted material is made available to anyone wishing to use, * modify, copy, or redistribute it subject to the terms and conditions * of the GNU Lesser General Public License v.2.1. * * You should have received a copy of the GNU Lesser General Public License * along with this program; if not, write to the Free Software Foundation, * Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include "lib/misc/lib.h" #include "lib/metadata/metadata.h" #include "lib/activate/activate.h" #include "lib/mm/memlock.h" #include "lib/display/display.h" #include "fs.h" #include "lib/misc/lvm-exec.h" #include "lib/misc/lvm-file.h" #include "lib/misc/lvm-string.h" #include "lib/commands/toolcontext.h" #include "dev_manager.h" #include "lib/datastruct/str_list.h" #include "lib/config/config.h" #include "lib/metadata/segtype.h" #include "lib/misc/sharedlib.h" #include "lib/metadata/metadata.h" #include "lib/misc/lvm-signal.h" #include #include #include #define _skip(fmt, args...) log_very_verbose("Skipping: " fmt , ## args) int list_segment_modules(struct dm_pool *mem, const struct lv_segment *seg, struct dm_list *modules) { unsigned int s; struct lv_segment *seg2, *snap_seg; struct dm_list *snh; if (seg->segtype->ops->modules_needed && !seg->segtype->ops->modules_needed(mem, seg, modules)) { log_error("module string allocation failed"); return 0; } if (lv_is_origin(seg->lv)) dm_list_iterate(snh, &seg->lv->snapshot_segs) if (!list_lv_modules(mem, dm_list_struct_base(snh, struct lv_segment, origin_list)->cow, modules)) return_0; if (lv_is_cow(seg->lv)) { snap_seg = find_snapshot(seg->lv); if (snap_seg->segtype->ops->modules_needed && !snap_seg->segtype->ops->modules_needed(mem, snap_seg, modules)) { log_error("snap_seg module string allocation failed"); return 0; } } for (s = 0; s < seg->area_count; s++) { switch (seg_type(seg, s)) { case AREA_LV: seg2 = find_seg_by_le(seg_lv(seg, s), seg_le(seg, s)); if (seg2 && !list_segment_modules(mem, seg2, modules)) return_0; break; case AREA_PV: case AREA_UNASSIGNED: ; } } return 1; } int list_lv_modules(struct dm_pool *mem, const struct logical_volume *lv, struct dm_list *modules) { struct lv_segment *seg; dm_list_iterate_items(seg, &lv->segments) if (!list_segment_modules(mem, seg, modules)) return_0; return 1; } static int _lv_passes_volumes_filter(struct cmd_context *cmd, const struct logical_volume *lv, const struct dm_config_node *cn, const int cfg_id) { const struct dm_config_value *cv; const char *str; static char config_path[PATH_MAX]; size_t len = strlen(lv->vg->name); config_def_get_path(config_path, sizeof(config_path), cfg_id); log_verbose("%s configuration setting defined: " "Checking the list to match %s.", config_path, display_lvname(lv)); for (cv = cn->v; cv; cv = cv->next) { if (cv->type == DM_CFG_EMPTY_ARRAY) goto out; if (cv->type != DM_CFG_STRING) { log_print_unless_silent("Ignoring invalid string in config file %s.", config_path); continue; } str = cv->v.str; if (!*str) { log_print_unless_silent("Ignoring empty string in config file %s.", config_path); continue; } /* Tag? */ if (*str == '@') { str++; if (!*str) { log_print_unless_silent("Ignoring empty tag in config file %s", config_path); continue; } /* If any host tag matches any LV or VG tag, activate */ if (!strcmp(str, "*")) { if (str_list_match_list(&cmd->tags, &lv->tags, NULL) || str_list_match_list(&cmd->tags, &lv->vg->tags, NULL)) return 1; continue; } /* If supplied tag matches LV or VG tag, activate */ if (str_list_match_item(&lv->tags, str) || str_list_match_item(&lv->vg->tags, str)) return 1; continue; } /* If supplied name is vgname[/lvname] */ if ((strncmp(str, lv->vg->name, len) == 0) && (!str[len] || ((str[len] == '/') && !strcmp(str + len + 1, lv->name)))) return 1; } out: log_verbose("No item supplied in %s configuration setting matches %s.", config_path, display_lvname(lv)); return 0; } int lv_passes_auto_activation_filter(struct cmd_context *cmd, struct logical_volume *lv) { const struct dm_config_node *cn; if (!(cn = find_config_tree_array(cmd, activation_auto_activation_volume_list_CFG, NULL))) { log_verbose("activation/auto_activation_volume_list configuration setting " "not defined: All logical volumes will be auto-activated."); return 1; } return _lv_passes_volumes_filter(cmd, lv, cn, activation_auto_activation_volume_list_CFG); } static int _passes_readonly_filter(struct cmd_context *cmd, const struct logical_volume *lv) { const struct dm_config_node *cn; if (!(cn = find_config_tree_array(cmd, activation_read_only_volume_list_CFG, NULL))) return 0; return _lv_passes_volumes_filter(cmd, lv, cn, activation_read_only_volume_list_CFG); } int lv_passes_readonly_filter(const struct logical_volume *lv) { return _passes_readonly_filter(lv->vg->cmd, lv); } #ifndef DEVMAPPER_SUPPORT void set_activation(int act, int silent) { static int warned = 0; if (warned || !act) return; log_warn("WARNING: Compiled without libdevmapper support. " "Can't enable activation."); warned = 1; } int activation(void) { return 0; } int library_version(char *version, size_t size) { return 0; } int driver_version(char *version, size_t size) { return 0; } int target_version(const char *target_name, uint32_t *maj, uint32_t *min, uint32_t *patchlevel) { return 0; } int target_present(struct cmd_context *cmd, const char *target_name, int use_modprobe) { return 0; } int lvm_dm_prefix_check(int major, int minor, const char *prefix) { return 0; } int lv_info(struct cmd_context *cmd, const struct logical_volume *lv, int use_layer, struct lvinfo *info, int with_open_count, int with_read_ahead) { return 0; } int lv_info_with_seg_status(struct cmd_context *cmd, const struct lv_segment *lv_seg, struct lv_with_info_and_seg_status *status, int with_open_count, int with_read_ahead) { return 0; } int lv_cache_status(const struct logical_volume *cache_lv, struct lv_status_cache **status) { return 0; } int lv_check_not_in_use(const struct logical_volume *lv, int error_if_used) { return 0; } int lv_snapshot_percent(const struct logical_volume *lv, dm_percent_t *percent) { return 0; } int lv_mirror_percent(struct cmd_context *cmd, const struct logical_volume *lv, int wait, dm_percent_t *percent, uint32_t *event_nr) { return 0; } int lv_raid_percent(const struct logical_volume *lv, dm_percent_t *percent) { return 0; } int lv_raid_data_offset(const struct logical_volume *lv, uint64_t *data_offset) { return 0; } int lv_raid_dev_health(const struct logical_volume *lv, char **dev_health) { return 0; } int lv_raid_dev_count(const struct logical_volume *lv, uint32_t *dev_cnt) { return 0; } int lv_raid_mismatch_count(const struct logical_volume *lv, uint64_t *cnt) { return 0; } int lv_raid_sync_action(const struct logical_volume *lv, char **sync_action) { return 0; } int lv_raid_message(const struct logical_volume *lv, const char *msg) { return 0; } int lv_raid_status(const struct logical_volume *lv, struct lv_status_raid **status) { return 0; } int lv_writecache_message(const struct logical_volume *lv, const char *msg) { return 0; } int lv_thin_pool_status(const struct logical_volume *lv, int flush, struct lv_status_thin_pool **thin_pool_status) { return 0; } int lv_thin_status(const struct logical_volume *lv, int flush, struct lv_status_thin **thin_status) { return 0; } int lv_thin_device_id(const struct logical_volume *lv, uint32_t *device_id) { return 0; } int lv_vdo_pool_status(const struct logical_volume *lv, int flush, struct lv_status_vdo **vdo_status) { return 0; } int lv_vdo_pool_percent(const struct logical_volume *lv, dm_percent_t *percent) { return 0; } int lv_vdo_pool_size_config(const struct logical_volume *lv, struct vdo_pool_size_config *cfg) { return 0; } int lvs_in_vg_activated(const struct volume_group *vg) { return 0; } int lvs_in_vg_opened(const struct volume_group *vg) { return 0; } int lv_suspend_if_active(struct cmd_context *cmd, const char *lvid_s, unsigned origin_only, unsigned exclusive, const struct logical_volume *lv, const struct logical_volume *lv_pre) { return 1; } int lv_resume(struct cmd_context *cmd, const char *lvid_s, unsigned origin_only, const struct logical_volume *lv) { return 1; } int lv_resume_if_active(struct cmd_context *cmd, const char *lvid_s, unsigned origin_only, unsigned exclusive, unsigned revert, const struct logical_volume *lv) { return 1; } int lv_deactivate(struct cmd_context *cmd, const char *lvid_s, const struct logical_volume *lv) { return 1; } int lv_activation_filter(struct cmd_context *cmd, const char *lvid_s, int *activate_lv, const struct logical_volume *lv) { return 1; } int lv_activate(struct cmd_context *cmd, const char *lvid_s, int exclusive, int noscan, int temporary, const struct logical_volume *lv) { return 1; } int lv_activate_with_filter(struct cmd_context *cmd, const char *lvid_s, int exclusive, int noscan, int temporary, const struct logical_volume *lv) { return 1; } int lv_mknodes(struct cmd_context *cmd, const struct logical_volume *lv) { return 1; } int lv_deactivate_any_missing_subdevs(const struct logical_volume *lv) { return 1; } int pv_uses_vg(struct physical_volume *pv, struct volume_group *vg) { return 0; } void activation_release(void) { } void activation_exit(void) { } int raid4_is_supported(struct cmd_context *cmd, const struct segment_type *segtype) { return 1; } int lv_is_active(const struct logical_volume *lv) { return 0; } int lv_check_transient(struct logical_volume *lv) { return 1; } int monitor_dev_for_events(struct cmd_context *cmd, const struct logical_volume *lv, const struct lv_activate_opts *laopts, int monitor) { return 1; } /* fs.c */ void fs_unlock(void) { } /* dev_manager.c */ #include "lib/activate/targets.h" int add_areas_line(struct dev_manager *dm, struct lv_segment *seg, struct dm_tree_node *node, uint32_t start_area, uint32_t areas) { return 0; } int device_is_usable(struct cmd_context *cmd, struct device *dev, struct dev_usable_check_params check, int *is_lv) { return 0; } int lv_has_target_type(struct dm_pool *mem, const struct logical_volume *lv, const char *layer, const char *target_type) { return 0; } #else /* DEVMAPPER_SUPPORT */ static int _activation = 1; void set_activation(int act, int silent) { if (act == _activation) return; _activation = act; if (_activation) log_verbose("Activation enabled. Device-mapper kernel " "driver will be used."); else if (!silent) log_warn("WARNING: Activation disabled. No device-mapper " "interaction will be attempted."); else log_verbose("Activation disabled. No device-mapper " "interaction will be attempted."); } int activation(void) { return _activation; } static int _passes_activation_filter(struct cmd_context *cmd, const struct logical_volume *lv) { const struct dm_config_node *cn; if (!(cn = find_config_tree_array(cmd, activation_volume_list_CFG, NULL))) { log_verbose("activation/volume_list configuration setting " "not defined: Checking only host tags for %s.", display_lvname(lv)); /* If no host tags defined, activate */ if (dm_list_empty(&cmd->tags)) return 1; /* If any host tag matches any LV or VG tag, activate */ if (str_list_match_list(&cmd->tags, &lv->tags, NULL) || str_list_match_list(&cmd->tags, &lv->vg->tags, NULL)) return 1; log_verbose("No host tag matches %s", display_lvname(lv)); /* Don't activate */ return 0; } return _lv_passes_volumes_filter(cmd, lv, cn, activation_volume_list_CFG); } int library_version(char *version, size_t size) { if (!activation()) return 0; return dm_get_library_version(version, size); } int driver_version(char *version, size_t size) { static char _vsn[80] = { 0 }; if (!activation()) return 0; log_very_verbose("Getting driver version"); if (!_vsn[0] && !dm_driver_version(_vsn, sizeof(_vsn))) return_0; dm_strncpy(version, _vsn, size); return 1; } int target_version(const char *target_name, uint32_t *maj, uint32_t *min, uint32_t *patchlevel) { int r = 0; struct dm_task *dmt; struct dm_versions *target, *last_target; log_very_verbose("Getting target version for %s", target_name); if (!(dmt = dm_task_create(DM_DEVICE_LIST_VERSIONS))) return_0; if (activation_checks() && !dm_task_enable_checks(dmt)) goto_out; if (!dm_task_run(dmt)) { log_debug_activation("Failed to get %s target version", target_name); /* Assume this was because LIST_VERSIONS isn't supported */ *maj = 0; *min = 0; *patchlevel = 0; r = 1; goto out; } target = dm_task_get_versions(dmt); do { last_target = target; if (!strcmp(target_name, target->name)) { r = 1; *maj = target->version[0]; *min = target->version[1]; *patchlevel = target->version[2]; goto out; } target = (struct dm_versions *)((char *) target + target->next); } while (last_target != target); out: if (r) log_very_verbose("Found %s target " "v%" PRIu32 ".%" PRIu32 ".%" PRIu32 ".", target_name, *maj, *min, *patchlevel); dm_task_destroy(dmt); return r; } int lvm_dm_prefix_check(int major, int minor, const char *prefix) { return dev_manager_check_prefix_dm_major_minor(major, minor, prefix); } /* Search modules.builtin file for built-in kernel module */ static int _check_modules_builtin(struct cmd_context *cmd, const char *target) { FILE *fp; char *line = NULL; size_t len; int r = 0; char path[PATH_MAX]; if (dm_snprintf(path, sizeof(path), "%s/%s/modules.builtin", MODULES_PATH, cmd->kernel_vsn) < 0) { log_debug_activation("Modules path %s/%s/modules.builtin is too long.", MODULES_PATH, cmd->kernel_vsn); return 0; } if (!(fp = fopen(path, "r"))) { if (errno != ENOENT) log_sys_debug("fopen", path); return 0; } while (getline(&line, &len, fp) > 0) if (strstr(line, target)) { log_debug_activation("Found %s as built-in kernel module.", target); r = 1; break; } free(line); if (fclose(fp)) log_sys_debug("fclose", path); return r; } int module_present(struct cmd_context *cmd, const char *target_name) { int ret = 0; #ifdef MODPROBE_CMD char module[128]; const char *argv[] = { MODPROBE_CMD, module, NULL }; unsigned maj, min; #endif struct stat st; char path[PATH_MAX]; int i = dm_snprintf(path, sizeof(path), "%smodule/dm_%s", dm_sysfs_dir(), target_name); if (i > 0) { while ((i > 0) && path[--i] != '/') /* stop on dm_ */ if (path[i] == '-') path[i] = '_'; /* replace '-' with '_' */ if ((lstat(path, &st) == 0) && S_ISDIR(st.st_mode)) { log_debug_activation("Module directory %s exists.", path); return 1; } if (path[i] == '/' && _check_modules_builtin(cmd, path + i + 1)) return 1; } #ifdef MODPROBE_CMD if ((strcmp(target_name, TARGET_NAME_VDO) == 0) && (sscanf(cmd->kernel_vsn, "%u.%u", &maj, &min) == 2) && ((maj < 6) || ((maj == 6) && (min < 9)))) argv[1] = MODULE_NAME_VDO; /* Kernels < 6.9 -> "kvdo" without dm- prefix */ else if (dm_snprintf(module, sizeof(module), "dm-%s", target_name) < 0) { log_error("module_present module name too long: %s", target_name); return 0; } ret = exec_cmd(cmd, argv, NULL, 0); #endif return ret; } int target_present_version(struct cmd_context *cmd, const char *target_name, int use_modprobe, uint32_t *maj, uint32_t *min, uint32_t *patchlevel) { if (!activation()) { log_error(INTERNAL_ERROR "Target present version called when activation is disabled."); return 0; } #ifdef MODPROBE_CMD if (use_modprobe) { if (target_version(target_name, maj, min, patchlevel)) return 1; if (!module_present(cmd, target_name)) return_0; } #endif return target_version(target_name, maj, min, patchlevel); } int target_present(struct cmd_context *cmd, const char *target_name, int use_modprobe) { uint32_t maj, min, patchlevel; return target_present_version(cmd, target_name, use_modprobe, &maj, &min, &patchlevel); } int get_device_list(const struct volume_group *vg, struct dm_list **devs, unsigned *devs_features) { if (!activation()) return 0; return dev_manager_get_device_list(NULL, devs, devs_features); } /* * When '*info' is NULL, returns 1 only when LV is active. * When '*info' != NULL, returns 1 when info structure is populated. */ static int _lv_info(struct cmd_context *cmd, const struct logical_volume *lv, int use_layer, struct lvinfo *info, const struct lv_segment *seg, struct lv_seg_status *seg_status, int with_open_count, int with_read_ahead, int with_name_check) { struct dm_info dminfo; /* * If open_count info is requested and we have to be sure our own udev * transactions are finished * For non-clustered locking type we are only interested for non-delete operation * in progress - as only those could lead to opened files */ if (with_open_count) { if (fs_has_non_delete_ops()) fs_unlock(); /* For non clustered - wait if there are non-delete ops */ } /* New thin-pool has no layer, but -tpool suffix needs to be queried */ if (!use_layer && lv_is_new_thin_pool(lv)) { /* Check if there isn't existing old thin pool mapping in the table */ if (!dev_manager_info(cmd, lv, NULL, 0, 0, 0, &dminfo, NULL, NULL)) return_0; if (!dminfo.exists) use_layer = 1; } if (seg_status) { /* TODO: for now it's mess with seg_status */ seg_status->seg = seg; } if (!dev_manager_info(cmd, lv, (use_layer) ? lv_layer(lv) : NULL, with_open_count, with_read_ahead, with_name_check, &dminfo, (info) ? &info->read_ahead : NULL, seg_status)) return_0; if (!info) return dminfo.exists; info->exists = dminfo.exists; info->suspended = dminfo.suspended; info->open_count = dminfo.open_count; info->major = dminfo.major; info->minor = dminfo.minor; info->read_only = dminfo.read_only; info->live_table = dminfo.live_table; info->inactive_table = dminfo.inactive_table; return 1; } /* * Returns 1 if info structure populated, else 0 on failure. * When lvinfo* is NULL, it returns 1 if the device is locally active, 0 otherwise. */ int lv_info(struct cmd_context *cmd, const struct logical_volume *lv, int use_layer, struct lvinfo *info, int with_open_count, int with_read_ahead) { if (!activation()) return 0; return _lv_info(cmd, lv, use_layer, info, NULL, NULL, with_open_count, with_read_ahead, 0); } int lv_info_with_name_check(struct cmd_context *cmd, const struct logical_volume *lv, int use_layer, struct lvinfo *info) { if (!activation()) return 0; return _lv_info(cmd, lv, use_layer, info, NULL, NULL, 0, 0, 1); } /* * Returns 1 if lv_with_info_and_seg_status info structure populated, * else 0 on failure or if device not active locally. * * When seg_status parsing had troubles it will set type to SEG_STATUS_UNKNOWN. * * Using usually one ioctl to obtain info and status. * More complex segment do collect info from one device, * but status from another device. * * TODO: further improve with more statuses (i.e. snapshot's origin/merge) */ int lv_info_with_seg_status(struct cmd_context *cmd, const struct lv_segment *lv_seg, struct lv_with_info_and_seg_status *status, int with_open_count, int with_read_ahead) { const struct logical_volume *olv, *lv = status->lv = lv_seg->lv; if (!activation()) return 0; if (lv_is_used_cache_pool(lv)) { /* INFO is not set as cache-pool cannot be active. * STATUS is collected from cache LV */ if (!(lv_seg = get_only_segment_using_this_lv(lv))) return_0; (void) _lv_info(cmd, lv_seg->lv, 1, NULL, lv_seg, &status->seg_status, 0, 0, 0); return 1; } if (lv_is_thin_pool(lv)) { /* Always collect status for '-tpool' */ if (_lv_info(cmd, lv, 1, &status->info, lv_seg, &status->seg_status, 0, 0, 0) && (status->seg_status.type == SEG_STATUS_THIN_POOL)) { /* There is -tpool device, but query 'active' state of 'fake' thin-pool */ if (!_lv_info(cmd, lv, 0, NULL, NULL, NULL, 0, 0, 0) && !status->seg_status.thin_pool->needs_check) status->info.exists = 0; /* So pool LV is not active */ } return 1; } if (lv_is_external_origin(lv)) { if (!_lv_info(cmd, lv, 0, &status->info, NULL, NULL, with_open_count, with_read_ahead, 0)) return_0; (void) _lv_info(cmd, lv, 1, NULL, lv_seg, &status->seg_status, 0, 0, 0); return 1; } if (lv_is_origin(lv)) { /* Query segment status for 'layered' (-real) device most of the time, * only for merging snapshot, query its progress. * TODO: single LV may need couple status to be exposed at once.... * but this needs more logical background */ /* Show INFO for actual origin and grab status for merging origin */ if (!_lv_info(cmd, lv, 0, &status->info, lv_seg, lv_is_merging_origin(lv) ? &status->seg_status : NULL, with_open_count, with_read_ahead, 0)) return_0; if (status->info.exists && (status->seg_status.type != SEG_STATUS_SNAPSHOT)) /* Not merging */ /* Grab STATUS from layered -real */ (void) _lv_info(cmd, lv, 1, NULL, lv_seg, &status->seg_status, 0, 0, 0); return 1; } if (lv_is_cow(lv)) { if (lv_is_merging_cow(lv)) { olv = origin_from_cow(lv); if (!_lv_info(cmd, olv, 0, &status->info, first_seg(olv), &status->seg_status, with_open_count, with_read_ahead, 0)) return_0; if (status->seg_status.type == SEG_STATUS_SNAPSHOT || (lv_is_thin_volume(olv) && (status->seg_status.type == SEG_STATUS_THIN))) { log_debug_activation("Snapshot merge is in progress, querying status of %s instead.", display_lvname(lv)); /* * When merge is in progress, query merging origin LV instead. * COW volume is already mapped as error target in this case. */ return 1; } /* Merge not yet started, still a snapshot... */ } /* Hadle fictional lvm2 snapshot and query snapshotX volume */ lv_seg = find_snapshot(lv); } if (lv_is_vdo(lv)) { if (!_lv_info(cmd, lv, 0, &status->info, NULL, NULL, with_open_count, with_read_ahead, 0)) return_0; if (status->info.exists) { /* Status for VDO pool */ (void) _lv_info(cmd, seg_lv(lv_seg, 0), 1, NULL, first_seg(seg_lv(lv_seg, 0)), &status->seg_status, 0, 0, 0); /* Use VDO pool segtype result for VDO segtype */ status->seg_status.seg = lv_seg; } return 1; } if (lv_is_vdo_pool(lv)) { /* Always collect status for '-vpool' */ if (_lv_info(cmd, lv, 1, &status->info, lv_seg, &status->seg_status, 0, 0, 0) && (status->seg_status.type == SEG_STATUS_VDO_POOL)) { /* There is -tpool device, but query 'active' state of 'fake' vdo-pool */ if (!_lv_info(cmd, lv, 0, NULL, NULL, NULL, 0, 0, 0)) status->info.exists = 0; /* So VDO pool LV is not active */ } return 1; } return _lv_info(cmd, lv, 0, &status->info, lv_seg, &status->seg_status, with_open_count, with_read_ahead, 0); } #define OPEN_COUNT_CHECK_RETRIES 25 #define OPEN_COUNT_CHECK_USLEEP_DELAY 200000 /* Only report error if error_if_used is set */ /* Returns 0 if in use, 1 if it is unused, 2 when it is not present in table */ int lv_check_not_in_use(const struct logical_volume *lv, int error_if_used) { struct lvinfo info; unsigned int open_count_check_retries; if (!lv_info(lv->vg->cmd, lv, 0, &info, 1, 0) || !info.exists) return 2; else if (!info.open_count) return 1; /* If sysfs is not used, use open_count information only. */ if (dm_sysfs_dir()) { if (dm_device_has_holders(info.major, info.minor)) { if (error_if_used) log_error("Logical volume %s is used by another device.", display_lvname(lv)); else log_debug_activation("Logical volume %s is used by another device.", display_lvname(lv)); return 0; } if (dm_device_has_mounted_fs(info.major, info.minor)) { if (error_if_used) log_error("Logical volume %s contains a filesystem in use.", display_lvname(lv)); else log_debug_activation("Logical volume %s contains a filesystem in use.", display_lvname(lv)); return 0; } } open_count_check_retries = retry_deactivation() ? OPEN_COUNT_CHECK_RETRIES : 1; while (open_count_check_retries--) { if (interruptible_usleep(OPEN_COUNT_CHECK_USLEEP_DELAY)) break; /* interrupted */ log_debug_activation("Retrying open_count check for %s.", display_lvname(lv)); if (!lv_info(lv->vg->cmd, lv, 0, &info, 1, 0) || !info.exists) { stack; /* device dissappeared? */ return 1; } else if (!info.open_count) return 1; } if (error_if_used) log_error("Logical volume %s in use.", display_lvname(lv)); else log_debug_activation("Logical volume %s in use.", display_lvname(lv)); return 0; } /* * Returns 1 if percent set, else 0 on failure. */ int lv_check_transient(struct logical_volume *lv) { int r; struct dev_manager *dm; if (!activation()) return 0; log_debug_activation("Checking transient status for LV %s.", display_lvname(lv)); if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1))) return_0; if (!(r = dev_manager_transient(dm, lv))) stack; dev_manager_destroy(dm); return r; } /* * Returns 1 if percent set, else 0 on failure. */ int lv_snapshot_percent(const struct logical_volume *lv, dm_percent_t *percent) { int r; struct dev_manager *dm; if (!lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0)) return 0; log_debug_activation("Checking snapshot percent for LV %s.", display_lvname(lv)); if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1))) return_0; if (!(r = dev_manager_snapshot_percent(dm, lv, percent))) stack; dev_manager_destroy(dm); return r; } /* FIXME Merge with snapshot_percent */ int lv_mirror_percent(struct cmd_context *cmd, const struct logical_volume *lv, int wait, dm_percent_t *percent, uint32_t *event_nr) { int r; struct dev_manager *dm; /* If mirrored LV is temporarily shrinked to 1 area (= linear), * it should be considered in-sync. */ if (dm_list_size(&lv->segments) == 1 && first_seg(lv)->area_count == 1) { *percent = DM_PERCENT_100; return 1; } if (!lv_info(cmd, lv, 0, NULL, 0, 0)) return 0; log_debug_activation("Checking mirror percent for LV %s.", display_lvname(lv)); if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1))) return_0; if (!(r = dev_manager_mirror_percent(dm, lv, wait, percent, event_nr))) stack; dev_manager_destroy(dm); return r; } int lv_raid_percent(const struct logical_volume *lv, dm_percent_t *percent) { return lv_mirror_percent(lv->vg->cmd, lv, 0, percent, NULL); } int lv_raid_data_offset(const struct logical_volume *lv, uint64_t *data_offset) { struct lv_status_raid *raid_status; if (!lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0)) return 0; log_debug_activation("Checking raid data offset and dev sectors for LV %s/%s", lv->vg->name, lv->name); if (!lv_raid_status(lv, &raid_status)) return_0; *data_offset = raid_status->raid->data_offset; dm_pool_destroy(raid_status->mem); return 1; } int lv_raid_dev_health(const struct logical_volume *lv, char **dev_health) { int r = 1; struct lv_status_raid *raid_status; *dev_health = NULL; if (!lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0)) return 0; log_debug_activation("Checking raid device health for LV %s.", display_lvname(lv)); if (!lv_raid_status(lv, &raid_status)) return_0; if (!(*dev_health = dm_pool_strdup(lv->vg->cmd->mem, raid_status->raid->dev_health))) { stack; r = 0; } dm_pool_destroy(raid_status->mem); return r; } int lv_raid_dev_count(const struct logical_volume *lv, uint32_t *dev_cnt) { struct lv_status_raid *raid_status; *dev_cnt = 0; if (!lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0)) return 0; log_debug_activation("Checking raid device count for LV %s/%s", lv->vg->name, lv->name); if (!lv_raid_status(lv, &raid_status)) return_0; *dev_cnt = raid_status->raid->dev_count; dm_pool_destroy(raid_status->mem); return 1; } int lv_raid_mismatch_count(const struct logical_volume *lv, uint64_t *cnt) { struct lv_status_raid *raid_status; *cnt = 0; if (!lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0)) return 0; log_debug_activation("Checking raid mismatch count for LV %s.", display_lvname(lv)); if (!lv_raid_status(lv, &raid_status)) return_0; *cnt = raid_status->raid->mismatch_count; dm_pool_destroy(raid_status->mem); return 1; } int lv_raid_sync_action(const struct logical_volume *lv, char **sync_action) { struct lv_status_raid *raid_status; int r = 1; *sync_action = NULL; if (!lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0)) return 0; log_debug_activation("Checking raid sync_action for LV %s.", display_lvname(lv)); if (!lv_raid_status(lv, &raid_status)) return_0; /* status->sync_action can be NULL if dm-raid version < 1.5.0 */ if (!raid_status->raid->sync_action || !(*sync_action = dm_pool_strdup(lv->vg->cmd->mem, raid_status->raid->sync_action))) { stack; r = 0; } dm_pool_destroy(raid_status->mem); return r; } int lv_raid_message(const struct logical_volume *lv, const char *msg) { struct lv_status_raid *raid_status; struct dev_manager *dm = NULL; int r = 0; if (!seg_is_raid(first_seg(lv))) { /* * Make it easier for user to know what to do when * they are using thinpool. */ if (lv_is_thin_pool(lv) && (lv_is_raid(seg_lv(first_seg(lv), 0)) || lv_is_raid(first_seg(lv)->metadata_lv))) { log_error("Thin pool data or metadata volume " "must be specified. (E.g. \"%s_tdata\")", display_lvname(lv)); return 0; } log_error("%s must be a RAID logical volume to perform this action.", display_lvname(lv)); return 0; } if (!lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0)) { log_error("Unable to send message to an inactive logical volume."); return 0; } if (!lv_raid_status(lv, &raid_status)) return_0; if (!raid_status->raid->sync_action) { log_error("Kernel driver does not support this action: %s", msg); goto out; } /* * Note that 'dev_manager_raid_message' allows us to pass down any * currently valid message. However, this function restricts the * number of user available combinations to a minimum. Specifically, * "idle" -> "check" * "idle" -> "repair" * (The state automatically switches to "idle" when a sync process is * complete.) */ if (strcmp(msg, "check") && strcmp(msg, "repair")) { /* * MD allows "frozen" to operate in a toggling fashion. * We could allow this if we like... */ log_error("\"%s\" is not a supported sync operation.", msg); goto out; } if (strcmp(raid_status->raid->sync_action, "idle")) { log_error("%s state is currently \"%s\". Unable to switch to \"%s\".", display_lvname(lv), raid_status->raid->sync_action, msg); goto out; } if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1))) return_0; r = dev_manager_raid_message(dm, lv, msg); out: if (dm) dev_manager_destroy(dm); dm_pool_destroy(raid_status->mem); return r; } int lv_raid_status(const struct logical_volume *lv, struct lv_status_raid **status) { struct dev_manager *dm; int exists; if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1))) return_0; if (!dev_manager_raid_status(dm, lv, status, &exists)) { dev_manager_destroy(dm); if (exists) stack; return 0; } /* User has to call dm_pool_destroy(status->mem)! */ return 1; } int lv_writecache_message(const struct logical_volume *lv, const char *msg) { int r = 0; struct dev_manager *dm; if (!lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0)) { log_error("Unable to send message to an inactive logical volume."); return 0; } if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1))) return_0; r = dev_manager_writecache_message(dm, lv, msg); dev_manager_destroy(dm); return r; } /* * Return dm_status_cache for cache volume, accept also cache pool * * As there are too many variable for cache volumes, and it hard * to make good API - so let's obtain dm_status_cache and return * all info we have - user just has to release struct after its use. */ int lv_cache_status(const struct logical_volume *cache_lv, struct lv_status_cache **status) { struct dev_manager *dm; struct lv_segment *cache_seg; int exists; if (lv_is_cache_pool(cache_lv)) { if (dm_list_empty(&cache_lv->segs_using_this_lv) || !(cache_seg = get_only_segment_using_this_lv(cache_lv))) { log_error(INTERNAL_ERROR "Cannot check status for unused cache pool %s.", display_lvname(cache_lv)); return 0; } cache_lv = cache_seg->lv; } if (lv_is_pending_delete(cache_lv)) { log_error("Cannot check status for deleted cache volume %s.", display_lvname(cache_lv)); return 0; } if (!(dm = dev_manager_create(cache_lv->vg->cmd, cache_lv->vg->name, 1))) return_0; if (!dev_manager_cache_status(dm, cache_lv, status, &exists)) { dev_manager_destroy(dm); if (exists) stack; return 0; } /* User has to call dm_pool_destroy(status->mem)! */ return 1; } int lv_thin_pool_status(const struct logical_volume *lv, int flush, struct lv_status_thin_pool **thin_pool_status) { struct dev_manager *dm; int exists; if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1))) return_0; if (!dev_manager_thin_pool_status(dm, lv, flush, thin_pool_status, &exists)) { dev_manager_destroy(dm); if (exists) stack; return 0; } /* User has to call dm_pool_destroy(thin_pool_status->mem)! */ return 1; } int lv_thin_status(const struct logical_volume *lv, int flush, struct lv_status_thin **thin_status) { struct dev_manager *dm; int exists; if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1))) return_0; if (!dev_manager_thin_status(dm, lv, flush, thin_status, &exists)) { dev_manager_destroy(dm); if (exists) stack; return 0; } /* User has to call dm_pool_destroy(thin_status->mem)! */ return 1; } int lv_thin_device_id(const struct logical_volume *lv, uint32_t *device_id) { struct dev_manager *dm; int exists; int r; if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1))) return_0; if (!(r = dev_manager_thin_device_id(dm, lv, device_id, &exists))) if (exists) stack; dev_manager_destroy(dm); return r; } /* * lv_vdo_pool_status obtains status information about VDO pool * * If the 'params' string has been already retrieved, use it. * If the mempool already exists, use it. * */ int lv_vdo_pool_status(const struct logical_volume *lv, int flush, struct lv_status_vdo **vdo_status) { struct dev_manager *dm; int exists; if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, !lv_is_pvmove(lv)))) return_0; if (!dev_manager_vdo_pool_status(dm, lv, flush, vdo_status, &exists)) { dev_manager_destroy(dm); if (exists) stack; return 0; } /* User has to call dm_pool_destroy(vdo_status->mem) */ return 1; } int lv_vdo_pool_percent(const struct logical_volume *lv, dm_percent_t *percent) { struct lv_status_vdo *vdo_status; if (!lv_vdo_pool_status(lv, 0, &vdo_status)) return_0; *percent = vdo_status->usage; dm_pool_destroy(vdo_status->mem); return 1; } /* * lv_vdo_pool_size_config obtains size configuration from active VDO table line * * If the 'params' string has been already retrieved, use it. * If the mempool already exists, use it. * */ int lv_vdo_pool_size_config(const struct logical_volume *lv, struct vdo_pool_size_config *cfg) { struct dev_manager *dm; int r; if (!lv_info(lv->vg->cmd, lv, 1, NULL, 0, 0)) return 1; /* Inactive VDO pool -> no runtime config */ if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, !lv_is_pvmove(lv)))) return_0; r = dev_manager_vdo_pool_size_config(dm, lv, cfg); dev_manager_destroy(dm); return r; } static int _lv_active(struct cmd_context *cmd, const struct logical_volume *lv) { struct lvinfo info; if (!lv_info(cmd, lv, 0, &info, 0, 0)) { log_debug_activation("Cannot determine activation status of %s%s.", display_lvname(lv), activation() ? "" : " (no device driver)"); return 0; } return info.exists; } static int _lv_open_count(struct cmd_context *cmd, const struct logical_volume *lv) { struct lvinfo info; if (!lv_info(cmd, lv, 0, &info, 1, 0)) { stack; return -1; } return info.open_count; } static int _lv_activate_lv(const struct logical_volume *lv, struct lv_activate_opts *laopts) { int r; struct dev_manager *dm; if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, !lv_is_pvmove(lv)))) return_0; if (!(r = dev_manager_activate(dm, lv, laopts))) stack; dev_manager_destroy(dm); return r; } static int _lv_preload(const struct logical_volume *lv, struct lv_activate_opts *laopts, int *flush_required) { int r = 0; struct dev_manager *dm; int old_readonly = laopts->read_only; if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, !lv_is_pvmove(lv)))) goto_out; laopts->read_only = _passes_readonly_filter(lv->vg->cmd, lv); if (!(r = dev_manager_preload(dm, lv, laopts, flush_required))) stack; dev_manager_destroy(dm); laopts->read_only = old_readonly; out: return r; } static int _lv_deactivate(const struct logical_volume *lv) { int r; struct dev_manager *dm; if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1))) return_0; if (!(r = dev_manager_deactivate(dm, lv))) stack; dev_manager_destroy(dm); return r; } static int _lv_suspend_lv(const struct logical_volume *lv, struct lv_activate_opts *laopts, int lockfs, int flush_required) { int r; struct dev_manager *dm; laopts->read_only = _passes_readonly_filter(lv->vg->cmd, lv); /* * When we are asked to manipulate (normally suspend/resume) the PVMOVE * device directly, we don't want to touch the devices that use it. */ if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, !lv_is_pvmove(lv)))) return_0; if (!(r = dev_manager_suspend(dm, lv, laopts, lockfs, flush_required))) stack; dev_manager_destroy(dm); return r; } /* * These two functions return the number of visible LVs in the state, * or -1 on error. FIXME Check this. */ int lvs_in_vg_activated(const struct volume_group *vg) { struct lv_list *lvl; int count = 0; if (!activation()) return 0; dm_list_iterate_items(lvl, &vg->lvs) if (lv_is_visible(lvl->lv)) count += (_lv_active(vg->cmd, lvl->lv) == 1); log_debug_activation("Counted %d active LVs in VG %s", count, vg->name); return count; } int lvs_in_vg_opened(const struct volume_group *vg) { const struct lv_list *lvl; int count = 0; if (!activation()) return 0; dm_list_iterate_items(lvl, &vg->lvs) if (lv_is_visible(lvl->lv)) count += (_lv_open_count(vg->cmd, lvl->lv) > 0); log_debug_activation("Counted %d open LVs in VG %s.", count, vg->name); return count; } /* * Check if "raid4" @segtype is supported by kernel. * * if segment type is not raid4, return 1. */ int raid4_is_supported(struct cmd_context *cmd, const struct segment_type *segtype) { unsigned attrs = 0; if (segtype_is_raid4(segtype) && (!segtype->ops->target_present || !segtype->ops->target_present(cmd, NULL, &attrs) || !(attrs & RAID_FEATURE_RAID4))) { log_error("RAID module does not support RAID4."); return 0; } return 1; } /* * The VG lock must be held to call this function. * * Returns: 0 or 1 */ int lv_is_active(const struct logical_volume *lv) { return _lv_active(lv->vg->cmd, lv); } #ifdef DMEVENTD static struct dm_event_handler *_create_dm_event_handler(struct cmd_context *cmd, const char *dmuuid, const char *dso, const int timeout, enum dm_event_mask mask) { struct dm_event_handler *dmevh; if (!(dmevh = dm_event_handler_create())) return_NULL; if (!cmd->default_settings.dmeventd_executable) cmd->default_settings.dmeventd_executable = find_config_tree_str(cmd, dmeventd_executable_CFG, NULL); if (dm_event_handler_set_dmeventd_path(dmevh, cmd->default_settings.dmeventd_executable)) goto_bad; if (dso && dm_event_handler_set_dso(dmevh, dso)) goto_bad; if (dm_event_handler_set_uuid(dmevh, dmuuid)) goto_bad; dm_event_handler_set_timeout(dmevh, timeout); dm_event_handler_set_event_mask(dmevh, mask); return dmevh; bad: dm_event_handler_destroy(dmevh); return NULL; } char *get_monitor_dso_path(struct cmd_context *cmd, int id) { const char *libpath = find_config_tree_str(cmd, id, NULL); char path[PATH_MAX]; get_shared_library_path(cmd, libpath, path, sizeof(path)); return strdup(path); } static char *_build_target_uuid(struct cmd_context *cmd, const struct logical_volume *lv) { return build_dm_uuid(cmd->mem, lv, lv_layer(lv)); } static int _device_registered_with_dmeventd(struct cmd_context *cmd, const struct logical_volume *lv, const char **dso, int *pending, int *monitored) { char *uuid; enum dm_event_mask evmask; struct dm_event_handler *dmevh; int r; *pending = 0; *monitored = 0; if (!(uuid = _build_target_uuid(cmd, lv))) return_0; if (!(dmevh = _create_dm_event_handler(cmd, uuid, NULL, 0, DM_EVENT_ALL_ERRORS))) return_0; if ((r = dm_event_get_registered_device(dmevh, 0))) { if (r == -ENOENT) { r = 1; goto out; } r = 0; goto_out; } /* FIXME: why do we care which 'dso' is monitoring? */ if (dso && (*dso = dm_event_handler_get_dso(dmevh)) && !(*dso = dm_pool_strdup(cmd->mem, *dso))) { r = 0; goto_out; } evmask = dm_event_handler_get_event_mask(dmevh); if (evmask & DM_EVENT_REGISTRATION_PENDING) { *pending = 1; evmask &= ~DM_EVENT_REGISTRATION_PENDING; } *monitored = evmask; r = 1; out: dm_event_handler_destroy(dmevh); return r; } int target_registered_with_dmeventd(struct cmd_context *cmd, const char *dso, const struct logical_volume *lv, int *pending, int *monitored) { char *uuid; enum dm_event_mask evmask; struct dm_event_handler *dmevh; int r; *pending = 0; *monitored = 0; if (!dso) return_0; if (!(uuid = _build_target_uuid(cmd, lv))) return_0; if (!(dmevh = _create_dm_event_handler(cmd, uuid, dso, 0, DM_EVENT_ALL_ERRORS))) return_0; if ((r = dm_event_get_registered_device(dmevh, 0))) { if (r == -ENOENT) { r = 1; goto out; } r = 0; goto_out; } evmask = dm_event_handler_get_event_mask(dmevh); if (evmask & DM_EVENT_REGISTRATION_PENDING) { *pending = 1; evmask &= ~DM_EVENT_REGISTRATION_PENDING; } *monitored = evmask; r = 1; out: dm_event_handler_destroy(dmevh); return r; } int target_register_events(struct cmd_context *cmd, const char *dso, const struct logical_volume *lv, int evmask __attribute__((unused)), int set, int timeout) { char *uuid; struct dm_event_handler *dmevh; int r; if (!dso) return_0; /* We always monitor the "real" device, never the "snapshot-origin" itself. */ if (!(uuid = _build_target_uuid(cmd, lv))) return_0; if (!(dmevh = _create_dm_event_handler(cmd, uuid, dso, timeout, DM_EVENT_ALL_ERRORS | (timeout ? DM_EVENT_TIMEOUT : 0)))) return_0; r = set ? dm_event_register_handler(dmevh) : dm_event_unregister_handler(dmevh); dm_event_handler_destroy(dmevh); if (!r) return_0; log_verbose("%s %s for events", set ? "Monitored" : "Unmonitored", uuid); return 1; } #endif /* * Returns 0 if an attempt to (un)monitor the device failed. * Returns 1 otherwise. */ int monitor_dev_for_events(struct cmd_context *cmd, const struct logical_volume *lv, const struct lv_activate_opts *laopts, int monitor) { #ifdef DMEVENTD int i, pending = 0, monitored = 0; int r = 1; struct dm_list *snh, *snht; struct lv_segment *seg; struct lv_segment *log_seg; int (*monitor_fn) (struct lv_segment *s, int e); uint32_t s; const struct lv_activate_opts zlaopts = { 0 }; struct lv_activate_opts mirr_laopts = { .origin_only = 1 }; struct lvinfo info; const char *dso = NULL; int new_unmonitor; if (!laopts) laopts = &zlaopts; else mirr_laopts.read_only = laopts->read_only; /* skip dmeventd code altogether */ if (dmeventd_monitor_mode() == DMEVENTD_MONITOR_IGNORE) return 1; /* * Nothing to do if dmeventd configured not to be used. */ if (monitor && !dmeventd_monitor_mode()) return 1; if (sigint_caught()) return_0; /* * Activation of unused cache-pool activates metadata device as * a public LV for clearing purpose. * FIXME: * As VG lock is held across whole operation unmonitored volume * is usually OK since dmeventd couldn't do anything. * However in case command would have crashed, such LV is * left unmonitored and may potentially require dmeventd. */ if (lv_is_cache_pool_data(lv) || lv_is_cache_pool_metadata(lv)) { if (!(seg = find_pool_seg(first_seg(lv)))) return_0; if (!lv_is_used_cache_pool(seg->lv)) { log_debug_activation("Skipping %smonitor of %s.%s", (monitor) ? "" : "un", display_lvname(lv), (monitor) ? " Cache pool activation for clearing only." : ""); return 1; } } /* * Allow to unmonitor thin pool via explicit pool unmonitor * or unmonitor before the last thin pool user deactivation * Skip unmonitor, if invoked via deactivation of thin volume * and there is another thin pool user (open_count > 1) * FIXME think about watch ruler influence. */ if (laopts->skip_in_use && lv_is_thin_pool(lv) && lv_info(lv->vg->cmd, lv, 1, &info, 1, 0) && (info.open_count > 1)) { log_debug_activation("Skipping unmonitor of opened %s (open:%d)", display_lvname(lv), info.open_count); return 1; } /* Do not monitor snapshot that already covers origin */ if (monitor && lv_is_cow_covering_origin(lv)) { log_debug_activation("Skipping monitor of snapshot larger " "then origin %s.", display_lvname(lv)); return 1; } /* * In case of a snapshot device, we monitor lv->snapshot->lv, * not the actual LV itself. */ if (lv_is_cow(lv) && (laopts->no_merging || !lv_is_merging_cow(lv) || lv_has_target_type(lv->vg->cmd->mem, lv, NULL, TARGET_NAME_SNAPSHOT))) { if (!(r = monitor_dev_for_events(cmd, lv->snapshot->lv, NULL, monitor))) stack; return r; } /* * In case this LV is a snapshot origin, we instead monitor * each of its respective snapshots. The origin itself may * also need to be monitored if it is a mirror, for example, * so fall through to process it afterwards. * Before monitoring snapshots verify origin is active as with * external origin only read-only -real device can be active. */ if (!laopts->origin_only && lv_is_origin(lv) && lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0)) dm_list_iterate_safe(snh, snht, &lv->snapshot_segs) if (!monitor_dev_for_events(cmd, dm_list_struct_base(snh, struct lv_segment, origin_list)->cow, NULL, monitor)) { stack; r = 0; } /* * If the volume is mirrored and its log is also mirrored, monitor * the log volume as well. */ if ((seg = first_seg(lv)) != NULL && seg->log_lv != NULL && (log_seg = first_seg(seg->log_lv)) != NULL && seg_is_mirrored(log_seg)) if (!monitor_dev_for_events(cmd, seg->log_lv, NULL, monitor)) { stack; r = 0; } dm_list_iterate_items(seg, &lv->segments) { if (sigint_caught()) { stack; r = 0; break; } /* Recurse for AREA_LV */ for (s = 0; s < seg->area_count; s++) { if (seg_type(seg, s) != AREA_LV) continue; if (!monitor_dev_for_events(cmd, seg_lv(seg, s), NULL, monitor)) { stack; r = 0; } } /* * If requested unmonitoring of thin volume, preserve skip_in_use flag. * * FIXME: code here looks like _lv_postorder() */ if (seg->pool_lv && !monitor_dev_for_events(cmd, seg->pool_lv, (!monitor) ? laopts : NULL, monitor)) { stack; r = 0; } if (seg->external_lv && !monitor_dev_for_events(cmd, seg->external_lv, (!monitor) ? laopts : NULL, monitor)) { stack; r = 0; } if (seg->metadata_lv && !monitor_dev_for_events(cmd, seg->metadata_lv, NULL, monitor)) { stack; r = 0; } if (!seg_monitored(seg) || (seg->status & PVMOVE) || !seg->segtype->ops->target_monitored) /* doesn't support registration */ continue; if (!monitor) { /* When unmonitoring, obtain existing dso being used. */ if (!_device_registered_with_dmeventd(cmd, seg_is_snapshot(seg) ? seg->cow : seg->lv, &dso, &pending, &monitored)) { log_warn("WARNING: Failed to %smonitor %s.", monitor ? "" : "un", display_lvname(seg_is_snapshot(seg) ? seg->cow : seg->lv)); return 0; } } else if (!seg->segtype->ops->target_monitored(seg, &pending, &monitored)) { log_warn("WARNING: Failed to %smonitor %s.", monitor ? "" : "un", display_lvname(seg->lv)); return 0; } /* FIXME: We should really try again if pending */ monitored = (pending) ? 0 : monitored; monitor_fn = NULL; new_unmonitor = 0; if (monitor) { if (monitored) log_verbose("%s already monitored.", display_lvname(lv)); else if (seg->segtype->ops->target_monitor_events) { log_very_verbose("Monitoring %s with %s.%s", display_lvname(lv), seg->segtype->dso, test_mode() ? " [Test mode: skipping this]" : ""); monitor_fn = seg->segtype->ops->target_monitor_events; } } else { if (!monitored) log_verbose("%s already not monitored.", display_lvname(lv)); else if (dso && *dso) { /* * Divert unmonitor away from code that depends on the new segment * type instead of the existing one if it's changing. */ log_verbose("Not monitoring %s with %s%s", display_lvname(lv), dso, test_mode() ? " [Test mode: skipping this]" : ""); new_unmonitor = 1; } } /* FIXME Test mode should really continue a bit further. */ if (test_mode()) continue; if (new_unmonitor) { if (!target_register_events(cmd, dso, seg_is_snapshot(seg) ? seg->cow : lv, 0, 0, 10)) { log_warn("WARNING: %s: segment unmonitoring failed.", display_lvname(lv)); return 0; } } else if (monitor_fn) { /* FIXME specify events */ if (!monitor_fn(seg, 0)) { log_warn("WARNING: %s: %s segment monitoring function failed.", display_lvname(lv), lvseg_name(seg)); return 0; } } else continue; if (!vg_write_lock_held() && lv_is_mirror(lv)) { mirr_laopts.exclusive = lv_is_active(lv) ? 1 : 0; /* * Commands vgchange and lvchange do use read-only lock when changing * monitoring (--monitor y|n). All other use cases hold 'write-lock' * so they skip this dm mirror table refreshing step. */ if (!_lv_activate_lv(lv, &mirr_laopts)) { stack; r = 0; } } /* Check [un]monitor results */ /* Try a couple times if pending, but not forever... */ for (i = 0;; i++) { pending = 0; if (!seg->segtype->ops->target_monitored(seg, &pending, &monitored)) { stack; r = 0; break; } if (!pending || i >= 40) break; log_very_verbose("%s %smonitoring still pending: waiting...", display_lvname(lv), monitor ? "" : "un"); if (interruptible_usleep(10000 * i)) { stack; r = 0; break; } } if (r) r = (monitored && monitor) || (!monitored && !monitor); } if (!r && !error_message_produced()) log_warn("WARNING: %sonitoring %s failed.", monitor ? "M" : "Not m", display_lvname(lv)); return r; #else return 1; #endif } struct detached_lv_data { const struct volume_group *vg_pre; struct lv_activate_opts *laopts; int *flush_required; }; static int _preload_detached_lv(struct logical_volume *lv, void *data) { struct detached_lv_data *detached = data; struct logical_volume *lv_pre; /* Check if the LV was 'hidden' (non-toplevel) in committed metadata * and becomes 'visible' (toplevel) in precommitted metadata */ if (!lv_is_visible(lv) && (lv_pre = find_lv_in_vg_by_lvid(detached->vg_pre, &lv->lvid)) && lv_is_visible(lv_pre)) { log_debug_activation("Preloading detached hidden volume %s as visible volume %s.", display_lvname(lv), display_lvname(lv_pre)); if (!_lv_preload(lv_pre, detached->laopts, detached->flush_required)) return_0; } return 1; } static int _lv_suspend(struct cmd_context *cmd, const char *lvid_s, struct lv_activate_opts *laopts, int error_if_not_suspended, const struct logical_volume *lv, const struct logical_volume *lv_pre) { const struct logical_volume *pvmove_lv = NULL; struct logical_volume *lv_pre_tmp, *lv_tmp; struct seg_list *sl; struct lv_segment *snap_seg; struct lvinfo info; int r = 0, lockfs = 0, flush_required = 0; struct detached_lv_data detached; struct dm_pool *mem = NULL; struct dm_list suspend_lvs; struct lv_list *lvl; int found; if (!activation()) return 1; if (test_mode()) { _skip("Suspending %s%s.", display_lvname(lv), laopts->origin_only ? " origin without snapshots" : ""); r = 1; goto out; } if (!lv_info(cmd, lv, laopts->origin_only, &info, 0, 0)) goto_out; if (!info.exists || info.suspended) { if (!error_if_not_suspended) { r = 1; if (info.suspended) critical_section_inc(cmd, "already suspended"); } goto out; } lv_calculate_readahead(lv, NULL); /* Ignore origin_only unless LV is origin in both old and new metadata */ /* or LV is thin or thin pool volume */ if (laopts->origin_only && !lv_is_thin_volume(lv) && !lv_is_thin_pool(lv) && !(lv_is_origin(lv) && lv_is_origin(lv_pre))) { log_debug_activation("Not using origin only for suspend of %s.", display_lvname(lv)); laopts->origin_only = 0; } /* * Preload devices for the LV. * If the PVMOVE LV is being removed, it's only present in the old * metadata and not the new, so we must explicitly add the new * tables for all the changed LVs here, as the relationships * are not found by walking the new metadata. */ if (lv_is_locked(lv) && !lv_is_locked(lv_pre) && (pvmove_lv = find_pvmove_lv_in_lv(lv))) { /* Preload all the LVs above the PVMOVE LV */ dm_list_iterate_items(sl, &pvmove_lv->segs_using_this_lv) { if (!(lv_pre_tmp = find_lv(lv_pre->vg, sl->seg->lv->name))) { log_error(INTERNAL_ERROR "LV %s missing from preload metadata.", display_lvname(sl->seg->lv)); goto out; } if (!_lv_preload(lv_pre_tmp, laopts, &flush_required)) goto_out; } /* Now preload the PVMOVE LV itself */ if (!(lv_pre_tmp = find_lv(lv_pre->vg, pvmove_lv->name))) { log_error(INTERNAL_ERROR "LV %s missing from preload metadata.", display_lvname(pvmove_lv)); goto out; } if (!_lv_preload(lv_pre_tmp, laopts, &flush_required)) goto_out; /* Suspending 1st. LV above PVMOVE suspends whole tree */ dm_list_iterate_items(sl, &pvmove_lv->segs_using_this_lv) { lv = sl->seg->lv; break; } } else { if (!_lv_preload(lv_pre, laopts, &flush_required)) /* FIXME Revert preloading */ goto_out; /* * Search for existing LVs that have become detached and preload them. */ detached.vg_pre = lv_pre->vg; detached.laopts = laopts; detached.flush_required = &flush_required; if (!for_each_sub_lv((struct logical_volume *)lv, &_preload_detached_lv, &detached)) goto_out; /* * Preload any snapshots that are being removed. */ if (!laopts->origin_only && lv_is_origin(lv)) { dm_list_iterate_items_gen(snap_seg, &lv->snapshot_segs, origin_list) { if (!(lv_pre_tmp = find_lv_in_vg_by_lvid(lv_pre->vg, &snap_seg->cow->lvid))) { log_error(INTERNAL_ERROR "LV %s (%s) missing from preload metadata.", display_lvname(snap_seg->cow), snap_seg->cow->lvid.id[1].uuid); goto out; } if (!lv_is_cow(lv_pre_tmp) && !_lv_preload(lv_pre_tmp, laopts, &flush_required)) goto_out; } } } /* Flush is ATM required for the tested cases * NOTE: Mirror repair requires noflush for proper repair! * TODO: Relax this limiting condition further */ if (!flush_required && (lv_is_pvmove(lv) || pvmove_lv || (!lv_is_mirror(lv) && !lv_is_thin_volume(lv) && !lv_is_thin_pool(lv) && !lv_is_vdo(lv) && !lv_is_vdo_pool(lv)))) { log_debug_activation("Requiring flush for LV %s.", display_lvname(lv)); flush_required = 1; } if (!monitor_dev_for_events(cmd, lv, laopts, 0)) /* FIXME Consider aborting here */ stack; /* Require fs synchronization when taking a thin snapshot */ if (laopts->origin_only && lv_is_thin_volume(lv) && lv_is_thin_volume(lv_pre)) lockfs = 1; /* Require fs synchronization when taking a thick snapshot */ else if (!laopts->origin_only && (lv_is_origin(lv_pre) || lv_is_cow(lv_pre))) lockfs = 1; /* Require fs synchronization when converting a non-thin LV to a thin LV or * a non/thin LV with/out external origin to a thin LV with external origin LV. */ else if (!laopts->origin_only && lv_is_thin_volume(lv_pre) && /* new LV is a Thin */ (!lv_is_thin_volume(lv) || /* and either the existing LV is NOT a Thin */ (first_seg(lv_pre)->external_lv && /* or the existing LV IS Thin and the new LV is Thin with the external origin */ (!first_seg(lv)->external_lv || /* and check if existing Thin is either without the external origin */ memcmp(&first_seg(lv_pre)->external_lv->lvid.id[1], /* or it uses a different external origin */ &first_seg(lv)->external_lv->lvid.id[1], ID_LEN) != 0)))) lockfs = 1; if (!lv_is_locked(lv) && lv_is_locked(lv_pre) && (pvmove_lv = find_pvmove_lv_in_lv(lv_pre))) { /* * When starting PVMOVE, suspend participating LVs first * with committed metadata by looking at precommited pvmove list. * In committed metadata these LVs are not connected in any way. * * TODO: prepare list of LVs needed to be suspended and pass them * via 'struct laopts' directly to _lv_suspend_lv() and handle this * with a single 'dmtree' call. */ if (!(mem = dm_pool_create("suspend_lvs", 128))) goto_out; /* Prepare list of all LVs for suspend ahead */ dm_list_init(&suspend_lvs); dm_list_iterate_items(sl, &pvmove_lv->segs_using_this_lv) { lv_tmp = sl->seg->lv; if (lv_is_cow(lv_tmp)) /* Never suspend COW, always has to be origin */ lv_tmp = origin_from_cow(lv_tmp); found = 0; dm_list_iterate_items(lvl, &suspend_lvs) if (strcmp(lvl->lv->name, lv_tmp->name) == 0) { found = 1; break; } if (found) continue; /* LV is already in the list */ if (!(lvl = dm_pool_alloc(mem, sizeof(*lvl)))) { log_error("lv_list alloc failed."); goto out; } /* Look for precommitted LV name in commmitted VG */ if (!(lvl->lv = find_lv(lv->vg, lv_tmp->name))) { log_error(INTERNAL_ERROR "LV %s missing from preload metadata.", display_lvname(lv_tmp)); goto out; } dm_list_add(&suspend_lvs, &lvl->list); } critical_section_inc(cmd, "suspending"); dm_list_iterate_items(lvl, &suspend_lvs) if (!_lv_suspend_lv(lvl->lv, laopts, lockfs, 1)) { critical_section_dec(cmd, "failed suspend"); goto_out; /* FIXME: resume on recovery path? */ } } else { /* Standard suspend */ critical_section_inc(cmd, "suspending"); if (!_lv_suspend_lv(lv, laopts, lockfs, flush_required)) { critical_section_dec(cmd, "failed suspend"); goto_out; } } r = 1; out: if (mem) dm_pool_destroy(mem); return r; } /* * In a cluster, set exclusive to indicate that only one node is using the * device. Any preloaded tables may then use non-clustered targets. * * Returns success if the device is not active */ int lv_suspend_if_active(struct cmd_context *cmd, const char *lvid_s, unsigned origin_only, unsigned exclusive, const struct logical_volume *lv, const struct logical_volume *lv_pre) { struct lv_activate_opts laopts = { .exclusive = exclusive, .origin_only = origin_only }; return _lv_suspend(cmd, lvid_s, &laopts, 0, lv, lv_pre); } static int _check_suspended_lv(struct logical_volume *lv, void *data) { struct lvinfo info; if (lv_info(lv->vg->cmd, lv, 0, &info, 0, 0) && info.exists && info.suspended) { log_debug_activation("Found suspended LV %s in critical section().", display_lvname(lv)); return 0; /* There is suspended subLV in the tree */ } if (lv_layer(lv) && lv_info(lv->vg->cmd, lv, 1, &info, 0, 0) && info.exists && info.suspended) { log_debug_activation("Found suspended layered LV %s in critical section().", display_lvname(lv)); return 0; /* There is suspended subLV in the tree */ } return 1; } static int _lv_resume(struct cmd_context *cmd, const char *lvid_s, struct lv_activate_opts *laopts, int error_if_not_active, const struct logical_volume *lv) { struct dm_list *snh; struct lvinfo info; int r = 0; if (!activation()) return 1; if (!lv_is_origin(lv) && !lv_is_thin_volume(lv) && !lv_is_thin_pool(lv)) laopts->origin_only = 0; if (test_mode()) { _skip("Resuming %s%s%s.", display_lvname(lv), laopts->origin_only ? " without snapshots" : "", laopts->revert ? " (reverting)" : ""); r = 1; goto out; } log_debug_activation("Resuming LV %s%s%s%s.", display_lvname(lv), error_if_not_active ? "" : " if active", laopts->origin_only ? (lv_is_thin_pool(lv) ? " pool only" : lv_is_thin_volume(lv) ? " thin only" : " without snapshots") : "", laopts->revert ? " (reverting)" : ""); if (laopts->revert) goto needs_resume; if (!lv_info(cmd, lv, laopts->origin_only, &info, 0, 0)) goto_out; if (!info.exists || !info.suspended) { if (error_if_not_active) goto_out; /* ATM only thin-pool with origin-only suspend does not really suspend anything * it's used only for message passing to thin-pool */ if (laopts->origin_only && lv_is_thin_pool(lv)) critical_section_dec(cmd, "resumed"); if (!info.suspended && critical_section()) { /* Validation check if any subLV is suspended */ if (!laopts->origin_only && lv_is_origin(lv)) { /* Check all snapshots for this origin LV */ dm_list_iterate(snh, &lv->snapshot_segs) if (!_check_suspended_lv(dm_list_struct_base(snh, struct lv_segment, origin_list)->cow, NULL)) goto needs_resume; /* Found suspended snapshot */ } if ((r = for_each_sub_lv((struct logical_volume *)lv, &_check_suspended_lv, NULL))) goto out; /* Nothing was found suspended */ } else { r = 1; goto out; } } needs_resume: laopts->read_only = _passes_readonly_filter(cmd, lv); laopts->resuming = 1; if (!_lv_activate_lv(lv, laopts)) goto_out; critical_section_dec(cmd, "resumed"); if (!monitor_dev_for_events(cmd, lv, laopts, 1)) stack; r = 1; out: return r; } /* * In a cluster, set exclusive to indicate that only one node is using the * device. Any tables loaded may then use non-clustered targets. * * @origin_only * @exclusive This parameter only has an affect in cluster-context. * It forces local target type to be used (instead of * cluster-aware type). * Returns success if the device is not active */ int lv_resume_if_active(struct cmd_context *cmd, const char *lvid_s, unsigned origin_only, unsigned exclusive, unsigned revert, const struct logical_volume *lv) { struct lv_activate_opts laopts = { .exclusive = exclusive, .origin_only = origin_only, .revert = revert }; return _lv_resume(cmd, lvid_s, &laopts, 0, lv); } int lv_resume(struct cmd_context *cmd, const char *lvid_s, unsigned origin_only, const struct logical_volume *lv) { struct lv_activate_opts laopts = { .origin_only = origin_only, }; return _lv_resume(cmd, lvid_s, &laopts, 1, lv); } static int _lv_has_open_snapshots(const struct logical_volume *lv) { struct lv_segment *snap_seg; int r = 0; dm_list_iterate_items_gen(snap_seg, &lv->snapshot_segs, origin_list) if (!lv_check_not_in_use(snap_seg->cow, 1)) r++; if (r) log_error("LV %s has open %d snapshot(s), not deactivating.", display_lvname(lv), r); return r; } int lv_deactivate(struct cmd_context *cmd, const char *lvid_s, const struct logical_volume *lv) { struct lvinfo info; const struct lv_activate_opts laopts = { .skip_in_use = 1 }; struct dm_list *snh; int r = 0; if (!activation()) return 1; if (test_mode()) { _skip("Deactivating %s.", display_lvname(lv)); r = 1; goto out; } log_debug_activation("Deactivating %s.", display_lvname(lv)); if (lv_is_visible(lv) || lv_is_virtual_origin(lv) || lv_is_merging_thin_snapshot(lv)) { switch (lv_check_not_in_use(lv, 1)) { case 0: goto_out; case 2: goto no_exists; } if (lv_is_origin(lv) && _lv_has_open_snapshots(lv)) goto_out; } else { if (!lv_info(cmd, lv, 0, &info, 0, 0)) goto_out; if (!info.exists) { no_exists: r = 1; /* Check attached snapshot segments are also inactive */ dm_list_iterate(snh, &lv->snapshot_segs) { if (!lv_info(cmd, dm_list_struct_base(snh, struct lv_segment, origin_list)->cow, 0, &info, 0, 0)) goto_out; if (info.exists) { r = 0; /* Snapshot left in table? */ break; } } if (lv_is_vdo_pool(lv)) { /* If someone has remove 'linear' mapping over VDO device * we may still be able to deactivate the rest of the tree * i.e. in test-suite we simulate this via 'dmsetup remove' */ if (!lv_info(cmd, lv, 1, &info, 1, 0)) goto_out; if (info.exists && !info.open_count) r = 0; /* Unused VDO device left in table? */ } if (r) goto out; } } if (!monitor_dev_for_events(cmd, lv, &laopts, 0)) stack; critical_section_inc(cmd, "deactivating"); r = _lv_deactivate(lv); /* * Remove any transiently activated error * devices which arean't used any more. */ if (r && lv_is_raid(lv) && !lv_deactivate_any_missing_subdevs(lv)) { log_error("Failed to remove temporary SubLVs from %s", display_lvname(lv)); r = 0; } critical_section_dec(cmd, "deactivated"); if (!lv_info(cmd, lv, 0, &info, 0, 0) || info.exists) { /* Turn into log_error, but we do not log error */ log_debug_activation("Deactivated volume is still %s present.", display_lvname(lv)); r = 0; } out: return r; } /* Test if LV passes filter */ int lv_activation_filter(struct cmd_context *cmd, const char *lvid_s, int *activate, const struct logical_volume *lv) { if (!activation()) { *activate = 1; return 1; } if (!_passes_activation_filter(cmd, lv)) { log_verbose("Not activating %s since it does not pass " "activation filter.", display_lvname(lv)); *activate = 0; } else *activate = 1; return 1; } static int _lv_activate(struct cmd_context *cmd, const char *lvid_s, struct lv_activate_opts *laopts, int filter, const struct logical_volume *lv) { struct lvinfo info; int r = 0; if (!activation()) return 1; if (filter && !_passes_activation_filter(cmd, lv)) { log_verbose("Not activating %s since it does not pass " "activation filter.", display_lvname(lv)); r = 1; goto out; } if ((cmd->partial_activation || cmd->degraded_activation) && lv_is_partial(lv) && lv_is_raid(lv) && lv_raid_has_integrity((struct logical_volume *)lv)) { cmd->partial_activation = 0; cmd->degraded_activation = 0; log_print_unless_silent("No degraded or partial activation for raid with integrity."); } if ((!lv->vg->cmd->partial_activation) && lv_is_partial(lv)) { if (!lv_is_raid_type(lv) || !partial_raid_lv_supports_degraded_activation(lv)) { log_error("Refusing activation of partial LV %s. " "Use '--activationmode partial' to override.", display_lvname(lv)); goto out; } if (!lv->vg->cmd->degraded_activation) { log_error("Refusing activation of partial LV %s. " "Try '--activationmode degraded'.", display_lvname(lv)); goto out; } } if ((cmd->partial_activation || cmd->degraded_activation) && lv_is_writecache(lv)) { struct logical_volume *lv_fast = first_seg(lv)->writecache; if (lv_is_partial(lv) || (lv_fast && lv_is_partial(lv_fast))) { log_error("Cannot use partial or degraded activation with writecache."); goto out; } } if (lv_has_unknown_segments(lv)) { log_error("Refusing activation of LV %s containing " "an unrecognised segment.", display_lvname(lv)); goto out; } if (lv_raid_has_visible_sublvs(lv)) { log_error("Refusing activation of RAID LV %s with " "visible SubLVs.", display_lvname(lv)); goto out; } if (test_mode()) { _skip("Activating %s.", display_lvname(lv)); r = 1; goto out; } /* Component LV activation is enforced to be 'read-only' */ /* TODO: should not apply for LVs in maintenance mode */ if (!lv_is_visible(lv) && lv_is_component(lv)) { laopts->read_only = 1; laopts->component_lv = lv; } else if (lv_is_pool_metadata_spare(lv)) { laopts->component_lv = lv; } else if (filter) laopts->read_only = _passes_readonly_filter(cmd, lv); log_debug_activation("Activating %s%s%s%s%s.", display_lvname(lv), laopts->exclusive ? " exclusively" : "", laopts->read_only ? " read-only" : "", laopts->noscan ? " noscan" : "", laopts->temporary ? " temporary" : ""); if (!lv_info_with_name_check(cmd, lv, 0, &info)) goto_out; /* * Nothing to do? */ if (info.exists && !info.suspended && info.live_table && (info.read_only == read_only_lv(lv, laopts, NULL))) { r = 1; log_debug_activation("LV %s is already active.", display_lvname(lv)); goto out; } lv_calculate_readahead(lv, NULL); critical_section_inc(cmd, "activating"); if (!(r = _lv_activate_lv(lv, laopts))) stack; critical_section_dec(cmd, "activated"); if (r && !monitor_dev_for_events(cmd, lv, laopts, 1)) stack; out: return r; } /* Activate LV */ int lv_activate(struct cmd_context *cmd, const char *lvid_s, int exclusive, int noscan, int temporary, const struct logical_volume *lv) { struct lv_activate_opts laopts = { .exclusive = exclusive, .noscan = noscan, .temporary = temporary }; if (!_lv_activate(cmd, lvid_s, &laopts, 0, lv)) return_0; return 1; } /* Activate LV only if it passes filter */ int lv_activate_with_filter(struct cmd_context *cmd, const char *lvid_s, int exclusive, int noscan, int temporary, const struct logical_volume *lv) { struct lv_activate_opts laopts = { .exclusive = exclusive, .noscan = noscan, .temporary = temporary }; if (!_lv_activate(cmd, lvid_s, &laopts, 1, lv)) return_0; return 1; } int lv_mknodes(struct cmd_context *cmd, const struct logical_volume *lv) { int r; if (!lv) { r = dm_mknodes(NULL); fs_unlock(); return r; } if (!activation()) return 1; r = dev_manager_mknodes(lv); fs_unlock(); return r; } /* Remove any existing, closed mapped device by @name */ static int _remove_dm_dev_by_name(const char *name) { int r = 0; struct dm_task *dmt; struct dm_info info; if (!(dmt = dm_task_create(DM_DEVICE_INFO))) return_0; /* Check, if the device exists. */ if (dm_task_set_name(dmt, name) && dm_task_run(dmt) && dm_task_get_info(dmt, &info)) { dm_task_destroy(dmt); /* Ignore non-existing or open dm devices */ if (!info.exists || info.open_count) return 1; if (!(dmt = dm_task_create(DM_DEVICE_REMOVE))) return_0; if (dm_task_set_name(dmt, name)) r = dm_task_run(dmt); } dm_task_destroy(dmt); return r; } /* Work all segments of @lv removing any existing, closed "*-missing_N_0" sub devices. */ static int _lv_remove_any_missing_subdevs(struct logical_volume *lv) { char name[NAME_LEN]; struct lv_segment *seg; uint32_t seg_no = 0; if (lv) { dm_list_iterate_items(seg, &lv->segments) { if (dm_snprintf(name, sizeof(name), "%s-%s-missing_%u_0", seg->lv->vg->name, seg->lv->name, seg_no) < 0) return_0; if (!_remove_dm_dev_by_name(name)) return_0; seg_no++; } } return 1; } /* Remove any "*-missing_*" sub devices added by the activation layer for an rmate/rimage missing PV mapping */ int lv_deactivate_any_missing_subdevs(const struct logical_volume *lv) { uint32_t s; struct lv_segment *seg = first_seg(lv); for (s = 0; s < seg->area_count; s++) { if (seg_type(seg, s) == AREA_LV && !_lv_remove_any_missing_subdevs(seg_lv(seg, s))) return_0; if (seg->meta_areas && seg_metatype(seg, s) == AREA_LV && !_lv_remove_any_missing_subdevs(seg_metalv(seg, s))) return_0; } return 1; } /* * Does PV use VG somewhere in its construction? * Returns 1 on failure. */ int pv_uses_vg(struct physical_volume *pv, struct volume_group *vg) { if (!activation() || !pv->dev) return 0; if (!dm_is_dm_major(MAJOR(pv->dev->dev))) return 0; return dev_manager_device_uses_vg(pv->dev, vg); } void activation_release(void) { if (critical_section()) /* May leak stacked operation */ log_error("Releasing activation in critical section."); fs_unlock(); /* Implicit dev_manager_release(); */ } void activation_exit(void) { activation_release(); dev_manager_exit(); } #endif static int _component_cb(struct logical_volume *lv, void *data) { struct logical_volume **component_lv = (struct logical_volume **) data; if (lv_is_locked(lv) || lv_is_pvmove(lv) ||/* ignoring */ /* thin-pool is special and it's using layered device */ (lv_is_thin_pool(lv) && thin_pool_is_active(lv))) return -1; /* External origin is activated through thinLV and uses -real suffix. * Note: for old clustered logic we would need to check for all thins */ if ((lv_is_external_origin(lv) && lv_info(lv->vg->cmd, lv, 1, NULL, 0, 0)) || lv_is_active(lv)) { if (!lv_is_component(lv) || lv_is_visible(lv)) return -1; /* skip whole subtree */ log_debug_activation("Found active component LV %s.", display_lvname(lv)); *component_lv = lv; return 0; /* break any further processing */ } return 1; } /* * Finds out for any LV if any of its component LVs are active. * Function first checks if an existing LV is visible and active eventually * it's lock holding LV is already active. In such case sub LV cannot be * actived alone and no further checking is needed. * * Returns active component LV if there is such. */ const struct logical_volume *lv_component_is_active(const struct logical_volume *lv) { const struct logical_volume *component_lv = NULL; const struct logical_volume *holder_lv = lv_lock_holder(lv); if ((holder_lv != lv) && lv_is_active(holder_lv)) return NULL; /* Lock holding LV is active, do not check components */ if (_component_cb((struct logical_volume *) lv, &holder_lv) == 1) (void) for_each_sub_lv((struct logical_volume *) lv, _component_cb, (void*) &component_lv); return component_lv; } /* * Finds out if any LV above is active, as stacked device tree can be composed of * chained set of LVs. * * Returns active holder LV if there is such. */ const struct logical_volume *lv_holder_is_active(const struct logical_volume *lv) { const struct logical_volume *holder; const struct seg_list *sl; if (lv_is_locked(lv) || lv_is_pvmove(lv)) return NULL; /* Skip pvmove/locked LV tracking */ dm_list_iterate_items(sl, &lv->segs_using_this_lv) { /* Recursive call for upper-stack holder */ if ((holder = lv_holder_is_active(sl->seg->lv))) return holder; if (lv_is_active(sl->seg->lv)) { log_debug_activation("Found active holder LV %s.", display_lvname(sl->seg->lv)); return sl->seg->lv; } } return NULL; } static int _deactivate_sub_lv_cb(struct logical_volume *lv, void *data) { struct logical_volume **slv = data; if (lv_is_thin_pool(lv) || lv_is_external_origin(lv)) return -1; if (!deactivate_lv(lv->vg->cmd, lv)) { *slv = lv; return 0; } return 1; } /* * Deactivates LV toghether with explicit deactivation call made also for all its component LVs. */ int deactivate_lv_with_sub_lv(const struct logical_volume *lv) { struct logical_volume *flv = NULL; if (!deactivate_lv(lv->vg->cmd, lv)) { log_error("Cannot deactivate logical volume %s.", display_lvname(lv)); return 0; } if (!for_each_sub_lv((struct logical_volume *)lv, _deactivate_sub_lv_cb, &flv)) { log_error("Cannot deactivate subvolume %s of logical volume %s.", (flv) ? display_lvname(flv) : "", display_lvname(lv)); return 0; } return 1; } int activate_lv(struct cmd_context *cmd, const struct logical_volume *lv) { const struct logical_volume *active_lv; int ret; /* * When trying activating component LV, make sure none of sub component * LV or LVs that are using it are active. */ if (!lv_is_visible(lv)) active_lv = lv_holder_is_active(lv); else active_lv = lv_component_is_active(lv); if (active_lv) { log_error("Activation of logical volume %s is prohibited while logical volume %s is active.", display_lvname(lv), display_lvname(active_lv)); ret = 0; goto out; } ret = lv_activate_with_filter(cmd, NULL, 0, (lv->status & LV_NOSCAN) ? 1 : 0, (lv->status & LV_TEMPORARY) ? 1 : 0, lv_committed(lv)); out: return ret; } int deactivate_lv(struct cmd_context *cmd, const struct logical_volume *lv) { int ret; ret = lv_deactivate(cmd, NULL, lv_committed(lv)); return ret; } int suspend_lv(struct cmd_context *cmd, const struct logical_volume *lv) { int ret; critical_section_inc(cmd, "locking for suspend"); ret = lv_suspend_if_active(cmd, NULL, 0, 0, lv_committed(lv), lv); return ret; } int suspend_lv_origin(struct cmd_context *cmd, const struct logical_volume *lv) { int ret; critical_section_inc(cmd, "locking for suspend"); ret = lv_suspend_if_active(cmd, NULL, 1, 0, lv_committed(lv), lv); return ret; } int resume_lv(struct cmd_context *cmd, const struct logical_volume *lv) { int ret; ret = lv_resume_if_active(cmd, NULL, 0, 0, 0, lv_committed(lv)); critical_section_dec(cmd, "unlocking on resume"); return ret; } int resume_lv_origin(struct cmd_context *cmd, const struct logical_volume *lv) { int ret; ret = lv_resume_if_active(cmd, NULL, 1, 0, 0, lv_committed(lv)); critical_section_dec(cmd, "unlocking on resume"); return ret; } int revert_lv(struct cmd_context *cmd, const struct logical_volume *lv) { int ret; ret = lv_resume_if_active(cmd, NULL, 0, 0, 1, lv_committed(lv)); critical_section_dec(cmd, "unlocking on revert"); return ret; }