mirror of
git://sourceware.org/git/lvm2.git
synced 2024-12-21 13:34:40 +03:00
3028 lines
77 KiB
C
3028 lines
77 KiB
C
/*
|
|
* Copyright (C) 2001-2004 Sistina Software, Inc. All rights reserved.
|
|
* Copyright (C) 2004-2018 Red Hat, Inc. All rights reserved.
|
|
*
|
|
* This file is part of LVM2.
|
|
*
|
|
* This copyrighted material is made available to anyone wishing to use,
|
|
* modify, copy, or redistribute it subject to the terms and conditions
|
|
* of the GNU Lesser General Public License v.2.1.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public License
|
|
* along with this program; if not, write to the Free Software Foundation,
|
|
* Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "lib/misc/lib.h"
|
|
#include "lib/metadata/metadata.h"
|
|
#include "lib/activate/activate.h"
|
|
#include "lib/mm/memlock.h"
|
|
#include "lib/display/display.h"
|
|
#include "fs.h"
|
|
#include "lib/misc/lvm-exec.h"
|
|
#include "lib/misc/lvm-file.h"
|
|
#include "lib/misc/lvm-string.h"
|
|
#include "lib/commands/toolcontext.h"
|
|
#include "dev_manager.h"
|
|
#include "lib/datastruct/str_list.h"
|
|
#include "lib/config/config.h"
|
|
#include "lib/metadata/segtype.h"
|
|
#include "lib/misc/sharedlib.h"
|
|
#include "lib/metadata/metadata.h"
|
|
#include "lib/misc/lvm-signal.h"
|
|
|
|
#include <limits.h>
|
|
#include <fcntl.h>
|
|
#include <unistd.h>
|
|
|
|
#define _skip(fmt, args...) log_very_verbose("Skipping: " fmt , ## args)
|
|
|
|
int list_segment_modules(struct dm_pool *mem, const struct lv_segment *seg,
|
|
struct dm_list *modules)
|
|
{
|
|
unsigned int s;
|
|
struct lv_segment *seg2, *snap_seg;
|
|
struct dm_list *snh;
|
|
|
|
if (seg->segtype->ops->modules_needed &&
|
|
!seg->segtype->ops->modules_needed(mem, seg, modules)) {
|
|
log_error("module string allocation failed");
|
|
return 0;
|
|
}
|
|
|
|
if (lv_is_origin(seg->lv))
|
|
dm_list_iterate(snh, &seg->lv->snapshot_segs)
|
|
if (!list_lv_modules(mem,
|
|
dm_list_struct_base(snh,
|
|
struct lv_segment,
|
|
origin_list)->cow,
|
|
modules))
|
|
return_0;
|
|
|
|
if (lv_is_cow(seg->lv)) {
|
|
snap_seg = find_snapshot(seg->lv);
|
|
if (snap_seg->segtype->ops->modules_needed &&
|
|
!snap_seg->segtype->ops->modules_needed(mem, snap_seg,
|
|
modules)) {
|
|
log_error("snap_seg module string allocation failed");
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
for (s = 0; s < seg->area_count; s++) {
|
|
switch (seg_type(seg, s)) {
|
|
case AREA_LV:
|
|
seg2 = find_seg_by_le(seg_lv(seg, s), seg_le(seg, s));
|
|
if (seg2 && !list_segment_modules(mem, seg2, modules))
|
|
return_0;
|
|
break;
|
|
case AREA_PV:
|
|
case AREA_UNASSIGNED:
|
|
;
|
|
}
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
int list_lv_modules(struct dm_pool *mem, const struct logical_volume *lv,
|
|
struct dm_list *modules)
|
|
{
|
|
struct lv_segment *seg;
|
|
|
|
dm_list_iterate_items(seg, &lv->segments)
|
|
if (!list_segment_modules(mem, seg, modules))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int _lv_passes_volumes_filter(struct cmd_context *cmd, const struct logical_volume *lv,
|
|
const struct dm_config_node *cn, const int cfg_id)
|
|
{
|
|
const struct dm_config_value *cv;
|
|
const char *str;
|
|
static char config_path[PATH_MAX];
|
|
size_t len = strlen(lv->vg->name);
|
|
|
|
config_def_get_path(config_path, sizeof(config_path), cfg_id);
|
|
log_verbose("%s configuration setting defined: "
|
|
"Checking the list to match %s.",
|
|
config_path, display_lvname(lv));
|
|
|
|
for (cv = cn->v; cv; cv = cv->next) {
|
|
if (cv->type == DM_CFG_EMPTY_ARRAY)
|
|
goto out;
|
|
if (cv->type != DM_CFG_STRING) {
|
|
log_print_unless_silent("Ignoring invalid string in config file %s.",
|
|
config_path);
|
|
continue;
|
|
}
|
|
str = cv->v.str;
|
|
if (!*str) {
|
|
log_print_unless_silent("Ignoring empty string in config file %s.",
|
|
config_path);
|
|
continue;
|
|
}
|
|
|
|
/* Tag? */
|
|
if (*str == '@') {
|
|
str++;
|
|
if (!*str) {
|
|
log_print_unless_silent("Ignoring empty tag in config file %s",
|
|
config_path);
|
|
continue;
|
|
}
|
|
/* If any host tag matches any LV or VG tag, activate */
|
|
if (!strcmp(str, "*")) {
|
|
if (str_list_match_list(&cmd->tags, &lv->tags, NULL)
|
|
|| str_list_match_list(&cmd->tags,
|
|
&lv->vg->tags, NULL))
|
|
return 1;
|
|
|
|
continue;
|
|
}
|
|
/* If supplied tag matches LV or VG tag, activate */
|
|
if (str_list_match_item(&lv->tags, str) ||
|
|
str_list_match_item(&lv->vg->tags, str))
|
|
return 1;
|
|
|
|
continue;
|
|
}
|
|
|
|
/* If supplied name is vgname[/lvname] */
|
|
if ((strncmp(str, lv->vg->name, len) == 0) &&
|
|
(!str[len] ||
|
|
((str[len] == '/') &&
|
|
!strcmp(str + len + 1, lv->name))))
|
|
return 1;
|
|
}
|
|
|
|
out:
|
|
log_verbose("No item supplied in %s configuration setting matches %s.",
|
|
config_path, display_lvname(lv));
|
|
|
|
return 0;
|
|
}
|
|
|
|
int lv_passes_auto_activation_filter(struct cmd_context *cmd, struct logical_volume *lv)
|
|
{
|
|
const struct dm_config_node *cn;
|
|
|
|
if (!(cn = find_config_tree_array(cmd, activation_auto_activation_volume_list_CFG, NULL))) {
|
|
log_verbose("activation/auto_activation_volume_list configuration setting "
|
|
"not defined: All logical volumes will be auto-activated.");
|
|
return 1;
|
|
}
|
|
|
|
return _lv_passes_volumes_filter(cmd, lv, cn, activation_auto_activation_volume_list_CFG);
|
|
}
|
|
|
|
static int _passes_readonly_filter(struct cmd_context *cmd,
|
|
const struct logical_volume *lv)
|
|
{
|
|
const struct dm_config_node *cn;
|
|
|
|
if (!(cn = find_config_tree_array(cmd, activation_read_only_volume_list_CFG, NULL)))
|
|
return 0;
|
|
|
|
return _lv_passes_volumes_filter(cmd, lv, cn, activation_read_only_volume_list_CFG);
|
|
}
|
|
|
|
int lv_passes_readonly_filter(const struct logical_volume *lv)
|
|
{
|
|
return _passes_readonly_filter(lv->vg->cmd, lv);
|
|
}
|
|
|
|
#ifndef DEVMAPPER_SUPPORT
|
|
void set_activation(int act, int silent)
|
|
{
|
|
static int warned = 0;
|
|
|
|
if (warned || !act)
|
|
return;
|
|
|
|
log_warn("WARNING: Compiled without libdevmapper support. "
|
|
"Can't enable activation.");
|
|
|
|
warned = 1;
|
|
}
|
|
int activation(void)
|
|
{
|
|
return 0;
|
|
}
|
|
int library_version(char *version, size_t size)
|
|
{
|
|
return 0;
|
|
}
|
|
int driver_version(char *version, size_t size)
|
|
{
|
|
return 0;
|
|
}
|
|
int target_version(const char *target_name, uint32_t *maj,
|
|
uint32_t *min, uint32_t *patchlevel)
|
|
{
|
|
return 0;
|
|
}
|
|
int target_present(struct cmd_context *cmd, const char *target_name,
|
|
int use_modprobe)
|
|
{
|
|
return 0;
|
|
}
|
|
int lvm_dm_prefix_check(int major, int minor, const char *prefix)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_info(struct cmd_context *cmd, const struct logical_volume *lv, int use_layer,
|
|
struct lvinfo *info, int with_open_count, int with_read_ahead)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_info_with_seg_status(struct cmd_context *cmd,
|
|
const struct lv_segment *lv_seg,
|
|
struct lv_with_info_and_seg_status *status,
|
|
int with_open_count, int with_read_ahead)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_cache_status(const struct logical_volume *cache_lv,
|
|
struct lv_status_cache **status)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_check_not_in_use(const struct logical_volume *lv, int error_if_used)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_snapshot_percent(const struct logical_volume *lv, dm_percent_t *percent)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_mirror_percent(struct cmd_context *cmd, const struct logical_volume *lv,
|
|
int wait, dm_percent_t *percent, uint32_t *event_nr)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_raid_percent(const struct logical_volume *lv, dm_percent_t *percent)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_raid_data_offset(const struct logical_volume *lv, uint64_t *data_offset)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_raid_dev_health(const struct logical_volume *lv, char **dev_health)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_raid_dev_count(const struct logical_volume *lv, uint32_t *dev_cnt)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_raid_mismatch_count(const struct logical_volume *lv, uint64_t *cnt)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_raid_sync_action(const struct logical_volume *lv, char **sync_action)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_raid_message(const struct logical_volume *lv, const char *msg)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_raid_status(const struct logical_volume *lv, struct lv_status_raid **status)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_writecache_message(const struct logical_volume *lv, const char *msg)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_thin_pool_status(const struct logical_volume *lv, int flush,
|
|
struct lv_status_thin_pool **thin_pool_status)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_thin_status(const struct logical_volume *lv, int flush,
|
|
struct lv_status_thin **thin_status)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_thin_device_id(const struct logical_volume *lv, uint32_t *device_id)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_vdo_pool_status(const struct logical_volume *lv, int flush,
|
|
struct lv_status_vdo **vdo_status)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_vdo_pool_percent(const struct logical_volume *lv, dm_percent_t *percent)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_vdo_pool_size_config(const struct logical_volume *lv,
|
|
struct vdo_pool_size_config *cfg)
|
|
{
|
|
return 0;
|
|
}
|
|
int lvs_in_vg_activated(const struct volume_group *vg)
|
|
{
|
|
return 0;
|
|
}
|
|
int lvs_in_vg_opened(const struct volume_group *vg)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_suspend_if_active(struct cmd_context *cmd, const char *lvid_s, unsigned origin_only, unsigned exclusive,
|
|
const struct logical_volume *lv, const struct logical_volume *lv_pre)
|
|
{
|
|
return 1;
|
|
}
|
|
int lv_resume(struct cmd_context *cmd, const char *lvid_s, unsigned origin_only, const struct logical_volume *lv)
|
|
{
|
|
return 1;
|
|
}
|
|
int lv_resume_if_active(struct cmd_context *cmd, const char *lvid_s, unsigned origin_only,
|
|
unsigned exclusive, unsigned revert, const struct logical_volume *lv)
|
|
{
|
|
return 1;
|
|
}
|
|
int lv_deactivate(struct cmd_context *cmd, const char *lvid_s, const struct logical_volume *lv)
|
|
{
|
|
return 1;
|
|
}
|
|
int lv_activation_filter(struct cmd_context *cmd, const char *lvid_s,
|
|
int *activate_lv, const struct logical_volume *lv)
|
|
{
|
|
return 1;
|
|
}
|
|
int lv_activate(struct cmd_context *cmd, const char *lvid_s, int exclusive, int noscan,
|
|
int temporary, const struct logical_volume *lv)
|
|
{
|
|
return 1;
|
|
}
|
|
int lv_activate_with_filter(struct cmd_context *cmd, const char *lvid_s, int exclusive,
|
|
int noscan, int temporary, const struct logical_volume *lv)
|
|
{
|
|
return 1;
|
|
}
|
|
int lv_mknodes(struct cmd_context *cmd, const struct logical_volume *lv)
|
|
{
|
|
return 1;
|
|
}
|
|
int lv_deactivate_any_missing_subdevs(const struct logical_volume *lv)
|
|
{
|
|
return 1;
|
|
}
|
|
int pv_uses_vg(struct physical_volume *pv,
|
|
struct volume_group *vg)
|
|
{
|
|
return 0;
|
|
}
|
|
void activation_release(void)
|
|
{
|
|
}
|
|
void activation_exit(void)
|
|
{
|
|
}
|
|
|
|
int raid4_is_supported(struct cmd_context *cmd, const struct segment_type *segtype)
|
|
{
|
|
return 1;
|
|
}
|
|
|
|
int lv_is_active(const struct logical_volume *lv)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_check_transient(struct logical_volume *lv)
|
|
{
|
|
return 1;
|
|
}
|
|
int monitor_dev_for_events(struct cmd_context *cmd, const struct logical_volume *lv,
|
|
const struct lv_activate_opts *laopts, int monitor)
|
|
{
|
|
return 1;
|
|
}
|
|
/* fs.c */
|
|
void fs_unlock(void)
|
|
{
|
|
}
|
|
/* dev_manager.c */
|
|
#include "lib/activate/targets.h"
|
|
int add_areas_line(struct dev_manager *dm, struct lv_segment *seg,
|
|
struct dm_tree_node *node, uint32_t start_area,
|
|
uint32_t areas)
|
|
{
|
|
return 0;
|
|
}
|
|
int dm_device_is_usable(struct cmd_context *cmd, struct device *dev, struct dev_usable_check_params check, int *is_lv)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_has_target_type(struct dm_pool *mem, const struct logical_volume *lv,
|
|
const char *layer, const char *target_type)
|
|
{
|
|
return 0;
|
|
}
|
|
#else /* DEVMAPPER_SUPPORT */
|
|
|
|
static int _activation = 1;
|
|
|
|
void set_activation(int act, int silent)
|
|
{
|
|
if (act == _activation)
|
|
return;
|
|
|
|
_activation = act;
|
|
if (_activation)
|
|
log_verbose("Activation enabled. Device-mapper kernel "
|
|
"driver will be used.");
|
|
else if (!silent)
|
|
log_warn("WARNING: Activation disabled. No device-mapper "
|
|
"interaction will be attempted.");
|
|
else
|
|
log_verbose("Activation disabled. No device-mapper "
|
|
"interaction will be attempted.");
|
|
}
|
|
|
|
int activation(void)
|
|
{
|
|
return _activation;
|
|
}
|
|
|
|
static int _passes_activation_filter(struct cmd_context *cmd,
|
|
const struct logical_volume *lv)
|
|
{
|
|
const struct dm_config_node *cn;
|
|
|
|
if (!(cn = find_config_tree_array(cmd, activation_volume_list_CFG, NULL))) {
|
|
log_verbose("activation/volume_list configuration setting "
|
|
"not defined: Checking only host tags for %s.",
|
|
display_lvname(lv));
|
|
|
|
/* If no host tags defined, activate */
|
|
if (dm_list_empty(&cmd->tags))
|
|
return 1;
|
|
|
|
/* If any host tag matches any LV or VG tag, activate */
|
|
if (str_list_match_list(&cmd->tags, &lv->tags, NULL) ||
|
|
str_list_match_list(&cmd->tags, &lv->vg->tags, NULL))
|
|
return 1;
|
|
|
|
log_verbose("No host tag matches %s", display_lvname(lv));
|
|
|
|
/* Don't activate */
|
|
return 0;
|
|
}
|
|
|
|
return _lv_passes_volumes_filter(cmd, lv, cn, activation_volume_list_CFG);
|
|
}
|
|
|
|
int library_version(char *version, size_t size)
|
|
{
|
|
if (!activation())
|
|
return 0;
|
|
|
|
return dm_get_library_version(version, size);
|
|
}
|
|
|
|
int driver_version(char *version, size_t size)
|
|
{
|
|
static char _vsn[80] = { 0 };
|
|
|
|
if (!activation())
|
|
return 0;
|
|
|
|
log_very_verbose("Getting driver version");
|
|
|
|
if (!_vsn[0] &&
|
|
!dm_driver_version(_vsn, sizeof(_vsn)))
|
|
return_0;
|
|
|
|
dm_strncpy(version, _vsn, size);
|
|
|
|
return 1;
|
|
}
|
|
|
|
int target_version(const char *target_name, uint32_t *maj,
|
|
uint32_t *min, uint32_t *patchlevel)
|
|
{
|
|
int r = 0;
|
|
struct dm_task *dmt;
|
|
struct dm_versions *target, *last_target;
|
|
|
|
log_very_verbose("Getting target version for %s", target_name);
|
|
if (!(dmt = dm_task_create(DM_DEVICE_LIST_VERSIONS)))
|
|
return_0;
|
|
|
|
if (activation_checks() && !dm_task_enable_checks(dmt))
|
|
goto_out;
|
|
|
|
if (!dm_task_run(dmt)) {
|
|
log_debug_activation("Failed to get %s target version", target_name);
|
|
/* Assume this was because LIST_VERSIONS isn't supported */
|
|
*maj = 0;
|
|
*min = 0;
|
|
*patchlevel = 0;
|
|
r = 1;
|
|
goto out;
|
|
}
|
|
|
|
target = dm_task_get_versions(dmt);
|
|
|
|
do {
|
|
last_target = target;
|
|
|
|
if (!strcmp(target_name, target->name)) {
|
|
r = 1;
|
|
*maj = target->version[0];
|
|
*min = target->version[1];
|
|
*patchlevel = target->version[2];
|
|
goto out;
|
|
}
|
|
|
|
target = (struct dm_versions *)((char *) target + target->next);
|
|
} while (last_target != target);
|
|
|
|
out:
|
|
if (r)
|
|
log_very_verbose("Found %s target "
|
|
"v%" PRIu32 ".%" PRIu32 ".%" PRIu32 ".",
|
|
target_name, *maj, *min, *patchlevel);
|
|
|
|
dm_task_destroy(dmt);
|
|
|
|
return r;
|
|
}
|
|
|
|
int lvm_dm_prefix_check(int major, int minor, const char *prefix)
|
|
{
|
|
return dev_manager_check_prefix_dm_major_minor(major, minor, prefix);
|
|
}
|
|
|
|
/* Search modules.builtin file for built-in kernel module */
|
|
static int _check_modules_builtin(struct cmd_context *cmd, const char *target)
|
|
{
|
|
FILE *fp;
|
|
char *line = NULL;
|
|
size_t len;
|
|
int r = 0;
|
|
char path[PATH_MAX];
|
|
|
|
if (dm_snprintf(path, sizeof(path), "%s/%s/modules.builtin",
|
|
MODULES_PATH, cmd->kernel_vsn) < 0) {
|
|
log_debug_activation("Modules path %s/%s/modules.builtin is too long.",
|
|
MODULES_PATH, cmd->kernel_vsn);
|
|
return 0;
|
|
}
|
|
|
|
if (!(fp = fopen(path, "r"))) {
|
|
if (errno != ENOENT)
|
|
log_sys_debug("fopen", path);
|
|
return 0;
|
|
}
|
|
|
|
while (getline(&line, &len, fp) > 0)
|
|
if (strstr(line, target)) {
|
|
log_debug_activation("Found %s as built-in kernel module.", target);
|
|
r = 1;
|
|
break;
|
|
}
|
|
|
|
free(line);
|
|
|
|
if (fclose(fp))
|
|
log_sys_debug("fclose", path);
|
|
|
|
return r;
|
|
}
|
|
|
|
int module_present(struct cmd_context *cmd, const char *target_name)
|
|
{
|
|
int ret = 0;
|
|
#ifdef MODPROBE_CMD
|
|
char module[128];
|
|
const char *argv[] = { MODPROBE_CMD, module, NULL };
|
|
unsigned maj, min;
|
|
#endif
|
|
struct stat st;
|
|
char path[PATH_MAX];
|
|
int i = dm_snprintf(path, sizeof(path), "%smodule/dm_%s",
|
|
dm_sysfs_dir(), target_name);
|
|
|
|
if (i > 0) {
|
|
while ((i > 0) && path[--i] != '/') /* stop on dm_ */
|
|
if (path[i] == '-')
|
|
path[i] = '_'; /* replace '-' with '_' */
|
|
|
|
if ((lstat(path, &st) == 0) && S_ISDIR(st.st_mode)) {
|
|
log_debug_activation("Module directory %s exists.", path);
|
|
return 1;
|
|
}
|
|
|
|
if (path[i] == '/' && _check_modules_builtin(cmd, path + i + 1))
|
|
return 1;
|
|
}
|
|
|
|
#ifdef MODPROBE_CMD
|
|
if ((strcmp(target_name, TARGET_NAME_VDO) == 0) &&
|
|
(sscanf(cmd->kernel_vsn, "%u.%u", &maj, &min) == 2) &&
|
|
((maj < 6) || ((maj == 6) && (min < 9))))
|
|
argv[1] = MODULE_NAME_VDO; /* Kernels < 6.9 -> "kvdo" without dm- prefix */
|
|
else if (dm_snprintf(module, sizeof(module), "dm-%s", target_name) < 0) {
|
|
log_error("module_present module name too long: %s",
|
|
target_name);
|
|
return 0;
|
|
}
|
|
|
|
ret = exec_cmd(cmd, argv, NULL, 0);
|
|
#endif
|
|
return ret;
|
|
}
|
|
|
|
int target_present_version(struct cmd_context *cmd, const char *target_name,
|
|
int use_modprobe,
|
|
uint32_t *maj, uint32_t *min, uint32_t *patchlevel)
|
|
{
|
|
if (!activation()) {
|
|
log_error(INTERNAL_ERROR "Target present version called when activation is disabled.");
|
|
return 0;
|
|
}
|
|
#ifdef MODPROBE_CMD
|
|
if (use_modprobe) {
|
|
if (target_version(target_name, maj, min, patchlevel))
|
|
return 1;
|
|
|
|
if (!module_present(cmd, target_name))
|
|
return_0;
|
|
}
|
|
#endif
|
|
return target_version(target_name, maj, min, patchlevel);
|
|
}
|
|
|
|
int target_present(struct cmd_context *cmd, const char *target_name,
|
|
int use_modprobe)
|
|
{
|
|
uint32_t maj, min, patchlevel;
|
|
|
|
return target_present_version(cmd, target_name, use_modprobe,
|
|
&maj, &min, &patchlevel);
|
|
}
|
|
|
|
int get_dm_active_devices(const struct volume_group *vg, struct dm_list **devs,
|
|
unsigned *devs_features)
|
|
{
|
|
if (!activation())
|
|
return 0;
|
|
|
|
return dev_manager_get_dm_active_devices(NULL, devs, devs_features);
|
|
}
|
|
|
|
/*
|
|
* When '*info' is NULL, returns 1 only when LV is active.
|
|
* When '*info' != NULL, returns 1 when info structure is populated.
|
|
*/
|
|
static int _lv_info(struct cmd_context *cmd, const struct logical_volume *lv,
|
|
int use_layer, struct lvinfo *info,
|
|
const struct lv_segment *seg,
|
|
struct lv_seg_status *seg_status,
|
|
int with_open_count, int with_read_ahead, int with_name_check)
|
|
{
|
|
struct dm_info dminfo;
|
|
|
|
/*
|
|
* If open_count info is requested and we have to be sure our own udev
|
|
* transactions are finished
|
|
* For non-clustered locking type we are only interested for non-delete operation
|
|
* in progress - as only those could lead to opened files
|
|
*/
|
|
if (with_open_count) {
|
|
if (fs_has_non_delete_ops())
|
|
fs_unlock(); /* For non clustered - wait if there are non-delete ops */
|
|
}
|
|
|
|
/* New thin-pool has no layer, but -tpool suffix needs to be queried */
|
|
if (!use_layer && lv_is_new_thin_pool(lv)) {
|
|
/* Check if there isn't existing old thin pool mapping in the table */
|
|
if (!dev_manager_info(cmd, lv, NULL, 0, 0, 0, &dminfo, NULL, NULL))
|
|
return_0;
|
|
if (!dminfo.exists)
|
|
use_layer = 1;
|
|
}
|
|
|
|
if (seg_status) {
|
|
/* TODO: for now it's mess with seg_status */
|
|
seg_status->seg = seg;
|
|
}
|
|
|
|
if (!dev_manager_info(cmd, lv,
|
|
(use_layer) ? lv_layer(lv) : NULL,
|
|
with_open_count, with_read_ahead, with_name_check,
|
|
&dminfo,
|
|
(info) ? &info->read_ahead : NULL,
|
|
seg_status))
|
|
return_0;
|
|
|
|
if (!info)
|
|
return dminfo.exists;
|
|
|
|
info->exists = dminfo.exists;
|
|
info->suspended = dminfo.suspended;
|
|
info->open_count = dminfo.open_count;
|
|
info->major = dminfo.major;
|
|
info->minor = dminfo.minor;
|
|
info->read_only = dminfo.read_only;
|
|
info->live_table = dminfo.live_table;
|
|
info->inactive_table = dminfo.inactive_table;
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* Returns 1 if info structure populated, else 0 on failure.
|
|
* When lvinfo* is NULL, it returns 1 if the device is locally active, 0 otherwise.
|
|
*/
|
|
int lv_info(struct cmd_context *cmd, const struct logical_volume *lv, int use_layer,
|
|
struct lvinfo *info, int with_open_count, int with_read_ahead)
|
|
{
|
|
if (!activation())
|
|
return 0;
|
|
|
|
return _lv_info(cmd, lv, use_layer, info, NULL, NULL, with_open_count, with_read_ahead, 0);
|
|
}
|
|
|
|
int lv_info_with_name_check(struct cmd_context *cmd, const struct logical_volume *lv,
|
|
int use_layer, struct lvinfo *info)
|
|
{
|
|
if (!activation())
|
|
return 0;
|
|
|
|
return _lv_info(cmd, lv, use_layer, info, NULL, NULL, 0, 0, 1);
|
|
}
|
|
|
|
/*
|
|
* Returns 1 if lv_with_info_and_seg_status info structure populated,
|
|
* else 0 on failure or if device not active locally.
|
|
*
|
|
* When seg_status parsing had troubles it will set type to SEG_STATUS_UNKNOWN.
|
|
*
|
|
* Using usually one ioctl to obtain info and status.
|
|
* More complex segment do collect info from one device,
|
|
* but status from another device.
|
|
*
|
|
* TODO: further improve with more statuses (i.e. snapshot's origin/merge)
|
|
*/
|
|
int lv_info_with_seg_status(struct cmd_context *cmd,
|
|
const struct lv_segment *lv_seg,
|
|
struct lv_with_info_and_seg_status *status,
|
|
int with_open_count, int with_read_ahead)
|
|
{
|
|
const struct logical_volume *olv, *lv = status->lv = lv_seg->lv;
|
|
|
|
if (!activation())
|
|
return 0;
|
|
|
|
if (lv_is_used_cache_pool(lv)) {
|
|
/* INFO is not set as cache-pool cannot be active.
|
|
* STATUS is collected from cache LV */
|
|
if (!(lv_seg = get_only_segment_using_this_lv(lv)))
|
|
return_0;
|
|
(void) _lv_info(cmd, lv_seg->lv, 1, NULL, lv_seg, &status->seg_status, 0, 0, 0);
|
|
return 1;
|
|
}
|
|
|
|
if (lv_is_thin_pool(lv)) {
|
|
/* Always collect status for '-tpool' */
|
|
if (_lv_info(cmd, lv, 1, &status->info, lv_seg, &status->seg_status, 0, 0, 0) &&
|
|
(status->seg_status.type == SEG_STATUS_THIN_POOL)) {
|
|
/* There is -tpool device, but query 'active' state of 'fake' thin-pool */
|
|
if (!_lv_info(cmd, lv, 0, NULL, NULL, NULL, 0, 0, 0) &&
|
|
!status->seg_status.thin_pool->needs_check)
|
|
status->info.exists = 0; /* So pool LV is not active */
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
if (lv_is_external_origin(lv)) {
|
|
if (!_lv_info(cmd, lv, 0, &status->info, NULL, NULL,
|
|
with_open_count, with_read_ahead, 0))
|
|
return_0;
|
|
|
|
(void) _lv_info(cmd, lv, 1, NULL, lv_seg, &status->seg_status, 0, 0, 0);
|
|
return 1;
|
|
}
|
|
|
|
if (lv_is_origin(lv)) {
|
|
/* Query segment status for 'layered' (-real) device most of the time,
|
|
* only for merging snapshot, query its progress.
|
|
* TODO: single LV may need couple status to be exposed at once....
|
|
* but this needs more logical background
|
|
*/
|
|
/* Show INFO for actual origin and grab status for merging origin */
|
|
if (!_lv_info(cmd, lv, 0, &status->info, lv_seg,
|
|
lv_is_merging_origin(lv) ? &status->seg_status : NULL,
|
|
with_open_count, with_read_ahead, 0))
|
|
return_0;
|
|
|
|
if (status->info.exists &&
|
|
(status->seg_status.type != SEG_STATUS_SNAPSHOT)) /* Not merging */
|
|
/* Grab STATUS from layered -real */
|
|
(void) _lv_info(cmd, lv, 1, NULL, lv_seg, &status->seg_status, 0, 0, 0);
|
|
return 1;
|
|
}
|
|
|
|
if (lv_is_cow(lv)) {
|
|
if (lv_is_merging_cow(lv)) {
|
|
olv = origin_from_cow(lv);
|
|
|
|
if (!_lv_info(cmd, olv, 0, &status->info, first_seg(olv), &status->seg_status,
|
|
with_open_count, with_read_ahead, 0))
|
|
return_0;
|
|
|
|
if (status->seg_status.type == SEG_STATUS_SNAPSHOT ||
|
|
(lv_is_thin_volume(olv) && (status->seg_status.type == SEG_STATUS_THIN))) {
|
|
log_debug_activation("Snapshot merge is in progress, querying status of %s instead.",
|
|
display_lvname(lv));
|
|
/*
|
|
* When merge is in progress, query merging origin LV instead.
|
|
* COW volume is already mapped as error target in this case.
|
|
*/
|
|
return 1;
|
|
}
|
|
|
|
/* Merge not yet started, still a snapshot... */
|
|
}
|
|
/* Hadle fictional lvm2 snapshot and query snapshotX volume */
|
|
lv_seg = find_snapshot(lv);
|
|
}
|
|
|
|
if (lv_is_vdo(lv)) {
|
|
if (!_lv_info(cmd, lv, 0, &status->info, NULL, NULL,
|
|
with_open_count, with_read_ahead, 0))
|
|
return_0;
|
|
if (status->info.exists) {
|
|
/* Status for VDO pool */
|
|
(void) _lv_info(cmd, seg_lv(lv_seg, 0), 1, NULL,
|
|
first_seg(seg_lv(lv_seg, 0)),
|
|
&status->seg_status, 0, 0, 0);
|
|
/* Use VDO pool segtype result for VDO segtype */
|
|
status->seg_status.seg = lv_seg;
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
if (lv_is_vdo_pool(lv)) {
|
|
/* Always collect status for '-vpool' */
|
|
if (_lv_info(cmd, lv, 1, &status->info, lv_seg, &status->seg_status, 0, 0, 0) &&
|
|
(status->seg_status.type == SEG_STATUS_VDO_POOL)) {
|
|
/* There is -tpool device, but query 'active' state of 'fake' vdo-pool */
|
|
if (!_lv_info(cmd, lv, 0, NULL, NULL, NULL, 0, 0, 0))
|
|
status->info.exists = 0; /* So VDO pool LV is not active */
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
return _lv_info(cmd, lv, 0, &status->info, lv_seg, &status->seg_status,
|
|
with_open_count, with_read_ahead, 0);
|
|
}
|
|
|
|
#define OPEN_COUNT_CHECK_RETRIES 25
|
|
#define OPEN_COUNT_CHECK_USLEEP_DELAY 200000
|
|
|
|
/* Only report error if error_if_used is set */
|
|
/* Returns 0 if in use, 1 if it is unused, 2 when it is not present in table */
|
|
int lv_check_not_in_use(const struct logical_volume *lv, int error_if_used)
|
|
{
|
|
struct lvinfo info;
|
|
unsigned int open_count_check_retries;
|
|
|
|
if (!lv_info(lv->vg->cmd, lv, 0, &info, 1, 0) || !info.exists)
|
|
return 2;
|
|
else if (!info.open_count)
|
|
return 1;
|
|
|
|
/* If sysfs is not used, use open_count information only. */
|
|
if (dm_sysfs_dir()) {
|
|
if (dm_device_has_holders(info.major, info.minor)) {
|
|
if (error_if_used)
|
|
log_error("Logical volume %s is used by another device.",
|
|
display_lvname(lv));
|
|
else
|
|
log_debug_activation("Logical volume %s is used by another device.",
|
|
display_lvname(lv));
|
|
return 0;
|
|
}
|
|
|
|
if (dm_device_has_mounted_fs(info.major, info.minor)) {
|
|
if (error_if_used)
|
|
log_error("Logical volume %s contains a filesystem in use.",
|
|
display_lvname(lv));
|
|
else
|
|
log_debug_activation("Logical volume %s contains a filesystem in use.",
|
|
display_lvname(lv));
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
open_count_check_retries = retry_deactivation() ? OPEN_COUNT_CHECK_RETRIES : 1;
|
|
while (open_count_check_retries--) {
|
|
if (interruptible_usleep(OPEN_COUNT_CHECK_USLEEP_DELAY))
|
|
break; /* interrupted */
|
|
|
|
log_debug_activation("Retrying open_count check for %s.",
|
|
display_lvname(lv));
|
|
if (!lv_info(lv->vg->cmd, lv, 0, &info, 1, 0) || !info.exists) {
|
|
stack; /* device dissappeared? */
|
|
return 1;
|
|
} else if (!info.open_count)
|
|
return 1;
|
|
}
|
|
|
|
if (error_if_used)
|
|
log_error("Logical volume %s in use.", display_lvname(lv));
|
|
else
|
|
log_debug_activation("Logical volume %s in use.", display_lvname(lv));
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Returns 1 if percent set, else 0 on failure.
|
|
*/
|
|
int lv_check_transient(struct logical_volume *lv)
|
|
{
|
|
int r;
|
|
struct dev_manager *dm;
|
|
|
|
if (!activation())
|
|
return 0;
|
|
|
|
log_debug_activation("Checking transient status for LV %s.",
|
|
display_lvname(lv));
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1)))
|
|
return_0;
|
|
|
|
if (!(r = dev_manager_transient(dm, lv)))
|
|
stack;
|
|
|
|
dev_manager_destroy(dm);
|
|
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* Returns 1 if percent set, else 0 on failure.
|
|
*/
|
|
int lv_snapshot_percent(const struct logical_volume *lv, dm_percent_t *percent)
|
|
{
|
|
int r;
|
|
struct dev_manager *dm;
|
|
|
|
if (!lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0))
|
|
return 0;
|
|
|
|
log_debug_activation("Checking snapshot percent for LV %s.",
|
|
display_lvname(lv));
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1)))
|
|
return_0;
|
|
|
|
if (!(r = dev_manager_snapshot_percent(dm, lv, percent)))
|
|
stack;
|
|
|
|
dev_manager_destroy(dm);
|
|
|
|
return r;
|
|
}
|
|
|
|
/* FIXME Merge with snapshot_percent */
|
|
int lv_mirror_percent(struct cmd_context *cmd, const struct logical_volume *lv,
|
|
int wait, dm_percent_t *percent, uint32_t *event_nr)
|
|
{
|
|
int r;
|
|
struct dev_manager *dm;
|
|
|
|
/* If mirrored LV is temporarily shrinked to 1 area (= linear),
|
|
* it should be considered in-sync. */
|
|
if (dm_list_size(&lv->segments) == 1 && first_seg(lv)->area_count == 1) {
|
|
*percent = DM_PERCENT_100;
|
|
return 1;
|
|
}
|
|
|
|
if (!lv_info(cmd, lv, 0, NULL, 0, 0))
|
|
return 0;
|
|
|
|
log_debug_activation("Checking mirror percent for LV %s.",
|
|
display_lvname(lv));
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1)))
|
|
return_0;
|
|
|
|
if (!(r = dev_manager_mirror_percent(dm, lv, wait, percent, event_nr)))
|
|
stack;
|
|
|
|
dev_manager_destroy(dm);
|
|
|
|
return r;
|
|
}
|
|
|
|
int lv_raid_percent(const struct logical_volume *lv, dm_percent_t *percent)
|
|
{
|
|
return lv_mirror_percent(lv->vg->cmd, lv, 0, percent, NULL);
|
|
}
|
|
|
|
int lv_raid_data_offset(const struct logical_volume *lv, uint64_t *data_offset)
|
|
{
|
|
struct lv_status_raid *raid_status;
|
|
|
|
if (!lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0))
|
|
return 0;
|
|
|
|
log_debug_activation("Checking raid data offset and dev sectors for LV %s/%s",
|
|
lv->vg->name, lv->name);
|
|
|
|
if (!lv_raid_status(lv, &raid_status))
|
|
return_0;
|
|
|
|
*data_offset = raid_status->raid->data_offset;
|
|
|
|
dm_pool_destroy(raid_status->mem);
|
|
|
|
return 1;
|
|
}
|
|
|
|
int lv_raid_dev_health(const struct logical_volume *lv, char **dev_health)
|
|
{
|
|
int r = 1;
|
|
struct lv_status_raid *raid_status;
|
|
|
|
*dev_health = NULL;
|
|
|
|
if (!lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0))
|
|
return 0;
|
|
|
|
log_debug_activation("Checking raid device health for LV %s.",
|
|
display_lvname(lv));
|
|
|
|
if (!lv_raid_status(lv, &raid_status))
|
|
return_0;
|
|
|
|
if (!(*dev_health = dm_pool_strdup(lv->vg->cmd->mem,
|
|
raid_status->raid->dev_health))) {
|
|
stack;
|
|
r = 0;
|
|
}
|
|
|
|
dm_pool_destroy(raid_status->mem);
|
|
|
|
return r;
|
|
}
|
|
|
|
int lv_raid_dev_count(const struct logical_volume *lv, uint32_t *dev_cnt)
|
|
{
|
|
struct lv_status_raid *raid_status;
|
|
|
|
*dev_cnt = 0;
|
|
|
|
if (!lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0))
|
|
return 0;
|
|
|
|
log_debug_activation("Checking raid device count for LV %s/%s",
|
|
lv->vg->name, lv->name);
|
|
|
|
if (!lv_raid_status(lv, &raid_status))
|
|
return_0;
|
|
|
|
*dev_cnt = raid_status->raid->dev_count;
|
|
|
|
dm_pool_destroy(raid_status->mem);
|
|
|
|
return 1;
|
|
}
|
|
|
|
int lv_raid_mismatch_count(const struct logical_volume *lv, uint64_t *cnt)
|
|
{
|
|
struct lv_status_raid *raid_status;
|
|
|
|
*cnt = 0;
|
|
|
|
if (!lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0))
|
|
return 0;
|
|
|
|
log_debug_activation("Checking raid mismatch count for LV %s.",
|
|
display_lvname(lv));
|
|
|
|
if (!lv_raid_status(lv, &raid_status))
|
|
return_0;
|
|
|
|
*cnt = raid_status->raid->mismatch_count;
|
|
|
|
dm_pool_destroy(raid_status->mem);
|
|
|
|
return 1;
|
|
}
|
|
|
|
int lv_raid_sync_action(const struct logical_volume *lv, char **sync_action)
|
|
{
|
|
struct lv_status_raid *raid_status;
|
|
int r = 1;
|
|
|
|
*sync_action = NULL;
|
|
|
|
if (!lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0))
|
|
return 0;
|
|
|
|
log_debug_activation("Checking raid sync_action for LV %s.",
|
|
display_lvname(lv));
|
|
|
|
if (!lv_raid_status(lv, &raid_status))
|
|
return_0;
|
|
|
|
/* status->sync_action can be NULL if dm-raid version < 1.5.0 */
|
|
if (!raid_status->raid->sync_action ||
|
|
!(*sync_action = dm_pool_strdup(lv->vg->cmd->mem,
|
|
raid_status->raid->sync_action))) {
|
|
stack;
|
|
r = 0;
|
|
}
|
|
|
|
dm_pool_destroy(raid_status->mem);
|
|
|
|
return r;
|
|
}
|
|
|
|
int lv_raid_message(const struct logical_volume *lv, const char *msg)
|
|
{
|
|
struct lv_status_raid *raid_status;
|
|
struct dev_manager *dm = NULL;
|
|
int r = 0;
|
|
|
|
if (!seg_is_raid(first_seg(lv))) {
|
|
/*
|
|
* Make it easier for user to know what to do when
|
|
* they are using thinpool.
|
|
*/
|
|
if (lv_is_thin_pool(lv) &&
|
|
(lv_is_raid(seg_lv(first_seg(lv), 0)) ||
|
|
lv_is_raid(first_seg(lv)->metadata_lv))) {
|
|
log_error("Thin pool data or metadata volume "
|
|
"must be specified. (E.g. \"%s_tdata\")",
|
|
display_lvname(lv));
|
|
return 0;
|
|
}
|
|
log_error("%s must be a RAID logical volume to perform this action.",
|
|
display_lvname(lv));
|
|
return 0;
|
|
}
|
|
|
|
if (!lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0)) {
|
|
log_error("Unable to send message to an inactive logical volume.");
|
|
return 0;
|
|
}
|
|
|
|
if (!lv_raid_status(lv, &raid_status))
|
|
return_0;
|
|
|
|
if (!raid_status->raid->sync_action) {
|
|
log_error("Kernel driver does not support this action: %s", msg);
|
|
goto out;
|
|
}
|
|
|
|
/*
|
|
* Note that 'dev_manager_raid_message' allows us to pass down any
|
|
* currently valid message. However, this function restricts the
|
|
* number of user available combinations to a minimum. Specifically,
|
|
* "idle" -> "check"
|
|
* "idle" -> "repair"
|
|
* (The state automatically switches to "idle" when a sync process is
|
|
* complete.)
|
|
*/
|
|
if (strcmp(msg, "check") && strcmp(msg, "repair")) {
|
|
/*
|
|
* MD allows "frozen" to operate in a toggling fashion.
|
|
* We could allow this if we like...
|
|
*/
|
|
log_error("\"%s\" is not a supported sync operation.", msg);
|
|
goto out;
|
|
}
|
|
if (strcmp(raid_status->raid->sync_action, "idle")) {
|
|
log_error("%s state is currently \"%s\". Unable to switch to \"%s\".",
|
|
display_lvname(lv), raid_status->raid->sync_action, msg);
|
|
goto out;
|
|
}
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1)))
|
|
return_0;
|
|
|
|
r = dev_manager_raid_message(dm, lv, msg);
|
|
out:
|
|
if (dm)
|
|
dev_manager_destroy(dm);
|
|
dm_pool_destroy(raid_status->mem);
|
|
|
|
return r;
|
|
}
|
|
|
|
int lv_raid_status(const struct logical_volume *lv, struct lv_status_raid **status)
|
|
{
|
|
struct dev_manager *dm;
|
|
int exists;
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1)))
|
|
return_0;
|
|
|
|
if (!dev_manager_raid_status(dm, lv, status, &exists)) {
|
|
dev_manager_destroy(dm);
|
|
if (exists)
|
|
stack;
|
|
return 0;
|
|
}
|
|
/* User has to call dm_pool_destroy(status->mem)! */
|
|
|
|
return 1;
|
|
}
|
|
|
|
int lv_writecache_message(const struct logical_volume *lv, const char *msg)
|
|
{
|
|
int r = 0;
|
|
struct dev_manager *dm;
|
|
|
|
if (!lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0)) {
|
|
log_error("Unable to send message to an inactive logical volume.");
|
|
return 0;
|
|
}
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1)))
|
|
return_0;
|
|
|
|
r = dev_manager_writecache_message(dm, lv, msg);
|
|
|
|
dev_manager_destroy(dm);
|
|
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* Return dm_status_cache for cache volume, accept also cache pool
|
|
*
|
|
* As there are too many variable for cache volumes, and it hard
|
|
* to make good API - so let's obtain dm_status_cache and return
|
|
* all info we have - user just has to release struct after its use.
|
|
*/
|
|
int lv_cache_status(const struct logical_volume *cache_lv,
|
|
struct lv_status_cache **status)
|
|
{
|
|
struct dev_manager *dm;
|
|
struct lv_segment *cache_seg;
|
|
int exists;
|
|
|
|
if (lv_is_cache_pool(cache_lv)) {
|
|
if (dm_list_empty(&cache_lv->segs_using_this_lv) ||
|
|
!(cache_seg = get_only_segment_using_this_lv(cache_lv))) {
|
|
log_error(INTERNAL_ERROR "Cannot check status for unused cache pool %s.",
|
|
display_lvname(cache_lv));
|
|
return 0;
|
|
}
|
|
cache_lv = cache_seg->lv;
|
|
}
|
|
|
|
if (lv_is_pending_delete(cache_lv)) {
|
|
log_error("Cannot check status for deleted cache volume %s.",
|
|
display_lvname(cache_lv));
|
|
return 0;
|
|
}
|
|
|
|
if (!(dm = dev_manager_create(cache_lv->vg->cmd, cache_lv->vg->name, 1)))
|
|
return_0;
|
|
|
|
if (!dev_manager_cache_status(dm, cache_lv, status, &exists)) {
|
|
dev_manager_destroy(dm);
|
|
if (exists)
|
|
stack;
|
|
return 0;
|
|
}
|
|
/* User has to call dm_pool_destroy(status->mem)! */
|
|
|
|
return 1;
|
|
}
|
|
|
|
int lv_thin_pool_status(const struct logical_volume *lv, int flush,
|
|
struct lv_status_thin_pool **thin_pool_status)
|
|
{
|
|
struct dev_manager *dm;
|
|
int exists;
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1)))
|
|
return_0;
|
|
|
|
if (!dev_manager_thin_pool_status(dm, lv, flush, thin_pool_status, &exists)) {
|
|
dev_manager_destroy(dm);
|
|
if (exists)
|
|
stack;
|
|
return 0;
|
|
}
|
|
|
|
/* User has to call dm_pool_destroy(thin_pool_status->mem)! */
|
|
|
|
return 1;
|
|
}
|
|
|
|
int lv_thin_status(const struct logical_volume *lv, int flush,
|
|
struct lv_status_thin **thin_status)
|
|
{
|
|
struct dev_manager *dm;
|
|
int exists;
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1)))
|
|
return_0;
|
|
|
|
if (!dev_manager_thin_status(dm, lv, flush, thin_status, &exists)) {
|
|
dev_manager_destroy(dm);
|
|
if (exists)
|
|
stack;
|
|
return 0;
|
|
}
|
|
|
|
/* User has to call dm_pool_destroy(thin_status->mem)! */
|
|
|
|
return 1;
|
|
}
|
|
|
|
int lv_thin_device_id(const struct logical_volume *lv, uint32_t *device_id)
|
|
{
|
|
struct dev_manager *dm;
|
|
int exists;
|
|
int r;
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1)))
|
|
return_0;
|
|
|
|
if (!(r = dev_manager_thin_device_id(dm, lv, device_id, &exists)))
|
|
if (exists)
|
|
stack;
|
|
|
|
dev_manager_destroy(dm);
|
|
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* lv_vdo_pool_status obtains status information about VDO pool
|
|
*
|
|
* If the 'params' string has been already retrieved, use it.
|
|
* If the mempool already exists, use it.
|
|
*
|
|
*/
|
|
int lv_vdo_pool_status(const struct logical_volume *lv, int flush,
|
|
struct lv_status_vdo **vdo_status)
|
|
{
|
|
struct dev_manager *dm;
|
|
int exists;
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, !lv_is_pvmove(lv))))
|
|
return_0;
|
|
|
|
if (!dev_manager_vdo_pool_status(dm, lv, flush, vdo_status, &exists)) {
|
|
dev_manager_destroy(dm);
|
|
if (exists)
|
|
stack;
|
|
return 0;
|
|
}
|
|
|
|
/* User has to call dm_pool_destroy(vdo_status->mem) */
|
|
|
|
return 1;
|
|
}
|
|
|
|
int lv_vdo_pool_percent(const struct logical_volume *lv, dm_percent_t *percent)
|
|
{
|
|
struct lv_status_vdo *vdo_status;
|
|
|
|
if (!lv_vdo_pool_status(lv, 0, &vdo_status))
|
|
return_0;
|
|
|
|
*percent = vdo_status->usage;
|
|
dm_pool_destroy(vdo_status->mem);
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* lv_vdo_pool_size_config obtains size configuration from active VDO table line
|
|
*
|
|
* If the 'params' string has been already retrieved, use it.
|
|
* If the mempool already exists, use it.
|
|
*
|
|
*/
|
|
int lv_vdo_pool_size_config(const struct logical_volume *lv,
|
|
struct vdo_pool_size_config *cfg)
|
|
{
|
|
struct dev_manager *dm;
|
|
int r;
|
|
|
|
if (!lv_info(lv->vg->cmd, lv, 1, NULL, 0, 0))
|
|
return 1; /* Inactive VDO pool -> no runtime config */
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, !lv_is_pvmove(lv))))
|
|
return_0;
|
|
|
|
r = dev_manager_vdo_pool_size_config(dm, lv, cfg);
|
|
|
|
dev_manager_destroy(dm);
|
|
|
|
return r;
|
|
}
|
|
|
|
static int _lv_active(struct cmd_context *cmd, const struct logical_volume *lv)
|
|
{
|
|
struct lvinfo info;
|
|
|
|
if (!lv_info(cmd, lv, 0, &info, 0, 0)) {
|
|
log_debug_activation("Cannot determine activation status of %s%s.",
|
|
display_lvname(lv),
|
|
activation() ? "" : " (no device driver)");
|
|
return 0;
|
|
}
|
|
|
|
return info.exists;
|
|
}
|
|
|
|
static int _lv_open_count(struct cmd_context *cmd, const struct logical_volume *lv)
|
|
{
|
|
struct lvinfo info;
|
|
|
|
if (!lv_info(cmd, lv, 0, &info, 1, 0)) {
|
|
stack;
|
|
return -1;
|
|
}
|
|
|
|
return info.open_count;
|
|
}
|
|
|
|
static int _lv_activate_lv(const struct logical_volume *lv, struct lv_activate_opts *laopts)
|
|
{
|
|
int r;
|
|
struct dev_manager *dm;
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, !lv_is_pvmove(lv))))
|
|
return_0;
|
|
|
|
if (!(r = dev_manager_activate(dm, lv, laopts)))
|
|
stack;
|
|
|
|
dev_manager_destroy(dm);
|
|
return r;
|
|
}
|
|
|
|
static int _lv_preload(const struct logical_volume *lv, struct lv_activate_opts *laopts,
|
|
int *flush_required)
|
|
{
|
|
int r = 0;
|
|
struct dev_manager *dm;
|
|
int old_readonly = laopts->read_only;
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, !lv_is_pvmove(lv))))
|
|
goto_out;
|
|
|
|
laopts->read_only = _passes_readonly_filter(lv->vg->cmd, lv);
|
|
|
|
if (!(r = dev_manager_preload(dm, lv, laopts, flush_required)))
|
|
stack;
|
|
|
|
dev_manager_destroy(dm);
|
|
|
|
laopts->read_only = old_readonly;
|
|
out:
|
|
return r;
|
|
}
|
|
|
|
static int _lv_deactivate(const struct logical_volume *lv)
|
|
{
|
|
int r;
|
|
struct dev_manager *dm;
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, 1)))
|
|
return_0;
|
|
|
|
if (!(r = dev_manager_deactivate(dm, lv)))
|
|
stack;
|
|
|
|
dev_manager_destroy(dm);
|
|
return r;
|
|
}
|
|
|
|
static int _lv_suspend_lv(const struct logical_volume *lv, struct lv_activate_opts *laopts,
|
|
int lockfs, int flush_required)
|
|
{
|
|
int r;
|
|
struct dev_manager *dm;
|
|
|
|
laopts->read_only = _passes_readonly_filter(lv->vg->cmd, lv);
|
|
|
|
/*
|
|
* When we are asked to manipulate (normally suspend/resume) the PVMOVE
|
|
* device directly, we don't want to touch the devices that use it.
|
|
*/
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name, !lv_is_pvmove(lv))))
|
|
return_0;
|
|
|
|
if (!(r = dev_manager_suspend(dm, lv, laopts, lockfs, flush_required)))
|
|
stack;
|
|
|
|
dev_manager_destroy(dm);
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* These two functions return the number of visible LVs in the state,
|
|
* or -1 on error. FIXME Check this.
|
|
*/
|
|
int lvs_in_vg_activated(const struct volume_group *vg)
|
|
{
|
|
struct lv_list *lvl;
|
|
int count = 0;
|
|
|
|
if (!activation())
|
|
return 0;
|
|
|
|
dm_list_iterate_items(lvl, &vg->lvs)
|
|
if (lv_is_visible(lvl->lv))
|
|
count += (_lv_active(vg->cmd, lvl->lv) == 1);
|
|
|
|
log_debug_activation("Counted %d active LVs in VG %s", count, vg->name);
|
|
|
|
return count;
|
|
}
|
|
|
|
int lvs_in_vg_opened(const struct volume_group *vg)
|
|
{
|
|
const struct lv_list *lvl;
|
|
int count = 0;
|
|
|
|
if (!activation())
|
|
return 0;
|
|
|
|
dm_list_iterate_items(lvl, &vg->lvs)
|
|
if (lv_is_visible(lvl->lv))
|
|
count += (_lv_open_count(vg->cmd, lvl->lv) > 0);
|
|
|
|
log_debug_activation("Counted %d open LVs in VG %s.", count, vg->name);
|
|
|
|
return count;
|
|
}
|
|
|
|
/*
|
|
* Check if "raid4" @segtype is supported by kernel.
|
|
*
|
|
* if segment type is not raid4, return 1.
|
|
*/
|
|
int raid4_is_supported(struct cmd_context *cmd, const struct segment_type *segtype)
|
|
{
|
|
unsigned attrs = 0;
|
|
|
|
if (segtype_is_raid4(segtype) &&
|
|
(!segtype->ops->target_present ||
|
|
!segtype->ops->target_present(cmd, NULL, &attrs) ||
|
|
!(attrs & RAID_FEATURE_RAID4))) {
|
|
log_error("RAID module does not support RAID4.");
|
|
return 0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* The VG lock must be held to call this function.
|
|
*
|
|
* Returns: 0 or 1
|
|
*/
|
|
int lv_is_active(const struct logical_volume *lv)
|
|
{
|
|
return _lv_active(lv->vg->cmd, lv);
|
|
}
|
|
|
|
#ifdef DMEVENTD
|
|
static struct dm_event_handler *_create_dm_event_handler(struct cmd_context *cmd, const char *dmuuid, const char *dso,
|
|
const int timeout, enum dm_event_mask mask)
|
|
{
|
|
struct dm_event_handler *dmevh;
|
|
|
|
if (!(dmevh = dm_event_handler_create()))
|
|
return_NULL;
|
|
|
|
if (!cmd->default_settings.dmeventd_executable)
|
|
cmd->default_settings.dmeventd_executable = find_config_tree_str(cmd, dmeventd_executable_CFG, NULL);
|
|
|
|
if (dm_event_handler_set_dmeventd_path(dmevh, cmd->default_settings.dmeventd_executable))
|
|
goto_bad;
|
|
|
|
if (dso && dm_event_handler_set_dso(dmevh, dso))
|
|
goto_bad;
|
|
|
|
if (dm_event_handler_set_uuid(dmevh, dmuuid))
|
|
goto_bad;
|
|
|
|
dm_event_handler_set_timeout(dmevh, timeout);
|
|
dm_event_handler_set_event_mask(dmevh, mask);
|
|
|
|
return dmevh;
|
|
|
|
bad:
|
|
dm_event_handler_destroy(dmevh);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
char *get_monitor_dso_path(struct cmd_context *cmd, int id)
|
|
{
|
|
const char *libpath = find_config_tree_str(cmd, id, NULL);
|
|
char path[PATH_MAX];
|
|
|
|
get_shared_library_path(cmd, libpath, path, sizeof(path));
|
|
|
|
return strdup(path);
|
|
}
|
|
|
|
static char *_build_target_uuid(struct cmd_context *cmd, const struct logical_volume *lv)
|
|
{
|
|
return build_dm_uuid(cmd->mem, lv, lv_layer(lv));
|
|
}
|
|
|
|
static int _device_registered_with_dmeventd(struct cmd_context *cmd,
|
|
const struct logical_volume *lv,
|
|
const char **dso,
|
|
int *pending, int *monitored)
|
|
{
|
|
char *uuid;
|
|
enum dm_event_mask evmask;
|
|
struct dm_event_handler *dmevh;
|
|
int r;
|
|
|
|
*pending = 0;
|
|
*monitored = 0;
|
|
|
|
if (!(uuid = _build_target_uuid(cmd, lv)))
|
|
return_0;
|
|
|
|
if (!(dmevh = _create_dm_event_handler(cmd, uuid, NULL, 0, DM_EVENT_ALL_ERRORS)))
|
|
return_0;
|
|
|
|
if ((r = dm_event_get_registered_device(dmevh, 0))) {
|
|
if (r == -ENOENT) {
|
|
r = 1;
|
|
goto out;
|
|
}
|
|
r = 0;
|
|
goto_out;
|
|
}
|
|
|
|
/* FIXME: why do we care which 'dso' is monitoring? */
|
|
if (dso && (*dso = dm_event_handler_get_dso(dmevh)) &&
|
|
!(*dso = dm_pool_strdup(cmd->mem, *dso))) {
|
|
r = 0;
|
|
goto_out;
|
|
}
|
|
|
|
evmask = dm_event_handler_get_event_mask(dmevh);
|
|
if (evmask & DM_EVENT_REGISTRATION_PENDING) {
|
|
*pending = 1;
|
|
evmask &= ~DM_EVENT_REGISTRATION_PENDING;
|
|
}
|
|
|
|
*monitored = evmask;
|
|
r = 1;
|
|
out:
|
|
dm_event_handler_destroy(dmevh);
|
|
|
|
return r;
|
|
}
|
|
|
|
int target_registered_with_dmeventd(struct cmd_context *cmd, const char *dso,
|
|
const struct logical_volume *lv,
|
|
int *pending, int *monitored)
|
|
{
|
|
char *uuid;
|
|
enum dm_event_mask evmask;
|
|
struct dm_event_handler *dmevh;
|
|
int r;
|
|
|
|
*pending = 0;
|
|
*monitored = 0;
|
|
|
|
if (!dso)
|
|
return_0;
|
|
|
|
if (!(uuid = _build_target_uuid(cmd, lv)))
|
|
return_0;
|
|
|
|
if (!(dmevh = _create_dm_event_handler(cmd, uuid, dso, 0, DM_EVENT_ALL_ERRORS)))
|
|
return_0;
|
|
|
|
if ((r = dm_event_get_registered_device(dmevh, 0))) {
|
|
if (r == -ENOENT) {
|
|
r = 1;
|
|
goto out;
|
|
}
|
|
r = 0;
|
|
goto_out;
|
|
}
|
|
|
|
evmask = dm_event_handler_get_event_mask(dmevh);
|
|
if (evmask & DM_EVENT_REGISTRATION_PENDING) {
|
|
*pending = 1;
|
|
evmask &= ~DM_EVENT_REGISTRATION_PENDING;
|
|
}
|
|
|
|
*monitored = evmask;
|
|
r = 1;
|
|
out:
|
|
dm_event_handler_destroy(dmevh);
|
|
|
|
return r;
|
|
}
|
|
|
|
int target_register_events(struct cmd_context *cmd, const char *dso, const struct logical_volume *lv,
|
|
int evmask __attribute__((unused)), int set, int timeout)
|
|
{
|
|
char *uuid;
|
|
struct dm_event_handler *dmevh;
|
|
int r;
|
|
|
|
if (!dso)
|
|
return_0;
|
|
|
|
/* We always monitor the "real" device, never the "snapshot-origin" itself. */
|
|
if (!(uuid = _build_target_uuid(cmd, lv)))
|
|
return_0;
|
|
|
|
if (!(dmevh = _create_dm_event_handler(cmd, uuid, dso, timeout,
|
|
DM_EVENT_ALL_ERRORS | (timeout ? DM_EVENT_TIMEOUT : 0))))
|
|
return_0;
|
|
|
|
r = set ? dm_event_register_handler(dmevh) : dm_event_unregister_handler(dmevh);
|
|
|
|
dm_event_handler_destroy(dmevh);
|
|
|
|
if (!r)
|
|
return_0;
|
|
|
|
log_verbose("%s %s for events", set ? "Monitored" : "Unmonitored", uuid);
|
|
|
|
return 1;
|
|
}
|
|
|
|
#endif
|
|
|
|
/*
|
|
* Returns 0 if an attempt to (un)monitor the device failed.
|
|
* Returns 1 otherwise.
|
|
*/
|
|
int monitor_dev_for_events(struct cmd_context *cmd, const struct logical_volume *lv,
|
|
const struct lv_activate_opts *laopts, int monitor)
|
|
{
|
|
#ifdef DMEVENTD
|
|
int i, pending = 0, monitored = 0;
|
|
int r = 1;
|
|
struct dm_list *snh, *snht;
|
|
struct lv_segment *seg;
|
|
struct lv_segment *log_seg;
|
|
int (*monitor_fn) (struct lv_segment *s, int e);
|
|
uint32_t s;
|
|
const struct lv_activate_opts zlaopts = { 0 };
|
|
struct lv_activate_opts mirr_laopts = { .origin_only = 1 };
|
|
struct lvinfo info;
|
|
const char *dso = NULL;
|
|
int new_unmonitor;
|
|
|
|
if (!laopts)
|
|
laopts = &zlaopts;
|
|
else
|
|
mirr_laopts.read_only = laopts->read_only;
|
|
|
|
/* skip dmeventd code altogether */
|
|
if (dmeventd_monitor_mode() == DMEVENTD_MONITOR_IGNORE)
|
|
return 1;
|
|
|
|
/*
|
|
* Nothing to do if dmeventd configured not to be used.
|
|
*/
|
|
if (monitor && !dmeventd_monitor_mode())
|
|
return 1;
|
|
|
|
if (sigint_caught())
|
|
return_0;
|
|
|
|
/*
|
|
* Activation of unused cache-pool activates metadata device as
|
|
* a public LV for clearing purpose.
|
|
* FIXME:
|
|
* As VG lock is held across whole operation unmonitored volume
|
|
* is usually OK since dmeventd couldn't do anything.
|
|
* However in case command would have crashed, such LV is
|
|
* left unmonitored and may potentially require dmeventd.
|
|
*/
|
|
if (lv_is_cache_pool_data(lv) || lv_is_cache_pool_metadata(lv)) {
|
|
if (!(seg = find_pool_seg(first_seg(lv))))
|
|
return_0;
|
|
if (!lv_is_used_cache_pool(seg->lv)) {
|
|
log_debug_activation("Skipping %smonitor of %s.%s",
|
|
(monitor) ? "" : "un", display_lvname(lv),
|
|
(monitor) ? " Cache pool activation for clearing only." : "");
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Allow to unmonitor thin pool via explicit pool unmonitor
|
|
* or unmonitor before the last thin pool user deactivation
|
|
* Skip unmonitor, if invoked via deactivation of thin volume
|
|
* and there is another thin pool user (open_count > 1)
|
|
* FIXME think about watch ruler influence.
|
|
*/
|
|
if (laopts->skip_in_use && lv_is_thin_pool(lv) &&
|
|
lv_info(lv->vg->cmd, lv, 1, &info, 1, 0) && (info.open_count > 1)) {
|
|
log_debug_activation("Skipping unmonitor of opened %s (open:%d)",
|
|
display_lvname(lv), info.open_count);
|
|
return 1;
|
|
}
|
|
|
|
/* Do not monitor snapshot that already covers origin */
|
|
if (monitor && lv_is_cow_covering_origin(lv)) {
|
|
log_debug_activation("Skipping monitor of snapshot larger "
|
|
"then origin %s.", display_lvname(lv));
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* In case of a snapshot device, we monitor lv->snapshot->lv,
|
|
* not the actual LV itself.
|
|
*/
|
|
if (lv_is_cow(lv) && (laopts->no_merging || !lv_is_merging_cow(lv) ||
|
|
lv_has_target_type(lv->vg->cmd->mem, lv, NULL, TARGET_NAME_SNAPSHOT))) {
|
|
if (!(r = monitor_dev_for_events(cmd, lv->snapshot->lv, NULL, monitor)))
|
|
stack;
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* In case this LV is a snapshot origin, we instead monitor
|
|
* each of its respective snapshots. The origin itself may
|
|
* also need to be monitored if it is a mirror, for example,
|
|
* so fall through to process it afterwards.
|
|
* Before monitoring snapshots verify origin is active as with
|
|
* external origin only read-only -real device can be active.
|
|
*/
|
|
if (!laopts->origin_only && lv_is_origin(lv) && lv_info(lv->vg->cmd, lv, 0, NULL, 0, 0))
|
|
dm_list_iterate_safe(snh, snht, &lv->snapshot_segs)
|
|
if (!monitor_dev_for_events(cmd, dm_list_struct_base(snh,
|
|
struct lv_segment, origin_list)->cow, NULL, monitor)) {
|
|
stack;
|
|
r = 0;
|
|
}
|
|
|
|
/*
|
|
* If the volume is mirrored and its log is also mirrored, monitor
|
|
* the log volume as well.
|
|
*/
|
|
if ((seg = first_seg(lv)) != NULL && seg->log_lv != NULL &&
|
|
(log_seg = first_seg(seg->log_lv)) != NULL &&
|
|
seg_is_mirrored(log_seg))
|
|
if (!monitor_dev_for_events(cmd, seg->log_lv, NULL, monitor)) {
|
|
stack;
|
|
r = 0;
|
|
}
|
|
|
|
dm_list_iterate_items(seg, &lv->segments) {
|
|
if (sigint_caught()) {
|
|
stack;
|
|
r = 0;
|
|
break;
|
|
}
|
|
|
|
/* Recurse for AREA_LV */
|
|
for (s = 0; s < seg->area_count; s++) {
|
|
if (seg_type(seg, s) != AREA_LV)
|
|
continue;
|
|
if (!monitor_dev_for_events(cmd, seg_lv(seg, s), NULL,
|
|
monitor)) {
|
|
stack;
|
|
r = 0;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* If requested unmonitoring of thin volume, preserve skip_in_use flag.
|
|
*
|
|
* FIXME: code here looks like _lv_postorder()
|
|
*/
|
|
if (seg->pool_lv &&
|
|
!monitor_dev_for_events(cmd, seg->pool_lv,
|
|
(!monitor) ? laopts : NULL, monitor)) {
|
|
stack;
|
|
r = 0;
|
|
}
|
|
|
|
if (seg->external_lv &&
|
|
!monitor_dev_for_events(cmd, seg->external_lv,
|
|
(!monitor) ? laopts : NULL, monitor)) {
|
|
stack;
|
|
r = 0;
|
|
}
|
|
|
|
if (seg->metadata_lv &&
|
|
!monitor_dev_for_events(cmd, seg->metadata_lv, NULL, monitor)) {
|
|
stack;
|
|
r = 0;
|
|
}
|
|
|
|
if (!seg_monitored(seg) ||
|
|
(seg->status & PVMOVE) ||
|
|
!seg->segtype->ops->target_monitored) /* doesn't support registration */
|
|
continue;
|
|
|
|
if (!monitor) {
|
|
/* When unmonitoring, obtain existing dso being used. */
|
|
if (!_device_registered_with_dmeventd(cmd, seg_is_snapshot(seg) ? seg->cow : seg->lv,
|
|
&dso, &pending, &monitored)) {
|
|
log_warn("WARNING: Failed to %smonitor %s.",
|
|
monitor ? "" : "un",
|
|
display_lvname(seg_is_snapshot(seg) ? seg->cow : seg->lv));
|
|
return 0;
|
|
}
|
|
} else if (!seg->segtype->ops->target_monitored(seg, &pending, &monitored)) {
|
|
log_warn("WARNING: Failed to %smonitor %s.",
|
|
monitor ? "" : "un",
|
|
display_lvname(seg->lv));
|
|
return 0;
|
|
}
|
|
|
|
/* FIXME: We should really try again if pending */
|
|
monitored = (pending) ? 0 : monitored;
|
|
|
|
monitor_fn = NULL;
|
|
new_unmonitor = 0;
|
|
|
|
if (monitor) {
|
|
if (monitored)
|
|
log_verbose("%s already monitored.", display_lvname(lv));
|
|
else if (seg->segtype->ops->target_monitor_events) {
|
|
log_very_verbose("Monitoring %s with %s.%s", display_lvname(lv),
|
|
seg->segtype->dso,
|
|
test_mode() ? " [Test mode: skipping this]" : "");
|
|
monitor_fn = seg->segtype->ops->target_monitor_events;
|
|
}
|
|
} else {
|
|
if (!monitored)
|
|
log_verbose("%s already not monitored.", display_lvname(lv));
|
|
else if (dso && *dso) {
|
|
/*
|
|
* Divert unmonitor away from code that depends on the new segment
|
|
* type instead of the existing one if it's changing.
|
|
*/
|
|
log_verbose("Not monitoring %s with %s%s", display_lvname(lv), dso, test_mode() ? " [Test mode: skipping this]" : "");
|
|
new_unmonitor = 1;
|
|
}
|
|
}
|
|
|
|
/* FIXME Test mode should really continue a bit further. */
|
|
if (test_mode())
|
|
continue;
|
|
|
|
if (new_unmonitor) {
|
|
if (!target_register_events(cmd, dso, seg_is_snapshot(seg) ? seg->cow : lv, 0, 0, 10)) {
|
|
log_warn("WARNING: %s: segment unmonitoring failed.",
|
|
display_lvname(lv));
|
|
return 0;
|
|
}
|
|
} else if (monitor_fn) {
|
|
/* FIXME specify events */
|
|
if (!monitor_fn(seg, 0)) {
|
|
log_warn("WARNING: %s: %s segment monitoring function failed.",
|
|
display_lvname(lv), lvseg_name(seg));
|
|
return 0;
|
|
}
|
|
} else
|
|
continue;
|
|
|
|
if (!vg_write_lock_held() && lv_is_mirror(lv)) {
|
|
mirr_laopts.exclusive = lv_is_active(lv) ? 1 : 0;
|
|
/*
|
|
* Commands vgchange and lvchange do use read-only lock when changing
|
|
* monitoring (--monitor y|n). All other use cases hold 'write-lock'
|
|
* so they skip this dm mirror table refreshing step.
|
|
*/
|
|
if (!_lv_activate_lv(lv, &mirr_laopts)) {
|
|
stack;
|
|
r = 0;
|
|
}
|
|
}
|
|
|
|
/* Check [un]monitor results */
|
|
/* Try a couple times if pending, but not forever... */
|
|
for (i = 0;; i++) {
|
|
pending = 0;
|
|
if (!seg->segtype->ops->target_monitored(seg, &pending, &monitored)) {
|
|
stack;
|
|
r = 0;
|
|
break;
|
|
}
|
|
if (!pending || i >= 40)
|
|
break;
|
|
log_very_verbose("%s %smonitoring still pending: waiting...",
|
|
display_lvname(lv), monitor ? "" : "un");
|
|
if (interruptible_usleep(10000 * i)) {
|
|
stack;
|
|
r = 0;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (r)
|
|
r = (monitored && monitor) || (!monitored && !monitor);
|
|
}
|
|
|
|
if (!r && !error_message_produced())
|
|
log_warn("WARNING: %sonitoring %s failed.", monitor ? "M" : "Not m",
|
|
display_lvname(lv));
|
|
return r;
|
|
#else
|
|
return 1;
|
|
#endif
|
|
}
|
|
|
|
struct detached_lv_data {
|
|
const struct volume_group *vg_pre;
|
|
struct lv_activate_opts *laopts;
|
|
int *flush_required;
|
|
};
|
|
|
|
static int _preload_detached_lv(struct logical_volume *lv, void *data)
|
|
{
|
|
struct detached_lv_data *detached = data;
|
|
struct logical_volume *lv_pre;
|
|
|
|
/* Check if the LV was 'hidden' (non-toplevel) in committed metadata
|
|
* and becomes 'visible' (toplevel) in precommitted metadata */
|
|
if (!lv_is_visible(lv) &&
|
|
(lv_pre = find_lv_in_vg_by_lvid(detached->vg_pre, &lv->lvid)) &&
|
|
lv_is_visible(lv_pre)) {
|
|
log_debug_activation("Preloading detached hidden volume %s as visible volume %s.",
|
|
display_lvname(lv), display_lvname(lv_pre));
|
|
if (!_lv_preload(lv_pre, detached->laopts, detached->flush_required))
|
|
return_0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int _lv_suspend(struct cmd_context *cmd, const char *lvid_s,
|
|
struct lv_activate_opts *laopts, int error_if_not_suspended,
|
|
const struct logical_volume *lv, const struct logical_volume *lv_pre)
|
|
{
|
|
const struct logical_volume *pvmove_lv = NULL;
|
|
struct logical_volume *lv_pre_tmp, *lv_tmp;
|
|
struct seg_list *sl;
|
|
struct lv_segment *snap_seg;
|
|
struct lvinfo info;
|
|
int r = 0, lockfs = 0, flush_required = 0;
|
|
struct detached_lv_data detached;
|
|
struct dm_pool *mem = NULL;
|
|
struct dm_list suspend_lvs;
|
|
struct lv_list *lvl;
|
|
int found;
|
|
|
|
if (!activation())
|
|
return 1;
|
|
|
|
if (test_mode()) {
|
|
_skip("Suspending %s%s.", display_lvname(lv),
|
|
laopts->origin_only ? " origin without snapshots" : "");
|
|
r = 1;
|
|
goto out;
|
|
}
|
|
|
|
if (!lv_info(cmd, lv, laopts->origin_only, &info, 0, 0))
|
|
goto_out;
|
|
|
|
if (!info.exists || info.suspended) {
|
|
if (!error_if_not_suspended) {
|
|
r = 1;
|
|
if (info.suspended)
|
|
critical_section_inc(cmd, "already suspended");
|
|
}
|
|
goto out;
|
|
}
|
|
|
|
lv_calculate_readahead(lv, NULL);
|
|
|
|
/* Ignore origin_only unless LV is origin in both old and new metadata */
|
|
/* or LV is thin or thin pool volume */
|
|
if (laopts->origin_only &&
|
|
!lv_is_thin_volume(lv) && !lv_is_thin_pool(lv) &&
|
|
!(lv_is_origin(lv) && lv_is_origin(lv_pre))) {
|
|
log_debug_activation("Not using origin only for suspend of %s.",
|
|
display_lvname(lv));
|
|
laopts->origin_only = 0;
|
|
}
|
|
|
|
/*
|
|
* Preload devices for the LV.
|
|
* If the PVMOVE LV is being removed, it's only present in the old
|
|
* metadata and not the new, so we must explicitly add the new
|
|
* tables for all the changed LVs here, as the relationships
|
|
* are not found by walking the new metadata.
|
|
*/
|
|
if (lv_is_locked(lv) && !lv_is_locked(lv_pre) &&
|
|
(pvmove_lv = find_pvmove_lv_in_lv(lv))) {
|
|
/* Preload all the LVs above the PVMOVE LV */
|
|
dm_list_iterate_items(sl, &pvmove_lv->segs_using_this_lv) {
|
|
if (!(lv_pre_tmp = find_lv(lv_pre->vg, sl->seg->lv->name))) {
|
|
log_error(INTERNAL_ERROR "LV %s missing from preload metadata.",
|
|
display_lvname(sl->seg->lv));
|
|
goto out;
|
|
}
|
|
if (!_lv_preload(lv_pre_tmp, laopts, &flush_required))
|
|
goto_out;
|
|
}
|
|
/* Now preload the PVMOVE LV itself */
|
|
if (!(lv_pre_tmp = find_lv(lv_pre->vg, pvmove_lv->name))) {
|
|
log_error(INTERNAL_ERROR "LV %s missing from preload metadata.",
|
|
display_lvname(pvmove_lv));
|
|
goto out;
|
|
}
|
|
if (!_lv_preload(lv_pre_tmp, laopts, &flush_required))
|
|
goto_out;
|
|
|
|
/* Suspending 1st. LV above PVMOVE suspends whole tree */
|
|
dm_list_iterate_items(sl, &pvmove_lv->segs_using_this_lv) {
|
|
lv = sl->seg->lv;
|
|
break;
|
|
}
|
|
} else {
|
|
if (!_lv_preload(lv_pre, laopts, &flush_required))
|
|
/* FIXME Revert preloading */
|
|
goto_out;
|
|
|
|
/*
|
|
* Search for existing LVs that have become detached and preload them.
|
|
*/
|
|
detached.vg_pre = lv_pre->vg;
|
|
detached.laopts = laopts;
|
|
detached.flush_required = &flush_required;
|
|
|
|
if (!for_each_sub_lv((struct logical_volume *)lv, &_preload_detached_lv, &detached))
|
|
goto_out;
|
|
|
|
/*
|
|
* Preload any snapshots that are being removed.
|
|
*/
|
|
if (!laopts->origin_only && lv_is_origin(lv)) {
|
|
dm_list_iterate_items_gen(snap_seg, &lv->snapshot_segs, origin_list) {
|
|
if (!(lv_pre_tmp = find_lv_in_vg_by_lvid(lv_pre->vg, &snap_seg->cow->lvid))) {
|
|
log_error(INTERNAL_ERROR "LV %s (%s) missing from preload metadata.",
|
|
display_lvname(snap_seg->cow),
|
|
snap_seg->cow->lvid.id[1].uuid);
|
|
goto out;
|
|
}
|
|
if (!lv_is_cow(lv_pre_tmp) &&
|
|
!_lv_preload(lv_pre_tmp, laopts, &flush_required))
|
|
goto_out;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Flush is ATM required for the tested cases
|
|
* NOTE: Mirror repair requires noflush for proper repair!
|
|
* TODO: Relax this limiting condition further */
|
|
if (!flush_required &&
|
|
(lv_is_pvmove(lv) || pvmove_lv ||
|
|
(!lv_is_mirror(lv) &&
|
|
!lv_is_thin_volume(lv) &&
|
|
!lv_is_thin_pool(lv) &&
|
|
!lv_is_vdo(lv) &&
|
|
!lv_is_vdo_pool(lv)))) {
|
|
log_debug_activation("Requiring flush for LV %s.", display_lvname(lv));
|
|
flush_required = 1;
|
|
}
|
|
|
|
if (!monitor_dev_for_events(cmd, lv, laopts, 0))
|
|
/* FIXME Consider aborting here */
|
|
stack;
|
|
|
|
/* Require fs synchronization when taking a thin snapshot */
|
|
if (laopts->origin_only && lv_is_thin_volume(lv) && lv_is_thin_volume(lv_pre))
|
|
lockfs = 1;
|
|
/* Require fs synchronization when taking a thick snapshot */
|
|
else if (!laopts->origin_only &&
|
|
(lv_is_origin(lv_pre) || lv_is_cow(lv_pre)))
|
|
lockfs = 1;
|
|
/* Require fs synchronization when converting a non-thin LV to a thin LV or
|
|
* a non/thin LV with/out external origin to a thin LV with external origin LV. */
|
|
else if (!laopts->origin_only &&
|
|
lv_is_thin_volume(lv_pre) && /* new LV is a Thin */
|
|
(!lv_is_thin_volume(lv) || /* and either the existing LV is NOT a Thin */
|
|
(first_seg(lv_pre)->external_lv && /* or the existing LV IS Thin and the new LV is Thin with the external origin */
|
|
(!first_seg(lv)->external_lv || /* and check if existing Thin is either without the external origin */
|
|
memcmp(&first_seg(lv_pre)->external_lv->lvid.id[1], /* or it uses a different external origin */
|
|
&first_seg(lv)->external_lv->lvid.id[1], ID_LEN) != 0))))
|
|
lockfs = 1;
|
|
|
|
if (!lv_is_locked(lv) && lv_is_locked(lv_pre) &&
|
|
(pvmove_lv = find_pvmove_lv_in_lv(lv_pre))) {
|
|
/*
|
|
* When starting PVMOVE, suspend participating LVs first
|
|
* with committed metadata by looking at precommited pvmove list.
|
|
* In committed metadata these LVs are not connected in any way.
|
|
*
|
|
* TODO: prepare list of LVs needed to be suspended and pass them
|
|
* via 'struct laopts' directly to _lv_suspend_lv() and handle this
|
|
* with a single 'dmtree' call.
|
|
*/
|
|
if (!(mem = dm_pool_create("suspend_lvs", 128)))
|
|
goto_out;
|
|
|
|
/* Prepare list of all LVs for suspend ahead */
|
|
dm_list_init(&suspend_lvs);
|
|
dm_list_iterate_items(sl, &pvmove_lv->segs_using_this_lv) {
|
|
lv_tmp = sl->seg->lv;
|
|
if (lv_is_cow(lv_tmp))
|
|
/* Never suspend COW, always has to be origin */
|
|
lv_tmp = origin_from_cow(lv_tmp);
|
|
found = 0;
|
|
dm_list_iterate_items(lvl, &suspend_lvs)
|
|
if (strcmp(lvl->lv->name, lv_tmp->name) == 0) {
|
|
found = 1;
|
|
break;
|
|
}
|
|
if (found)
|
|
continue; /* LV is already in the list */
|
|
if (!(lvl = dm_pool_alloc(mem, sizeof(*lvl)))) {
|
|
log_error("lv_list alloc failed.");
|
|
goto out;
|
|
}
|
|
/* Look for precommitted LV name in commmitted VG */
|
|
if (!(lvl->lv = find_lv(lv->vg, lv_tmp->name))) {
|
|
log_error(INTERNAL_ERROR "LV %s missing from preload metadata.",
|
|
display_lvname(lv_tmp));
|
|
goto out;
|
|
}
|
|
dm_list_add(&suspend_lvs, &lvl->list);
|
|
}
|
|
|
|
critical_section_inc(cmd, "suspending");
|
|
|
|
dm_list_iterate_items(lvl, &suspend_lvs)
|
|
if (!_lv_suspend_lv(lvl->lv, laopts, lockfs, 1)) {
|
|
critical_section_dec(cmd, "failed suspend");
|
|
goto_out; /* FIXME: resume on recovery path? */
|
|
}
|
|
|
|
} else { /* Standard suspend */
|
|
critical_section_inc(cmd, "suspending");
|
|
|
|
if (!_lv_suspend_lv(lv, laopts, lockfs, flush_required)) {
|
|
critical_section_dec(cmd, "failed suspend");
|
|
goto_out;
|
|
}
|
|
}
|
|
|
|
r = 1;
|
|
out:
|
|
if (mem)
|
|
dm_pool_destroy(mem);
|
|
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* In a cluster, set exclusive to indicate that only one node is using the
|
|
* device. Any preloaded tables may then use non-clustered targets.
|
|
*
|
|
* Returns success if the device is not active
|
|
*/
|
|
int lv_suspend_if_active(struct cmd_context *cmd, const char *lvid_s, unsigned origin_only, unsigned exclusive,
|
|
const struct logical_volume *lv, const struct logical_volume *lv_pre)
|
|
{
|
|
struct lv_activate_opts laopts = {
|
|
.exclusive = exclusive,
|
|
.origin_only = origin_only
|
|
};
|
|
|
|
return _lv_suspend(cmd, lvid_s, &laopts, 0, lv, lv_pre);
|
|
}
|
|
|
|
static int _check_suspended_lv(struct logical_volume *lv, void *data)
|
|
{
|
|
struct lvinfo info;
|
|
|
|
if (lv_info(lv->vg->cmd, lv, 0, &info, 0, 0) && info.exists && info.suspended) {
|
|
log_debug_activation("Found suspended LV %s in critical section().", display_lvname(lv));
|
|
return 0; /* There is suspended subLV in the tree */
|
|
}
|
|
|
|
if (lv_layer(lv) && lv_info(lv->vg->cmd, lv, 1, &info, 0, 0) && info.exists && info.suspended) {
|
|
log_debug_activation("Found suspended layered LV %s in critical section().", display_lvname(lv));
|
|
return 0; /* There is suspended subLV in the tree */
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int _lv_resume(struct cmd_context *cmd, const char *lvid_s,
|
|
struct lv_activate_opts *laopts, int error_if_not_active,
|
|
const struct logical_volume *lv)
|
|
{
|
|
struct dm_list *snh;
|
|
struct lvinfo info;
|
|
int r = 0;
|
|
|
|
if (!activation())
|
|
return 1;
|
|
|
|
if (!lv_is_origin(lv) && !lv_is_thin_volume(lv) && !lv_is_thin_pool(lv))
|
|
laopts->origin_only = 0;
|
|
|
|
if (test_mode()) {
|
|
_skip("Resuming %s%s%s.", display_lvname(lv),
|
|
laopts->origin_only ? " without snapshots" : "",
|
|
laopts->revert ? " (reverting)" : "");
|
|
r = 1;
|
|
goto out;
|
|
}
|
|
|
|
log_debug_activation("Resuming LV %s%s%s%s.", display_lvname(lv),
|
|
error_if_not_active ? "" : " if active",
|
|
laopts->origin_only ?
|
|
(lv_is_thin_pool(lv) ? " pool only" :
|
|
lv_is_thin_volume(lv) ? " thin only" : " without snapshots") : "",
|
|
laopts->revert ? " (reverting)" : "");
|
|
|
|
if (laopts->revert)
|
|
goto needs_resume;
|
|
|
|
if (!lv_info(cmd, lv, laopts->origin_only, &info, 0, 0))
|
|
goto_out;
|
|
|
|
if (!info.exists || !info.suspended) {
|
|
if (error_if_not_active)
|
|
goto_out;
|
|
|
|
/* ATM only thin-pool with origin-only suspend does not really suspend anything
|
|
* it's used only for message passing to thin-pool */
|
|
if (laopts->origin_only && lv_is_thin_pool(lv))
|
|
critical_section_dec(cmd, "resumed");
|
|
|
|
if (!info.suspended && critical_section()) {
|
|
/* Validation check if any subLV is suspended */
|
|
if (!laopts->origin_only && lv_is_origin(lv)) {
|
|
/* Check all snapshots for this origin LV */
|
|
dm_list_iterate(snh, &lv->snapshot_segs)
|
|
if (!_check_suspended_lv(dm_list_struct_base(snh, struct lv_segment, origin_list)->cow, NULL))
|
|
goto needs_resume; /* Found suspended snapshot */
|
|
}
|
|
if ((r = for_each_sub_lv((struct logical_volume *)lv, &_check_suspended_lv, NULL)))
|
|
goto out; /* Nothing was found suspended */
|
|
} else {
|
|
r = 1;
|
|
goto out;
|
|
}
|
|
}
|
|
needs_resume:
|
|
laopts->read_only = _passes_readonly_filter(cmd, lv);
|
|
laopts->resuming = 1;
|
|
|
|
if (!_lv_activate_lv(lv, laopts))
|
|
goto_out;
|
|
|
|
critical_section_dec(cmd, "resumed");
|
|
|
|
if (!monitor_dev_for_events(cmd, lv, laopts, 1))
|
|
stack;
|
|
|
|
r = 1;
|
|
out:
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* In a cluster, set exclusive to indicate that only one node is using the
|
|
* device. Any tables loaded may then use non-clustered targets.
|
|
*
|
|
* @origin_only
|
|
* @exclusive This parameter only has an affect in cluster-context.
|
|
* It forces local target type to be used (instead of
|
|
* cluster-aware type).
|
|
* Returns success if the device is not active
|
|
*/
|
|
int lv_resume_if_active(struct cmd_context *cmd, const char *lvid_s,
|
|
unsigned origin_only, unsigned exclusive,
|
|
unsigned revert, const struct logical_volume *lv)
|
|
{
|
|
struct lv_activate_opts laopts = {
|
|
.exclusive = exclusive,
|
|
.origin_only = origin_only,
|
|
.revert = revert
|
|
};
|
|
|
|
return _lv_resume(cmd, lvid_s, &laopts, 0, lv);
|
|
}
|
|
|
|
int lv_resume(struct cmd_context *cmd, const char *lvid_s, unsigned origin_only,
|
|
const struct logical_volume *lv)
|
|
{
|
|
struct lv_activate_opts laopts = { .origin_only = origin_only, };
|
|
|
|
return _lv_resume(cmd, lvid_s, &laopts, 1, lv);
|
|
}
|
|
|
|
static int _lv_has_open_snapshots(const struct logical_volume *lv)
|
|
{
|
|
struct lv_segment *snap_seg;
|
|
int r = 0;
|
|
|
|
dm_list_iterate_items_gen(snap_seg, &lv->snapshot_segs, origin_list)
|
|
if (!lv_check_not_in_use(snap_seg->cow, 1))
|
|
r++;
|
|
|
|
if (r)
|
|
log_error("LV %s has open %d snapshot(s), not deactivating.",
|
|
display_lvname(lv), r);
|
|
|
|
return r;
|
|
}
|
|
|
|
int lv_deactivate(struct cmd_context *cmd, const char *lvid_s, const struct logical_volume *lv)
|
|
{
|
|
struct lvinfo info;
|
|
const struct lv_activate_opts laopts = { .skip_in_use = 1 };
|
|
struct dm_list *snh;
|
|
int r = 0;
|
|
|
|
if (!activation())
|
|
return 1;
|
|
|
|
if (test_mode()) {
|
|
_skip("Deactivating %s.", display_lvname(lv));
|
|
r = 1;
|
|
goto out;
|
|
}
|
|
|
|
log_debug_activation("Deactivating %s.", display_lvname(lv));
|
|
|
|
if (lv_is_visible(lv) || lv_is_virtual_origin(lv) ||
|
|
lv_is_merging_thin_snapshot(lv)) {
|
|
switch (lv_check_not_in_use(lv, 1)) {
|
|
case 0: goto_out;
|
|
case 2: goto no_exists;
|
|
}
|
|
|
|
if (lv_is_origin(lv) && _lv_has_open_snapshots(lv))
|
|
goto_out;
|
|
} else {
|
|
if (!lv_info(cmd, lv, 0, &info, 0, 0))
|
|
goto_out;
|
|
|
|
if (!info.exists) {
|
|
no_exists:
|
|
r = 1;
|
|
/* Check attached snapshot segments are also inactive */
|
|
dm_list_iterate(snh, &lv->snapshot_segs) {
|
|
if (!lv_info(cmd, dm_list_struct_base(snh, struct lv_segment, origin_list)->cow,
|
|
0, &info, 0, 0))
|
|
goto_out;
|
|
if (info.exists) {
|
|
r = 0; /* Snapshot left in table? */
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (lv_is_vdo_pool(lv)) {
|
|
/* If someone has remove 'linear' mapping over VDO device
|
|
* we may still be able to deactivate the rest of the tree
|
|
* i.e. in test-suite we simulate this via 'dmsetup remove' */
|
|
if (!lv_info(cmd, lv, 1, &info, 1, 0))
|
|
goto_out;
|
|
|
|
if (info.exists && !info.open_count)
|
|
r = 0; /* Unused VDO device left in table? */
|
|
}
|
|
|
|
if (r)
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
if (!monitor_dev_for_events(cmd, lv, &laopts, 0))
|
|
stack;
|
|
|
|
critical_section_inc(cmd, "deactivating");
|
|
r = _lv_deactivate(lv);
|
|
|
|
/*
|
|
* Remove any transiently activated error
|
|
* devices which arean't used any more.
|
|
*/
|
|
if (r && lv_is_raid(lv) && !lv_deactivate_any_missing_subdevs(lv)) {
|
|
log_error("Failed to remove temporary SubLVs from %s",
|
|
display_lvname(lv));
|
|
r = 0;
|
|
}
|
|
critical_section_dec(cmd, "deactivated");
|
|
|
|
if (!lv_info(cmd, lv, 0, &info, 0, 0) || info.exists) {
|
|
/* Turn into log_error, but we do not log error */
|
|
log_debug_activation("Deactivated volume is still %s present.",
|
|
display_lvname(lv));
|
|
r = 0;
|
|
}
|
|
|
|
out:
|
|
|
|
return r;
|
|
}
|
|
|
|
/* Test if LV passes filter */
|
|
int lv_activation_filter(struct cmd_context *cmd, const char *lvid_s,
|
|
int *activate, const struct logical_volume *lv)
|
|
{
|
|
if (!activation()) {
|
|
*activate = 1;
|
|
return 1;
|
|
}
|
|
|
|
if (!_passes_activation_filter(cmd, lv)) {
|
|
log_verbose("Not activating %s since it does not pass "
|
|
"activation filter.", display_lvname(lv));
|
|
*activate = 0;
|
|
} else
|
|
*activate = 1;
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int _lv_activate(struct cmd_context *cmd, const char *lvid_s,
|
|
struct lv_activate_opts *laopts, int filter,
|
|
const struct logical_volume *lv)
|
|
{
|
|
struct lvinfo info;
|
|
int r = 0;
|
|
|
|
if (!activation())
|
|
return 1;
|
|
|
|
if (filter && !_passes_activation_filter(cmd, lv)) {
|
|
log_verbose("Not activating %s since it does not pass "
|
|
"activation filter.", display_lvname(lv));
|
|
r = 1;
|
|
goto out;
|
|
}
|
|
|
|
if ((cmd->partial_activation || cmd->degraded_activation) &&
|
|
lv_is_partial(lv) && lv_is_raid(lv) && lv_raid_has_integrity((struct logical_volume *)lv)) {
|
|
cmd->partial_activation = 0;
|
|
cmd->degraded_activation = 0;
|
|
log_print_unless_silent("No degraded or partial activation for raid with integrity.");
|
|
}
|
|
|
|
if ((!lv->vg->cmd->partial_activation) && lv_is_partial(lv)) {
|
|
if (!lv_is_raid_type(lv) || !partial_raid_lv_supports_degraded_activation(lv)) {
|
|
log_error("Refusing activation of partial LV %s. "
|
|
"Use '--activationmode partial' to override.",
|
|
display_lvname(lv));
|
|
goto out;
|
|
}
|
|
|
|
if (!lv->vg->cmd->degraded_activation) {
|
|
log_error("Refusing activation of partial LV %s. "
|
|
"Try '--activationmode degraded'.",
|
|
display_lvname(lv));
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
if ((cmd->partial_activation || cmd->degraded_activation) && lv_is_writecache(lv)) {
|
|
struct logical_volume *lv_fast = first_seg(lv)->writecache;
|
|
if (lv_is_partial(lv) || (lv_fast && lv_is_partial(lv_fast))) {
|
|
log_error("Cannot use partial or degraded activation with writecache.");
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
if (lv_has_unknown_segments(lv)) {
|
|
log_error("Refusing activation of LV %s containing "
|
|
"an unrecognised segment.", display_lvname(lv));
|
|
goto out;
|
|
}
|
|
|
|
if (lv_raid_has_visible_sublvs(lv)) {
|
|
log_error("Refusing activation of RAID LV %s with "
|
|
"visible SubLVs.", display_lvname(lv));
|
|
goto out;
|
|
}
|
|
|
|
if (test_mode()) {
|
|
_skip("Activating %s.", display_lvname(lv));
|
|
r = 1;
|
|
goto out;
|
|
}
|
|
|
|
/* Component LV activation is enforced to be 'read-only' */
|
|
/* TODO: should not apply for LVs in maintenance mode */
|
|
if (!lv_is_visible(lv) && lv_is_component(lv)) {
|
|
laopts->read_only = 1;
|
|
laopts->component_lv = lv;
|
|
} else if (lv_is_pool_metadata_spare(lv)) {
|
|
laopts->component_lv = lv;
|
|
} else if (filter)
|
|
laopts->read_only = _passes_readonly_filter(cmd, lv);
|
|
|
|
log_debug_activation("Activating %s%s%s%s%s.", display_lvname(lv),
|
|
laopts->exclusive ? " exclusively" : "",
|
|
laopts->read_only ? " read-only" : "",
|
|
laopts->noscan ? " noscan" : "",
|
|
laopts->temporary ? " temporary" : "");
|
|
|
|
if (!lv_info_with_name_check(cmd, lv, 0, &info))
|
|
goto_out;
|
|
|
|
/*
|
|
* Nothing to do?
|
|
*/
|
|
if (info.exists && !info.suspended && info.live_table &&
|
|
(info.read_only == read_only_lv(lv, laopts, NULL))) {
|
|
r = 1;
|
|
log_debug_activation("LV %s is already active.", display_lvname(lv));
|
|
goto out;
|
|
}
|
|
|
|
lv_calculate_readahead(lv, NULL);
|
|
|
|
critical_section_inc(cmd, "activating");
|
|
if (!(r = _lv_activate_lv(lv, laopts)))
|
|
stack;
|
|
critical_section_dec(cmd, "activated");
|
|
|
|
if (r && !monitor_dev_for_events(cmd, lv, laopts, 1))
|
|
stack;
|
|
out:
|
|
return r;
|
|
}
|
|
|
|
/* Activate LV */
|
|
int lv_activate(struct cmd_context *cmd, const char *lvid_s, int exclusive,
|
|
int noscan, int temporary, const struct logical_volume *lv)
|
|
{
|
|
struct lv_activate_opts laopts = { .exclusive = exclusive,
|
|
.noscan = noscan,
|
|
.temporary = temporary };
|
|
|
|
if (!_lv_activate(cmd, lvid_s, &laopts, 0, lv))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
/* Activate LV only if it passes filter */
|
|
int lv_activate_with_filter(struct cmd_context *cmd, const char *lvid_s, int exclusive,
|
|
int noscan, int temporary, const struct logical_volume *lv)
|
|
{
|
|
struct lv_activate_opts laopts = { .exclusive = exclusive,
|
|
.noscan = noscan,
|
|
.temporary = temporary };
|
|
|
|
if (!_lv_activate(cmd, lvid_s, &laopts, 1, lv))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
int lv_mknodes(struct cmd_context *cmd, const struct logical_volume *lv)
|
|
{
|
|
int r;
|
|
|
|
if (!lv) {
|
|
r = dm_mknodes(NULL);
|
|
fs_unlock();
|
|
return r;
|
|
}
|
|
|
|
if (!activation())
|
|
return 1;
|
|
|
|
r = dev_manager_mknodes(lv);
|
|
|
|
fs_unlock();
|
|
|
|
return r;
|
|
}
|
|
|
|
/* Remove any existing, closed mapped device by @name */
|
|
static int _remove_dm_dev_by_name(const char *name)
|
|
{
|
|
int r = 0;
|
|
struct dm_task *dmt;
|
|
struct dm_info info;
|
|
|
|
if (!(dmt = dm_task_create(DM_DEVICE_INFO)))
|
|
return_0;
|
|
|
|
/* Check, if the device exists. */
|
|
if (dm_task_set_name(dmt, name) && dm_task_run(dmt) && dm_task_get_info(dmt, &info)) {
|
|
dm_task_destroy(dmt);
|
|
|
|
/* Ignore non-existing or open dm devices */
|
|
if (!info.exists || info.open_count)
|
|
return 1;
|
|
|
|
if (!(dmt = dm_task_create(DM_DEVICE_REMOVE)))
|
|
return_0;
|
|
|
|
if (dm_task_set_name(dmt, name))
|
|
r = dm_task_run(dmt);
|
|
}
|
|
|
|
dm_task_destroy(dmt);
|
|
|
|
return r;
|
|
}
|
|
|
|
/* Work all segments of @lv removing any existing, closed "*-missing_N_0" sub devices. */
|
|
static int _lv_remove_any_missing_subdevs(struct logical_volume *lv)
|
|
{
|
|
char name[NAME_LEN];
|
|
struct lv_segment *seg;
|
|
uint32_t seg_no = 0;
|
|
|
|
if (lv) {
|
|
dm_list_iterate_items(seg, &lv->segments) {
|
|
if (dm_snprintf(name, sizeof(name), "%s-%s-missing_%u_0", seg->lv->vg->name, seg->lv->name, seg_no) < 0)
|
|
return_0;
|
|
if (!_remove_dm_dev_by_name(name))
|
|
return_0;
|
|
|
|
seg_no++;
|
|
}
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
/* Remove any "*-missing_*" sub devices added by the activation layer for an rmate/rimage missing PV mapping */
|
|
int lv_deactivate_any_missing_subdevs(const struct logical_volume *lv)
|
|
{
|
|
uint32_t s;
|
|
struct lv_segment *seg = first_seg(lv);
|
|
|
|
for (s = 0; s < seg->area_count; s++) {
|
|
if (seg_type(seg, s) == AREA_LV &&
|
|
!_lv_remove_any_missing_subdevs(seg_lv(seg, s)))
|
|
return_0;
|
|
if (seg->meta_areas && seg_metatype(seg, s) == AREA_LV &&
|
|
!_lv_remove_any_missing_subdevs(seg_metalv(seg, s)))
|
|
return_0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* Does PV use VG somewhere in its construction?
|
|
* Returns 1 on failure.
|
|
*/
|
|
int pv_uses_vg(struct physical_volume *pv,
|
|
struct volume_group *vg)
|
|
{
|
|
if (!activation() || !pv->dev)
|
|
return 0;
|
|
|
|
if (!dm_is_dm_major(MAJOR(pv->dev->dev)))
|
|
return 0;
|
|
|
|
return dev_manager_device_uses_vg(pv->dev, vg);
|
|
}
|
|
|
|
void activation_release(void)
|
|
{
|
|
if (critical_section())
|
|
/* May leak stacked operation */
|
|
log_error("Releasing activation in critical section.");
|
|
|
|
fs_unlock(); /* Implicit dev_manager_release(); */
|
|
}
|
|
|
|
void activation_exit(void)
|
|
{
|
|
activation_release();
|
|
dev_manager_exit();
|
|
}
|
|
#endif
|
|
|
|
static int _component_cb(struct logical_volume *lv, void *data)
|
|
{
|
|
struct logical_volume **component_lv = (struct logical_volume **) data;
|
|
|
|
if (lv_is_locked(lv) || lv_is_pvmove(lv) ||/* ignoring */
|
|
/* thin-pool is special and it's using layered device */
|
|
(lv_is_thin_pool(lv) && thin_pool_is_active(lv)))
|
|
return -1;
|
|
|
|
/* External origin is activated through thinLV and uses -real suffix.
|
|
* Note: for old clustered logic we would need to check for all thins */
|
|
if ((lv_is_external_origin(lv) && lv_info(lv->vg->cmd, lv, 1, NULL, 0, 0)) ||
|
|
lv_is_active(lv)) {
|
|
if (!lv_is_component(lv) || lv_is_visible(lv))
|
|
return -1; /* skip whole subtree */
|
|
|
|
log_debug_activation("Found active component LV %s.", display_lvname(lv));
|
|
*component_lv = lv;
|
|
return 0; /* break any further processing */
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* Finds out for any LV if any of its component LVs are active.
|
|
* Function first checks if an existing LV is visible and active eventually
|
|
* it's lock holding LV is already active. In such case sub LV cannot be
|
|
* actived alone and no further checking is needed.
|
|
*
|
|
* Returns active component LV if there is such.
|
|
*/
|
|
const struct logical_volume *lv_component_is_active(const struct logical_volume *lv)
|
|
{
|
|
const struct logical_volume *component_lv = NULL;
|
|
const struct logical_volume *holder_lv = lv_lock_holder(lv);
|
|
|
|
if ((holder_lv != lv) && lv_is_active(holder_lv))
|
|
return NULL; /* Lock holding LV is active, do not check components */
|
|
|
|
if (_component_cb((struct logical_volume *) lv, &holder_lv) == 1)
|
|
(void) for_each_sub_lv((struct logical_volume *) lv, _component_cb,
|
|
(void*) &component_lv);
|
|
|
|
return component_lv;
|
|
}
|
|
|
|
/*
|
|
* Finds out if any LV above is active, as stacked device tree can be composed of
|
|
* chained set of LVs.
|
|
*
|
|
* Returns active holder LV if there is such.
|
|
*/
|
|
const struct logical_volume *lv_holder_is_active(const struct logical_volume *lv)
|
|
{
|
|
const struct logical_volume *holder;
|
|
const struct seg_list *sl;
|
|
|
|
if (lv_is_locked(lv) || lv_is_pvmove(lv))
|
|
return NULL; /* Skip pvmove/locked LV tracking */
|
|
|
|
dm_list_iterate_items(sl, &lv->segs_using_this_lv) {
|
|
/* Recursive call for upper-stack holder */
|
|
if ((holder = lv_holder_is_active(sl->seg->lv)))
|
|
return holder;
|
|
|
|
if (lv_is_active(sl->seg->lv)) {
|
|
log_debug_activation("Found active holder LV %s.", display_lvname(sl->seg->lv));
|
|
return sl->seg->lv;
|
|
}
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static int _deactivate_sub_lv_cb(struct logical_volume *lv, void *data)
|
|
{
|
|
struct logical_volume **slv = data;
|
|
|
|
if (lv_is_thin_pool(lv) || lv_is_external_origin(lv))
|
|
return -1;
|
|
|
|
if (!deactivate_lv(lv->vg->cmd, lv)) {
|
|
*slv = lv;
|
|
return 0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* Deactivates LV toghether with explicit deactivation call made also for all its component LVs.
|
|
*/
|
|
int deactivate_lv_with_sub_lv(const struct logical_volume *lv)
|
|
{
|
|
struct logical_volume *flv = NULL;
|
|
|
|
if (!deactivate_lv(lv->vg->cmd, lv)) {
|
|
log_error("Cannot deactivate logical volume %s.",
|
|
display_lvname(lv));
|
|
return 0;
|
|
}
|
|
|
|
if (!for_each_sub_lv((struct logical_volume *)lv, _deactivate_sub_lv_cb, &flv)) {
|
|
log_error("Cannot deactivate subvolume %s of logical volume %s.",
|
|
(flv) ? display_lvname(flv) : "", display_lvname(lv));
|
|
return 0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
int activate_lv(struct cmd_context *cmd, const struct logical_volume *lv)
|
|
{
|
|
const struct logical_volume *active_lv;
|
|
int ret;
|
|
|
|
/*
|
|
* When trying activating component LV, make sure none of sub component
|
|
* LV or LVs that are using it are active.
|
|
*/
|
|
if (!lv_is_visible(lv))
|
|
active_lv = lv_holder_is_active(lv);
|
|
else
|
|
active_lv = lv_component_is_active(lv);
|
|
|
|
if (active_lv) {
|
|
log_error("Activation of logical volume %s is prohibited while logical volume %s is active.",
|
|
display_lvname(lv), display_lvname(active_lv));
|
|
ret = 0;
|
|
goto out;
|
|
}
|
|
|
|
ret = lv_activate_with_filter(cmd, NULL, 0,
|
|
(lv->status & LV_NOSCAN) ? 1 : 0,
|
|
(lv->status & LV_TEMPORARY) ? 1 : 0,
|
|
lv_committed(lv));
|
|
out:
|
|
return ret;
|
|
}
|
|
|
|
int deactivate_lv(struct cmd_context *cmd, const struct logical_volume *lv)
|
|
{
|
|
int ret;
|
|
|
|
ret = lv_deactivate(cmd, NULL, lv_committed(lv));
|
|
|
|
return ret;
|
|
}
|
|
|
|
int suspend_lv(struct cmd_context *cmd, const struct logical_volume *lv)
|
|
{
|
|
int ret;
|
|
|
|
critical_section_inc(cmd, "locking for suspend");
|
|
|
|
ret = lv_suspend_if_active(cmd, NULL, 0, 0, lv_committed(lv), lv);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int suspend_lv_origin(struct cmd_context *cmd, const struct logical_volume *lv)
|
|
{
|
|
int ret;
|
|
|
|
critical_section_inc(cmd, "locking for suspend");
|
|
|
|
ret = lv_suspend_if_active(cmd, NULL, 1, 0, lv_committed(lv), lv);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int resume_lv(struct cmd_context *cmd, const struct logical_volume *lv)
|
|
{
|
|
int ret;
|
|
|
|
ret = lv_resume_if_active(cmd, NULL, 0, 0, 0, lv_committed(lv));
|
|
|
|
critical_section_dec(cmd, "unlocking on resume");
|
|
|
|
return ret;
|
|
}
|
|
|
|
int resume_lv_origin(struct cmd_context *cmd, const struct logical_volume *lv)
|
|
{
|
|
int ret;
|
|
|
|
ret = lv_resume_if_active(cmd, NULL, 1, 0, 0, lv_committed(lv));
|
|
|
|
critical_section_dec(cmd, "unlocking on resume");
|
|
|
|
return ret;
|
|
}
|
|
|
|
int revert_lv(struct cmd_context *cmd, const struct logical_volume *lv)
|
|
{
|
|
int ret;
|
|
|
|
ret = lv_resume_if_active(cmd, NULL, 0, 0, 1, lv_committed(lv));
|
|
|
|
critical_section_dec(cmd, "unlocking on revert");
|
|
|
|
return ret;
|
|
}
|