mirror of
git://sourceware.org/git/lvm2.git
synced 2025-01-03 05:18:29 +03:00
27ff8813da
activated. In order to achieve this, we need to be able to query whether the origin is active exclusively (a condition of being able to add an exclusive snapshot). Once we are able to query the exclusive activation of an LV, we can safely create/activate the snapshot. A change to 'hold_lock' was also made so that a request to aquire a WRITE lock did not replace an EX lock, which is already a form of write lock.
1464 lines
31 KiB
C
1464 lines
31 KiB
C
/*
|
|
* Copyright (C) 2001-2004 Sistina Software, Inc. All rights reserved.
|
|
* Copyright (C) 2004-2009 Red Hat, Inc. All rights reserved.
|
|
*
|
|
* This file is part of LVM2.
|
|
*
|
|
* This copyrighted material is made available to anyone wishing to use,
|
|
* modify, copy, or redistribute it subject to the terms and conditions
|
|
* of the GNU Lesser General Public License v.2.1.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public License
|
|
* along with this program; if not, write to the Free Software Foundation,
|
|
* Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
*/
|
|
|
|
#include "lib.h"
|
|
#include "metadata.h"
|
|
#include "activate.h"
|
|
#include "memlock.h"
|
|
#include "display.h"
|
|
#include "fs.h"
|
|
#include "lvm-exec.h"
|
|
#include "lvm-file.h"
|
|
#include "lvm-string.h"
|
|
#include "toolcontext.h"
|
|
#include "dev_manager.h"
|
|
#include "str_list.h"
|
|
#include "config.h"
|
|
#include "filter.h"
|
|
#include "segtype.h"
|
|
#include "sharedlib.h"
|
|
|
|
#include <limits.h>
|
|
#include <fcntl.h>
|
|
#include <unistd.h>
|
|
|
|
#define _skip(fmt, args...) log_very_verbose("Skipping: " fmt , ## args)
|
|
|
|
int lvm1_present(struct cmd_context *cmd)
|
|
{
|
|
char path[PATH_MAX];
|
|
|
|
if (dm_snprintf(path, sizeof(path), "%s/lvm/global", cmd->proc_dir)
|
|
< 0) {
|
|
log_error("LVM1 proc global snprintf failed");
|
|
return 0;
|
|
}
|
|
|
|
if (path_exists(path))
|
|
return 1;
|
|
else
|
|
return 0;
|
|
}
|
|
|
|
int list_segment_modules(struct dm_pool *mem, const struct lv_segment *seg,
|
|
struct dm_list *modules)
|
|
{
|
|
unsigned int s;
|
|
struct lv_segment *seg2, *snap_seg;
|
|
struct dm_list *snh;
|
|
|
|
if (seg->segtype->ops->modules_needed &&
|
|
!seg->segtype->ops->modules_needed(mem, seg, modules)) {
|
|
log_error("module string allocation failed");
|
|
return 0;
|
|
}
|
|
|
|
if (lv_is_origin(seg->lv))
|
|
dm_list_iterate(snh, &seg->lv->snapshot_segs)
|
|
if (!list_lv_modules(mem,
|
|
dm_list_struct_base(snh,
|
|
struct lv_segment,
|
|
origin_list)->cow,
|
|
modules))
|
|
return_0;
|
|
|
|
if (lv_is_cow(seg->lv)) {
|
|
snap_seg = find_cow(seg->lv);
|
|
if (snap_seg->segtype->ops->modules_needed &&
|
|
!snap_seg->segtype->ops->modules_needed(mem, snap_seg,
|
|
modules)) {
|
|
log_error("snap_seg module string allocation failed");
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
for (s = 0; s < seg->area_count; s++) {
|
|
switch (seg_type(seg, s)) {
|
|
case AREA_LV:
|
|
seg2 = find_seg_by_le(seg_lv(seg, s), seg_le(seg, s));
|
|
if (seg2 && !list_segment_modules(mem, seg2, modules))
|
|
return_0;
|
|
break;
|
|
case AREA_PV:
|
|
case AREA_UNASSIGNED:
|
|
;
|
|
}
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
int list_lv_modules(struct dm_pool *mem, const struct logical_volume *lv,
|
|
struct dm_list *modules)
|
|
{
|
|
struct lv_segment *seg;
|
|
|
|
dm_list_iterate_items(seg, &lv->segments)
|
|
if (!list_segment_modules(mem, seg, modules))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
#ifndef DEVMAPPER_SUPPORT
|
|
void set_activation(int act)
|
|
{
|
|
static int warned = 0;
|
|
|
|
if (warned || !act)
|
|
return;
|
|
|
|
log_error("Compiled without libdevmapper support. "
|
|
"Can't enable activation.");
|
|
|
|
warned = 1;
|
|
}
|
|
int activation(void)
|
|
{
|
|
return 0;
|
|
}
|
|
int library_version(char *version, size_t size)
|
|
{
|
|
return 0;
|
|
}
|
|
int driver_version(char *version, size_t size)
|
|
{
|
|
return 0;
|
|
}
|
|
int target_version(const char *target_name, uint32_t *maj,
|
|
uint32_t *min, uint32_t *patchlevel)
|
|
{
|
|
return 0;
|
|
}
|
|
int target_present(struct cmd_context *cmd, const char *target_name,
|
|
int use_modprobe)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_info(struct cmd_context *cmd, const struct logical_volume *lv, unsigned origin_only,
|
|
struct lvinfo *info, int with_open_count, int with_read_ahead)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_info_by_lvid(struct cmd_context *cmd, const char *lvid_s,
|
|
unsigned origin_only,
|
|
struct lvinfo *info, int with_open_count, int with_read_ahead)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_snapshot_percent(const struct logical_volume *lv, percent_t *percent)
|
|
{
|
|
return 0;
|
|
}
|
|
int lv_mirror_percent(struct cmd_context *cmd, struct logical_volume *lv,
|
|
int wait, percent_t *percent, uint32_t *event_nr)
|
|
{
|
|
return 0;
|
|
}
|
|
int lvs_in_vg_activated(struct volume_group *vg)
|
|
{
|
|
return 0;
|
|
}
|
|
int lvs_in_vg_opened(struct volume_group *vg)
|
|
{
|
|
return 0;
|
|
}
|
|
/******
|
|
int lv_suspend(struct cmd_context *cmd, const char *lvid_s)
|
|
{
|
|
return 1;
|
|
}
|
|
*******/
|
|
int lv_suspend_if_active(struct cmd_context *cmd, const char *lvid_s)
|
|
{
|
|
return 1;
|
|
}
|
|
int lv_resume(struct cmd_context *cmd, const char *lvid_s)
|
|
{
|
|
return 1;
|
|
}
|
|
int lv_resume_if_active(struct cmd_context *cmd, const char *lvid_s)
|
|
{
|
|
return 1;
|
|
}
|
|
int lv_deactivate(struct cmd_context *cmd, const char *lvid_s)
|
|
{
|
|
return 1;
|
|
}
|
|
int lv_activation_filter(struct cmd_context *cmd, const char *lvid_s,
|
|
int *activate_lv)
|
|
{
|
|
return 1;
|
|
}
|
|
int lv_activate(struct cmd_context *cmd, const char *lvid_s, int exclusive)
|
|
{
|
|
return 1;
|
|
}
|
|
int lv_activate_with_filter(struct cmd_context *cmd, const char *lvid_s, int exclusive)
|
|
{
|
|
return 1;
|
|
}
|
|
|
|
int lv_mknodes(struct cmd_context *cmd, const struct logical_volume *lv)
|
|
{
|
|
return 1;
|
|
}
|
|
|
|
int pv_uses_vg(struct physical_volume *pv,
|
|
struct volume_group *vg)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
void activation_release(void)
|
|
{
|
|
return;
|
|
}
|
|
|
|
void activation_exit(void)
|
|
{
|
|
return;
|
|
}
|
|
|
|
#else /* DEVMAPPER_SUPPORT */
|
|
|
|
static int _activation = 1;
|
|
|
|
void set_activation(int act)
|
|
{
|
|
if (act == _activation)
|
|
return;
|
|
|
|
_activation = act;
|
|
if (_activation)
|
|
log_verbose("Activation enabled. Device-mapper kernel "
|
|
"driver will be used.");
|
|
else
|
|
log_warn("WARNING: Activation disabled. No device-mapper "
|
|
"interaction will be attempted.");
|
|
}
|
|
|
|
int activation(void)
|
|
{
|
|
return _activation;
|
|
}
|
|
|
|
static int _passes_activation_filter(struct cmd_context *cmd,
|
|
struct logical_volume *lv)
|
|
{
|
|
const struct config_node *cn;
|
|
const struct config_value *cv;
|
|
const char *str;
|
|
char path[PATH_MAX];
|
|
|
|
if (!(cn = find_config_tree_node(cmd, "activation/volume_list"))) {
|
|
log_verbose("activation/volume_list configuration setting "
|
|
"not defined, checking only host tags for %s/%s",
|
|
lv->vg->name, lv->name);
|
|
|
|
/* If no host tags defined, activate */
|
|
if (dm_list_empty(&cmd->tags))
|
|
return 1;
|
|
|
|
/* If any host tag matches any LV or VG tag, activate */
|
|
if (str_list_match_list(&cmd->tags, &lv->tags, NULL) ||
|
|
str_list_match_list(&cmd->tags, &lv->vg->tags, NULL))
|
|
return 1;
|
|
|
|
log_verbose("No host tag matches %s/%s",
|
|
lv->vg->name, lv->name);
|
|
|
|
/* Don't activate */
|
|
return 0;
|
|
}
|
|
else
|
|
log_verbose("activation/volume_list configuration setting "
|
|
"defined, checking the list to match %s/%s",
|
|
lv->vg->name, lv->name);
|
|
|
|
for (cv = cn->v; cv; cv = cv->next) {
|
|
if (cv->type != CFG_STRING) {
|
|
log_error("Ignoring invalid string in config file "
|
|
"activation/volume_list");
|
|
continue;
|
|
}
|
|
str = cv->v.str;
|
|
if (!*str) {
|
|
log_error("Ignoring empty string in config file "
|
|
"activation/volume_list");
|
|
continue;
|
|
}
|
|
|
|
|
|
/* Tag? */
|
|
if (*str == '@') {
|
|
str++;
|
|
if (!*str) {
|
|
log_error("Ignoring empty tag in config file "
|
|
"activation/volume_list");
|
|
continue;
|
|
}
|
|
/* If any host tag matches any LV or VG tag, activate */
|
|
if (!strcmp(str, "*")) {
|
|
if (str_list_match_list(&cmd->tags, &lv->tags, NULL)
|
|
|| str_list_match_list(&cmd->tags,
|
|
&lv->vg->tags, NULL))
|
|
return 1;
|
|
else
|
|
continue;
|
|
}
|
|
/* If supplied tag matches LV or VG tag, activate */
|
|
if (str_list_match_item(&lv->tags, str) ||
|
|
str_list_match_item(&lv->vg->tags, str))
|
|
return 1;
|
|
else
|
|
continue;
|
|
}
|
|
if (!strchr(str, '/')) {
|
|
/* vgname supplied */
|
|
if (!strcmp(str, lv->vg->name))
|
|
return 1;
|
|
else
|
|
continue;
|
|
}
|
|
/* vgname/lvname */
|
|
if (dm_snprintf(path, sizeof(path), "%s/%s", lv->vg->name,
|
|
lv->name) < 0) {
|
|
log_error("dm_snprintf error from %s/%s", lv->vg->name,
|
|
lv->name);
|
|
continue;
|
|
}
|
|
if (!strcmp(path, str))
|
|
return 1;
|
|
}
|
|
|
|
log_verbose("No item supplied in activation/volume_list configuration "
|
|
"setting matches %s/%s", lv->vg->name, lv->name);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int library_version(char *version, size_t size)
|
|
{
|
|
if (!activation())
|
|
return 0;
|
|
|
|
return dm_get_library_version(version, size);
|
|
}
|
|
|
|
int driver_version(char *version, size_t size)
|
|
{
|
|
if (!activation())
|
|
return 0;
|
|
|
|
log_very_verbose("Getting driver version");
|
|
|
|
return dm_driver_version(version, size);
|
|
}
|
|
|
|
int target_version(const char *target_name, uint32_t *maj,
|
|
uint32_t *min, uint32_t *patchlevel)
|
|
{
|
|
int r = 0;
|
|
struct dm_task *dmt;
|
|
struct dm_versions *target, *last_target;
|
|
|
|
log_very_verbose("Getting target version for %s", target_name);
|
|
if (!(dmt = dm_task_create(DM_DEVICE_LIST_VERSIONS)))
|
|
return_0;
|
|
|
|
if (!dm_task_run(dmt)) {
|
|
log_debug("Failed to get %s target version", target_name);
|
|
/* Assume this was because LIST_VERSIONS isn't supported */
|
|
return 1;
|
|
}
|
|
|
|
target = dm_task_get_versions(dmt);
|
|
|
|
do {
|
|
last_target = target;
|
|
|
|
if (!strcmp(target_name, target->name)) {
|
|
r = 1;
|
|
*maj = target->version[0];
|
|
*min = target->version[1];
|
|
*patchlevel = target->version[2];
|
|
goto out;
|
|
}
|
|
|
|
target = (struct dm_versions *)((char *) target + target->next);
|
|
} while (last_target != target);
|
|
|
|
out:
|
|
dm_task_destroy(dmt);
|
|
|
|
return r;
|
|
}
|
|
|
|
int module_present(struct cmd_context *cmd, const char *target_name)
|
|
{
|
|
int ret = 0;
|
|
#ifdef MODPROBE_CMD
|
|
char module[128];
|
|
const char *argv[3];
|
|
|
|
if (dm_snprintf(module, sizeof(module), "dm-%s", target_name) < 0) {
|
|
log_error("module_present module name too long: %s",
|
|
target_name);
|
|
return 0;
|
|
}
|
|
|
|
argv[0] = MODPROBE_CMD;
|
|
argv[1] = module;
|
|
argv[2] = NULL;
|
|
|
|
ret = exec_cmd(cmd, argv, NULL, 0);
|
|
#endif
|
|
return ret;
|
|
}
|
|
|
|
int target_present(struct cmd_context *cmd, const char *target_name,
|
|
int use_modprobe)
|
|
{
|
|
uint32_t maj, min, patchlevel;
|
|
|
|
if (!activation())
|
|
return 0;
|
|
|
|
#ifdef MODPROBE_CMD
|
|
if (use_modprobe) {
|
|
if (target_version(target_name, &maj, &min, &patchlevel))
|
|
return 1;
|
|
|
|
if (!module_present(cmd, target_name))
|
|
return_0;
|
|
}
|
|
#endif
|
|
|
|
return target_version(target_name, &maj, &min, &patchlevel);
|
|
}
|
|
|
|
/*
|
|
* Returns 1 if info structure populated, else 0 on failure.
|
|
*/
|
|
int lv_info(struct cmd_context *cmd, const struct logical_volume *lv, unsigned origin_only,
|
|
struct lvinfo *info, int with_open_count, int with_read_ahead)
|
|
{
|
|
struct dm_info dminfo;
|
|
|
|
if (!activation())
|
|
return 0;
|
|
/*
|
|
* If open_count info is requested and we have to be sure our own udev
|
|
* transactions are finished
|
|
* For non-clustered locking type we are only interested for non-delete operation
|
|
* in progress - as only those could lead to opened files
|
|
*/
|
|
if (with_open_count) {
|
|
if (locking_is_clustered())
|
|
sync_local_dev_names(cmd); /* Wait to have udev in sync */
|
|
else if (fs_has_non_delete_ops())
|
|
fs_unlock(); /* For non clustered - wait if there are non-delete ops */
|
|
}
|
|
|
|
if (!dev_manager_info(lv->vg->cmd->mem, lv, origin_only ? "real" : NULL, with_open_count,
|
|
with_read_ahead, &dminfo, &info->read_ahead))
|
|
return_0;
|
|
|
|
info->exists = dminfo.exists;
|
|
info->suspended = dminfo.suspended;
|
|
info->open_count = dminfo.open_count;
|
|
info->major = dminfo.major;
|
|
info->minor = dminfo.minor;
|
|
info->read_only = dminfo.read_only;
|
|
info->live_table = dminfo.live_table;
|
|
info->inactive_table = dminfo.inactive_table;
|
|
|
|
return 1;
|
|
}
|
|
|
|
int lv_info_by_lvid(struct cmd_context *cmd, const char *lvid_s,
|
|
unsigned origin_only,
|
|
struct lvinfo *info, int with_open_count, int with_read_ahead)
|
|
{
|
|
int r;
|
|
struct logical_volume *lv;
|
|
|
|
if (!(lv = lv_from_lvid(cmd, lvid_s, 0)))
|
|
return 0;
|
|
|
|
if (!lv_is_origin(lv))
|
|
origin_only = 0;
|
|
|
|
r = lv_info(cmd, lv, origin_only, info, with_open_count, with_read_ahead);
|
|
free_vg(lv->vg);
|
|
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* Returns 1 if percent set, else 0 on failure.
|
|
*/
|
|
int lv_check_transient(struct logical_volume *lv)
|
|
{
|
|
int r;
|
|
struct dev_manager *dm;
|
|
|
|
if (!activation())
|
|
return 0;
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name)))
|
|
return_0;
|
|
|
|
if (!(r = dev_manager_transient(dm, lv)))
|
|
stack;
|
|
|
|
dev_manager_destroy(dm);
|
|
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* Returns 1 if percent set, else 0 on failure.
|
|
*/
|
|
int lv_snapshot_percent(const struct logical_volume *lv, percent_t *percent)
|
|
{
|
|
int r;
|
|
struct dev_manager *dm;
|
|
|
|
if (!activation())
|
|
return 0;
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name)))
|
|
return_0;
|
|
|
|
if (!(r = dev_manager_snapshot_percent(dm, lv, percent)))
|
|
stack;
|
|
|
|
dev_manager_destroy(dm);
|
|
|
|
return r;
|
|
}
|
|
|
|
/* FIXME Merge with snapshot_percent */
|
|
int lv_mirror_percent(struct cmd_context *cmd, struct logical_volume *lv,
|
|
int wait, percent_t *percent, uint32_t *event_nr)
|
|
{
|
|
int r;
|
|
struct dev_manager *dm;
|
|
struct lvinfo info;
|
|
|
|
/* If mirrored LV is temporarily shrinked to 1 area (= linear),
|
|
* it should be considered in-sync. */
|
|
if (dm_list_size(&lv->segments) == 1 && first_seg(lv)->area_count == 1) {
|
|
*percent = PERCENT_100;
|
|
return 1;
|
|
}
|
|
|
|
if (!activation())
|
|
return 0;
|
|
|
|
if (!lv_info(cmd, lv, 0, &info, 0, 0))
|
|
return_0;
|
|
|
|
if (!info.exists)
|
|
return 0;
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name)))
|
|
return_0;
|
|
|
|
if (!(r = dev_manager_mirror_percent(dm, lv, wait, percent, event_nr)))
|
|
stack;
|
|
|
|
dev_manager_destroy(dm);
|
|
|
|
return r;
|
|
}
|
|
|
|
static int _lv_active(struct cmd_context *cmd, struct logical_volume *lv)
|
|
{
|
|
struct lvinfo info;
|
|
|
|
if (!lv_info(cmd, lv, 0, &info, 0, 0)) {
|
|
stack;
|
|
return -1;
|
|
}
|
|
|
|
return info.exists;
|
|
}
|
|
|
|
static int _lv_open_count(struct cmd_context *cmd, struct logical_volume *lv)
|
|
{
|
|
struct lvinfo info;
|
|
|
|
if (!lv_info(cmd, lv, 0, &info, 1, 0)) {
|
|
stack;
|
|
return -1;
|
|
}
|
|
|
|
return info.open_count;
|
|
}
|
|
|
|
static int _lv_activate_lv(struct logical_volume *lv, unsigned origin_only)
|
|
{
|
|
int r;
|
|
struct dev_manager *dm;
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name)))
|
|
return_0;
|
|
|
|
if (!(r = dev_manager_activate(dm, lv, origin_only)))
|
|
stack;
|
|
|
|
dev_manager_destroy(dm);
|
|
return r;
|
|
}
|
|
|
|
static int _lv_preload(struct logical_volume *lv, unsigned origin_only, int *flush_required)
|
|
{
|
|
int r;
|
|
struct dev_manager *dm;
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name)))
|
|
return_0;
|
|
|
|
if (!(r = dev_manager_preload(dm, lv, origin_only, flush_required)))
|
|
stack;
|
|
|
|
dev_manager_destroy(dm);
|
|
return r;
|
|
}
|
|
|
|
static int _lv_deactivate(struct logical_volume *lv)
|
|
{
|
|
int r;
|
|
struct dev_manager *dm;
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name)))
|
|
return_0;
|
|
|
|
if (!(r = dev_manager_deactivate(dm, lv)))
|
|
stack;
|
|
|
|
dev_manager_destroy(dm);
|
|
return r;
|
|
}
|
|
|
|
static int _lv_suspend_lv(struct logical_volume *lv, unsigned origin_only, int lockfs, int flush_required)
|
|
{
|
|
int r;
|
|
struct dev_manager *dm;
|
|
|
|
if (!(dm = dev_manager_create(lv->vg->cmd, lv->vg->name)))
|
|
return_0;
|
|
|
|
if (!(r = dev_manager_suspend(dm, lv, origin_only, lockfs, flush_required)))
|
|
stack;
|
|
|
|
dev_manager_destroy(dm);
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* These two functions return the number of visible LVs in the state,
|
|
* or -1 on error.
|
|
*/
|
|
int lvs_in_vg_activated(struct volume_group *vg)
|
|
{
|
|
struct lv_list *lvl;
|
|
int count = 0;
|
|
|
|
if (!activation())
|
|
return 0;
|
|
|
|
dm_list_iterate_items(lvl, &vg->lvs) {
|
|
if (lv_is_visible(lvl->lv))
|
|
count += (_lv_active(vg->cmd, lvl->lv) == 1);
|
|
}
|
|
|
|
return count;
|
|
}
|
|
|
|
int lvs_in_vg_opened(const struct volume_group *vg)
|
|
{
|
|
const struct lv_list *lvl;
|
|
int count = 0;
|
|
|
|
if (!activation())
|
|
return 0;
|
|
|
|
dm_list_iterate_items(lvl, &vg->lvs) {
|
|
if (lv_is_visible(lvl->lv))
|
|
count += (_lv_open_count(vg->cmd, lvl->lv) > 0);
|
|
}
|
|
|
|
return count;
|
|
}
|
|
|
|
/*
|
|
* _lv_is_active
|
|
* @lv: logical volume being queried
|
|
* @locally: set if active locally (when provided)
|
|
* @exclusive: set if active exclusively (when provided)
|
|
*
|
|
* Determine whether an LV is active locally or in a cluster.
|
|
* In addition to the return code which indicates whether or
|
|
* not the LV is active somewhere, two other values are set
|
|
* to yield more information about the status of the activation:
|
|
* return locally exclusively status
|
|
* ====== ======= =========== ======
|
|
* 0 0 0 not active
|
|
* 1 0 0 active remotely
|
|
* 1 0 1 exclusive remotely
|
|
* 1 1 0 active locally and possibly remotely
|
|
* 1 1 1 exclusive locally (or local && !cluster)
|
|
* The VG lock must be held to call this function.
|
|
*
|
|
* Returns: 0 or 1
|
|
*/
|
|
static int _lv_is_active(struct logical_volume *lv,
|
|
int *locally, int *exclusive)
|
|
{
|
|
int r, l, e; /* remote, local, and exclusive */
|
|
|
|
r = l = e = 0;
|
|
|
|
if (_lv_active(lv->vg->cmd, lv))
|
|
l = 1;
|
|
|
|
if (!vg_is_clustered(lv->vg)) {
|
|
e = 1; /* exclusive by definition */
|
|
goto out;
|
|
}
|
|
|
|
/* Active locally, and the caller doesn't care about exclusive */
|
|
if (l && !exclusive)
|
|
goto out;
|
|
|
|
if ((r = remote_lock_held(lv->lvid.s, &e)) >= 0)
|
|
goto out;
|
|
|
|
/*
|
|
* If lock query is not supported (due to interfacing with old
|
|
* code), then we cannot evaluate exclusivity properly.
|
|
*
|
|
* Old users of this function will never be affected by this,
|
|
* since they are only concerned about active vs. not active.
|
|
* New users of this function who specifically ask for 'exclusive'
|
|
* will be given an error message.
|
|
*/
|
|
if (l) {
|
|
if (exclusive)
|
|
log_error("Unable to determine exclusivity of %s",
|
|
lv->name);
|
|
goto out;
|
|
}
|
|
|
|
if (activate_lv_excl(lv->vg->cmd, lv)) {
|
|
if (!deactivate_lv(lv->vg->cmd, lv))
|
|
stack;
|
|
return 0;
|
|
}
|
|
|
|
out:
|
|
if (locally)
|
|
*locally = l;
|
|
if (exclusive)
|
|
*exclusive = e;
|
|
|
|
log_very_verbose("%s/%s is %sactive%s%s",
|
|
lv->vg->name, lv->name,
|
|
(r || l) ? "" : "not ",
|
|
(exclusive && e) ? " exclusive" : "",
|
|
e ? (l ? " locally" : " remotely") : "");
|
|
|
|
return r || l;
|
|
}
|
|
|
|
int lv_is_active(struct logical_volume *lv)
|
|
{
|
|
return _lv_is_active(lv, NULL, NULL);
|
|
}
|
|
|
|
/*
|
|
int lv_is_active_locally(struct logical_volume *lv)
|
|
{
|
|
int l;
|
|
return _lv_is_active(lv, &l, NULL) && l;
|
|
}
|
|
*/
|
|
|
|
int lv_is_active_exclusive_locally(struct logical_volume *lv)
|
|
{
|
|
int l, e;
|
|
return _lv_is_active(lv, &l, &e) && l && e;
|
|
}
|
|
|
|
int lv_is_active_exclusive_remotely(struct logical_volume *lv)
|
|
{
|
|
int l, e;
|
|
return _lv_is_active(lv, &l, &e) && !l && e;
|
|
}
|
|
|
|
#ifdef DMEVENTD
|
|
static struct dm_event_handler *_create_dm_event_handler(struct cmd_context *cmd, const char *dmuuid, const char *dso,
|
|
const int timeout, enum dm_event_mask mask)
|
|
{
|
|
struct dm_event_handler *dmevh;
|
|
|
|
if (!(dmevh = dm_event_handler_create()))
|
|
return_NULL;
|
|
|
|
if (dm_event_handler_set_dmeventd_path(dmevh, find_config_tree_str(cmd, "dmeventd/executable", NULL)))
|
|
goto_bad;
|
|
|
|
if (dm_event_handler_set_dso(dmevh, dso))
|
|
goto_bad;
|
|
|
|
if (dm_event_handler_set_uuid(dmevh, dmuuid))
|
|
goto_bad;
|
|
|
|
dm_event_handler_set_timeout(dmevh, timeout);
|
|
dm_event_handler_set_event_mask(dmevh, mask);
|
|
|
|
return dmevh;
|
|
|
|
bad:
|
|
dm_event_handler_destroy(dmevh);
|
|
return NULL;
|
|
}
|
|
|
|
char *get_monitor_dso_path(struct cmd_context *cmd, const char *libpath)
|
|
{
|
|
char *path;
|
|
|
|
if (!(path = dm_pool_alloc(cmd->mem, PATH_MAX))) {
|
|
log_error("Failed to allocate dmeventd library path.");
|
|
return NULL;
|
|
}
|
|
|
|
get_shared_library_path(cmd, libpath, path, PATH_MAX);
|
|
|
|
return path;
|
|
}
|
|
|
|
int target_registered_with_dmeventd(struct cmd_context *cmd, const char *dso,
|
|
struct logical_volume *lv, int *pending)
|
|
{
|
|
char *uuid;
|
|
enum dm_event_mask evmask = 0;
|
|
struct dm_event_handler *dmevh;
|
|
|
|
*pending = 0;
|
|
|
|
if (!dso)
|
|
return_0;
|
|
|
|
/* We always monitor the "real" device, never the "snapshot-origin" itself. */
|
|
if (!(uuid = build_dm_uuid(cmd->mem, lv->lvid.s, lv_is_origin(lv) ? "real" : NULL)))
|
|
return_0;
|
|
|
|
if (!(dmevh = _create_dm_event_handler(cmd, uuid, dso, 0, DM_EVENT_ALL_ERRORS)))
|
|
return_0;
|
|
|
|
if (dm_event_get_registered_device(dmevh, 0)) {
|
|
dm_event_handler_destroy(dmevh);
|
|
return 0;
|
|
}
|
|
|
|
evmask = dm_event_handler_get_event_mask(dmevh);
|
|
if (evmask & DM_EVENT_REGISTRATION_PENDING) {
|
|
*pending = 1;
|
|
evmask &= ~DM_EVENT_REGISTRATION_PENDING;
|
|
}
|
|
|
|
dm_event_handler_destroy(dmevh);
|
|
|
|
return evmask;
|
|
}
|
|
|
|
int target_register_events(struct cmd_context *cmd, const char *dso, struct logical_volume *lv,
|
|
int evmask __attribute__((unused)), int set, int timeout)
|
|
{
|
|
char *uuid;
|
|
struct dm_event_handler *dmevh;
|
|
int r;
|
|
|
|
if (!dso)
|
|
return_0;
|
|
|
|
/* We always monitor the "real" device, never the "snapshot-origin" itself. */
|
|
if (!(uuid = build_dm_uuid(cmd->mem, lv->lvid.s, lv_is_origin(lv) ? "real" : NULL)))
|
|
return_0;
|
|
|
|
if (!(dmevh = _create_dm_event_handler(cmd, uuid, dso, timeout,
|
|
DM_EVENT_ALL_ERRORS | (timeout ? DM_EVENT_TIMEOUT : 0))))
|
|
return_0;
|
|
|
|
r = set ? dm_event_register_handler(dmevh) : dm_event_unregister_handler(dmevh);
|
|
|
|
dm_event_handler_destroy(dmevh);
|
|
|
|
if (!r)
|
|
return_0;
|
|
|
|
log_info("%s %s for events", set ? "Monitored" : "Unmonitored", uuid);
|
|
|
|
return 1;
|
|
}
|
|
|
|
#endif
|
|
|
|
/*
|
|
* Returns 0 if an attempt to (un)monitor the device failed.
|
|
* Returns 1 otherwise.
|
|
*/
|
|
int monitor_dev_for_events(struct cmd_context *cmd, struct logical_volume *lv,
|
|
unsigned origin_only, int monitor)
|
|
{
|
|
#ifdef DMEVENTD
|
|
int i, pending = 0, monitored;
|
|
int r = 1;
|
|
struct dm_list *tmp, *snh, *snht;
|
|
struct lv_segment *seg;
|
|
struct lv_segment *log_seg;
|
|
int (*monitor_fn) (struct lv_segment *s, int e);
|
|
uint32_t s;
|
|
|
|
/* skip dmeventd code altogether */
|
|
if (dmeventd_monitor_mode() == DMEVENTD_MONITOR_IGNORE)
|
|
return 1;
|
|
|
|
/*
|
|
* Nothing to do if dmeventd configured not to be used.
|
|
*/
|
|
if (monitor && !dmeventd_monitor_mode())
|
|
return 1;
|
|
|
|
/*
|
|
* In case of a snapshot device, we monitor lv->snapshot->lv,
|
|
* not the actual LV itself.
|
|
*/
|
|
if (lv_is_cow(lv) && !lv_is_merging_cow(lv))
|
|
return monitor_dev_for_events(cmd, lv->snapshot->lv, 0, monitor);
|
|
|
|
/*
|
|
* In case this LV is a snapshot origin, we instead monitor
|
|
* each of its respective snapshots. The origin itself may
|
|
* also need to be monitored if it is a mirror, for example.
|
|
*/
|
|
if (!origin_only && lv_is_origin(lv))
|
|
dm_list_iterate_safe(snh, snht, &lv->snapshot_segs)
|
|
if (!monitor_dev_for_events(cmd, dm_list_struct_base(snh,
|
|
struct lv_segment, origin_list)->cow, 0, monitor))
|
|
r = 0;
|
|
|
|
/*
|
|
* If the volume is mirrored and its log is also mirrored, monitor
|
|
* the log volume as well.
|
|
*/
|
|
if ((seg = first_seg(lv)) != NULL && seg->log_lv != NULL &&
|
|
(log_seg = first_seg(seg->log_lv)) != NULL &&
|
|
seg_is_mirrored(log_seg))
|
|
if (!monitor_dev_for_events(cmd, seg->log_lv, 0, monitor))
|
|
r = 0;
|
|
|
|
dm_list_iterate(tmp, &lv->segments) {
|
|
seg = dm_list_item(tmp, struct lv_segment);
|
|
|
|
/* Recurse for AREA_LV */
|
|
for (s = 0; s < seg->area_count; s++) {
|
|
if (seg_type(seg, s) != AREA_LV)
|
|
continue;
|
|
if (!monitor_dev_for_events(cmd, seg_lv(seg, s), 0,
|
|
monitor)) {
|
|
log_error("Failed to %smonitor %s",
|
|
monitor ? "" : "un",
|
|
seg_lv(seg, s)->name);
|
|
r = 0;
|
|
}
|
|
}
|
|
|
|
if (!seg_monitored(seg) || (seg->status & PVMOVE))
|
|
continue;
|
|
|
|
monitor_fn = NULL;
|
|
|
|
/* Check monitoring status */
|
|
if (seg->segtype->ops->target_monitored)
|
|
monitored = seg->segtype->ops->target_monitored(seg, &pending);
|
|
else
|
|
continue; /* segtype doesn't support registration */
|
|
|
|
/*
|
|
* FIXME: We should really try again if pending
|
|
*/
|
|
monitored = (pending) ? 0 : monitored;
|
|
|
|
if (monitor) {
|
|
if (monitored)
|
|
log_verbose("%s/%s already monitored.", lv->vg->name, lv->name);
|
|
else if (seg->segtype->ops->target_monitor_events)
|
|
monitor_fn = seg->segtype->ops->target_monitor_events;
|
|
} else {
|
|
if (!monitored)
|
|
log_verbose("%s/%s already not monitored.", lv->vg->name, lv->name);
|
|
else if (seg->segtype->ops->target_unmonitor_events)
|
|
monitor_fn = seg->segtype->ops->target_unmonitor_events;
|
|
}
|
|
|
|
/* Do [un]monitor */
|
|
if (!monitor_fn)
|
|
continue;
|
|
|
|
log_verbose("%sonitoring %s/%s%s", monitor ? "M" : "Not m", lv->vg->name, lv->name,
|
|
test_mode() ? " [Test mode: skipping this]" : "");
|
|
|
|
/* FIXME Test mode should really continue a bit further. */
|
|
if (test_mode())
|
|
continue;
|
|
|
|
/* FIXME specify events */
|
|
if (!monitor_fn(seg, 0)) {
|
|
log_error("%s/%s: %s segment monitoring function failed.",
|
|
lv->vg->name, lv->name, seg->segtype->name);
|
|
return 0;
|
|
}
|
|
|
|
/* Check [un]monitor results */
|
|
/* Try a couple times if pending, but not forever... */
|
|
for (i = 0; i < 10; i++) {
|
|
pending = 0;
|
|
monitored = seg->segtype->ops->target_monitored(seg, &pending);
|
|
if (pending ||
|
|
(!monitored && monitor) ||
|
|
(monitored && !monitor))
|
|
log_very_verbose("%s/%s %smonitoring still pending: waiting...",
|
|
lv->vg->name, lv->name, monitor ? "" : "un");
|
|
else
|
|
break;
|
|
sleep(1);
|
|
}
|
|
|
|
if (r)
|
|
r = (monitored && monitor) || (!monitored && !monitor);
|
|
}
|
|
|
|
return r;
|
|
#else
|
|
return 1;
|
|
#endif
|
|
}
|
|
|
|
static int _lv_suspend(struct cmd_context *cmd, const char *lvid_s,
|
|
unsigned origin_only, int error_if_not_suspended)
|
|
{
|
|
struct logical_volume *lv = NULL, *lv_pre = NULL;
|
|
struct lvinfo info;
|
|
int r = 0, lockfs = 0, flush_required = 0;
|
|
|
|
if (!activation())
|
|
return 1;
|
|
|
|
if (!(lv = lv_from_lvid(cmd, lvid_s, 0)))
|
|
goto_out;
|
|
|
|
/* Use precommitted metadata if present */
|
|
if (!(lv_pre = lv_from_lvid(cmd, lvid_s, 1)))
|
|
goto_out;
|
|
|
|
/* Ignore origin_only unless LV is origin in both old and new metadata */
|
|
if (!lv_is_origin(lv) || !lv_is_origin(lv_pre))
|
|
origin_only = 0;
|
|
|
|
if (test_mode()) {
|
|
_skip("Suspending %s%s.", lv->name, origin_only ? " origin without snapshots" : "");
|
|
r = 1;
|
|
goto out;
|
|
}
|
|
|
|
if (!lv_info(cmd, lv, origin_only, &info, 0, 0))
|
|
goto_out;
|
|
|
|
if (!info.exists || info.suspended) {
|
|
if (!error_if_not_suspended) {
|
|
r = 1;
|
|
if (info.suspended)
|
|
memlock_inc(cmd);
|
|
}
|
|
goto out;
|
|
}
|
|
|
|
if (!lv_read_replicator_vgs(lv))
|
|
goto_out;
|
|
|
|
lv_calculate_readahead(lv, NULL);
|
|
|
|
/* If VG was precommitted, preload devices for the LV */
|
|
if ((lv_pre->vg->status & PRECOMMITTED)) {
|
|
if (!_lv_preload(lv_pre, origin_only, &flush_required)) {
|
|
/* FIXME Revert preloading */
|
|
goto_out;
|
|
}
|
|
}
|
|
|
|
if (!monitor_dev_for_events(cmd, lv, origin_only, 0))
|
|
/* FIXME Consider aborting here */
|
|
stack;
|
|
|
|
memlock_inc(cmd);
|
|
|
|
if (!origin_only &&
|
|
(lv_is_origin(lv_pre) || lv_is_cow(lv_pre)))
|
|
lockfs = 1;
|
|
|
|
if (!_lv_suspend_lv(lv, origin_only, lockfs, flush_required)) {
|
|
memlock_dec(cmd);
|
|
fs_unlock();
|
|
goto out;
|
|
}
|
|
|
|
r = 1;
|
|
out:
|
|
if (lv_pre)
|
|
free_vg(lv_pre->vg);
|
|
if (lv) {
|
|
lv_release_replicator_vgs(lv);
|
|
free_vg(lv->vg);
|
|
}
|
|
|
|
return r;
|
|
}
|
|
|
|
/* Returns success if the device is not active */
|
|
int lv_suspend_if_active(struct cmd_context *cmd, const char *lvid_s, unsigned origin_only)
|
|
{
|
|
return _lv_suspend(cmd, lvid_s, origin_only, 0);
|
|
}
|
|
|
|
/* No longer used */
|
|
/***********
|
|
int lv_suspend(struct cmd_context *cmd, const char *lvid_s)
|
|
{
|
|
return _lv_suspend(cmd, lvid_s, 1);
|
|
}
|
|
***********/
|
|
|
|
static int _lv_resume(struct cmd_context *cmd, const char *lvid_s,
|
|
unsigned origin_only,
|
|
int error_if_not_active)
|
|
{
|
|
struct logical_volume *lv;
|
|
struct lvinfo info;
|
|
int r = 0;
|
|
|
|
if (!activation())
|
|
return 1;
|
|
|
|
if (!(lv = lv_from_lvid(cmd, lvid_s, 0)))
|
|
goto_out;
|
|
|
|
if (!lv_is_origin(lv))
|
|
origin_only = 0;
|
|
|
|
if (test_mode()) {
|
|
_skip("Resuming %s%s.", lv->name, origin_only ? " without snapshots" : "");
|
|
r = 1;
|
|
goto out;
|
|
}
|
|
|
|
if (!lv_info(cmd, lv, origin_only, &info, 0, 0))
|
|
goto_out;
|
|
|
|
if (!info.exists || !info.suspended) {
|
|
if (error_if_not_active)
|
|
goto_out;
|
|
r = 1;
|
|
goto out;
|
|
}
|
|
|
|
if (!_lv_activate_lv(lv, origin_only))
|
|
goto_out;
|
|
|
|
memlock_dec(cmd);
|
|
|
|
if (!monitor_dev_for_events(cmd, lv, origin_only, 1))
|
|
stack;
|
|
|
|
r = 1;
|
|
out:
|
|
if (lv)
|
|
free_vg(lv->vg);
|
|
|
|
return r;
|
|
}
|
|
|
|
/* Returns success if the device is not active */
|
|
int lv_resume_if_active(struct cmd_context *cmd, const char *lvid_s, unsigned origin_only)
|
|
{
|
|
return _lv_resume(cmd, lvid_s, origin_only, 0);
|
|
}
|
|
|
|
int lv_resume(struct cmd_context *cmd, const char *lvid_s, unsigned origin_only)
|
|
{
|
|
return _lv_resume(cmd, lvid_s, origin_only, 1);
|
|
}
|
|
|
|
static int _lv_has_open_snapshots(struct logical_volume *lv)
|
|
{
|
|
struct lv_segment *snap_seg;
|
|
struct lvinfo info;
|
|
int r = 0;
|
|
|
|
dm_list_iterate_items_gen(snap_seg, &lv->snapshot_segs, origin_list) {
|
|
if (!lv_info(lv->vg->cmd, snap_seg->cow, 0, &info, 1, 0)) {
|
|
r = 1;
|
|
continue;
|
|
}
|
|
|
|
if (info.exists && info.open_count) {
|
|
log_error("LV %s/%s has open snapshot %s: "
|
|
"not deactivating", lv->vg->name, lv->name,
|
|
snap_seg->cow->name);
|
|
r = 1;
|
|
}
|
|
}
|
|
|
|
return r;
|
|
}
|
|
|
|
int lv_deactivate(struct cmd_context *cmd, const char *lvid_s)
|
|
{
|
|
struct logical_volume *lv;
|
|
struct lvinfo info;
|
|
int r = 0;
|
|
|
|
if (!activation())
|
|
return 1;
|
|
|
|
if (!(lv = lv_from_lvid(cmd, lvid_s, 0)))
|
|
goto out;
|
|
|
|
if (test_mode()) {
|
|
_skip("Deactivating '%s'.", lv->name);
|
|
r = 1;
|
|
goto out;
|
|
}
|
|
|
|
if (!lv_info(cmd, lv, 0, &info, 1, 0))
|
|
goto_out;
|
|
|
|
if (!info.exists) {
|
|
r = 1;
|
|
goto out;
|
|
}
|
|
|
|
if (lv_is_visible(lv)) {
|
|
if (info.open_count) {
|
|
log_error("LV %s/%s in use: not deactivating",
|
|
lv->vg->name, lv->name);
|
|
goto out;
|
|
}
|
|
if (lv_is_origin(lv) && _lv_has_open_snapshots(lv))
|
|
goto_out;
|
|
}
|
|
|
|
if (!lv_read_replicator_vgs(lv))
|
|
goto_out;
|
|
|
|
lv_calculate_readahead(lv, NULL);
|
|
|
|
if (!monitor_dev_for_events(cmd, lv, 0, 0))
|
|
stack;
|
|
|
|
memlock_inc(cmd);
|
|
r = _lv_deactivate(lv);
|
|
memlock_dec(cmd);
|
|
|
|
if (!lv_info(cmd, lv, 0, &info, 0, 0) || info.exists)
|
|
r = 0;
|
|
out:
|
|
if (lv) {
|
|
lv_release_replicator_vgs(lv);
|
|
free_vg(lv->vg);
|
|
}
|
|
|
|
return r;
|
|
}
|
|
|
|
/* Test if LV passes filter */
|
|
int lv_activation_filter(struct cmd_context *cmd, const char *lvid_s,
|
|
int *activate_lv)
|
|
{
|
|
struct logical_volume *lv;
|
|
int r = 0;
|
|
|
|
if (!activation()) {
|
|
*activate_lv = 1;
|
|
return 1;
|
|
}
|
|
|
|
if (!(lv = lv_from_lvid(cmd, lvid_s, 0)))
|
|
goto out;
|
|
|
|
if (!_passes_activation_filter(cmd, lv)) {
|
|
log_verbose("Not activating %s/%s since it does not pass "
|
|
"activation filter.", lv->vg->name, lv->name);
|
|
*activate_lv = 0;
|
|
} else
|
|
*activate_lv = 1;
|
|
r = 1;
|
|
out:
|
|
if (lv)
|
|
free_vg(lv->vg);
|
|
|
|
return r;
|
|
}
|
|
|
|
static int _lv_activate(struct cmd_context *cmd, const char *lvid_s,
|
|
int exclusive, int filter)
|
|
{
|
|
struct logical_volume *lv;
|
|
struct lvinfo info;
|
|
int r = 0;
|
|
|
|
if (!activation())
|
|
return 1;
|
|
|
|
if (!(lv = lv_from_lvid(cmd, lvid_s, 0)))
|
|
goto out;
|
|
|
|
if (filter && !_passes_activation_filter(cmd, lv)) {
|
|
log_error("Not activating %s/%s since it does not pass "
|
|
"activation filter.", lv->vg->name, lv->name);
|
|
goto out;
|
|
}
|
|
|
|
if ((!lv->vg->cmd->partial_activation) && (lv->status & PARTIAL_LV)) {
|
|
log_error("Refusing activation of partial LV %s. Use --partial to override.",
|
|
lv->name);
|
|
goto_out;
|
|
}
|
|
|
|
if (lv_has_unknown_segments(lv)) {
|
|
log_error("Refusing activation of LV %s containing "
|
|
"an unrecognised segment.", lv->name);
|
|
goto_out;
|
|
}
|
|
|
|
if (test_mode()) {
|
|
_skip("Activating '%s'.", lv->name);
|
|
r = 1;
|
|
goto out;
|
|
}
|
|
|
|
if (!lv_info(cmd, lv, 0, &info, 0, 0))
|
|
goto_out;
|
|
|
|
if (info.exists && !info.suspended && info.live_table) {
|
|
r = 1;
|
|
goto out;
|
|
}
|
|
|
|
if (!lv_read_replicator_vgs(lv))
|
|
goto_out;
|
|
|
|
lv_calculate_readahead(lv, NULL);
|
|
|
|
if (exclusive)
|
|
lv->status |= ACTIVATE_EXCL;
|
|
|
|
memlock_inc(cmd);
|
|
if (!(r = _lv_activate_lv(lv, 0)))
|
|
stack;
|
|
memlock_dec(cmd);
|
|
|
|
if (r && !monitor_dev_for_events(cmd, lv, 0, 1))
|
|
stack;
|
|
|
|
out:
|
|
if (lv) {
|
|
lv_release_replicator_vgs(lv);
|
|
free_vg(lv->vg);
|
|
}
|
|
|
|
return r;
|
|
}
|
|
|
|
/* Activate LV */
|
|
int lv_activate(struct cmd_context *cmd, const char *lvid_s, int exclusive)
|
|
{
|
|
if (!_lv_activate(cmd, lvid_s, exclusive, 0))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
/* Activate LV only if it passes filter */
|
|
int lv_activate_with_filter(struct cmd_context *cmd, const char *lvid_s, int exclusive)
|
|
{
|
|
if (!_lv_activate(cmd, lvid_s, exclusive, 1))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
int lv_mknodes(struct cmd_context *cmd, const struct logical_volume *lv)
|
|
{
|
|
int r = 1;
|
|
|
|
if (!lv) {
|
|
r = dm_mknodes(NULL);
|
|
fs_unlock();
|
|
return r;
|
|
}
|
|
|
|
if (!activation())
|
|
return 1;
|
|
|
|
r = dev_manager_mknodes(lv);
|
|
|
|
fs_unlock();
|
|
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* Does PV use VG somewhere in its construction?
|
|
* Returns 1 on failure.
|
|
*/
|
|
int pv_uses_vg(struct physical_volume *pv,
|
|
struct volume_group *vg)
|
|
{
|
|
if (!activation())
|
|
return 0;
|
|
|
|
if (!dm_is_dm_major(MAJOR(pv->dev->dev)))
|
|
return 0;
|
|
|
|
return dev_manager_device_uses_vg(pv->dev, vg);
|
|
}
|
|
|
|
void activation_release(void)
|
|
{
|
|
dev_manager_release();
|
|
}
|
|
|
|
void activation_exit(void)
|
|
{
|
|
dev_manager_exit();
|
|
}
|
|
#endif
|