mirror of
git://sourceware.org/git/lvm2.git
synced 2024-12-22 17:35:59 +03:00
a30832cedd
Commit 02f6f4902f
introduced a bug that caused
RAID devices to fail to activate if the device for a single sub-LV failed.
The special case of LVM mirror was handled, but not LVM RAID.
EXAMPLE:
[root@bp-01 ~]# devices vg
LV Copy% Devices
lv 100.00 lv_rimage_0(0),lv_rimage_1(0)
[lv_rimage_0] /dev/sde1(1)
[lv_rimage_1] /dev/sdh1(1)
[lv_rmeta_0] /dev/sde1(0)
[lv_rmeta_1] /dev/sdh1(0)
[root@bp-01 ~]# vgchange -an vg
0 logical volume(s) in volume group "vg" now active
[root@bp-01 ~]# off.sh sdh
Turning off sdh
[root@bp-01 ~]# vgchange -ay vg --partial
Partial mode. Incomplete logical volumes will be processed.
Couldn't find device with uuid fbI0YO-GX7x-firU-Vy5o-vzwx-vAKZ-feRxfF.
Cannot activate vg/lv_rimage_1: all segments missing.
0 logical volume(s) in volume group "vg" now active
AFTER this patch:
[root@bp-01 ~]# vgchange -ay vg --partial
Partial mode. Incomplete logical volumes will be processed.
Couldn't find device with uuid fbI0YO-GX7x-firU-Vy5o-vzwx-vAKZ-feRxfF.
1 logical volume(s) in volume group "vg" now active
[root@bp-01 ~]# devices vg
Couldn't find device with uuid fbI0YO-GX7x-firU-Vy5o-vzwx-vAKZ-feRxfF.
LV Copy% Devices
lv 100.00 lv_rimage_0(0),lv_rimage_1(0)
[lv_rimage_0] /dev/sde1(1)
[lv_rimage_1] unknown device(1)
[lv_rmeta_0] /dev/sde1(0)
[lv_rmeta_1] unknown device(0)
[root@bp-01 ~]# dmsetup table vg-lv; dmsetup status vg-lv
0 1024000 raid raid1 3 0 region_size 1024 2 253:2 253:3 - -
0 1024000 raid raid1 2 AD 1024000/1024000
No WHATSNEW update necessary because this is an intrarelease fix.
brassow
2190 lines
57 KiB
C
2190 lines
57 KiB
C
/*
|
|
* Copyright (C) 2002-2004 Sistina Software, Inc. All rights reserved.
|
|
* Copyright (C) 2004-2012 Red Hat, Inc. All rights reserved.
|
|
*
|
|
* This file is part of LVM2.
|
|
*
|
|
* This copyrighted material is made available to anyone wishing to use,
|
|
* modify, copy, or redistribute it subject to the terms and conditions
|
|
* of the GNU Lesser General Public License v.2.1.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public License
|
|
* along with this program; if not, write to the Free Software Foundation,
|
|
* Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
*/
|
|
|
|
#include "lib.h"
|
|
#include "str_list.h"
|
|
#include "dev_manager.h"
|
|
#include "lvm-string.h"
|
|
#include "fs.h"
|
|
#include "defaults.h"
|
|
#include "segtype.h"
|
|
#include "display.h"
|
|
#include "toolcontext.h"
|
|
#include "targets.h"
|
|
#include "config.h"
|
|
#include "filter.h"
|
|
#include "activate.h"
|
|
|
|
#include <limits.h>
|
|
#include <dirent.h>
|
|
|
|
#define MAX_TARGET_PARAMSIZE 50000
|
|
|
|
typedef enum {
|
|
PRELOAD,
|
|
ACTIVATE,
|
|
DEACTIVATE,
|
|
SUSPEND,
|
|
SUSPEND_WITH_LOCKFS,
|
|
CLEAN
|
|
} action_t;
|
|
|
|
struct dev_manager {
|
|
struct dm_pool *mem;
|
|
|
|
struct cmd_context *cmd;
|
|
|
|
void *target_state;
|
|
uint32_t pvmove_mirror_count;
|
|
int flush_required;
|
|
unsigned track_pvmove_deps;
|
|
|
|
char *vg_name;
|
|
};
|
|
|
|
struct lv_layer {
|
|
struct logical_volume *lv;
|
|
const char *old_name;
|
|
};
|
|
|
|
static const char _thin_layer[] = "tpool";
|
|
|
|
int read_only_lv(struct logical_volume *lv, struct lv_activate_opts *laopts)
|
|
{
|
|
return (laopts->read_only || !(lv->vg->status & LVM_WRITE) || !(lv->status & LVM_WRITE));
|
|
}
|
|
|
|
/*
|
|
* Low level device-layer operations.
|
|
*/
|
|
static struct dm_task *_setup_task(const char *name, const char *uuid,
|
|
uint32_t *event_nr, int task,
|
|
uint32_t major, uint32_t minor)
|
|
{
|
|
struct dm_task *dmt;
|
|
|
|
if (!(dmt = dm_task_create(task)))
|
|
return_NULL;
|
|
|
|
if (name && !dm_task_set_name(dmt, name))
|
|
goto_out;
|
|
|
|
if (uuid && *uuid && !dm_task_set_uuid(dmt, uuid))
|
|
goto_out;
|
|
|
|
if (event_nr && !dm_task_set_event_nr(dmt, *event_nr))
|
|
goto_out;
|
|
|
|
if (major && !dm_task_set_major_minor(dmt, major, minor, 1))
|
|
goto_out;
|
|
|
|
if (activation_checks() && !dm_task_enable_checks(dmt))
|
|
goto_out;
|
|
|
|
return dmt;
|
|
out:
|
|
dm_task_destroy(dmt);
|
|
return NULL;
|
|
}
|
|
|
|
static int _info_run(const char *name, const char *dlid, struct dm_info *info,
|
|
uint32_t *read_ahead, int mknodes, int with_open_count,
|
|
int with_read_ahead, uint32_t major, uint32_t minor)
|
|
{
|
|
int r = 0;
|
|
struct dm_task *dmt;
|
|
int dmtask;
|
|
|
|
dmtask = mknodes ? DM_DEVICE_MKNODES : DM_DEVICE_INFO;
|
|
|
|
if (!(dmt = _setup_task(mknodes ? name : NULL, dlid, 0, dmtask, major, minor)))
|
|
return_0;
|
|
|
|
if (!with_open_count)
|
|
if (!dm_task_no_open_count(dmt))
|
|
log_error("Failed to disable open_count");
|
|
|
|
if (!dm_task_run(dmt))
|
|
goto_out;
|
|
|
|
if (!dm_task_get_info(dmt, info))
|
|
goto_out;
|
|
|
|
if (with_read_ahead && info->exists) {
|
|
if (!dm_task_get_read_ahead(dmt, read_ahead))
|
|
goto_out;
|
|
} else if (read_ahead)
|
|
*read_ahead = DM_READ_AHEAD_NONE;
|
|
|
|
r = 1;
|
|
|
|
out:
|
|
dm_task_destroy(dmt);
|
|
return r;
|
|
}
|
|
|
|
int device_is_usable(struct device *dev)
|
|
{
|
|
struct dm_task *dmt;
|
|
struct dm_info info;
|
|
const char *name, *uuid;
|
|
uint64_t start, length;
|
|
char *target_type = NULL;
|
|
char *params, *vgname = NULL, *lvname, *layer;
|
|
void *next = NULL;
|
|
int only_error_target = 1;
|
|
int r = 0;
|
|
|
|
if (!(dmt = dm_task_create(DM_DEVICE_STATUS)))
|
|
return_0;
|
|
|
|
if (!dm_task_set_major_minor(dmt, MAJOR(dev->dev), MINOR(dev->dev), 1))
|
|
goto_out;
|
|
|
|
if (activation_checks() && !dm_task_enable_checks(dmt))
|
|
goto_out;
|
|
|
|
if (!dm_task_run(dmt)) {
|
|
log_error("Failed to get state of mapped device");
|
|
goto out;
|
|
}
|
|
|
|
if (!dm_task_get_info(dmt, &info))
|
|
goto_out;
|
|
|
|
if (!info.exists)
|
|
goto out;
|
|
|
|
name = dm_task_get_name(dmt);
|
|
uuid = dm_task_get_uuid(dmt);
|
|
|
|
if (!info.target_count) {
|
|
log_debug("%s: Empty device %s not usable.", dev_name(dev), name);
|
|
goto out;
|
|
}
|
|
|
|
if (info.suspended && ignore_suspended_devices()) {
|
|
log_debug("%s: Suspended device %s not usable.", dev_name(dev), name);
|
|
goto out;
|
|
}
|
|
|
|
/* FIXME Also check for mirror block_on_error and mpath no paths */
|
|
/* For now, we exclude all mirrors */
|
|
|
|
do {
|
|
next = dm_get_next_target(dmt, next, &start, &length,
|
|
&target_type, ¶ms);
|
|
/* Skip if target type doesn't match */
|
|
if (target_type && !strcmp(target_type, "mirror") && ignore_suspended_devices()) {
|
|
log_debug("%s: Mirror device %s not usable.", dev_name(dev), name);
|
|
goto out;
|
|
}
|
|
|
|
/*
|
|
* Snapshot origin could be sitting on top of a mirror which
|
|
* could be blocking I/O. Skip snapshot origins entirely for
|
|
* now.
|
|
*
|
|
* FIXME: rather than skipping origin, check if mirror is
|
|
* underneath and if the mirror is blocking I/O.
|
|
*/
|
|
if (target_type && !strcmp(target_type, "snapshot-origin") &&
|
|
ignore_suspended_devices()) {
|
|
log_debug("%s: Snapshot-origin device %s not usable.",
|
|
dev_name(dev), name);
|
|
goto out;
|
|
}
|
|
|
|
if (target_type && strcmp(target_type, "error"))
|
|
only_error_target = 0;
|
|
} while (next);
|
|
|
|
/* Skip devices consisting entirely of error targets. */
|
|
/* FIXME Deal with device stacked above error targets? */
|
|
if (only_error_target) {
|
|
log_debug("%s: Error device %s not usable.",
|
|
dev_name(dev), name);
|
|
goto out;
|
|
}
|
|
|
|
/* FIXME Also check dependencies? */
|
|
|
|
/* Check internal lvm devices */
|
|
if (uuid && !strncmp(uuid, UUID_PREFIX, sizeof(UUID_PREFIX) - 1)) {
|
|
if (!(vgname = dm_strdup(name)) ||
|
|
!dm_split_lvm_name(NULL, NULL, &vgname, &lvname, &layer))
|
|
goto_out;
|
|
|
|
if (lvname && (is_reserved_lvname(lvname) || *layer)) {
|
|
log_debug("%s: Reserved internal LV device %s/%s%s%s not usable.",
|
|
dev_name(dev), vgname, lvname, *layer ? "-" : "", layer);
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
r = 1;
|
|
|
|
out:
|
|
dm_free(vgname);
|
|
dm_task_destroy(dmt);
|
|
return r;
|
|
}
|
|
|
|
static int _info(const char *dlid, int with_open_count, int with_read_ahead,
|
|
struct dm_info *info, uint32_t *read_ahead)
|
|
{
|
|
int r = 0;
|
|
|
|
if ((r = _info_run(NULL, dlid, info, read_ahead, 0, with_open_count,
|
|
with_read_ahead, 0, 0)) && info->exists)
|
|
return 1;
|
|
else if ((r = _info_run(NULL, dlid + sizeof(UUID_PREFIX) - 1, info,
|
|
read_ahead, 0, with_open_count,
|
|
with_read_ahead, 0, 0)) && info->exists)
|
|
return 1;
|
|
|
|
return r;
|
|
}
|
|
|
|
static int _info_by_dev(uint32_t major, uint32_t minor, struct dm_info *info)
|
|
{
|
|
return _info_run(NULL, NULL, info, NULL, 0, 0, 0, major, minor);
|
|
}
|
|
|
|
int dev_manager_info(struct dm_pool *mem, const struct logical_volume *lv,
|
|
const char *layer,
|
|
int with_open_count, int with_read_ahead,
|
|
struct dm_info *info, uint32_t *read_ahead)
|
|
{
|
|
char *dlid, *name;
|
|
int r;
|
|
|
|
if (!(name = dm_build_dm_name(mem, lv->vg->name, lv->name, layer))) {
|
|
log_error("name build failed for %s", lv->name);
|
|
return 0;
|
|
}
|
|
|
|
if (!(dlid = build_dm_uuid(mem, lv->lvid.s, layer))) {
|
|
log_error("dlid build failed for %s", name);
|
|
return 0;
|
|
}
|
|
|
|
log_debug("Getting device info for %s [%s]", name, dlid);
|
|
r = _info(dlid, with_open_count, with_read_ahead, info, read_ahead);
|
|
|
|
dm_pool_free(mem, name);
|
|
return r;
|
|
}
|
|
|
|
static const struct dm_info *_cached_info(struct dm_pool *mem,
|
|
const struct logical_volume *lv,
|
|
struct dm_tree *dtree)
|
|
{
|
|
const char *dlid;
|
|
struct dm_tree_node *dnode;
|
|
const struct dm_info *dinfo;
|
|
|
|
if (!(dlid = build_dm_uuid(mem, lv->lvid.s, NULL))) {
|
|
log_error("dlid build failed for %s", lv->name);
|
|
return NULL;
|
|
}
|
|
|
|
/* An activating merging origin won't have a node in the tree yet */
|
|
if (!(dnode = dm_tree_find_node_by_uuid(dtree, dlid)))
|
|
return NULL;
|
|
|
|
if (!(dinfo = dm_tree_node_get_info(dnode))) {
|
|
log_error("failed to get info from tree node for %s", lv->name);
|
|
return NULL;
|
|
}
|
|
|
|
if (!dinfo->exists)
|
|
return NULL;
|
|
|
|
return dinfo;
|
|
}
|
|
|
|
#if 0
|
|
/* FIXME Interface must cope with multiple targets */
|
|
static int _status_run(const char *name, const char *uuid,
|
|
unsigned long long *s, unsigned long long *l,
|
|
char **t, uint32_t t_size, char **p, uint32_t p_size)
|
|
{
|
|
int r = 0;
|
|
struct dm_task *dmt;
|
|
struct dm_info info;
|
|
void *next = NULL;
|
|
uint64_t start, length;
|
|
char *type = NULL;
|
|
char *params = NULL;
|
|
|
|
if (!(dmt = _setup_task(name, uuid, 0, DM_DEVICE_STATUS, 0, 0)))
|
|
return_0;
|
|
|
|
if (!dm_task_no_open_count(dmt))
|
|
log_error("Failed to disable open_count");
|
|
|
|
if (!dm_task_run(dmt))
|
|
goto_out;
|
|
|
|
if (!dm_task_get_info(dmt, &info) || !info.exists)
|
|
goto_out;
|
|
|
|
do {
|
|
next = dm_get_next_target(dmt, next, &start, &length,
|
|
&type, ¶ms);
|
|
if (type) {
|
|
*s = start;
|
|
*l = length;
|
|
/* Make sure things are null terminated */
|
|
strncpy(*t, type, t_size);
|
|
(*t)[t_size - 1] = '\0';
|
|
strncpy(*p, params, p_size);
|
|
(*p)[p_size - 1] = '\0';
|
|
|
|
r = 1;
|
|
/* FIXME Cope with multiple targets! */
|
|
break;
|
|
}
|
|
|
|
} while (next);
|
|
|
|
out:
|
|
dm_task_destroy(dmt);
|
|
return r;
|
|
}
|
|
|
|
static int _status(const char *name, const char *uuid,
|
|
unsigned long long *start, unsigned long long *length,
|
|
char **type, uint32_t type_size, char **params,
|
|
uint32_t param_size) __attribute__ ((unused));
|
|
|
|
static int _status(const char *name, const char *uuid,
|
|
unsigned long long *start, unsigned long long *length,
|
|
char **type, uint32_t type_size, char **params,
|
|
uint32_t param_size)
|
|
{
|
|
if (uuid && *uuid) {
|
|
if (_status_run(NULL, uuid, start, length, type,
|
|
type_size, params, param_size) &&
|
|
*params)
|
|
return 1;
|
|
else if (_status_run(NULL, uuid + sizeof(UUID_PREFIX) - 1, start,
|
|
length, type, type_size, params,
|
|
param_size) &&
|
|
*params)
|
|
return 1;
|
|
}
|
|
|
|
if (name && _status_run(name, NULL, start, length, type, type_size,
|
|
params, param_size))
|
|
return 1;
|
|
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
int lv_has_target_type(struct dm_pool *mem, struct logical_volume *lv,
|
|
const char *layer, const char *target_type)
|
|
{
|
|
int r = 0;
|
|
char *dlid;
|
|
struct dm_task *dmt;
|
|
struct dm_info info;
|
|
void *next = NULL;
|
|
uint64_t start, length;
|
|
char *type = NULL;
|
|
char *params = NULL;
|
|
|
|
if (!(dlid = build_dm_uuid(mem, lv->lvid.s, layer)))
|
|
return_0;
|
|
|
|
if (!(dmt = _setup_task(NULL, dlid, 0,
|
|
DM_DEVICE_STATUS, 0, 0)))
|
|
goto_bad;
|
|
|
|
if (!dm_task_no_open_count(dmt))
|
|
log_error("Failed to disable open_count");
|
|
|
|
if (!dm_task_run(dmt))
|
|
goto_out;
|
|
|
|
if (!dm_task_get_info(dmt, &info) || !info.exists)
|
|
goto_out;
|
|
|
|
do {
|
|
next = dm_get_next_target(dmt, next, &start, &length,
|
|
&type, ¶ms);
|
|
if (type && strncmp(type, target_type,
|
|
strlen(target_type)) == 0) {
|
|
if (info.live_table)
|
|
r = 1;
|
|
break;
|
|
}
|
|
} while (next);
|
|
|
|
out:
|
|
dm_task_destroy(dmt);
|
|
bad:
|
|
dm_pool_free(mem, dlid);
|
|
|
|
return r;
|
|
}
|
|
|
|
int add_linear_area_to_dtree(struct dm_tree_node *node, uint64_t size, uint32_t extent_size, int use_linear_target, const char *vgname, const char *lvname)
|
|
{
|
|
uint32_t page_size;
|
|
|
|
/*
|
|
* Use striped or linear target?
|
|
*/
|
|
if (!use_linear_target) {
|
|
page_size = lvm_getpagesize() >> SECTOR_SHIFT;
|
|
|
|
/*
|
|
* We'll use the extent size as the stripe size.
|
|
* Extent size and page size are always powers of 2.
|
|
* The striped target requires that the stripe size is
|
|
* divisible by the page size.
|
|
*/
|
|
if (extent_size >= page_size) {
|
|
/* Use striped target */
|
|
if (!dm_tree_node_add_striped_target(node, size, extent_size))
|
|
return_0;
|
|
return 1;
|
|
} else
|
|
/* Some exotic cases are unsupported by striped. */
|
|
log_warn("WARNING: Using linear target for %s/%s: Striped requires extent size (%" PRIu32 " sectors) >= page size (%" PRIu32 ").",
|
|
vgname, lvname, extent_size, page_size);
|
|
}
|
|
|
|
/*
|
|
* Use linear target.
|
|
*/
|
|
if (!dm_tree_node_add_linear_target(node, size))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
static percent_range_t _combine_percent(percent_t a, percent_t b,
|
|
uint32_t numerator, uint32_t denominator)
|
|
{
|
|
if (a == PERCENT_MERGE_FAILED || b == PERCENT_MERGE_FAILED)
|
|
return PERCENT_MERGE_FAILED;
|
|
|
|
if (a == PERCENT_INVALID || b == PERCENT_INVALID)
|
|
return PERCENT_INVALID;
|
|
|
|
if (a == PERCENT_100 && b == PERCENT_100)
|
|
return PERCENT_100;
|
|
|
|
if (a == PERCENT_0 && b == PERCENT_0)
|
|
return PERCENT_0;
|
|
|
|
return make_percent(numerator, denominator);
|
|
}
|
|
|
|
static int _percent_run(struct dev_manager *dm, const char *name,
|
|
const char *dlid,
|
|
const char *target_type, int wait,
|
|
const struct logical_volume *lv, percent_t *overall_percent,
|
|
uint32_t *event_nr, int fail_if_percent_unsupported)
|
|
{
|
|
int r = 0;
|
|
struct dm_task *dmt;
|
|
struct dm_info info;
|
|
void *next = NULL;
|
|
uint64_t start, length;
|
|
char *type = NULL;
|
|
char *params = NULL;
|
|
const struct dm_list *segh = lv ? &lv->segments : NULL;
|
|
struct lv_segment *seg = NULL;
|
|
struct segment_type *segtype;
|
|
int first_time = 1;
|
|
percent_t percent;
|
|
|
|
uint64_t total_numerator = 0, total_denominator = 0;
|
|
|
|
*overall_percent = PERCENT_INVALID;
|
|
|
|
if (!(dmt = _setup_task(name, dlid, event_nr,
|
|
wait ? DM_DEVICE_WAITEVENT : DM_DEVICE_STATUS, 0, 0)))
|
|
return_0;
|
|
|
|
if (!dm_task_no_open_count(dmt))
|
|
log_error("Failed to disable open_count");
|
|
|
|
if (!dm_task_run(dmt))
|
|
goto_out;
|
|
|
|
if (!dm_task_get_info(dmt, &info) || !info.exists)
|
|
goto_out;
|
|
|
|
if (event_nr)
|
|
*event_nr = info.event_nr;
|
|
|
|
do {
|
|
next = dm_get_next_target(dmt, next, &start, &length, &type,
|
|
¶ms);
|
|
if (lv) {
|
|
if (!(segh = dm_list_next(&lv->segments, segh))) {
|
|
log_error("Number of segments in active LV %s "
|
|
"does not match metadata", lv->name);
|
|
goto out;
|
|
}
|
|
seg = dm_list_item(segh, struct lv_segment);
|
|
}
|
|
|
|
if (!type || !params)
|
|
continue;
|
|
|
|
if (!(segtype = get_segtype_from_string(dm->cmd, target_type)))
|
|
continue;
|
|
|
|
if (strcmp(type, target_type)) {
|
|
/* If kernel's type isn't an exact match is it compatible? */
|
|
if (!segtype->ops->target_status_compatible ||
|
|
!segtype->ops->target_status_compatible(type))
|
|
continue;
|
|
}
|
|
|
|
if (!segtype->ops->target_percent)
|
|
continue;
|
|
|
|
if (!segtype->ops->target_percent(&dm->target_state,
|
|
&percent, dm->mem,
|
|
dm->cmd, seg, params,
|
|
&total_numerator,
|
|
&total_denominator))
|
|
goto_out;
|
|
|
|
if (first_time) {
|
|
*overall_percent = percent;
|
|
first_time = 0;
|
|
} else
|
|
*overall_percent =
|
|
_combine_percent(*overall_percent, percent,
|
|
total_numerator, total_denominator);
|
|
} while (next);
|
|
|
|
if (lv && dm_list_next(&lv->segments, segh)) {
|
|
log_error("Number of segments in active LV %s does not "
|
|
"match metadata", lv->name);
|
|
goto out;
|
|
}
|
|
|
|
if (first_time) {
|
|
/* above ->target_percent() was not executed! */
|
|
/* FIXME why return PERCENT_100 et. al. in this case? */
|
|
*overall_percent = PERCENT_100;
|
|
if (fail_if_percent_unsupported)
|
|
goto_out;
|
|
}
|
|
|
|
log_debug("LV percent: %f", percent_to_float(*overall_percent));
|
|
r = 1;
|
|
|
|
out:
|
|
dm_task_destroy(dmt);
|
|
return r;
|
|
}
|
|
|
|
static int _percent(struct dev_manager *dm, const char *name, const char *dlid,
|
|
const char *target_type, int wait,
|
|
const struct logical_volume *lv, percent_t *percent,
|
|
uint32_t *event_nr, int fail_if_percent_unsupported)
|
|
{
|
|
if (dlid && *dlid) {
|
|
if (_percent_run(dm, NULL, dlid, target_type, wait, lv, percent,
|
|
event_nr, fail_if_percent_unsupported))
|
|
return 1;
|
|
else if (_percent_run(dm, NULL, dlid + sizeof(UUID_PREFIX) - 1,
|
|
target_type, wait, lv, percent,
|
|
event_nr, fail_if_percent_unsupported))
|
|
return 1;
|
|
}
|
|
|
|
if (name && _percent_run(dm, name, NULL, target_type, wait, lv, percent,
|
|
event_nr, fail_if_percent_unsupported))
|
|
return 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* FIXME Merge with the percent function */
|
|
int dev_manager_transient(struct dev_manager *dm, struct logical_volume *lv)
|
|
{
|
|
int r = 0;
|
|
struct dm_task *dmt;
|
|
struct dm_info info;
|
|
void *next = NULL;
|
|
uint64_t start, length;
|
|
char *type = NULL;
|
|
char *params = NULL;
|
|
char *dlid = NULL;
|
|
const char *layer = lv_is_origin(lv) ? "real" : NULL;
|
|
const struct dm_list *segh = &lv->segments;
|
|
struct lv_segment *seg = NULL;
|
|
|
|
if (!(dlid = build_dm_uuid(dm->mem, lv->lvid.s, layer)))
|
|
return_0;
|
|
|
|
if (!(dmt = _setup_task(0, dlid, NULL, DM_DEVICE_STATUS, 0, 0)))
|
|
return_0;
|
|
|
|
if (!dm_task_no_open_count(dmt))
|
|
log_error("Failed to disable open_count");
|
|
|
|
if (!dm_task_run(dmt))
|
|
goto_out;
|
|
|
|
if (!dm_task_get_info(dmt, &info) || !info.exists)
|
|
goto_out;
|
|
|
|
do {
|
|
next = dm_get_next_target(dmt, next, &start, &length, &type,
|
|
¶ms);
|
|
|
|
if (!(segh = dm_list_next(&lv->segments, segh))) {
|
|
log_error("Number of segments in active LV %s "
|
|
"does not match metadata", lv->name);
|
|
goto out;
|
|
}
|
|
seg = dm_list_item(segh, struct lv_segment);
|
|
|
|
if (!type || !params)
|
|
continue;
|
|
|
|
if (seg->segtype->ops->check_transient_status &&
|
|
!seg->segtype->ops->check_transient_status(seg, params))
|
|
goto_out;
|
|
|
|
} while (next);
|
|
|
|
if (dm_list_next(&lv->segments, segh)) {
|
|
log_error("Number of segments in active LV %s does not "
|
|
"match metadata", lv->name);
|
|
goto out;
|
|
}
|
|
|
|
r = 1;
|
|
|
|
out:
|
|
dm_task_destroy(dmt);
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* dev_manager implementation.
|
|
*/
|
|
struct dev_manager *dev_manager_create(struct cmd_context *cmd,
|
|
const char *vg_name,
|
|
unsigned track_pvmove_deps)
|
|
{
|
|
struct dm_pool *mem;
|
|
struct dev_manager *dm;
|
|
|
|
if (!(mem = dm_pool_create("dev_manager", 16 * 1024)))
|
|
return_NULL;
|
|
|
|
if (!(dm = dm_pool_zalloc(mem, sizeof(*dm))))
|
|
goto_bad;
|
|
|
|
dm->cmd = cmd;
|
|
dm->mem = mem;
|
|
|
|
if (!(dm->vg_name = dm_pool_strdup(dm->mem, vg_name)))
|
|
goto_bad;
|
|
|
|
/*
|
|
* When we manipulate (normally suspend/resume) the PVMOVE
|
|
* device directly, there's no need to touch the LVs above.
|
|
*/
|
|
dm->track_pvmove_deps = track_pvmove_deps;
|
|
|
|
dm->target_state = NULL;
|
|
|
|
dm_udev_set_sync_support(cmd->current_settings.udev_sync);
|
|
|
|
return dm;
|
|
|
|
bad:
|
|
dm_pool_destroy(mem);
|
|
return NULL;
|
|
}
|
|
|
|
void dev_manager_destroy(struct dev_manager *dm)
|
|
{
|
|
dm_pool_destroy(dm->mem);
|
|
}
|
|
|
|
void dev_manager_release(void)
|
|
{
|
|
dm_lib_release();
|
|
}
|
|
|
|
void dev_manager_exit(void)
|
|
{
|
|
dm_lib_exit();
|
|
}
|
|
|
|
int dev_manager_snapshot_percent(struct dev_manager *dm,
|
|
const struct logical_volume *lv,
|
|
percent_t *percent)
|
|
{
|
|
const struct logical_volume *snap_lv;
|
|
char *name;
|
|
const char *dlid;
|
|
int fail_if_percent_unsupported = 0;
|
|
|
|
if (lv_is_merging_origin(lv)) {
|
|
/*
|
|
* Set 'fail_if_percent_unsupported', otherwise passing
|
|
* unsupported LV types to _percent will lead to a default
|
|
* successful return with percent_range as PERCENT_100.
|
|
* - For a merging origin, this will result in a polldaemon
|
|
* that runs infinitely (because completion is PERCENT_0)
|
|
* - We unfortunately don't yet _know_ if a snapshot-merge
|
|
* target is active (activation is deferred if dev is open);
|
|
* so we can't short-circuit origin devices based purely on
|
|
* existing LVM LV attributes.
|
|
*/
|
|
fail_if_percent_unsupported = 1;
|
|
}
|
|
|
|
if (lv_is_merging_cow(lv)) {
|
|
/* must check percent of origin for a merging snapshot */
|
|
snap_lv = origin_from_cow(lv);
|
|
} else
|
|
snap_lv = lv;
|
|
|
|
/*
|
|
* Build a name for the top layer.
|
|
*/
|
|
if (!(name = dm_build_dm_name(dm->mem, snap_lv->vg->name, snap_lv->name, NULL)))
|
|
return_0;
|
|
|
|
if (!(dlid = build_dm_uuid(dm->mem, snap_lv->lvid.s, NULL)))
|
|
return_0;
|
|
|
|
/*
|
|
* Try and get some info on this device.
|
|
*/
|
|
log_debug("Getting device status percentage for %s", name);
|
|
if (!(_percent(dm, name, dlid, "snapshot", 0, NULL, percent,
|
|
NULL, fail_if_percent_unsupported)))
|
|
return_0;
|
|
|
|
/* If the snapshot isn't available, percent will be -1 */
|
|
return 1;
|
|
}
|
|
|
|
/* FIXME Merge with snapshot_percent, auto-detecting target type */
|
|
/* FIXME Cope with more than one target */
|
|
int dev_manager_mirror_percent(struct dev_manager *dm,
|
|
const struct logical_volume *lv, int wait,
|
|
percent_t *percent, uint32_t *event_nr)
|
|
{
|
|
char *name;
|
|
const char *dlid;
|
|
const char *target_type = first_seg(lv)->segtype->name;
|
|
const char *layer = (lv_is_origin(lv)) ? "real" : NULL;
|
|
|
|
/*
|
|
* Build a name for the top layer.
|
|
*/
|
|
if (!(name = dm_build_dm_name(dm->mem, lv->vg->name, lv->name, layer)))
|
|
return_0;
|
|
|
|
if (!(dlid = build_dm_uuid(dm->mem, lv->lvid.s, layer))) {
|
|
log_error("dlid build failed for %s", lv->name);
|
|
return 0;
|
|
}
|
|
|
|
log_debug("Getting device %s status percentage for %s",
|
|
target_type, name);
|
|
if (!(_percent(dm, name, dlid, target_type, wait, lv, percent,
|
|
event_nr, 0)))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
#if 0
|
|
log_very_verbose("%s %s", sus ? "Suspending" : "Resuming", name);
|
|
|
|
log_verbose("Loading %s", dl->name);
|
|
log_very_verbose("Activating %s read-only", dl->name);
|
|
log_very_verbose("Activated %s %s %03u:%03u", dl->name,
|
|
dl->dlid, dl->info.major, dl->info.minor);
|
|
|
|
if (_get_flag(dl, VISIBLE))
|
|
log_verbose("Removing %s", dl->name);
|
|
else
|
|
log_very_verbose("Removing %s", dl->name);
|
|
|
|
log_debug("Adding target: %" PRIu64 " %" PRIu64 " %s %s",
|
|
extent_size * seg->le, extent_size * seg->len, target, params);
|
|
|
|
log_debug("Adding target: 0 %" PRIu64 " snapshot-origin %s",
|
|
dl->lv->size, params);
|
|
log_debug("Adding target: 0 %" PRIu64 " snapshot %s", size, params);
|
|
log_debug("Getting device info for %s", dl->name);
|
|
|
|
/* Rename? */
|
|
if ((suffix = strrchr(dl->dlid + sizeof(UUID_PREFIX) - 1, '-')))
|
|
suffix++;
|
|
new_name = dm_build_dm_name(dm->mem, dm->vg_name, dl->lv->name,
|
|
suffix);
|
|
|
|
static int _belong_to_vg(const char *vgname, const char *name)
|
|
{
|
|
const char *v = vgname, *n = name;
|
|
|
|
while (*v) {
|
|
if ((*v != *n) || (*v == '-' && *(++n) != '-'))
|
|
return 0;
|
|
v++, n++;
|
|
}
|
|
|
|
if (*n == '-' && *(n + 1) != '-')
|
|
return 1;
|
|
else
|
|
return 0;
|
|
}
|
|
|
|
if (!(snap_seg = find_cow(lv)))
|
|
return 1;
|
|
|
|
old_origin = snap_seg->origin;
|
|
|
|
/* Was this the last active snapshot with this origin? */
|
|
dm_list_iterate_items(lvl, active_head) {
|
|
active = lvl->lv;
|
|
if ((snap_seg = find_cow(active)) &&
|
|
snap_seg->origin == old_origin) {
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
#endif
|
|
|
|
int dev_manager_thin_pool_status(struct dev_manager *dm,
|
|
const struct logical_volume *lv,
|
|
struct dm_status_thin_pool **status)
|
|
{
|
|
const char *dlid;
|
|
struct dm_task *dmt;
|
|
struct dm_info info;
|
|
uint64_t start, length;
|
|
char *type = NULL;
|
|
char *params = NULL;
|
|
int r = 0;
|
|
|
|
/* Build dlid for the thin pool layer */
|
|
if (!(dlid = build_dm_uuid(dm->mem, lv->lvid.s, _thin_layer)))
|
|
return_0;
|
|
|
|
log_debug("Getting thin pool device status for %s.", lv->name);
|
|
|
|
if (!(dmt = _setup_task(NULL, dlid, 0, DM_DEVICE_STATUS, 0, 0)))
|
|
return_0;
|
|
|
|
if (!dm_task_no_open_count(dmt))
|
|
log_error("Failed to disable open_count.");
|
|
|
|
if (!dm_task_run(dmt))
|
|
goto_out;
|
|
|
|
if (!dm_task_get_info(dmt, &info) || !info.exists)
|
|
goto_out;
|
|
|
|
dm_get_next_target(dmt, NULL, &start, &length, &type, ¶ms);
|
|
|
|
if (!dm_get_status_thin_pool(dm->mem, params, status))
|
|
goto_out;
|
|
|
|
r = 1;
|
|
out:
|
|
dm_task_destroy(dmt);
|
|
|
|
return r;
|
|
}
|
|
|
|
int dev_manager_thin_pool_percent(struct dev_manager *dm,
|
|
const struct logical_volume *lv,
|
|
int metadata, percent_t *percent)
|
|
{
|
|
char *name;
|
|
const char *dlid;
|
|
|
|
/* Build a name for the top layer */
|
|
if (!(name = dm_build_dm_name(dm->mem, lv->vg->name, lv->name,
|
|
_thin_layer)))
|
|
return_0;
|
|
|
|
if (!(dlid = build_dm_uuid(dm->mem, lv->lvid.s, _thin_layer)))
|
|
return_0;
|
|
|
|
log_debug("Getting device status percentage for %s", name);
|
|
if (!(_percent(dm, name, dlid, "thin-pool", 0,
|
|
(metadata) ? lv : NULL, percent, NULL, 1)))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
int dev_manager_thin_percent(struct dev_manager *dm,
|
|
const struct logical_volume *lv,
|
|
int mapped, percent_t *percent)
|
|
{
|
|
char *name;
|
|
const char *dlid;
|
|
const char *layer = lv_is_origin(lv) ? "real" : NULL;
|
|
|
|
/* Build a name for the top layer */
|
|
if (!(name = dm_build_dm_name(dm->mem, lv->vg->name, lv->name, layer)))
|
|
return_0;
|
|
|
|
if (!(dlid = build_dm_uuid(dm->mem, lv->lvid.s, layer)))
|
|
return_0;
|
|
|
|
log_debug("Getting device status percentage for %s", name);
|
|
if (!(_percent(dm, name, dlid, "thin", 0,
|
|
(mapped) ? NULL : lv, percent, NULL, 1)))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*************************/
|
|
/* NEW CODE STARTS HERE */
|
|
/*************************/
|
|
|
|
static int _dev_manager_lv_mknodes(const struct logical_volume *lv)
|
|
{
|
|
char *name;
|
|
|
|
if (!(name = dm_build_dm_name(lv->vg->cmd->mem, lv->vg->name,
|
|
lv->name, NULL)))
|
|
return_0;
|
|
|
|
return fs_add_lv(lv, name);
|
|
}
|
|
|
|
static int _dev_manager_lv_rmnodes(const struct logical_volume *lv)
|
|
{
|
|
return fs_del_lv(lv);
|
|
}
|
|
|
|
int dev_manager_mknodes(const struct logical_volume *lv)
|
|
{
|
|
struct dm_info dminfo;
|
|
char *name;
|
|
int r = 0;
|
|
|
|
if (!(name = dm_build_dm_name(lv->vg->cmd->mem, lv->vg->name, lv->name, NULL)))
|
|
return_0;
|
|
|
|
if ((r = _info_run(name, NULL, &dminfo, NULL, 1, 0, 0, 0, 0))) {
|
|
if (dminfo.exists) {
|
|
if (lv_is_visible(lv))
|
|
r = _dev_manager_lv_mknodes(lv);
|
|
} else
|
|
r = _dev_manager_lv_rmnodes(lv);
|
|
}
|
|
|
|
dm_pool_free(lv->vg->cmd->mem, name);
|
|
return r;
|
|
}
|
|
|
|
static uint16_t _get_udev_flags(struct dev_manager *dm, struct logical_volume *lv,
|
|
const char *layer)
|
|
{
|
|
uint16_t udev_flags = 0;
|
|
|
|
/*
|
|
* Instruct also libdevmapper to disable udev
|
|
* fallback in accordance to LVM2 settings.
|
|
*/
|
|
if (!dm->cmd->current_settings.udev_fallback)
|
|
udev_flags |= DM_UDEV_DISABLE_LIBRARY_FALLBACK;
|
|
|
|
/*
|
|
* Is this top-level and visible device?
|
|
* If not, create just the /dev/mapper content.
|
|
*/
|
|
/* FIXME: add target's method for this */
|
|
if (layer || !lv_is_visible(lv) || lv_is_thin_pool(lv))
|
|
udev_flags |= DM_UDEV_DISABLE_SUBSYSTEM_RULES_FLAG |
|
|
DM_UDEV_DISABLE_DISK_RULES_FLAG |
|
|
DM_UDEV_DISABLE_OTHER_RULES_FLAG;
|
|
/*
|
|
* There's no need for other udev rules to touch special LVs with
|
|
* reserved names. We don't need to populate /dev/disk here either.
|
|
* Even if they happen to be visible and top-level.
|
|
*/
|
|
else if (is_reserved_lvname(lv->name))
|
|
udev_flags |= DM_UDEV_DISABLE_DISK_RULES_FLAG |
|
|
DM_UDEV_DISABLE_OTHER_RULES_FLAG;
|
|
|
|
/*
|
|
* Snapshots and origins could have the same rule applied that will
|
|
* give symlinks exactly the same name (e.g. a name based on
|
|
* filesystem UUID). We give preference to origins to make such
|
|
* naming deterministic (e.g. symlinks in /dev/disk/by-uuid).
|
|
*/
|
|
if (lv_is_cow(lv))
|
|
udev_flags |= DM_UDEV_LOW_PRIORITY_FLAG;
|
|
|
|
/*
|
|
* Finally, add flags to disable /dev/mapper and /dev/<vgname> content
|
|
* to be created by udev if it is requested by user's configuration.
|
|
* This is basically an explicit fallback to old node/symlink creation
|
|
* without udev.
|
|
*/
|
|
if (!dm->cmd->current_settings.udev_rules)
|
|
udev_flags |= DM_UDEV_DISABLE_DM_RULES_FLAG |
|
|
DM_UDEV_DISABLE_SUBSYSTEM_RULES_FLAG;
|
|
|
|
return udev_flags;
|
|
}
|
|
|
|
static int _add_dev_to_dtree(struct dev_manager *dm, struct dm_tree *dtree,
|
|
struct logical_volume *lv, const char *layer)
|
|
{
|
|
char *dlid, *name;
|
|
struct dm_info info, info2;
|
|
|
|
if (!(name = dm_build_dm_name(dm->mem, lv->vg->name, lv->name, layer)))
|
|
return_0;
|
|
|
|
if (!(dlid = build_dm_uuid(dm->mem, lv->lvid.s, layer)))
|
|
return_0;
|
|
|
|
log_debug("Getting device info for %s [%s]", name, dlid);
|
|
if (!_info(dlid, 1, 0, &info, NULL)) {
|
|
log_error("Failed to get info for %s [%s].", name, dlid);
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* For top level volumes verify that existing device match
|
|
* requested major/minor and that major/minor pair is available for use
|
|
*/
|
|
if (!layer && lv->major != -1 && lv->minor != -1) {
|
|
/*
|
|
* FIXME compare info.major with lv->major if multiple major support
|
|
*/
|
|
if (info.exists && (info.minor != lv->minor)) {
|
|
log_error("Volume %s (%" PRIu32 ":%" PRIu32")"
|
|
" differs from already active device "
|
|
"(%" PRIu32 ":%" PRIu32")",
|
|
lv->name, lv->major, lv->minor, info.major, info.minor);
|
|
return 0;
|
|
}
|
|
if (!info.exists && _info_by_dev(lv->major, lv->minor, &info2) &&
|
|
info2.exists) {
|
|
log_error("The requested major:minor pair "
|
|
"(%" PRIu32 ":%" PRIu32") is already used",
|
|
lv->major, lv->minor);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
if (info.exists && !dm_tree_add_dev_with_udev_flags(dtree, info.major, info.minor,
|
|
_get_udev_flags(dm, lv, layer))) {
|
|
log_error("Failed to add device (%" PRIu32 ":%" PRIu32") to dtree",
|
|
info.major, info.minor);
|
|
return 0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* Add replicator devices
|
|
*
|
|
* Using _add_dev_to_dtree() directly instead of _add_lv_to_dtree()
|
|
* to avoid extra checks with extensions.
|
|
*/
|
|
static int _add_partial_replicator_to_dtree(struct dev_manager *dm,
|
|
struct dm_tree *dtree,
|
|
struct logical_volume *lv)
|
|
{
|
|
struct logical_volume *rlv = first_seg(lv)->replicator;
|
|
struct replicator_device *rdev;
|
|
struct replicator_site *rsite;
|
|
struct dm_tree_node *rep_node, *rdev_node;
|
|
const char *uuid;
|
|
|
|
if (!lv_is_active_replicator_dev(lv)) {
|
|
if (!_add_dev_to_dtree(dm, dtree, lv->rdevice->lv,
|
|
NULL))
|
|
return_0;
|
|
return 1;
|
|
}
|
|
|
|
/* Add _rlog and replicator device */
|
|
if (!_add_dev_to_dtree(dm, dtree, first_seg(rlv)->rlog_lv, NULL))
|
|
return_0;
|
|
|
|
if (!_add_dev_to_dtree(dm, dtree, rlv, NULL))
|
|
return_0;
|
|
|
|
if (!(uuid = build_dm_uuid(dm->mem, rlv->lvid.s, NULL)))
|
|
return_0;
|
|
|
|
rep_node = dm_tree_find_node_by_uuid(dtree, uuid);
|
|
|
|
/* Add all related devices for replicator */
|
|
dm_list_iterate_items(rsite, &rlv->rsites)
|
|
dm_list_iterate_items(rdev, &rsite->rdevices) {
|
|
if (rsite->state == REPLICATOR_STATE_ACTIVE) {
|
|
/* Add _rimage LV */
|
|
if (!_add_dev_to_dtree(dm, dtree, rdev->lv, NULL))
|
|
return_0;
|
|
|
|
/* Add replicator-dev LV, except of the already added one */
|
|
if ((lv != rdev->replicator_dev->lv) &&
|
|
!_add_dev_to_dtree(dm, dtree,
|
|
rdev->replicator_dev->lv, NULL))
|
|
return_0;
|
|
|
|
/* If replicator exists - try connect existing heads */
|
|
if (rep_node) {
|
|
uuid = build_dm_uuid(dm->mem,
|
|
rdev->replicator_dev->lv->lvid.s,
|
|
NULL);
|
|
if (!uuid)
|
|
return_0;
|
|
|
|
rdev_node = dm_tree_find_node_by_uuid(dtree, uuid);
|
|
if (rdev_node)
|
|
dm_tree_node_set_presuspend_node(rdev_node,
|
|
rep_node);
|
|
}
|
|
}
|
|
|
|
if (!rdev->rsite->vg_name)
|
|
continue;
|
|
|
|
if (!_add_dev_to_dtree(dm, dtree, rdev->lv, NULL))
|
|
return_0;
|
|
|
|
if (rdev->slog &&
|
|
!_add_dev_to_dtree(dm, dtree, rdev->slog, NULL))
|
|
return_0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* Add LV and any known dependencies
|
|
*/
|
|
static int _add_lv_to_dtree(struct dev_manager *dm, struct dm_tree *dtree,
|
|
struct logical_volume *lv, int origin_only)
|
|
{
|
|
uint32_t s;
|
|
struct seg_list *sl;
|
|
struct lv_segment *seg = first_seg(lv);
|
|
#if 0
|
|
/* FIXME For dm_tree_node_skip_children optimisation below */
|
|
struct dm_tree_node *thin_node;
|
|
const char *uuid;
|
|
#endif
|
|
|
|
if ((!origin_only || lv_is_thin_volume(lv)) &&
|
|
!_add_dev_to_dtree(dm, dtree, lv, NULL))
|
|
return_0;
|
|
|
|
/* FIXME Can we avoid doing this every time? */
|
|
if (!_add_dev_to_dtree(dm, dtree, lv, "real"))
|
|
return_0;
|
|
|
|
if (!origin_only && !_add_dev_to_dtree(dm, dtree, lv, "cow"))
|
|
return_0;
|
|
|
|
if ((lv->status & MIRRORED) && seg->log_lv &&
|
|
!_add_dev_to_dtree(dm, dtree, seg->log_lv, NULL))
|
|
return_0;
|
|
|
|
if (lv->status & RAID)
|
|
for (s = 0; s < seg->area_count; s++)
|
|
if (!_add_dev_to_dtree(dm, dtree,
|
|
seg_metalv(seg, s), NULL))
|
|
return_0;
|
|
|
|
/* Add any LVs referencing a PVMOVE LV unless told not to. */
|
|
if (dm->track_pvmove_deps && lv->status & PVMOVE)
|
|
dm_list_iterate_items(sl, &lv->segs_using_this_lv)
|
|
if (!_add_lv_to_dtree(dm, dtree, sl->seg->lv, origin_only))
|
|
return_0;
|
|
|
|
/* Adding LV head of replicator adds all other related devs */
|
|
if (lv_is_replicator_dev(lv) &&
|
|
!_add_partial_replicator_to_dtree(dm, dtree, lv))
|
|
return_0;
|
|
|
|
if (lv_is_thin_volume(lv)) {
|
|
#if 0
|
|
/* FIXME Implement dm_tree_node_skip_children optimisation */
|
|
if (origin_only) {
|
|
if (!(uuid = build_dm_uuid(dm->mem, lv->lvid.s, NULL)))
|
|
return_0;
|
|
if ((thin_node = dm_tree_find_node_by_uuid(dtree, uuid)))
|
|
dm_tree_node_skip_children(thin_node, 1);
|
|
}
|
|
#endif
|
|
/* Add thin pool LV layer */
|
|
lv = seg->pool_lv;
|
|
seg = first_seg(lv);
|
|
}
|
|
|
|
if (lv_is_thin_pool(lv)) {
|
|
if (!_add_lv_to_dtree(dm, dtree, seg->metadata_lv, 0))
|
|
return_0;
|
|
/* FIXME code from _create_partial_dtree() should be moved here */
|
|
if (!_add_lv_to_dtree(dm, dtree, seg_lv(seg, 0), 0))
|
|
return_0;
|
|
if (!_add_dev_to_dtree(dm, dtree, lv, _thin_layer))
|
|
return_0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
static struct dm_tree *_create_partial_dtree(struct dev_manager *dm, struct logical_volume *lv, int origin_only)
|
|
{
|
|
struct dm_tree *dtree;
|
|
struct dm_list *snh;
|
|
struct lv_segment *seg;
|
|
uint32_t s;
|
|
|
|
if (!(dtree = dm_tree_create())) {
|
|
log_debug("Partial dtree creation failed for %s.", lv->name);
|
|
return NULL;
|
|
}
|
|
|
|
if (!_add_lv_to_dtree(dm, dtree, lv, (lv_is_origin(lv) || lv_is_thin_volume(lv)) ? origin_only : 0))
|
|
goto_bad;
|
|
|
|
/* Add any snapshots of this LV */
|
|
if (!origin_only && lv_is_origin(lv))
|
|
dm_list_iterate(snh, &lv->snapshot_segs)
|
|
if (!_add_lv_to_dtree(dm, dtree, dm_list_struct_base(snh, struct lv_segment, origin_list)->cow, 0))
|
|
goto_bad;
|
|
|
|
/* Add any LVs used by segments in this LV */
|
|
dm_list_iterate_items(seg, &lv->segments)
|
|
for (s = 0; s < seg->area_count; s++)
|
|
if (seg_type(seg, s) == AREA_LV && seg_lv(seg, s)) {
|
|
if (!_add_lv_to_dtree(dm, dtree, seg_lv(seg, s), 0))
|
|
goto_bad;
|
|
}
|
|
|
|
return dtree;
|
|
|
|
bad:
|
|
dm_tree_free(dtree);
|
|
return NULL;
|
|
}
|
|
|
|
static char *_add_error_device(struct dev_manager *dm, struct dm_tree *dtree,
|
|
struct lv_segment *seg, int s)
|
|
{
|
|
char *id, *name;
|
|
char errid[32];
|
|
struct dm_tree_node *node;
|
|
struct lv_segment *seg_i;
|
|
int segno = -1, i = 0;
|
|
uint64_t size = seg->len * seg->lv->vg->extent_size;
|
|
|
|
dm_list_iterate_items(seg_i, &seg->lv->segments) {
|
|
if (seg == seg_i)
|
|
segno = i;
|
|
++i;
|
|
}
|
|
|
|
if (segno < 0) {
|
|
log_error("_add_error_device called with bad segment");
|
|
return NULL;
|
|
}
|
|
|
|
sprintf(errid, "missing_%d_%d", segno, s);
|
|
|
|
if (!(id = build_dm_uuid(dm->mem, seg->lv->lvid.s, errid)))
|
|
return_NULL;
|
|
|
|
if (!(name = dm_build_dm_name(dm->mem, seg->lv->vg->name,
|
|
seg->lv->name, errid)))
|
|
return_NULL;
|
|
if (!(node = dm_tree_add_new_dev(dtree, name, id, 0, 0, 0, 0, 0)))
|
|
return_NULL;
|
|
if (!dm_tree_node_add_error_target(node, size))
|
|
return_NULL;
|
|
|
|
return id;
|
|
}
|
|
|
|
static int _add_error_area(struct dev_manager *dm, struct dm_tree_node *node,
|
|
struct lv_segment *seg, int s)
|
|
{
|
|
char *dlid;
|
|
uint64_t extent_size = seg->lv->vg->extent_size;
|
|
|
|
if (!strcmp(dm->cmd->stripe_filler, "error")) {
|
|
/*
|
|
* FIXME, the tree pointer is first field of dm_tree_node, but
|
|
* we don't have the struct definition available.
|
|
*/
|
|
struct dm_tree **tree = (struct dm_tree **) node;
|
|
if (!(dlid = _add_error_device(dm, *tree, seg, s)))
|
|
return_0;
|
|
if (!dm_tree_node_add_target_area(node, NULL, dlid, extent_size * seg_le(seg, s)))
|
|
return_0;
|
|
} else
|
|
if (!dm_tree_node_add_target_area(node, dm->cmd->stripe_filler, NULL, UINT64_C(0)))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
int add_areas_line(struct dev_manager *dm, struct lv_segment *seg,
|
|
struct dm_tree_node *node, uint32_t start_area,
|
|
uint32_t areas)
|
|
{
|
|
uint64_t extent_size = seg->lv->vg->extent_size;
|
|
uint32_t s;
|
|
char *dlid;
|
|
struct stat info;
|
|
const char *name;
|
|
unsigned num_error_areas = 0;
|
|
unsigned num_existing_areas = 0;
|
|
|
|
/* FIXME Avoid repeating identical stat in dm_tree_node_add_target_area */
|
|
for (s = start_area; s < areas; s++) {
|
|
if ((seg_type(seg, s) == AREA_PV &&
|
|
(!seg_pvseg(seg, s) || !seg_pv(seg, s) || !seg_dev(seg, s) ||
|
|
!(name = dev_name(seg_dev(seg, s))) || !*name ||
|
|
stat(name, &info) < 0 || !S_ISBLK(info.st_mode))) ||
|
|
(seg_type(seg, s) == AREA_LV && !seg_lv(seg, s))) {
|
|
if (!seg->lv->vg->cmd->partial_activation) {
|
|
log_error("Aborting. LV %s is now incomplete "
|
|
"and --partial was not specified.", seg->lv->name);
|
|
return 0;
|
|
}
|
|
if (!_add_error_area(dm, node, seg, s))
|
|
return_0;
|
|
num_error_areas++;
|
|
} else if (seg_type(seg, s) == AREA_PV) {
|
|
if (!dm_tree_node_add_target_area(node, dev_name(seg_dev(seg, s)), NULL,
|
|
(seg_pv(seg, s)->pe_start + (extent_size * seg_pe(seg, s)))))
|
|
return_0;
|
|
num_existing_areas++;
|
|
} else if (seg_is_raid(seg)) {
|
|
/*
|
|
* RAID can handle unassigned areas. It simple puts
|
|
* '- -' in for the metadata/data device pair. This
|
|
* is a valid way to indicate to the RAID target that
|
|
* the device is missing.
|
|
*
|
|
* If an image is marked as VISIBLE_LV and !LVM_WRITE,
|
|
* it means the device has temporarily been extracted
|
|
* from the array. It may come back at a future date,
|
|
* so the bitmap must track differences. Again, '- -'
|
|
* is used in the CTR table.
|
|
*/
|
|
if ((seg_type(seg, s) == AREA_UNASSIGNED) ||
|
|
((seg_lv(seg, s)->status & VISIBLE_LV) &&
|
|
!(seg_lv(seg, s)->status & LVM_WRITE))) {
|
|
/* One each for metadata area and data area */
|
|
if (!dm_tree_node_add_null_area(node, 0) ||
|
|
!dm_tree_node_add_null_area(node, 0))
|
|
return_0;
|
|
continue;
|
|
}
|
|
if (!(dlid = build_dm_uuid(dm->mem, seg_metalv(seg, s)->lvid.s, NULL)))
|
|
return_0;
|
|
if (!dm_tree_node_add_target_area(node, NULL, dlid, extent_size * seg_metale(seg, s)))
|
|
return_0;
|
|
|
|
if (!(dlid = build_dm_uuid(dm->mem, seg_lv(seg, s)->lvid.s, NULL)))
|
|
return_0;
|
|
if (!dm_tree_node_add_target_area(node, NULL, dlid, extent_size * seg_le(seg, s)))
|
|
return_0;
|
|
} else if (seg_type(seg, s) == AREA_LV) {
|
|
|
|
if (!(dlid = build_dm_uuid(dm->mem, seg_lv(seg, s)->lvid.s, NULL)))
|
|
return_0;
|
|
if (!dm_tree_node_add_target_area(node, NULL, dlid, extent_size * seg_le(seg, s)))
|
|
return_0;
|
|
} else {
|
|
log_error(INTERNAL_ERROR "Unassigned area found in LV %s.",
|
|
seg->lv->name);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
if (num_error_areas) {
|
|
/* Thins currently do not support partial activation */
|
|
if (lv_is_thin_type(seg->lv)) {
|
|
log_error("Cannot activate %s%s: pool incomplete.",
|
|
seg->lv->vg->name, seg->lv->name);
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Mirrors activate LVs replaced with error targets and
|
|
* RAID can handle non-accessible sub-LVs.
|
|
*
|
|
* TODO: Can we eventually skip to activate such LVs ?
|
|
*/
|
|
if (!num_existing_areas &&
|
|
!strstr(seg->lv->name, "_rmeta_") &&
|
|
!strstr(seg->lv->name, "_rimage_") &&
|
|
!strstr(seg->lv->name, "_mimage_") &&
|
|
!((name = strstr(seg->lv->name, "_mlog")) && !name[5])) {
|
|
log_error("Cannot activate %s/%s: all segments missing.",
|
|
seg->lv->vg->name, seg->lv->name);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int _add_origin_target_to_dtree(struct dev_manager *dm,
|
|
struct dm_tree_node *dnode,
|
|
struct logical_volume *lv)
|
|
{
|
|
const char *real_dlid;
|
|
|
|
if (!(real_dlid = build_dm_uuid(dm->mem, lv->lvid.s, "real")))
|
|
return_0;
|
|
|
|
if (!dm_tree_node_add_snapshot_origin_target(dnode, lv->size, real_dlid))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int _add_snapshot_merge_target_to_dtree(struct dev_manager *dm,
|
|
struct dm_tree_node *dnode,
|
|
struct logical_volume *lv)
|
|
{
|
|
const char *origin_dlid, *cow_dlid, *merge_dlid;
|
|
struct lv_segment *merging_cow_seg = find_merging_cow(lv);
|
|
|
|
if (!(origin_dlid = build_dm_uuid(dm->mem, lv->lvid.s, "real")))
|
|
return_0;
|
|
|
|
if (!(cow_dlid = build_dm_uuid(dm->mem, merging_cow_seg->cow->lvid.s, "cow")))
|
|
return_0;
|
|
|
|
if (!(merge_dlid = build_dm_uuid(dm->mem, merging_cow_seg->cow->lvid.s, NULL)))
|
|
return_0;
|
|
|
|
if (!dm_tree_node_add_snapshot_merge_target(dnode, lv->size, origin_dlid,
|
|
cow_dlid, merge_dlid,
|
|
merging_cow_seg->chunk_size))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int _add_snapshot_target_to_dtree(struct dev_manager *dm,
|
|
struct dm_tree_node *dnode,
|
|
struct logical_volume *lv,
|
|
struct lv_activate_opts *laopts)
|
|
{
|
|
const char *origin_dlid;
|
|
const char *cow_dlid;
|
|
struct lv_segment *snap_seg;
|
|
uint64_t size;
|
|
|
|
if (!(snap_seg = find_cow(lv))) {
|
|
log_error("Couldn't find snapshot for '%s'.", lv->name);
|
|
return 0;
|
|
}
|
|
|
|
if (!(origin_dlid = build_dm_uuid(dm->mem, snap_seg->origin->lvid.s, "real")))
|
|
return_0;
|
|
|
|
if (!(cow_dlid = build_dm_uuid(dm->mem, snap_seg->cow->lvid.s, "cow")))
|
|
return_0;
|
|
|
|
size = (uint64_t) snap_seg->len * snap_seg->origin->vg->extent_size;
|
|
|
|
if (!laopts->no_merging && lv_is_merging_cow(lv)) {
|
|
/* cow is to be merged so load the error target */
|
|
if (!dm_tree_node_add_error_target(dnode, size))
|
|
return_0;
|
|
}
|
|
else if (!dm_tree_node_add_snapshot_target(dnode, size, origin_dlid,
|
|
cow_dlid, 1, snap_seg->chunk_size))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int _add_target_to_dtree(struct dev_manager *dm,
|
|
struct dm_tree_node *dnode,
|
|
struct lv_segment *seg,
|
|
struct lv_activate_opts *laopts)
|
|
{
|
|
uint64_t extent_size = seg->lv->vg->extent_size;
|
|
|
|
if (!seg->segtype->ops->add_target_line) {
|
|
log_error(INTERNAL_ERROR "_emit_target cannot handle "
|
|
"segment type %s", seg->segtype->name);
|
|
return 0;
|
|
}
|
|
|
|
return seg->segtype->ops->add_target_line(dm, dm->mem, dm->cmd,
|
|
&dm->target_state, seg,
|
|
laopts, dnode,
|
|
extent_size * seg->len,
|
|
&dm-> pvmove_mirror_count);
|
|
}
|
|
|
|
static int _add_new_lv_to_dtree(struct dev_manager *dm, struct dm_tree *dtree,
|
|
struct logical_volume *lv,
|
|
struct lv_activate_opts *laopts,
|
|
const char *layer);
|
|
|
|
/* Add all replicators' LVs */
|
|
static int _add_replicator_dev_target_to_dtree(struct dev_manager *dm,
|
|
struct dm_tree *dtree,
|
|
struct lv_segment *seg,
|
|
struct lv_activate_opts *laopts)
|
|
{
|
|
struct replicator_device *rdev;
|
|
struct replicator_site *rsite;
|
|
|
|
/* For inactive replicator add linear mapping */
|
|
if (!lv_is_active_replicator_dev(seg->lv)) {
|
|
if (!_add_new_lv_to_dtree(dm, dtree, seg->lv->rdevice->lv, laopts, NULL))
|
|
return_0;
|
|
return 1;
|
|
}
|
|
|
|
/* Add rlog and replicator nodes */
|
|
if (!seg->replicator ||
|
|
!first_seg(seg->replicator)->rlog_lv ||
|
|
!_add_new_lv_to_dtree(dm, dtree,
|
|
first_seg(seg->replicator)->rlog_lv,
|
|
laopts, NULL) ||
|
|
!_add_new_lv_to_dtree(dm, dtree, seg->replicator, laopts, NULL))
|
|
return_0;
|
|
|
|
/* Activation of one replicator_dev node activates all other nodes */
|
|
dm_list_iterate_items(rsite, &seg->replicator->rsites) {
|
|
dm_list_iterate_items(rdev, &rsite->rdevices) {
|
|
if (rdev->lv &&
|
|
!_add_new_lv_to_dtree(dm, dtree, rdev->lv,
|
|
laopts, NULL))
|
|
return_0;
|
|
|
|
if (rdev->slog &&
|
|
!_add_new_lv_to_dtree(dm, dtree, rdev->slog,
|
|
laopts, NULL))
|
|
return_0;
|
|
}
|
|
}
|
|
/* Add remaining replicator-dev nodes in the second loop
|
|
* to avoid multiple retries for inserting all elements */
|
|
dm_list_iterate_items(rsite, &seg->replicator->rsites) {
|
|
if (rsite->state != REPLICATOR_STATE_ACTIVE)
|
|
continue;
|
|
dm_list_iterate_items(rdev, &rsite->rdevices) {
|
|
if (rdev->replicator_dev->lv == seg->lv)
|
|
continue;
|
|
if (!rdev->replicator_dev->lv ||
|
|
!_add_new_lv_to_dtree(dm, dtree,
|
|
rdev->replicator_dev->lv,
|
|
laopts, NULL))
|
|
return_0;
|
|
}
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int _add_segment_to_dtree(struct dev_manager *dm,
|
|
struct dm_tree *dtree,
|
|
struct dm_tree_node *dnode,
|
|
struct lv_segment *seg,
|
|
struct lv_activate_opts *laopts,
|
|
const char *layer)
|
|
{
|
|
uint32_t s;
|
|
struct dm_list *snh;
|
|
struct lv_segment *seg_present;
|
|
const char *target_name;
|
|
struct lv_activate_opts lva;
|
|
|
|
/* Ensure required device-mapper targets are loaded */
|
|
seg_present = find_cow(seg->lv) ? : seg;
|
|
target_name = (seg_present->segtype->ops->target_name ?
|
|
seg_present->segtype->ops->target_name(seg_present, laopts) :
|
|
seg_present->segtype->name);
|
|
|
|
log_debug("Checking kernel supports %s segment type for %s%s%s",
|
|
target_name, seg->lv->name,
|
|
layer ? "-" : "", layer ? : "");
|
|
|
|
if (seg_present->segtype->ops->target_present &&
|
|
!seg_present->segtype->ops->target_present(seg_present->lv->vg->cmd,
|
|
seg_present, NULL)) {
|
|
log_error("Can't process LV %s: %s target support missing "
|
|
"from kernel?", seg->lv->name, target_name);
|
|
return 0;
|
|
}
|
|
|
|
/* Add mirror log */
|
|
if (seg->log_lv &&
|
|
!_add_new_lv_to_dtree(dm, dtree, seg->log_lv, laopts, NULL))
|
|
return_0;
|
|
|
|
if (seg_is_replicator_dev(seg)) {
|
|
if (!_add_replicator_dev_target_to_dtree(dm, dtree, seg, laopts))
|
|
return_0;
|
|
/* If this is a snapshot origin, add real LV */
|
|
/* If this is a snapshot origin + merging snapshot, add cow + real LV */
|
|
} else if (lv_is_origin(seg->lv) && !layer) {
|
|
if (!laopts->no_merging && lv_is_merging_origin(seg->lv)) {
|
|
if (!_add_new_lv_to_dtree(dm, dtree,
|
|
find_merging_cow(seg->lv)->cow, laopts, "cow"))
|
|
return_0;
|
|
/*
|
|
* Must also add "real" LV for use when
|
|
* snapshot-merge target is added
|
|
*/
|
|
}
|
|
if (!_add_new_lv_to_dtree(dm, dtree, seg->lv, laopts, "real"))
|
|
return_0;
|
|
} else if (lv_is_cow(seg->lv) && !layer) {
|
|
if (!_add_new_lv_to_dtree(dm, dtree, seg->lv, laopts, "cow"))
|
|
return_0;
|
|
} else if ((layer != _thin_layer) && seg_is_thin(seg)) {
|
|
lva = *laopts;
|
|
lva.real_pool = 1;
|
|
if (!_add_new_lv_to_dtree(dm, dtree, seg_is_thin_pool(seg) ?
|
|
seg->lv : seg->pool_lv, &lva, _thin_layer))
|
|
return_0;
|
|
} else {
|
|
if (seg_is_thin_pool(seg) &&
|
|
!_add_new_lv_to_dtree(dm, dtree, seg->metadata_lv, laopts, NULL))
|
|
return_0;
|
|
|
|
/* Add any LVs used by this segment */
|
|
for (s = 0; s < seg->area_count; s++) {
|
|
if ((seg_type(seg, s) == AREA_LV) &&
|
|
(!_add_new_lv_to_dtree(dm, dtree, seg_lv(seg, s),
|
|
laopts, NULL)))
|
|
return_0;
|
|
if (seg_is_raid(seg) &&
|
|
!_add_new_lv_to_dtree(dm, dtree, seg_metalv(seg, s),
|
|
laopts, NULL))
|
|
return_0;
|
|
}
|
|
}
|
|
|
|
/* Now we've added its dependencies, we can add the target itself */
|
|
if (lv_is_origin(seg->lv) && !layer) {
|
|
if (laopts->no_merging || !lv_is_merging_origin(seg->lv)) {
|
|
if (!_add_origin_target_to_dtree(dm, dnode, seg->lv))
|
|
return_0;
|
|
} else {
|
|
if (!_add_snapshot_merge_target_to_dtree(dm, dnode, seg->lv))
|
|
return_0;
|
|
}
|
|
} else if (lv_is_cow(seg->lv) && !layer) {
|
|
if (!_add_snapshot_target_to_dtree(dm, dnode, seg->lv, laopts))
|
|
return_0;
|
|
} else if (!_add_target_to_dtree(dm, dnode, seg, laopts))
|
|
return_0;
|
|
|
|
if (lv_is_origin(seg->lv) && !layer)
|
|
/* Add any snapshots of this LV */
|
|
dm_list_iterate(snh, &seg->lv->snapshot_segs)
|
|
if (!_add_new_lv_to_dtree(dm, dtree, dm_list_struct_base(snh, struct lv_segment, origin_list)->cow,
|
|
laopts, NULL))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int _set_udev_flags_for_children(struct dev_manager *dm,
|
|
struct volume_group *vg,
|
|
struct dm_tree_node *dnode)
|
|
{
|
|
char *p;
|
|
const char *uuid;
|
|
void *handle = NULL;
|
|
struct dm_tree_node *child;
|
|
const struct dm_info *info;
|
|
struct lv_list *lvl;
|
|
|
|
while ((child = dm_tree_next_child(&handle, dnode, 0))) {
|
|
/* Ignore root node */
|
|
if (!(info = dm_tree_node_get_info(child)) || !info->exists)
|
|
continue;
|
|
|
|
if (!(uuid = dm_tree_node_get_uuid(child))) {
|
|
log_error(INTERNAL_ERROR
|
|
"Failed to get uuid for %" PRIu32 ":%" PRIu32,
|
|
info->major, info->minor);
|
|
continue;
|
|
}
|
|
|
|
/* Ignore non-LVM devices */
|
|
if (!(p = strstr(uuid, UUID_PREFIX)))
|
|
continue;
|
|
p += strlen(UUID_PREFIX);
|
|
|
|
/* Ignore LVs that belong to different VGs (due to stacking) */
|
|
if (strncmp(p, (char *)vg->id.uuid, ID_LEN))
|
|
continue;
|
|
|
|
/* Ignore LVM devices with 'layer' suffixes */
|
|
if (strrchr(p, '-'))
|
|
continue;
|
|
|
|
if (!(lvl = find_lv_in_vg_by_lvid(vg, (const union lvid *)p))) {
|
|
log_error(INTERNAL_ERROR
|
|
"%s (%" PRIu32 ":%" PRIu32 ") not found in VG",
|
|
dm_tree_node_get_name(child),
|
|
info->major, info->minor);
|
|
return 0;
|
|
}
|
|
|
|
dm_tree_node_set_udev_flags(child,
|
|
_get_udev_flags(dm, lvl->lv, NULL));
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int _add_new_lv_to_dtree(struct dev_manager *dm, struct dm_tree *dtree,
|
|
struct logical_volume *lv, struct lv_activate_opts *laopts,
|
|
const char *layer)
|
|
{
|
|
struct lv_segment *seg;
|
|
struct lv_layer *lvlayer;
|
|
struct seg_list *sl;
|
|
struct dm_tree_node *dnode;
|
|
const struct dm_info *dinfo;
|
|
char *name, *dlid;
|
|
uint32_t max_stripe_size = UINT32_C(0);
|
|
uint32_t read_ahead = lv->read_ahead;
|
|
uint32_t read_ahead_flags = UINT32_C(0);
|
|
|
|
/* FIXME Seek a simpler way to lay out the snapshot-merge tree. */
|
|
|
|
if (lv_is_origin(lv) && lv_is_merging_origin(lv) && !layer) {
|
|
/*
|
|
* Clear merge attributes if merge isn't currently possible:
|
|
* either origin or merging snapshot are open
|
|
* - but use "snapshot-merge" if it is already in use
|
|
* - open_count is always retrieved (as of dm-ioctl 4.7.0)
|
|
* so just use the tree's existing nodes' info
|
|
*/
|
|
if (((dinfo = _cached_info(dm->mem, lv,
|
|
dtree)) && dinfo->open_count) ||
|
|
((dinfo = _cached_info(dm->mem, find_merging_cow(lv)->cow,
|
|
dtree)) && dinfo->open_count)) {
|
|
/* FIXME Is there anything simpler to check for instead? */
|
|
if (!lv_has_target_type(dm->mem, lv, NULL, "snapshot-merge"))
|
|
laopts->no_merging = 1;
|
|
}
|
|
}
|
|
|
|
if (!(name = dm_build_dm_name(dm->mem, lv->vg->name, lv->name, layer)))
|
|
return_0;
|
|
|
|
if (!(dlid = build_dm_uuid(dm->mem, lv->lvid.s, layer)))
|
|
return_0;
|
|
|
|
/* We've already processed this node if it already has a context ptr */
|
|
if ((dnode = dm_tree_find_node_by_uuid(dtree, dlid)) &&
|
|
dm_tree_node_get_context(dnode))
|
|
return 1;
|
|
|
|
if (!(lvlayer = dm_pool_alloc(dm->mem, sizeof(*lvlayer)))) {
|
|
log_error("_add_new_lv_to_dtree: pool alloc failed for %s %s.",
|
|
lv->name, layer);
|
|
return 0;
|
|
}
|
|
|
|
lvlayer->lv = lv;
|
|
|
|
/*
|
|
* Add LV to dtree.
|
|
* If we're working with precommitted metadata, clear any
|
|
* existing inactive table left behind.
|
|
* Major/minor settings only apply to the visible layer.
|
|
*/
|
|
/* FIXME Move the clear from here until later, so we can leave
|
|
* identical inactive tables untouched. (For pvmove.)
|
|
*/
|
|
if (!(dnode = dm_tree_add_new_dev_with_udev_flags(dtree, name, dlid,
|
|
layer ? UINT32_C(0) : (uint32_t) lv->major,
|
|
layer ? UINT32_C(0) : (uint32_t) lv->minor,
|
|
read_only_lv(lv, laopts),
|
|
((lv->vg->status & PRECOMMITTED) | laopts->revert) ? 1 : 0,
|
|
lvlayer,
|
|
_get_udev_flags(dm, lv, layer))))
|
|
return_0;
|
|
|
|
/* Store existing name so we can do rename later */
|
|
lvlayer->old_name = dm_tree_node_get_name(dnode);
|
|
|
|
/* Create table */
|
|
dm->pvmove_mirror_count = 0u;
|
|
dm_list_iterate_items(seg, &lv->segments) {
|
|
if (!_add_segment_to_dtree(dm, dtree, dnode, seg, laopts, layer))
|
|
return_0;
|
|
/* These aren't real segments in the LVM2 metadata */
|
|
if (lv_is_origin(lv) && !layer)
|
|
break;
|
|
if (!laopts->no_merging && lv_is_cow(lv) && !layer)
|
|
break;
|
|
if (max_stripe_size < seg->stripe_size * seg->area_count)
|
|
max_stripe_size = seg->stripe_size * seg->area_count;
|
|
}
|
|
|
|
if (read_ahead == DM_READ_AHEAD_AUTO) {
|
|
/* we need RA at least twice a whole stripe - see the comment in md/raid0.c */
|
|
read_ahead = max_stripe_size * 2;
|
|
if (!read_ahead)
|
|
lv_calculate_readahead(lv, &read_ahead);
|
|
read_ahead_flags = DM_READ_AHEAD_MINIMUM_FLAG;
|
|
}
|
|
|
|
dm_tree_node_set_read_ahead(dnode, read_ahead, read_ahead_flags);
|
|
|
|
/* Add any LVs referencing a PVMOVE LV unless told not to */
|
|
if (dm->track_pvmove_deps && (lv->status & PVMOVE))
|
|
dm_list_iterate_items(sl, &lv->segs_using_this_lv)
|
|
if (!_add_new_lv_to_dtree(dm, dtree, sl->seg->lv, laopts, NULL))
|
|
return_0;
|
|
|
|
if (!_set_udev_flags_for_children(dm, lv->vg, dnode))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
/* FIXME: symlinks should be created/destroyed at the same time
|
|
* as the kernel devices but we can't do that from within libdevmapper
|
|
* at present so we must walk the tree twice instead. */
|
|
|
|
/*
|
|
* Create LV symlinks for children of supplied root node.
|
|
*/
|
|
static int _create_lv_symlinks(struct dev_manager *dm, struct dm_tree_node *root)
|
|
{
|
|
void *handle = NULL;
|
|
struct dm_tree_node *child;
|
|
struct lv_layer *lvlayer;
|
|
char *old_vgname, *old_lvname, *old_layer;
|
|
char *new_vgname, *new_lvname, *new_layer;
|
|
const char *name;
|
|
int r = 1;
|
|
|
|
/* Nothing to do if udev fallback is disabled. */
|
|
if (!dm->cmd->current_settings.udev_fallback) {
|
|
fs_set_create();
|
|
return 1;
|
|
}
|
|
|
|
while ((child = dm_tree_next_child(&handle, root, 0))) {
|
|
if (!(lvlayer = dm_tree_node_get_context(child)))
|
|
continue;
|
|
|
|
/* Detect rename */
|
|
name = dm_tree_node_get_name(child);
|
|
|
|
if (name && lvlayer->old_name && *lvlayer->old_name && strcmp(name, lvlayer->old_name)) {
|
|
if (!dm_split_lvm_name(dm->mem, lvlayer->old_name, &old_vgname, &old_lvname, &old_layer)) {
|
|
log_error("_create_lv_symlinks: Couldn't split up old device name %s", lvlayer->old_name);
|
|
return 0;
|
|
}
|
|
if (!dm_split_lvm_name(dm->mem, name, &new_vgname, &new_lvname, &new_layer)) {
|
|
log_error("_create_lv_symlinks: Couldn't split up new device name %s", name);
|
|
return 0;
|
|
}
|
|
if (!fs_rename_lv(lvlayer->lv, name, old_vgname, old_lvname))
|
|
r = 0;
|
|
continue;
|
|
}
|
|
if (lv_is_visible(lvlayer->lv)) {
|
|
if (!_dev_manager_lv_mknodes(lvlayer->lv))
|
|
r = 0;
|
|
continue;
|
|
}
|
|
if (!_dev_manager_lv_rmnodes(lvlayer->lv))
|
|
r = 0;
|
|
}
|
|
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* Remove LV symlinks for children of supplied root node.
|
|
*/
|
|
static int _remove_lv_symlinks(struct dev_manager *dm, struct dm_tree_node *root)
|
|
{
|
|
void *handle = NULL;
|
|
struct dm_tree_node *child;
|
|
char *vgname, *lvname, *layer;
|
|
int r = 1;
|
|
|
|
/* Nothing to do if udev fallback is disabled. */
|
|
if (!dm->cmd->current_settings.udev_fallback)
|
|
return 1;
|
|
|
|
while ((child = dm_tree_next_child(&handle, root, 0))) {
|
|
if (!dm_split_lvm_name(dm->mem, dm_tree_node_get_name(child), &vgname, &lvname, &layer)) {
|
|
r = 0;
|
|
continue;
|
|
}
|
|
|
|
if (!*vgname)
|
|
continue;
|
|
|
|
/* only top level layer has symlinks */
|
|
if (*layer)
|
|
continue;
|
|
|
|
fs_del_lv_byname(dm->cmd->dev_dir, vgname, lvname,
|
|
dm->cmd->current_settings.udev_rules);
|
|
}
|
|
|
|
return r;
|
|
}
|
|
|
|
static int _clean_tree(struct dev_manager *dm, struct dm_tree_node *root, char *non_toplevel_tree_dlid)
|
|
{
|
|
void *handle = NULL;
|
|
struct dm_tree_node *child;
|
|
char *vgname, *lvname, *layer;
|
|
const char *name, *uuid;
|
|
|
|
while ((child = dm_tree_next_child(&handle, root, 0))) {
|
|
if (!(name = dm_tree_node_get_name(child)))
|
|
continue;
|
|
|
|
if (!(uuid = dm_tree_node_get_uuid(child)))
|
|
continue;
|
|
|
|
if (!dm_split_lvm_name(dm->mem, name, &vgname, &lvname, &layer)) {
|
|
log_error("_clean_tree: Couldn't split up device name %s.", name);
|
|
return 0;
|
|
}
|
|
|
|
/* Not meant to be top level? */
|
|
if (!*layer)
|
|
continue;
|
|
|
|
/* If operation was performed on a partial tree, don't remove it */
|
|
if (non_toplevel_tree_dlid && !strcmp(non_toplevel_tree_dlid, uuid))
|
|
continue;
|
|
|
|
if (!dm_tree_deactivate_children(root, uuid, strlen(uuid)))
|
|
return_0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int _tree_action(struct dev_manager *dm, struct logical_volume *lv,
|
|
struct lv_activate_opts *laopts, action_t action)
|
|
{
|
|
const size_t DLID_SIZE = ID_LEN + sizeof(UUID_PREFIX) - 1;
|
|
struct dm_tree *dtree;
|
|
struct dm_tree_node *root;
|
|
char *dlid;
|
|
int r = 0;
|
|
|
|
laopts->is_activate = (action == ACTIVATE);
|
|
|
|
if (!(dtree = _create_partial_dtree(dm, lv, laopts->origin_only)))
|
|
return_0;
|
|
|
|
if (!(root = dm_tree_find_node(dtree, 0, 0))) {
|
|
log_error("Lost dependency tree root node");
|
|
goto out_no_root;
|
|
}
|
|
|
|
/* Restore fs cookie */
|
|
dm_tree_set_cookie(root, fs_get_cookie());
|
|
|
|
if (!(dlid = build_dm_uuid(dm->mem, lv->lvid.s, (lv_is_origin(lv) && laopts->origin_only) ? "real" : NULL)))
|
|
goto_out;
|
|
|
|
/* Only process nodes with uuid of "LVM-" plus VG id. */
|
|
switch(action) {
|
|
case CLEAN:
|
|
/* Deactivate any unused non-toplevel nodes */
|
|
if (!_clean_tree(dm, root, laopts->origin_only ? dlid : NULL))
|
|
goto_out;
|
|
break;
|
|
case DEACTIVATE:
|
|
if (retry_deactivation())
|
|
dm_tree_retry_remove(root);
|
|
/* Deactivate LV and all devices it references that nothing else has open. */
|
|
if (!dm_tree_deactivate_children(root, dlid, DLID_SIZE))
|
|
goto_out;
|
|
if (!_remove_lv_symlinks(dm, root))
|
|
log_warn("Failed to remove all device symlinks associated with %s.", lv->name);
|
|
break;
|
|
case SUSPEND:
|
|
dm_tree_skip_lockfs(root);
|
|
if (!dm->flush_required && (lv->status & MIRRORED) && !(lv->status & PVMOVE))
|
|
dm_tree_use_no_flush_suspend(root);
|
|
/* Fall through */
|
|
case SUSPEND_WITH_LOCKFS:
|
|
if (!dm_tree_suspend_children(root, dlid, DLID_SIZE))
|
|
goto_out;
|
|
break;
|
|
case PRELOAD:
|
|
case ACTIVATE:
|
|
/* Add all required new devices to tree */
|
|
if (!_add_new_lv_to_dtree(dm, dtree, lv, laopts, (lv_is_origin(lv) && laopts->origin_only) ? "real" : NULL))
|
|
goto_out;
|
|
|
|
/* Preload any devices required before any suspensions */
|
|
if (!dm_tree_preload_children(root, dlid, DLID_SIZE))
|
|
goto_out;
|
|
|
|
if (dm_tree_node_size_changed(root))
|
|
dm->flush_required = 1;
|
|
|
|
if (action == ACTIVATE) {
|
|
if (!dm_tree_activate_children(root, dlid, DLID_SIZE))
|
|
goto_out;
|
|
if (!_create_lv_symlinks(dm, root))
|
|
log_warn("Failed to create symlinks for %s.", lv->name);
|
|
}
|
|
|
|
break;
|
|
default:
|
|
log_error("_tree_action: Action %u not supported.", action);
|
|
goto out;
|
|
}
|
|
|
|
r = 1;
|
|
|
|
out:
|
|
/* Save fs cookie for udev settle, do not wait here */
|
|
fs_set_cookie(dm_tree_get_cookie(root));
|
|
out_no_root:
|
|
dm_tree_free(dtree);
|
|
|
|
return r;
|
|
}
|
|
|
|
/* origin_only may only be set if we are resuming (not activating) an origin LV */
|
|
int dev_manager_activate(struct dev_manager *dm, struct logical_volume *lv,
|
|
struct lv_activate_opts *laopts)
|
|
{
|
|
if (!_tree_action(dm, lv, laopts, ACTIVATE))
|
|
return_0;
|
|
|
|
if (!_tree_action(dm, lv, laopts, CLEAN))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
/* origin_only may only be set if we are resuming (not activating) an origin LV */
|
|
int dev_manager_preload(struct dev_manager *dm, struct logical_volume *lv,
|
|
struct lv_activate_opts *laopts, int *flush_required)
|
|
{
|
|
if (!_tree_action(dm, lv, laopts, PRELOAD))
|
|
return_0;
|
|
|
|
*flush_required = dm->flush_required;
|
|
|
|
return 1;
|
|
}
|
|
|
|
int dev_manager_deactivate(struct dev_manager *dm, struct logical_volume *lv)
|
|
{
|
|
struct lv_activate_opts laopts = { 0 };
|
|
|
|
if (!_tree_action(dm, lv, &laopts, DEACTIVATE))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
int dev_manager_suspend(struct dev_manager *dm, struct logical_volume *lv,
|
|
struct lv_activate_opts *laopts, int lockfs, int flush_required)
|
|
{
|
|
dm->flush_required = flush_required;
|
|
|
|
if (!_tree_action(dm, lv, laopts, lockfs ? SUSPEND_WITH_LOCKFS : SUSPEND))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* Does device use VG somewhere in its construction?
|
|
* Returns 1 if uncertain.
|
|
*/
|
|
int dev_manager_device_uses_vg(struct device *dev,
|
|
struct volume_group *vg)
|
|
{
|
|
struct dm_tree *dtree;
|
|
struct dm_tree_node *root;
|
|
char dlid[sizeof(UUID_PREFIX) + sizeof(struct id) - 1] __attribute__((aligned(8)));
|
|
int r = 1;
|
|
|
|
if (!(dtree = dm_tree_create())) {
|
|
log_error("partial dtree creation failed");
|
|
return r;
|
|
}
|
|
|
|
if (!dm_tree_add_dev(dtree, (uint32_t) MAJOR(dev->dev), (uint32_t) MINOR(dev->dev))) {
|
|
log_error("Failed to add device %s (%" PRIu32 ":%" PRIu32") to dtree",
|
|
dev_name(dev), (uint32_t) MAJOR(dev->dev), (uint32_t) MINOR(dev->dev));
|
|
goto out;
|
|
}
|
|
|
|
memcpy(dlid, UUID_PREFIX, sizeof(UUID_PREFIX) - 1);
|
|
memcpy(dlid + sizeof(UUID_PREFIX) - 1, &vg->id.uuid[0], sizeof(vg->id));
|
|
|
|
if (!(root = dm_tree_find_node(dtree, 0, 0))) {
|
|
log_error("Lost dependency tree root node");
|
|
goto out;
|
|
}
|
|
|
|
if (dm_tree_children_use_uuid(root, dlid, sizeof(UUID_PREFIX) + sizeof(vg->id) - 1))
|
|
goto_out;
|
|
|
|
r = 0;
|
|
|
|
out:
|
|
dm_tree_free(dtree);
|
|
return r;
|
|
}
|