mirror of
git://sourceware.org/git/lvm2.git
synced 2024-12-21 13:34:40 +03:00
Add basic RAID segment type(s) support.
Implementation described in doc/lvm2-raid.txt. Basic support includes: - ability to create RAID 1/4/5/6 arrays - ability to delete RAID arrays - ability to display RAID arrays Notable missing features (not included in this patch): - ability to clean-up/repair failures - ability to convert RAID segment types - ability to monitor RAID segment types
This commit is contained in:
parent
c212019f3a
commit
cac52ca4ce
@ -1,5 +1,6 @@
|
|||||||
Version 2.02.87 -
|
Version 2.02.87 -
|
||||||
===============================
|
===============================
|
||||||
|
Add basic support for RAID 1/4/5/6 (i.e. create, remove, display)
|
||||||
Change DEFAULT_UDEV_SYNC to 1 so udev_sync is used even without any config.
|
Change DEFAULT_UDEV_SYNC to 1 so udev_sync is used even without any config.
|
||||||
Add systemd unit file to provide lvm2 monitoring.
|
Add systemd unit file to provide lvm2 monitoring.
|
||||||
Compare also file size to detect changed config file.
|
Compare also file size to detect changed config file.
|
||||||
|
33
configure
vendored
33
configure
vendored
@ -614,11 +614,12 @@ STATICDIR
|
|||||||
SNAPSHOTS
|
SNAPSHOTS
|
||||||
SELINUX_PC
|
SELINUX_PC
|
||||||
SELINUX_LIBS
|
SELINUX_LIBS
|
||||||
|
REPLICATORS
|
||||||
READLINE_LIBS
|
READLINE_LIBS
|
||||||
|
RAID
|
||||||
PTHREAD_LIBS
|
PTHREAD_LIBS
|
||||||
POOL
|
POOL
|
||||||
PKGCONFIG
|
PKGCONFIG
|
||||||
REPLICATORS
|
|
||||||
OCFDIR
|
OCFDIR
|
||||||
OCF
|
OCF
|
||||||
MIRRORS
|
MIRRORS
|
||||||
@ -797,6 +798,7 @@ with_pool
|
|||||||
with_cluster
|
with_cluster
|
||||||
with_snapshots
|
with_snapshots
|
||||||
with_mirrors
|
with_mirrors
|
||||||
|
with_raid
|
||||||
with_replicators
|
with_replicators
|
||||||
enable_readline
|
enable_readline
|
||||||
enable_realtime
|
enable_realtime
|
||||||
@ -1543,6 +1545,8 @@ Optional Packages:
|
|||||||
[[TYPE=internal]]
|
[[TYPE=internal]]
|
||||||
--with-mirrors=TYPE mirror support: internal/shared/none
|
--with-mirrors=TYPE mirror support: internal/shared/none
|
||||||
[[TYPE=internal]]
|
[[TYPE=internal]]
|
||||||
|
--with-raid=TYPE mirror support: internal/shared/none
|
||||||
|
[[TYPE=internal]]
|
||||||
--with-replicators=TYPE replicator support: internal/shared/none
|
--with-replicators=TYPE replicator support: internal/shared/none
|
||||||
[[TYPE=none]]
|
[[TYPE=none]]
|
||||||
--with-ocfdir=DIR install OCF files in DIR
|
--with-ocfdir=DIR install OCF files in DIR
|
||||||
@ -6885,6 +6889,31 @@ $as_echo "#define MIRRORED_INTERNAL 1" >>confdefs.h
|
|||||||
|
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
################################################################################
|
||||||
|
{ $as_echo "$as_me:${as_lineno-$LINENO}: checking whether to include raid" >&5
|
||||||
|
$as_echo_n "checking whether to include raid... " >&6; }
|
||||||
|
|
||||||
|
# Check whether --with-raid was given.
|
||||||
|
if test "${with_raid+set}" = set; then :
|
||||||
|
withval=$with_raid; RAID=$withval
|
||||||
|
else
|
||||||
|
RAID=internal
|
||||||
|
fi
|
||||||
|
|
||||||
|
{ $as_echo "$as_me:${as_lineno-$LINENO}: result: $RAID" >&5
|
||||||
|
$as_echo "$RAID" >&6; }
|
||||||
|
|
||||||
|
if [ "x$RAID" != xnone -a "x$RAID" != xinternal -a "x$RAID" != xshared ];
|
||||||
|
then as_fn_error $? "--with-raid parameter invalid
|
||||||
|
" "$LINENO" 5
|
||||||
|
fi;
|
||||||
|
|
||||||
|
if test x$RAID = xinternal; then
|
||||||
|
|
||||||
|
$as_echo "#define RAID_INTERNAL 1" >>confdefs.h
|
||||||
|
|
||||||
|
fi
|
||||||
|
|
||||||
################################################################################
|
################################################################################
|
||||||
{ $as_echo "$as_me:${as_lineno-$LINENO}: checking whether to include replicators" >&5
|
{ $as_echo "$as_me:${as_lineno-$LINENO}: checking whether to include replicators" >&5
|
||||||
$as_echo_n "checking whether to include replicators... " >&6; }
|
$as_echo_n "checking whether to include replicators... " >&6; }
|
||||||
@ -9169,6 +9198,7 @@ fi
|
|||||||
################################################################################
|
################################################################################
|
||||||
if [ \( "x$LVM1" = xshared -o "x$POOL" = xshared -o "x$CLUSTER" = xshared \
|
if [ \( "x$LVM1" = xshared -o "x$POOL" = xshared -o "x$CLUSTER" = xshared \
|
||||||
-o "x$SNAPSHOTS" = xshared -o "x$MIRRORS" = xshared \
|
-o "x$SNAPSHOTS" = xshared -o "x$MIRRORS" = xshared \
|
||||||
|
-o "x$RAID" = xshared \
|
||||||
\) -a "x$STATIC_LINK" = xyes ];
|
\) -a "x$STATIC_LINK" = xyes ];
|
||||||
then as_fn_error $? "Features cannot be 'shared' when building statically
|
then as_fn_error $? "Features cannot be 'shared' when building statically
|
||||||
" "$LINENO" 5
|
" "$LINENO" 5
|
||||||
@ -10380,6 +10410,7 @@ LVM_LIBAPI=`echo "$VER" | $AWK -F '[()]' '{print $2}'`
|
|||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
################################################################################
|
################################################################################
|
||||||
|
26
configure.in
26
configure.in
@ -340,6 +340,26 @@ if test x$MIRRORS = xinternal; then
|
|||||||
AC_DEFINE([MIRRORED_INTERNAL], 1, [Define to 1 to include built-in support for mirrors.])
|
AC_DEFINE([MIRRORED_INTERNAL], 1, [Define to 1 to include built-in support for mirrors.])
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
################################################################################
|
||||||
|
dnl -- raid inclusion type
|
||||||
|
AC_MSG_CHECKING(whether to include raid)
|
||||||
|
AC_ARG_WITH(raid,
|
||||||
|
AC_HELP_STRING([--with-raid=TYPE],
|
||||||
|
[mirror support: internal/shared/none
|
||||||
|
[[TYPE=internal]]]),
|
||||||
|
RAID=$withval, RAID=internal)
|
||||||
|
AC_MSG_RESULT($RAID)
|
||||||
|
|
||||||
|
if [[ "x$RAID" != xnone -a "x$RAID" != xinternal -a "x$RAID" != xshared ]];
|
||||||
|
then AC_MSG_ERROR(
|
||||||
|
--with-raid parameter invalid
|
||||||
|
)
|
||||||
|
fi;
|
||||||
|
|
||||||
|
if test x$RAID = xinternal; then
|
||||||
|
AC_DEFINE([RAID_INTERNAL], 1, [Define to 1 to include built-in support for raid.])
|
||||||
|
fi
|
||||||
|
|
||||||
################################################################################
|
################################################################################
|
||||||
dnl -- asynchronous volume replicator inclusion type
|
dnl -- asynchronous volume replicator inclusion type
|
||||||
AC_MSG_CHECKING(whether to include replicators)
|
AC_MSG_CHECKING(whether to include replicators)
|
||||||
@ -961,6 +981,7 @@ AC_CHECK_LIB(dl, dlopen, [
|
|||||||
dnl -- Check for shared/static conflicts
|
dnl -- Check for shared/static conflicts
|
||||||
if [[ \( "x$LVM1" = xshared -o "x$POOL" = xshared -o "x$CLUSTER" = xshared \
|
if [[ \( "x$LVM1" = xshared -o "x$POOL" = xshared -o "x$CLUSTER" = xshared \
|
||||||
-o "x$SNAPSHOTS" = xshared -o "x$MIRRORS" = xshared \
|
-o "x$SNAPSHOTS" = xshared -o "x$MIRRORS" = xshared \
|
||||||
|
-o "x$RAID" = xshared \
|
||||||
\) -a "x$STATIC_LINK" = xyes ]];
|
\) -a "x$STATIC_LINK" = xyes ]];
|
||||||
then AC_MSG_ERROR(
|
then AC_MSG_ERROR(
|
||||||
Features cannot be 'shared' when building statically
|
Features cannot be 'shared' when building statically
|
||||||
@ -1346,16 +1367,17 @@ AC_SUBST(LVM_PATCHLEVEL)
|
|||||||
AC_SUBST(LVM_RELEASE)
|
AC_SUBST(LVM_RELEASE)
|
||||||
AC_SUBST(LVM_RELEASE_DATE)
|
AC_SUBST(LVM_RELEASE_DATE)
|
||||||
AC_SUBST(MIRRORS)
|
AC_SUBST(MIRRORS)
|
||||||
|
AC_SUBST(MSGFMT)
|
||||||
AC_SUBST(OCF)
|
AC_SUBST(OCF)
|
||||||
AC_SUBST(OCFDIR)
|
AC_SUBST(OCFDIR)
|
||||||
AC_SUBST(REPLICATORS)
|
|
||||||
AC_SUBST(MSGFMT)
|
|
||||||
AC_SUBST(PKGCONFIG)
|
AC_SUBST(PKGCONFIG)
|
||||||
AC_SUBST(POOL)
|
AC_SUBST(POOL)
|
||||||
AC_SUBST(PTHREAD_LIBS)
|
AC_SUBST(PTHREAD_LIBS)
|
||||||
AC_SUBST(QUORUM_CFLAGS)
|
AC_SUBST(QUORUM_CFLAGS)
|
||||||
AC_SUBST(QUORUM_LIBS)
|
AC_SUBST(QUORUM_LIBS)
|
||||||
|
AC_SUBST(RAID)
|
||||||
AC_SUBST(READLINE_LIBS)
|
AC_SUBST(READLINE_LIBS)
|
||||||
|
AC_SUBST(REPLICATORS)
|
||||||
AC_SUBST(SACKPT_CFLAGS)
|
AC_SUBST(SACKPT_CFLAGS)
|
||||||
AC_SUBST(SACKPT_LIBS)
|
AC_SUBST(SACKPT_LIBS)
|
||||||
AC_SUBST(SALCK_CFLAGS)
|
AC_SUBST(SALCK_CFLAGS)
|
||||||
|
@ -474,6 +474,26 @@ activation {
|
|||||||
# "auto" - Use default value chosen by kernel.
|
# "auto" - Use default value chosen by kernel.
|
||||||
readahead = "auto"
|
readahead = "auto"
|
||||||
|
|
||||||
|
# 'mirror_segtype_default' defines which segtype will be used when the
|
||||||
|
# shorthand '-m' option is used for mirroring. The possible options are:
|
||||||
|
#
|
||||||
|
# "mirror" - The original RAID1 implementation provided by LVM2/DM. It is
|
||||||
|
# characterized by a flexible log solution (core, disk, mirrored)
|
||||||
|
# and by the necessity to block I/O while reconfiguring in the
|
||||||
|
# event of a failure. Snapshots of this type of RAID1 can be
|
||||||
|
# problematic.
|
||||||
|
#
|
||||||
|
# "raid1" - This implementation leverages MD's RAID1 personality through
|
||||||
|
# device-mapper. It is characterized by a lack of log options.
|
||||||
|
# (A log is always allocated for every device and they are placed
|
||||||
|
# on the same device as the image - no separate devices are
|
||||||
|
# required.) This mirror implementation does not require I/O
|
||||||
|
# to be blocked in the kernel in the event of a failure.
|
||||||
|
#
|
||||||
|
# Specify the '--type <mirror|raid1>' option to override this default
|
||||||
|
# setting.
|
||||||
|
mirror_segtype_default = "mirror"
|
||||||
|
|
||||||
# 'mirror_image_fault_policy' and 'mirror_log_fault_policy' define
|
# 'mirror_image_fault_policy' and 'mirror_log_fault_policy' define
|
||||||
# how a device failure affecting a mirror is handled.
|
# how a device failure affecting a mirror is handled.
|
||||||
# A mirror is composed of mirror images (copies) and a log.
|
# A mirror is composed of mirror images (copies) and a log.
|
||||||
|
@ -32,6 +32,10 @@ ifeq ("@MIRRORS@", "shared")
|
|||||||
SUBDIRS += mirror
|
SUBDIRS += mirror
|
||||||
endif
|
endif
|
||||||
|
|
||||||
|
ifeq ("@RAID@", "shared")
|
||||||
|
SUBDIRS += raid
|
||||||
|
endif
|
||||||
|
|
||||||
ifeq ("@REPLICATORS@", "shared")
|
ifeq ("@REPLICATORS@", "shared")
|
||||||
SUBDIRS += replicator
|
SUBDIRS += replicator
|
||||||
endif
|
endif
|
||||||
@ -140,6 +144,10 @@ ifeq ("@MIRRORS@", "internal")
|
|||||||
SOURCES += mirror/mirrored.c
|
SOURCES += mirror/mirrored.c
|
||||||
endif
|
endif
|
||||||
|
|
||||||
|
ifeq ("@RAID@", "internal")
|
||||||
|
SOURCES += raid/raid.c
|
||||||
|
endif
|
||||||
|
|
||||||
ifeq ("@REPLICATORS@", "internal")
|
ifeq ("@REPLICATORS@", "internal")
|
||||||
SOURCES += replicator/replicator.c
|
SOURCES += replicator/replicator.c
|
||||||
endif
|
endif
|
||||||
@ -170,6 +178,7 @@ ifeq ($(MAKECMDGOALS),distclean)
|
|||||||
format_pool \
|
format_pool \
|
||||||
snapshot \
|
snapshot \
|
||||||
mirror \
|
mirror \
|
||||||
|
raid \
|
||||||
replicator \
|
replicator \
|
||||||
locking
|
locking
|
||||||
endif
|
endif
|
||||||
|
@ -751,6 +751,7 @@ int dev_manager_mirror_percent(struct dev_manager *dm,
|
|||||||
{
|
{
|
||||||
char *name;
|
char *name;
|
||||||
const char *dlid;
|
const char *dlid;
|
||||||
|
const char *target_type = first_seg(lv)->segtype->name;
|
||||||
const char *layer = (lv_is_origin(lv)) ? "real" : NULL;
|
const char *layer = (lv_is_origin(lv)) ? "real" : NULL;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -766,8 +767,9 @@ int dev_manager_mirror_percent(struct dev_manager *dm,
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
log_debug("Getting device mirror status percentage for %s", name);
|
log_debug("Getting device %s status percentage for %s",
|
||||||
if (!(_percent(dm, name, dlid, "mirror", wait, lv, percent,
|
target_type, name);
|
||||||
|
if (!(_percent(dm, name, dlid, target_type, wait, lv, percent,
|
||||||
event_nr, 0)))
|
event_nr, 0)))
|
||||||
return_0;
|
return_0;
|
||||||
|
|
||||||
@ -1216,6 +1218,15 @@ int add_areas_line(struct dev_manager *dm, struct lv_segment *seg,
|
|||||||
(seg_pv(seg, s)->pe_start + (extent_size * seg_pe(seg, s)))))
|
(seg_pv(seg, s)->pe_start + (extent_size * seg_pe(seg, s)))))
|
||||||
return_0;
|
return_0;
|
||||||
} else if (seg_type(seg, s) == AREA_LV) {
|
} else if (seg_type(seg, s) == AREA_LV) {
|
||||||
|
if (seg_is_raid(seg)) {
|
||||||
|
dlid = build_dm_uuid(dm->mem,
|
||||||
|
seg_metalv(seg, s)->lvid.s,
|
||||||
|
NULL);
|
||||||
|
if (!dlid)
|
||||||
|
return_0;
|
||||||
|
dm_tree_node_add_target_area(node, NULL, dlid,
|
||||||
|
extent_size * seg_metale(seg, s));
|
||||||
|
}
|
||||||
if (!(dlid = build_dm_uuid(dm->mem, seg_lv(seg, s)->lvid.s, NULL)))
|
if (!(dlid = build_dm_uuid(dm->mem, seg_lv(seg, s)->lvid.s, NULL)))
|
||||||
return_0;
|
return_0;
|
||||||
if (!dm_tree_node_add_target_area(node, NULL, dlid, extent_size * seg_le(seg, s)))
|
if (!dm_tree_node_add_target_area(node, NULL, dlid, extent_size * seg_le(seg, s)))
|
||||||
@ -1444,11 +1455,16 @@ static int _add_segment_to_dtree(struct dev_manager *dm,
|
|||||||
return_0;
|
return_0;
|
||||||
} else {
|
} else {
|
||||||
/* Add any LVs used by this segment */
|
/* Add any LVs used by this segment */
|
||||||
for (s = 0; s < seg->area_count; s++)
|
for (s = 0; s < seg->area_count; s++) {
|
||||||
if ((seg_type(seg, s) == AREA_LV) &&
|
if ((seg_type(seg, s) == AREA_LV) &&
|
||||||
(!_add_new_lv_to_dtree(dm, dtree, seg_lv(seg, s),
|
(!_add_new_lv_to_dtree(dm, dtree, seg_lv(seg, s),
|
||||||
laopts, NULL)))
|
laopts, NULL)))
|
||||||
return_0;
|
return_0;
|
||||||
|
if (seg_is_raid(seg) &&
|
||||||
|
!_add_new_lv_to_dtree(dm, dtree, seg_metalv(seg, s),
|
||||||
|
laopts, NULL))
|
||||||
|
return_0;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Now we've added its dependencies, we can add the target itself */
|
/* Now we've added its dependencies, we can add the target itself */
|
||||||
|
@ -988,32 +988,40 @@ static int _init_single_segtype(struct cmd_context *cmd,
|
|||||||
|
|
||||||
static int _init_segtypes(struct cmd_context *cmd)
|
static int _init_segtypes(struct cmd_context *cmd)
|
||||||
{
|
{
|
||||||
|
int i;
|
||||||
struct segment_type *segtype;
|
struct segment_type *segtype;
|
||||||
struct segtype_library seglib = { .cmd = cmd };
|
struct segtype_library seglib = { .cmd = cmd };
|
||||||
|
struct segment_type *(*init_segtype_array[])(struct cmd_context *cmd) = {
|
||||||
|
init_striped_segtype,
|
||||||
|
init_zero_segtype,
|
||||||
|
init_error_segtype,
|
||||||
|
init_free_segtype,
|
||||||
|
#ifdef RAID_INTERNAL
|
||||||
|
init_raid1_segtype,
|
||||||
|
init_raid4_segtype,
|
||||||
|
init_raid5_segtype,
|
||||||
|
init_raid5_la_segtype,
|
||||||
|
init_raid5_ra_segtype,
|
||||||
|
init_raid5_ls_segtype,
|
||||||
|
init_raid5_rs_segtype,
|
||||||
|
init_raid6_segtype,
|
||||||
|
init_raid6_zr_segtype,
|
||||||
|
init_raid6_nr_segtype,
|
||||||
|
init_raid6_nc_segtype,
|
||||||
|
#endif
|
||||||
|
NULL
|
||||||
|
};
|
||||||
|
|
||||||
#ifdef HAVE_LIBDL
|
#ifdef HAVE_LIBDL
|
||||||
const struct config_node *cn;
|
const struct config_node *cn;
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
if (!(segtype = init_striped_segtype(cmd)))
|
for (i = 0; init_segtype_array[i]; i++) {
|
||||||
return 0;
|
if (!(segtype = init_segtype_array[i](cmd)))
|
||||||
segtype->library = NULL;
|
return 0;
|
||||||
dm_list_add(&cmd->segtypes, &segtype->list);
|
segtype->library = NULL;
|
||||||
|
dm_list_add(&cmd->segtypes, &segtype->list);
|
||||||
if (!(segtype = init_zero_segtype(cmd)))
|
}
|
||||||
return 0;
|
|
||||||
segtype->library = NULL;
|
|
||||||
dm_list_add(&cmd->segtypes, &segtype->list);
|
|
||||||
|
|
||||||
if (!(segtype = init_error_segtype(cmd)))
|
|
||||||
return 0;
|
|
||||||
segtype->library = NULL;
|
|
||||||
dm_list_add(&cmd->segtypes, &segtype->list);
|
|
||||||
|
|
||||||
if (!(segtype = init_free_segtype(cmd)))
|
|
||||||
return 0;
|
|
||||||
segtype->library = NULL;
|
|
||||||
dm_list_add(&cmd->segtypes, &segtype->list);
|
|
||||||
|
|
||||||
#ifdef SNAPSHOT_INTERNAL
|
#ifdef SNAPSHOT_INTERNAL
|
||||||
if (!(segtype = init_snapshot_segtype(cmd)))
|
if (!(segtype = init_snapshot_segtype(cmd)))
|
||||||
|
@ -49,6 +49,7 @@
|
|||||||
#define DEFAULT_USE_MLOCKALL 0
|
#define DEFAULT_USE_MLOCKALL 0
|
||||||
#define DEFAULT_METADATA_READ_ONLY 0
|
#define DEFAULT_METADATA_READ_ONLY 0
|
||||||
|
|
||||||
|
#define DEFAULT_MIRROR_SEGTYPE "mirror"
|
||||||
#define DEFAULT_MIRRORLOG "disk"
|
#define DEFAULT_MIRRORLOG "disk"
|
||||||
#define DEFAULT_MIRROR_LOG_FAULT_POLICY "allocate"
|
#define DEFAULT_MIRROR_LOG_FAULT_POLICY "allocate"
|
||||||
#define DEFAULT_MIRROR_IMAGE_FAULT_POLICY "remove"
|
#define DEFAULT_MIRROR_IMAGE_FAULT_POLICY "remove"
|
||||||
|
@ -544,10 +544,25 @@ int out_areas(struct formatter *f, const struct lv_segment *seg,
|
|||||||
(s == seg->area_count - 1) ? "" : ",");
|
(s == seg->area_count - 1) ? "" : ",");
|
||||||
break;
|
break;
|
||||||
case AREA_LV:
|
case AREA_LV:
|
||||||
outf(f, "\"%s\", %u%s",
|
if (!(seg->status & RAID)) {
|
||||||
seg_lv(seg, s)->name,
|
outf(f, "\"%s\", %u%s",
|
||||||
seg_le(seg, s),
|
seg_lv(seg, s)->name,
|
||||||
|
seg_le(seg, s),
|
||||||
|
(s == seg->area_count - 1) ? "" : ",");
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* RAID devices are laid-out in metadata/data pairs */
|
||||||
|
if (!(seg_lv(seg, s)->status & RAID_IMAGE) ||
|
||||||
|
!(seg_metalv(seg, s)->status & RAID_META)) {
|
||||||
|
log_error("RAID segment has non-RAID areas");
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
outf(f, "\"%s\", \"%s\"%s",
|
||||||
|
seg_metalv(seg, s)->name, seg_lv(seg, s)->name,
|
||||||
(s == seg->area_count - 1) ? "" : ",");
|
(s == seg->area_count - 1) ? "" : ",");
|
||||||
|
|
||||||
break;
|
break;
|
||||||
case AREA_UNASSIGNED:
|
case AREA_UNASSIGNED:
|
||||||
return 0;
|
return 0;
|
||||||
|
@ -56,6 +56,9 @@ static const struct flag _lv_flags[] = {
|
|||||||
{PVMOVE, "PVMOVE", STATUS_FLAG},
|
{PVMOVE, "PVMOVE", STATUS_FLAG},
|
||||||
{LOCKED, "LOCKED", STATUS_FLAG},
|
{LOCKED, "LOCKED", STATUS_FLAG},
|
||||||
{LV_NOTSYNCED, "NOTSYNCED", STATUS_FLAG},
|
{LV_NOTSYNCED, "NOTSYNCED", STATUS_FLAG},
|
||||||
|
{RAID, NULL, 0},
|
||||||
|
{RAID_META, NULL, 0},
|
||||||
|
{RAID_IMAGE, NULL, 0},
|
||||||
{MIRROR_IMAGE, NULL, 0},
|
{MIRROR_IMAGE, NULL, 0},
|
||||||
{MIRROR_LOG, NULL, 0},
|
{MIRROR_LOG, NULL, 0},
|
||||||
{MIRRORED, NULL, 0},
|
{MIRRORED, NULL, 0},
|
||||||
|
@ -365,10 +365,13 @@ static int _read_segment(struct dm_pool *mem, struct volume_group *vg,
|
|||||||
if (seg_is_mirrored(seg))
|
if (seg_is_mirrored(seg))
|
||||||
lv->status |= MIRRORED;
|
lv->status |= MIRRORED;
|
||||||
|
|
||||||
|
if (seg_is_raid(seg))
|
||||||
|
lv->status |= RAID;
|
||||||
|
|
||||||
if (seg_is_virtual(seg))
|
if (seg_is_virtual(seg))
|
||||||
lv->status |= VIRTUAL;
|
lv->status |= VIRTUAL;
|
||||||
|
|
||||||
if (_is_converting(lv))
|
if (!seg_is_raid(seg) && _is_converting(lv))
|
||||||
lv->status |= CONVERTING;
|
lv->status |= CONVERTING;
|
||||||
|
|
||||||
return 1;
|
return 1;
|
||||||
|
@ -35,6 +35,9 @@ typedef enum {
|
|||||||
NEXT_AREA
|
NEXT_AREA
|
||||||
} area_use_t;
|
} area_use_t;
|
||||||
|
|
||||||
|
/* FIXME: remove RAID_METADATA_AREA_LEN macro after defining 'raid_log_extents'*/
|
||||||
|
#define RAID_METADATA_AREA_LEN 1
|
||||||
|
|
||||||
/* FIXME These ended up getting used differently from first intended. Refactor. */
|
/* FIXME These ended up getting used differently from first intended. Refactor. */
|
||||||
#define A_CONTIGUOUS 0x01
|
#define A_CONTIGUOUS 0x01
|
||||||
#define A_CLING 0x02
|
#define A_CLING 0x02
|
||||||
@ -215,6 +218,11 @@ struct lv_segment *alloc_lv_segment(struct dm_pool *mem,
|
|||||||
struct lv_segment *seg;
|
struct lv_segment *seg;
|
||||||
uint32_t areas_sz = area_count * sizeof(*seg->areas);
|
uint32_t areas_sz = area_count * sizeof(*seg->areas);
|
||||||
|
|
||||||
|
if (!segtype) {
|
||||||
|
log_error(INTERNAL_ERROR "alloc_lv_segment: Missing segtype.");
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
|
||||||
if (!(seg = dm_pool_zalloc(mem, sizeof(*seg))))
|
if (!(seg = dm_pool_zalloc(mem, sizeof(*seg))))
|
||||||
return_NULL;
|
return_NULL;
|
||||||
|
|
||||||
@ -223,9 +231,10 @@ struct lv_segment *alloc_lv_segment(struct dm_pool *mem,
|
|||||||
return_NULL;
|
return_NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!segtype) {
|
if (segtype_is_raid(segtype) &&
|
||||||
log_error("alloc_lv_segment: Missing segtype.");
|
!(seg->meta_areas = dm_pool_zalloc(mem, areas_sz))) {
|
||||||
return NULL;
|
dm_pool_free(mem, seg); /* frees everything alloced since seg */
|
||||||
|
return_NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
seg->segtype = segtype;
|
seg->segtype = segtype;
|
||||||
@ -293,6 +302,27 @@ void release_lv_segment_area(struct lv_segment *seg, uint32_t s,
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (seg_lv(seg, s)->status & RAID_IMAGE) {
|
||||||
|
/*
|
||||||
|
* FIXME: Use lv_reduce not lv_remove
|
||||||
|
* We use lv_remove for now, because I haven't figured out
|
||||||
|
* why lv_reduce won't remove the LV.
|
||||||
|
lv_reduce(seg_lv(seg, s), area_reduction);
|
||||||
|
*/
|
||||||
|
if (area_reduction != seg->area_len) {
|
||||||
|
log_error("Unable to reduce RAID LV - operation not implemented.");
|
||||||
|
return;
|
||||||
|
} else
|
||||||
|
lv_remove(seg_lv(seg, s));
|
||||||
|
|
||||||
|
/* Remove metadata area if image has been removed */
|
||||||
|
if (area_reduction == seg->area_len) {
|
||||||
|
lv_reduce(seg_metalv(seg, s),
|
||||||
|
seg_metalv(seg, s)->le_count);
|
||||||
|
}
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
if (area_reduction == seg->area_len) {
|
if (area_reduction == seg->area_len) {
|
||||||
log_very_verbose("Remove %s:%" PRIu32 "[%" PRIu32 "] from "
|
log_very_verbose("Remove %s:%" PRIu32 "[%" PRIu32 "] from "
|
||||||
"the top of LV %s:%" PRIu32,
|
"the top of LV %s:%" PRIu32,
|
||||||
@ -375,9 +405,19 @@ int set_lv_segment_area_lv(struct lv_segment *seg, uint32_t area_num,
|
|||||||
log_very_verbose("Stack %s:%" PRIu32 "[%" PRIu32 "] on LV %s:%" PRIu32,
|
log_very_verbose("Stack %s:%" PRIu32 "[%" PRIu32 "] on LV %s:%" PRIu32,
|
||||||
seg->lv->name, seg->le, area_num, lv->name, le);
|
seg->lv->name, seg->le, area_num, lv->name, le);
|
||||||
|
|
||||||
seg->areas[area_num].type = AREA_LV;
|
if (status & RAID_META) {
|
||||||
seg_lv(seg, area_num) = lv;
|
seg->meta_areas[area_num].type = AREA_LV;
|
||||||
seg_le(seg, area_num) = le;
|
seg_metalv(seg, area_num) = lv;
|
||||||
|
if (le) {
|
||||||
|
log_error(INTERNAL_ERROR "Meta le != 0");
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
seg_metale(seg, area_num) = 0;
|
||||||
|
} else {
|
||||||
|
seg->areas[area_num].type = AREA_LV;
|
||||||
|
seg_lv(seg, area_num) = lv;
|
||||||
|
seg_le(seg, area_num) = le;
|
||||||
|
}
|
||||||
lv->status |= status;
|
lv->status |= status;
|
||||||
|
|
||||||
if (!add_seg_to_segs_using_this_lv(lv, seg))
|
if (!add_seg_to_segs_using_this_lv(lv, seg))
|
||||||
@ -559,14 +599,25 @@ struct alloc_handle {
|
|||||||
alloc_policy_t alloc; /* Overall policy */
|
alloc_policy_t alloc; /* Overall policy */
|
||||||
uint32_t new_extents; /* Number of new extents required */
|
uint32_t new_extents; /* Number of new extents required */
|
||||||
uint32_t area_count; /* Number of parallel areas */
|
uint32_t area_count; /* Number of parallel areas */
|
||||||
|
uint32_t parity_count; /* Adds to area_count, but not area_multiple */
|
||||||
uint32_t area_multiple; /* seg->len = area_len * area_multiple */
|
uint32_t area_multiple; /* seg->len = area_len * area_multiple */
|
||||||
uint32_t log_area_count; /* Number of parallel logs */
|
uint32_t log_area_count; /* Number of parallel logs */
|
||||||
uint32_t log_len; /* Length of log */
|
uint32_t metadata_area_count; /* Number of parallel metadata areas */
|
||||||
|
uint32_t log_len; /* Length of log/metadata_area */
|
||||||
uint32_t region_size; /* Mirror region size */
|
uint32_t region_size; /* Mirror region size */
|
||||||
uint32_t total_area_len; /* Total number of parallel extents */
|
uint32_t total_area_len; /* Total number of parallel extents */
|
||||||
|
|
||||||
unsigned maximise_cling;
|
unsigned maximise_cling;
|
||||||
unsigned mirror_logs_separate; /* Must mirror logs be on separate PVs? */
|
unsigned mirror_logs_separate; /* Force mirror logs on separate PVs? */
|
||||||
|
|
||||||
|
/*
|
||||||
|
* RAID devices require a metadata area that accompanies each
|
||||||
|
* device. During initial creation, it is best to look for space
|
||||||
|
* that is new_extents + log_len and then split that between two
|
||||||
|
* allocated areas when found. 'alloc_and_split_meta' indicates
|
||||||
|
* that this is the desired dynamic.
|
||||||
|
*/
|
||||||
|
unsigned alloc_and_split_meta;
|
||||||
|
|
||||||
const struct config_node *cling_tag_list_cn;
|
const struct config_node *cling_tag_list_cn;
|
||||||
|
|
||||||
@ -631,13 +682,14 @@ static struct alloc_handle *_alloc_init(struct cmd_context *cmd,
|
|||||||
uint32_t new_extents,
|
uint32_t new_extents,
|
||||||
uint32_t mirrors,
|
uint32_t mirrors,
|
||||||
uint32_t stripes,
|
uint32_t stripes,
|
||||||
uint32_t log_area_count,
|
uint32_t metadata_area_count,
|
||||||
uint32_t extent_size,
|
uint32_t extent_size,
|
||||||
uint32_t region_size,
|
uint32_t region_size,
|
||||||
struct dm_list *parallel_areas)
|
struct dm_list *parallel_areas)
|
||||||
{
|
{
|
||||||
struct alloc_handle *ah;
|
struct alloc_handle *ah;
|
||||||
uint32_t s, area_count;
|
uint32_t s, area_count, alloc_count;
|
||||||
|
size_t size = 0;
|
||||||
|
|
||||||
/* FIXME Caller should ensure this */
|
/* FIXME Caller should ensure this */
|
||||||
if (mirrors && !stripes)
|
if (mirrors && !stripes)
|
||||||
@ -650,7 +702,18 @@ static struct alloc_handle *_alloc_init(struct cmd_context *cmd,
|
|||||||
else
|
else
|
||||||
area_count = stripes;
|
area_count = stripes;
|
||||||
|
|
||||||
if (!(ah = dm_pool_zalloc(mem, sizeof(*ah) + sizeof(ah->alloced_areas[0]) * (area_count + log_area_count)))) {
|
size = sizeof(*ah);
|
||||||
|
alloc_count = area_count + segtype->parity_devs;
|
||||||
|
if (segtype_is_raid(segtype) && metadata_area_count)
|
||||||
|
/* RAID has a meta area for each device */
|
||||||
|
alloc_count *= 2;
|
||||||
|
else
|
||||||
|
/* mirrors specify their exact log count */
|
||||||
|
alloc_count += metadata_area_count;
|
||||||
|
|
||||||
|
size += sizeof(ah->alloced_areas[0]) * alloc_count;
|
||||||
|
|
||||||
|
if (!(ah = dm_pool_zalloc(mem, size))) {
|
||||||
log_error("allocation handle allocation failed");
|
log_error("allocation handle allocation failed");
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
@ -660,7 +723,7 @@ static struct alloc_handle *_alloc_init(struct cmd_context *cmd,
|
|||||||
if (segtype_is_virtual(segtype))
|
if (segtype_is_virtual(segtype))
|
||||||
return ah;
|
return ah;
|
||||||
|
|
||||||
if (!(area_count + log_area_count)) {
|
if (!(area_count + metadata_area_count)) {
|
||||||
log_error(INTERNAL_ERROR "_alloc_init called for non-virtual segment with no disk space.");
|
log_error(INTERNAL_ERROR "_alloc_init called for non-virtual segment with no disk space.");
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
@ -672,14 +735,35 @@ static struct alloc_handle *_alloc_init(struct cmd_context *cmd,
|
|||||||
|
|
||||||
ah->new_extents = new_extents;
|
ah->new_extents = new_extents;
|
||||||
ah->area_count = area_count;
|
ah->area_count = area_count;
|
||||||
ah->log_area_count = log_area_count;
|
ah->parity_count = segtype->parity_devs;
|
||||||
ah->region_size = region_size;
|
ah->region_size = region_size;
|
||||||
ah->alloc = alloc;
|
ah->alloc = alloc;
|
||||||
ah->area_multiple = _calc_area_multiple(segtype, area_count, stripes);
|
ah->area_multiple = _calc_area_multiple(segtype, area_count, stripes);
|
||||||
|
|
||||||
ah->log_len = log_area_count ? mirror_log_extents(ah->region_size, extent_size, ah->new_extents / ah->area_multiple) : 0;
|
if (segtype_is_raid(segtype)) {
|
||||||
|
if (metadata_area_count) {
|
||||||
|
if (metadata_area_count != area_count)
|
||||||
|
log_error(INTERNAL_ERROR
|
||||||
|
"Bad metadata_area_count");
|
||||||
|
ah->metadata_area_count = area_count;
|
||||||
|
ah->alloc_and_split_meta = 1;
|
||||||
|
|
||||||
for (s = 0; s < ah->area_count + ah->log_area_count; s++)
|
ah->log_len = RAID_METADATA_AREA_LEN;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* We need 'log_len' extents for each
|
||||||
|
* RAID device's metadata_area
|
||||||
|
*/
|
||||||
|
ah->new_extents += (ah->log_len * ah->area_multiple);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
ah->log_area_count = metadata_area_count;
|
||||||
|
ah->log_len = !metadata_area_count ? 0 :
|
||||||
|
mirror_log_extents(ah->region_size, extent_size,
|
||||||
|
ah->new_extents / ah->area_multiple);
|
||||||
|
}
|
||||||
|
|
||||||
|
for (s = 0; s < alloc_count; s++)
|
||||||
dm_list_init(&ah->alloced_areas[s]);
|
dm_list_init(&ah->alloced_areas[s]);
|
||||||
|
|
||||||
ah->parallel_areas = parallel_areas;
|
ah->parallel_areas = parallel_areas;
|
||||||
@ -700,9 +784,13 @@ void alloc_destroy(struct alloc_handle *ah)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/* Is there enough total space or should we give up immediately? */
|
/* Is there enough total space or should we give up immediately? */
|
||||||
static int _sufficient_pes_free(struct alloc_handle *ah, struct dm_list *pvms, uint32_t allocated, uint32_t extents_still_needed)
|
static int _sufficient_pes_free(struct alloc_handle *ah, struct dm_list *pvms,
|
||||||
|
uint32_t allocated, uint32_t extents_still_needed)
|
||||||
{
|
{
|
||||||
uint32_t total_extents_needed = (extents_still_needed - allocated) * ah->area_count / ah->area_multiple;
|
uint32_t area_extents_needed = (extents_still_needed - allocated) * ah->area_count / ah->area_multiple;
|
||||||
|
uint32_t parity_extents_needed = (extents_still_needed - allocated) * ah->parity_count / ah->area_multiple;
|
||||||
|
uint32_t metadata_extents_needed = ah->metadata_area_count * RAID_METADATA_AREA_LEN; /* One each */
|
||||||
|
uint32_t total_extents_needed = area_extents_needed + parity_extents_needed + metadata_extents_needed;
|
||||||
uint32_t free_pes = pv_maps_size(pvms);
|
uint32_t free_pes = pv_maps_size(pvms);
|
||||||
|
|
||||||
if (total_extents_needed > free_pes) {
|
if (total_extents_needed > free_pes) {
|
||||||
@ -874,9 +962,12 @@ static int _alloc_parallel_area(struct alloc_handle *ah, uint32_t max_to_allocat
|
|||||||
uint32_t area_len, len;
|
uint32_t area_len, len;
|
||||||
uint32_t s;
|
uint32_t s;
|
||||||
uint32_t ix_log_skip = 0; /* How many areas to skip in middle of array to reach log areas */
|
uint32_t ix_log_skip = 0; /* How many areas to skip in middle of array to reach log areas */
|
||||||
uint32_t total_area_count = ah->area_count + alloc_state->log_area_count_still_needed;
|
uint32_t total_area_count;
|
||||||
struct alloced_area *aa;
|
struct alloced_area *aa;
|
||||||
|
struct pv_area *pva;
|
||||||
|
|
||||||
|
total_area_count = ah->area_count + alloc_state->log_area_count_still_needed;
|
||||||
|
total_area_count += ah->parity_count;
|
||||||
if (!total_area_count) {
|
if (!total_area_count) {
|
||||||
log_error(INTERNAL_ERROR "_alloc_parallel_area called without any allocation to do.");
|
log_error(INTERNAL_ERROR "_alloc_parallel_area called without any allocation to do.");
|
||||||
return 1;
|
return 1;
|
||||||
@ -885,11 +976,13 @@ static int _alloc_parallel_area(struct alloc_handle *ah, uint32_t max_to_allocat
|
|||||||
area_len = max_to_allocate / ah->area_multiple;
|
area_len = max_to_allocate / ah->area_multiple;
|
||||||
|
|
||||||
/* Reduce area_len to the smallest of the areas */
|
/* Reduce area_len to the smallest of the areas */
|
||||||
for (s = 0; s < ah->area_count; s++)
|
for (s = 0; s < ah->area_count + ah->parity_count; s++)
|
||||||
if (area_len > alloc_state->areas[s].used)
|
if (area_len > alloc_state->areas[s].used)
|
||||||
area_len = alloc_state->areas[s].used;
|
area_len = alloc_state->areas[s].used;
|
||||||
|
|
||||||
if (!(aa = dm_pool_alloc(ah->mem, sizeof(*aa) * total_area_count))) {
|
len = (ah->alloc_and_split_meta) ? total_area_count * 2 : total_area_count;
|
||||||
|
len *= sizeof(*aa);
|
||||||
|
if (!(aa = dm_pool_alloc(ah->mem, len))) {
|
||||||
log_error("alloced_area allocation failed");
|
log_error("alloced_area allocation failed");
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
@ -901,24 +994,53 @@ static int _alloc_parallel_area(struct alloc_handle *ah, uint32_t max_to_allocat
|
|||||||
*/
|
*/
|
||||||
len = area_len;
|
len = area_len;
|
||||||
for (s = 0; s < total_area_count; s++) {
|
for (s = 0; s < total_area_count; s++) {
|
||||||
if (s == ah->area_count) {
|
if (s == (ah->area_count + ah->parity_count)) {
|
||||||
ix_log_skip = ix_log_offset - ah->area_count;
|
ix_log_skip = ix_log_offset - ah->area_count;
|
||||||
len = ah->log_len;
|
len = ah->log_len;
|
||||||
}
|
}
|
||||||
|
|
||||||
aa[s].pv = alloc_state->areas[s + ix_log_skip].pva->map->pv;
|
pva = alloc_state->areas[s + ix_log_skip].pva;
|
||||||
aa[s].pe = alloc_state->areas[s + ix_log_skip].pva->start;
|
if (ah->alloc_and_split_meta) {
|
||||||
aa[s].len = len;
|
/*
|
||||||
|
* The metadata area goes at the front of the allocated
|
||||||
|
* space for now, but could easily go at the end (or
|
||||||
|
* middle!).
|
||||||
|
*
|
||||||
|
* Even though we split these two from the same
|
||||||
|
* allocation, we store the images at the beginning
|
||||||
|
* of the areas array and the metadata at the end.
|
||||||
|
*/
|
||||||
|
s += ah->area_count + ah->parity_count;
|
||||||
|
aa[s].pv = pva->map->pv;
|
||||||
|
aa[s].pe = pva->start;
|
||||||
|
aa[s].len = ah->log_len;
|
||||||
|
|
||||||
|
log_debug("Allocating parallel metadata area %" PRIu32
|
||||||
|
" on %s start PE %" PRIu32
|
||||||
|
" length %" PRIu32 ".",
|
||||||
|
s, pv_dev_name(aa[s].pv), aa[s].pe,
|
||||||
|
ah->log_len);
|
||||||
|
|
||||||
|
consume_pv_area(pva, ah->log_len);
|
||||||
|
dm_list_add(&ah->alloced_areas[s], &aa[s].list);
|
||||||
|
s -= ah->area_count + ah->parity_count;
|
||||||
|
}
|
||||||
|
aa[s].pv = pva->map->pv;
|
||||||
|
aa[s].pe = pva->start;
|
||||||
|
aa[s].len = (ah->alloc_and_split_meta) ? len - ah->log_len : len;
|
||||||
|
|
||||||
log_debug("Allocating parallel area %" PRIu32
|
log_debug("Allocating parallel area %" PRIu32
|
||||||
" on %s start PE %" PRIu32 " length %" PRIu32 ".",
|
" on %s start PE %" PRIu32 " length %" PRIu32 ".",
|
||||||
s, dev_name(aa[s].pv->dev), aa[s].pe, len);
|
s, pv_dev_name(aa[s].pv), aa[s].pe, aa[s].len);
|
||||||
|
|
||||||
consume_pv_area(alloc_state->areas[s + ix_log_skip].pva, len);
|
consume_pv_area(pva, aa[s].len);
|
||||||
|
|
||||||
dm_list_add(&ah->alloced_areas[s], &aa[s].list);
|
dm_list_add(&ah->alloced_areas[s], &aa[s].list);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* Only need to alloc metadata from the first batch */
|
||||||
|
ah->alloc_and_split_meta = 0;
|
||||||
|
|
||||||
ah->total_area_len += area_len;
|
ah->total_area_len += area_len;
|
||||||
|
|
||||||
alloc_state->allocated += area_len * ah->area_multiple;
|
alloc_state->allocated += area_len * ah->area_multiple;
|
||||||
@ -1425,6 +1547,7 @@ static int _find_some_parallel_space(struct alloc_handle *ah, const struct alloc
|
|||||||
unsigned log_iteration_count = 0; /* extra iteration for logs on data devices */
|
unsigned log_iteration_count = 0; /* extra iteration for logs on data devices */
|
||||||
struct alloced_area *aa;
|
struct alloced_area *aa;
|
||||||
uint32_t s;
|
uint32_t s;
|
||||||
|
uint32_t devices_needed = ah->area_count + ah->parity_count;
|
||||||
|
|
||||||
/* ix_offset holds the number of parallel allocations that must be contiguous/cling */
|
/* ix_offset holds the number of parallel allocations that must be contiguous/cling */
|
||||||
if (alloc_parms->flags & (A_CONTIGUOUS | A_CLING) && alloc_parms->prev_lvseg)
|
if (alloc_parms->flags & (A_CONTIGUOUS | A_CLING) && alloc_parms->prev_lvseg)
|
||||||
@ -1442,15 +1565,15 @@ static int _find_some_parallel_space(struct alloc_handle *ah, const struct alloc
|
|||||||
log_debug("Still need %" PRIu32 " extents for %" PRIu32 " parallel areas and %" PRIu32 " log areas of %" PRIu32 " extents. "
|
log_debug("Still need %" PRIu32 " extents for %" PRIu32 " parallel areas and %" PRIu32 " log areas of %" PRIu32 " extents. "
|
||||||
"(Total %" PRIu32 " extents.)",
|
"(Total %" PRIu32 " extents.)",
|
||||||
(ah->new_extents - alloc_state->allocated) / ah->area_multiple,
|
(ah->new_extents - alloc_state->allocated) / ah->area_multiple,
|
||||||
ah->area_count, alloc_state->log_area_count_still_needed,
|
devices_needed, alloc_state->log_area_count_still_needed,
|
||||||
alloc_state->log_area_count_still_needed ? ah->log_len : 0,
|
alloc_state->log_area_count_still_needed ? ah->log_len : 0,
|
||||||
(ah->new_extents - alloc_state->allocated) * ah->area_count / ah->area_multiple +
|
(ah->new_extents - alloc_state->allocated) * devices_needed / ah->area_multiple +
|
||||||
alloc_state->log_area_count_still_needed * ah->log_len);
|
alloc_state->log_area_count_still_needed * ah->log_len);
|
||||||
|
|
||||||
/* ix holds the number of areas found on other PVs */
|
/* ix holds the number of areas found on other PVs */
|
||||||
do {
|
do {
|
||||||
if (log_iteration_count) {
|
if (log_iteration_count) {
|
||||||
log_debug("Found %u areas for %" PRIu32 " parallel areas and %" PRIu32 " log areas so far.", ix, ah->area_count, alloc_state->log_area_count_still_needed);
|
log_debug("Found %u areas for %" PRIu32 " parallel areas and %" PRIu32 " log areas so far.", ix, devices_needed, alloc_state->log_area_count_still_needed);
|
||||||
} else if (iteration_count)
|
} else if (iteration_count)
|
||||||
log_debug("Filled %u out of %u preferred areas so far.", preferred_count, ix_offset);
|
log_debug("Filled %u out of %u preferred areas so far.", preferred_count, ix_offset);
|
||||||
|
|
||||||
@ -1493,12 +1616,12 @@ static int _find_some_parallel_space(struct alloc_handle *ah, const struct alloc
|
|||||||
* not enough for the logs.
|
* not enough for the logs.
|
||||||
*/
|
*/
|
||||||
if (log_iteration_count) {
|
if (log_iteration_count) {
|
||||||
for (s = ah->area_count; s < ix + ix_offset; s++)
|
for (s = devices_needed; s < ix + ix_offset; s++)
|
||||||
if (alloc_state->areas[s].pva && alloc_state->areas[s].pva->map->pv == pvm->pv)
|
if (alloc_state->areas[s].pva && alloc_state->areas[s].pva->map->pv == pvm->pv)
|
||||||
goto next_pv;
|
goto next_pv;
|
||||||
/* On a second pass, avoid PVs already used in an uncommitted area */
|
/* On a second pass, avoid PVs already used in an uncommitted area */
|
||||||
} else if (iteration_count)
|
} else if (iteration_count)
|
||||||
for (s = 0; s < ah->area_count; s++)
|
for (s = 0; s < devices_needed; s++)
|
||||||
if (alloc_state->areas[s].pva && alloc_state->areas[s].pva->map->pv == pvm->pv)
|
if (alloc_state->areas[s].pva && alloc_state->areas[s].pva->map->pv == pvm->pv)
|
||||||
goto next_pv;
|
goto next_pv;
|
||||||
}
|
}
|
||||||
@ -1548,32 +1671,34 @@ static int _find_some_parallel_space(struct alloc_handle *ah, const struct alloc
|
|||||||
/* With cling and contiguous we stop if we found a match for *all* the areas */
|
/* With cling and contiguous we stop if we found a match for *all* the areas */
|
||||||
/* FIXME Rename these variables! */
|
/* FIXME Rename these variables! */
|
||||||
if ((alloc_parms->alloc == ALLOC_ANYWHERE &&
|
if ((alloc_parms->alloc == ALLOC_ANYWHERE &&
|
||||||
ix + ix_offset >= ah->area_count + alloc_state->log_area_count_still_needed) ||
|
ix + ix_offset >= devices_needed + alloc_state->log_area_count_still_needed) ||
|
||||||
(preferred_count == ix_offset &&
|
(preferred_count == ix_offset &&
|
||||||
(ix_offset == ah->area_count + alloc_state->log_area_count_still_needed)))
|
(ix_offset == devices_needed + alloc_state->log_area_count_still_needed))) {
|
||||||
|
log_error("Breaking: preferred_count = %d, ix_offset = %d, devices_needed = %d", preferred_count, ix_offset, devices_needed);
|
||||||
break;
|
break;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
} while ((alloc_parms->alloc == ALLOC_ANYWHERE && last_ix != ix && ix < ah->area_count + alloc_state->log_area_count_still_needed) ||
|
} while ((alloc_parms->alloc == ALLOC_ANYWHERE && last_ix != ix && ix < devices_needed + alloc_state->log_area_count_still_needed) ||
|
||||||
/* With cling_to_alloced, if there were gaps in the preferred areas, have a second iteration */
|
/* With cling_to_alloced, if there were gaps in the preferred areas, have a second iteration */
|
||||||
(alloc_parms->alloc == ALLOC_NORMAL && preferred_count &&
|
(alloc_parms->alloc == ALLOC_NORMAL && preferred_count &&
|
||||||
(preferred_count < ix_offset || alloc_state->log_area_count_still_needed) &&
|
(preferred_count < ix_offset || alloc_state->log_area_count_still_needed) &&
|
||||||
(alloc_parms->flags & A_CLING_TO_ALLOCED) && !iteration_count++) ||
|
(alloc_parms->flags & A_CLING_TO_ALLOCED) && !iteration_count++) ||
|
||||||
/* Extra iteration needed to fill log areas on PVs already used? */
|
/* Extra iteration needed to fill log areas on PVs already used? */
|
||||||
(alloc_parms->alloc == ALLOC_NORMAL && preferred_count == ix_offset && !ah->mirror_logs_separate &&
|
(alloc_parms->alloc == ALLOC_NORMAL && preferred_count == ix_offset && !ah->mirror_logs_separate &&
|
||||||
(ix + preferred_count >= ah->area_count) &&
|
(ix + preferred_count >= devices_needed) &&
|
||||||
(ix + preferred_count < ah->area_count + alloc_state->log_area_count_still_needed) && !log_iteration_count++));
|
(ix + preferred_count < devices_needed + alloc_state->log_area_count_still_needed) && !log_iteration_count++));
|
||||||
|
|
||||||
if (preferred_count < ix_offset && !(alloc_parms->flags & A_CLING_TO_ALLOCED))
|
if (preferred_count < ix_offset && !(alloc_parms->flags & A_CLING_TO_ALLOCED))
|
||||||
return 1;
|
return 1;
|
||||||
|
|
||||||
if (ix + preferred_count < ah->area_count + alloc_state->log_area_count_still_needed)
|
if (ix + preferred_count < devices_needed + alloc_state->log_area_count_still_needed)
|
||||||
return 1;
|
return 1;
|
||||||
|
|
||||||
/* Sort the areas so we allocate from the biggest */
|
/* Sort the areas so we allocate from the biggest */
|
||||||
if (log_iteration_count) {
|
if (log_iteration_count) {
|
||||||
if (ix > ah->area_count + 1) {
|
if (ix > devices_needed + 1) {
|
||||||
log_debug("Sorting %u log areas", ix - ah->area_count);
|
log_debug("Sorting %u log areas", ix - devices_needed);
|
||||||
qsort(alloc_state->areas + ah->area_count, ix - ah->area_count, sizeof(*alloc_state->areas),
|
qsort(alloc_state->areas + devices_needed, ix - devices_needed, sizeof(*alloc_state->areas),
|
||||||
_comp_area);
|
_comp_area);
|
||||||
}
|
}
|
||||||
} else if (ix > 1) {
|
} else if (ix > 1) {
|
||||||
@ -1584,7 +1709,7 @@ static int _find_some_parallel_space(struct alloc_handle *ah, const struct alloc
|
|||||||
|
|
||||||
/* If there are gaps in our preferred areas, fill then from the sorted part of the array */
|
/* If there are gaps in our preferred areas, fill then from the sorted part of the array */
|
||||||
if (preferred_count && preferred_count != ix_offset) {
|
if (preferred_count && preferred_count != ix_offset) {
|
||||||
for (s = 0; s < ah->area_count; s++)
|
for (s = 0; s < devices_needed; s++)
|
||||||
if (!alloc_state->areas[s].pva) {
|
if (!alloc_state->areas[s].pva) {
|
||||||
alloc_state->areas[s].pva = alloc_state->areas[ix_offset].pva;
|
alloc_state->areas[s].pva = alloc_state->areas[ix_offset].pva;
|
||||||
alloc_state->areas[s].used = alloc_state->areas[ix_offset].used;
|
alloc_state->areas[s].used = alloc_state->areas[ix_offset].used;
|
||||||
@ -1609,7 +1734,7 @@ static int _find_some_parallel_space(struct alloc_handle *ah, const struct alloc
|
|||||||
ix_log_offset = ix_offset + ix - too_small_for_log_count - ah->log_area_count;
|
ix_log_offset = ix_offset + ix - too_small_for_log_count - ah->log_area_count;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (ix + ix_offset < ah->area_count +
|
if (ix + ix_offset < devices_needed +
|
||||||
(alloc_state->log_area_count_still_needed ? alloc_state->log_area_count_still_needed +
|
(alloc_state->log_area_count_still_needed ? alloc_state->log_area_count_still_needed +
|
||||||
too_small_for_log_count : 0))
|
too_small_for_log_count : 0))
|
||||||
return 1;
|
return 1;
|
||||||
@ -1741,14 +1866,15 @@ static int _allocate(struct alloc_handle *ah,
|
|||||||
stack;
|
stack;
|
||||||
|
|
||||||
alloc_state.areas_size = dm_list_size(pvms);
|
alloc_state.areas_size = dm_list_size(pvms);
|
||||||
if (alloc_state.areas_size && alloc_state.areas_size < (ah->area_count + ah->log_area_count)) {
|
if (alloc_state.areas_size &&
|
||||||
|
alloc_state.areas_size < (ah->area_count + ah->parity_count + ah->log_area_count)) {
|
||||||
if (ah->alloc != ALLOC_ANYWHERE && ah->mirror_logs_separate) {
|
if (ah->alloc != ALLOC_ANYWHERE && ah->mirror_logs_separate) {
|
||||||
log_error("Not enough PVs with free space available "
|
log_error("Not enough PVs with free space available "
|
||||||
"for parallel allocation.");
|
"for parallel allocation.");
|
||||||
log_error("Consider --alloc anywhere if desperate.");
|
log_error("Consider --alloc anywhere if desperate.");
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
alloc_state.areas_size = ah->area_count + ah->log_area_count;
|
alloc_state.areas_size = ah->area_count + ah->parity_count + ah->log_area_count;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Upper bound if none of the PVs in prev_lvseg is in pvms */
|
/* Upper bound if none of the PVs in prev_lvseg is in pvms */
|
||||||
@ -1780,7 +1906,9 @@ static int _allocate(struct alloc_handle *ah,
|
|||||||
if (!_sufficient_pes_free(ah, pvms, alloc_state.allocated, ah->new_extents))
|
if (!_sufficient_pes_free(ah, pvms, alloc_state.allocated, ah->new_extents))
|
||||||
goto_out;
|
goto_out;
|
||||||
|
|
||||||
_init_alloc_parms(ah, &alloc_parms, alloc, prev_lvseg, can_split, alloc_state.allocated, ah->new_extents);
|
_init_alloc_parms(ah, &alloc_parms, alloc, prev_lvseg,
|
||||||
|
can_split, alloc_state.allocated,
|
||||||
|
ah->new_extents);
|
||||||
|
|
||||||
if (!_find_max_parallel_space_for_one_policy(ah, &alloc_parms, pvms, &alloc_state))
|
if (!_find_max_parallel_space_for_one_policy(ah, &alloc_parms, pvms, &alloc_state))
|
||||||
goto_out;
|
goto_out;
|
||||||
@ -2119,12 +2247,13 @@ int lv_add_log_segment(struct alloc_handle *ah, uint32_t first_area,
|
|||||||
|
|
||||||
static int _lv_insert_empty_sublvs(struct logical_volume *lv,
|
static int _lv_insert_empty_sublvs(struct logical_volume *lv,
|
||||||
const struct segment_type *segtype,
|
const struct segment_type *segtype,
|
||||||
uint32_t region_size,
|
uint32_t stripe_size, uint32_t region_size,
|
||||||
uint32_t devices)
|
uint32_t devices)
|
||||||
{
|
{
|
||||||
struct logical_volume *sub_lv;
|
struct logical_volume *sub_lv;
|
||||||
uint32_t i;
|
uint32_t i;
|
||||||
uint64_t status = 0;
|
uint64_t status = 0;
|
||||||
|
const char *layer_name;
|
||||||
size_t len = strlen(lv->name) + 32;
|
size_t len = strlen(lv->name) + 32;
|
||||||
char img_name[len];
|
char img_name[len];
|
||||||
struct lv_segment *mapseg;
|
struct lv_segment *mapseg;
|
||||||
@ -2135,15 +2264,22 @@ static int _lv_insert_empty_sublvs(struct logical_volume *lv,
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!segtype_is_mirrored(segtype))
|
if (segtype_is_raid(segtype)) {
|
||||||
|
lv->status |= RAID;
|
||||||
|
status = RAID_IMAGE;
|
||||||
|
layer_name = "rimage";
|
||||||
|
} else if (segtype_is_mirrored(segtype)) {
|
||||||
|
lv->status |= MIRRORED;
|
||||||
|
status = MIRROR_IMAGE;
|
||||||
|
layer_name = "mimage";
|
||||||
|
} else
|
||||||
return_0;
|
return_0;
|
||||||
lv->status |= MIRRORED;
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* First, create our top-level segment for our top-level LV
|
* First, create our top-level segment for our top-level LV
|
||||||
*/
|
*/
|
||||||
if (!(mapseg = alloc_lv_segment(lv->vg->cmd->mem, segtype,
|
if (!(mapseg = alloc_lv_segment(lv->vg->cmd->mem, segtype,
|
||||||
lv, 0, 0, lv->status, 0, NULL,
|
lv, 0, 0, lv->status, stripe_size, NULL,
|
||||||
devices, 0, 0, region_size, 0, NULL))) {
|
devices, 0, 0, region_size, 0, NULL))) {
|
||||||
log_error("Failed to create mapping segment for %s", lv->name);
|
log_error("Failed to create mapping segment for %s", lv->name);
|
||||||
return 0;
|
return 0;
|
||||||
@ -2152,17 +2288,34 @@ static int _lv_insert_empty_sublvs(struct logical_volume *lv,
|
|||||||
/*
|
/*
|
||||||
* Next, create all of our sub_lv's and link them in.
|
* Next, create all of our sub_lv's and link them in.
|
||||||
*/
|
*/
|
||||||
if (dm_snprintf(img_name, len, "%s%s", lv->name, "_mimage_%d") < 0)
|
|
||||||
return_0;
|
|
||||||
|
|
||||||
for (i = 0; i < devices; i++) {
|
for (i = 0; i < devices; i++) {
|
||||||
|
/* Data LVs */
|
||||||
|
if (dm_snprintf(img_name, len, "%s_%s_%u",
|
||||||
|
lv->name, layer_name, i) < 0)
|
||||||
|
return_0;
|
||||||
|
|
||||||
sub_lv = lv_create_empty(img_name, NULL,
|
sub_lv = lv_create_empty(img_name, NULL,
|
||||||
LVM_READ | LVM_WRITE | MIRROR_IMAGE,
|
LVM_READ | LVM_WRITE | status,
|
||||||
lv->alloc, lv->vg);
|
lv->alloc, lv->vg);
|
||||||
|
|
||||||
if (!sub_lv)
|
if (!sub_lv)
|
||||||
return_0;
|
return_0;
|
||||||
if (!set_lv_segment_area_lv(mapseg, i, sub_lv, 0, status))
|
if (!set_lv_segment_area_lv(mapseg, i, sub_lv, 0, status))
|
||||||
return_0;
|
return_0;
|
||||||
|
if (!segtype_is_raid(segtype))
|
||||||
|
continue;
|
||||||
|
|
||||||
|
/* RAID meta LVs */
|
||||||
|
if (dm_snprintf(img_name, len, "%s_rmeta_%u", lv->name, i) < 0)
|
||||||
|
return_0;
|
||||||
|
|
||||||
|
sub_lv = lv_create_empty(img_name, NULL,
|
||||||
|
LVM_READ | LVM_WRITE | RAID_META,
|
||||||
|
lv->alloc, lv->vg);
|
||||||
|
if (!sub_lv)
|
||||||
|
return_0;
|
||||||
|
if (!set_lv_segment_area_lv(mapseg, i, sub_lv, 0, RAID_META))
|
||||||
|
return_0;
|
||||||
}
|
}
|
||||||
dm_list_add(&lv->segments, &mapseg->list);
|
dm_list_add(&lv->segments, &mapseg->list);
|
||||||
|
|
||||||
@ -2174,31 +2327,101 @@ static int _lv_extend_layered_lv(struct alloc_handle *ah,
|
|||||||
uint32_t extents, uint32_t first_area,
|
uint32_t extents, uint32_t first_area,
|
||||||
uint32_t stripes, uint32_t stripe_size)
|
uint32_t stripes, uint32_t stripe_size)
|
||||||
{
|
{
|
||||||
struct logical_volume *sub_lv;
|
const struct segment_type *segtype;
|
||||||
|
struct logical_volume *sub_lv, *meta_lv;
|
||||||
struct lv_segment *seg;
|
struct lv_segment *seg;
|
||||||
uint32_t m, s;
|
uint32_t fa, s;
|
||||||
|
int clear_metadata = 0;
|
||||||
|
|
||||||
|
segtype = get_segtype_from_string(lv->vg->cmd, "striped");
|
||||||
|
|
||||||
|
/*
|
||||||
|
* The component devices of a "striped" LV all go in the same
|
||||||
|
* LV. However, RAID has an LV for each device - making the
|
||||||
|
* 'stripes' and 'stripe_size' parameters meaningless.
|
||||||
|
*/
|
||||||
|
if (seg_is_raid(first_seg(lv))) {
|
||||||
|
stripes = 1;
|
||||||
|
stripe_size = 0;
|
||||||
|
}
|
||||||
|
|
||||||
seg = first_seg(lv);
|
seg = first_seg(lv);
|
||||||
for (m = first_area, s = 0; s < seg->area_count; s++) {
|
for (fa = first_area, s = 0; s < seg->area_count; s++) {
|
||||||
if (is_temporary_mirror_layer(seg_lv(seg, s))) {
|
if (is_temporary_mirror_layer(seg_lv(seg, s))) {
|
||||||
if (!_lv_extend_layered_lv(ah, seg_lv(seg, s), extents,
|
if (!_lv_extend_layered_lv(ah, seg_lv(seg, s), extents,
|
||||||
m, stripes, stripe_size))
|
fa, stripes, stripe_size))
|
||||||
return_0;
|
return_0;
|
||||||
m += lv_mirror_count(seg_lv(seg, s));
|
fa += lv_mirror_count(seg_lv(seg, s));
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
sub_lv = seg_lv(seg, s);
|
sub_lv = seg_lv(seg, s);
|
||||||
if (!lv_add_segment(ah, m, stripes, sub_lv,
|
if (!lv_add_segment(ah, fa, stripes, sub_lv, segtype,
|
||||||
get_segtype_from_string(lv->vg->cmd,
|
|
||||||
"striped"),
|
|
||||||
stripe_size, sub_lv->status, 0)) {
|
stripe_size, sub_lv->status, 0)) {
|
||||||
log_error("Aborting. Failed to extend %s in %s.",
|
log_error("Aborting. Failed to extend %s in %s.",
|
||||||
sub_lv->name, lv->name);
|
sub_lv->name, lv->name);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
m += stripes;
|
|
||||||
|
/* Extend metadata LVs only on initial creation */
|
||||||
|
if (seg_is_raid(seg) && !lv->le_count) {
|
||||||
|
if (!seg->meta_areas) {
|
||||||
|
log_error("No meta_areas for RAID type");
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
meta_lv = seg_metalv(seg, s);
|
||||||
|
if (!lv_add_segment(ah, fa + seg->area_count, 1,
|
||||||
|
meta_lv, segtype, 0,
|
||||||
|
meta_lv->status, 0)) {
|
||||||
|
log_error("Failed to extend %s in %s.",
|
||||||
|
meta_lv->name, lv->name);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
lv_set_visible(meta_lv);
|
||||||
|
clear_metadata = 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
fa += stripes;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (clear_metadata) {
|
||||||
|
/*
|
||||||
|
* We must clear the metadata areas upon creation.
|
||||||
|
*/
|
||||||
|
if (!vg_write(meta_lv->vg) || !vg_commit(meta_lv->vg))
|
||||||
|
return_0;
|
||||||
|
|
||||||
|
for (s = 0; s < seg->area_count; s++) {
|
||||||
|
meta_lv = seg_metalv(seg, s);
|
||||||
|
if (!activate_lv(meta_lv->vg->cmd, meta_lv)) {
|
||||||
|
log_error("Failed to activate %s/%s for clearing",
|
||||||
|
meta_lv->vg->name, meta_lv->name);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
log_verbose("Clearing metadata area of %s/%s",
|
||||||
|
meta_lv->vg->name, meta_lv->name);
|
||||||
|
/*
|
||||||
|
* Rather than wiping meta_lv->size, we can simply
|
||||||
|
* wipe '1' to remove the superblock of any previous
|
||||||
|
* RAID devices. It is much quicker.
|
||||||
|
*/
|
||||||
|
if (!set_lv(meta_lv->vg->cmd, meta_lv, 1, 0)) {
|
||||||
|
log_error("Failed to zero %s/%s",
|
||||||
|
meta_lv->vg->name, meta_lv->name);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!deactivate_lv(meta_lv->vg->cmd, meta_lv)) {
|
||||||
|
log_error("Failed to deactivate %s/%s",
|
||||||
|
meta_lv->vg->name, meta_lv->name);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
lv_set_hidden(meta_lv);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
seg->area_len += extents;
|
seg->area_len += extents;
|
||||||
seg->len += extents;
|
seg->len += extents;
|
||||||
lv->le_count += extents;
|
lv->le_count += extents;
|
||||||
@ -2218,22 +2441,40 @@ int lv_extend(struct logical_volume *lv,
|
|||||||
struct dm_list *allocatable_pvs, alloc_policy_t alloc)
|
struct dm_list *allocatable_pvs, alloc_policy_t alloc)
|
||||||
{
|
{
|
||||||
int r = 1;
|
int r = 1;
|
||||||
|
int raid_logs = 0;
|
||||||
struct alloc_handle *ah;
|
struct alloc_handle *ah;
|
||||||
|
uint32_t dev_count = mirrors * stripes + segtype->parity_devs;
|
||||||
|
|
||||||
|
log_very_verbose("Extending segment type, %s", segtype->name);
|
||||||
|
|
||||||
if (segtype_is_virtual(segtype))
|
if (segtype_is_virtual(segtype))
|
||||||
return lv_add_virtual_segment(lv, 0u, extents, segtype);
|
return lv_add_virtual_segment(lv, 0u, extents, segtype);
|
||||||
|
|
||||||
if (!(ah = allocate_extents(lv->vg, lv, segtype, stripes, mirrors, 0, 0,
|
if (segtype_is_raid(segtype) && !lv->le_count)
|
||||||
extents, allocatable_pvs, alloc, NULL)))
|
raid_logs = mirrors * stripes;
|
||||||
|
|
||||||
|
if (!(ah = allocate_extents(lv->vg, lv, segtype, stripes, mirrors,
|
||||||
|
raid_logs, region_size, extents,
|
||||||
|
allocatable_pvs, alloc, NULL)))
|
||||||
return_0;
|
return_0;
|
||||||
|
|
||||||
if (!segtype_is_mirrored(segtype))
|
if (!segtype_is_mirrored(segtype) && !segtype_is_raid(segtype))
|
||||||
r = lv_add_segment(ah, 0, ah->area_count, lv, segtype,
|
r = lv_add_segment(ah, 0, ah->area_count, lv, segtype,
|
||||||
stripe_size, 0u, 0);
|
stripe_size, 0u, 0);
|
||||||
else {
|
else {
|
||||||
|
/*
|
||||||
|
* For RAID, all the devices are AREA_LV.
|
||||||
|
* However, for 'mirror on stripe' using non-RAID targets,
|
||||||
|
* the mirror legs are AREA_LV while the stripes underneath
|
||||||
|
* are AREA_PV. So if this is not RAID, reset dev_count to
|
||||||
|
* just 'mirrors' - the necessary sub_lv count.
|
||||||
|
*/
|
||||||
|
if (!segtype_is_raid(segtype))
|
||||||
|
dev_count = mirrors;
|
||||||
|
|
||||||
if (!lv->le_count &&
|
if (!lv->le_count &&
|
||||||
!_lv_insert_empty_sublvs(lv, segtype,
|
!_lv_insert_empty_sublvs(lv, segtype, stripe_size,
|
||||||
region_size, mirrors)) {
|
region_size, dev_count)) {
|
||||||
log_error("Failed to insert layer for %s", lv->name);
|
log_error("Failed to insert layer for %s", lv->name);
|
||||||
alloc_destroy(ah);
|
alloc_destroy(ah);
|
||||||
return 0;
|
return 0;
|
||||||
@ -2707,6 +2948,12 @@ int lv_remove_single(struct cmd_context *cmd, struct logical_volume *lv,
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (lv->status & (RAID_META | RAID_IMAGE)) {
|
||||||
|
log_error("Can't remove logical volume %s used as RAID device",
|
||||||
|
lv->name);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
if (lv->status & LOCKED) {
|
if (lv->status & LOCKED) {
|
||||||
log_error("Can't remove locked LV %s", lv->name);
|
log_error("Can't remove locked LV %s", lv->name);
|
||||||
return 0;
|
return 0;
|
||||||
@ -3498,8 +3745,11 @@ int lv_create_single(struct volume_group *vg,
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (lp->mirrors > 1 && !(vg->fid->fmt->features & FMT_SEGMENTS)) {
|
if ((segtype_is_mirrored(lp->segtype) ||
|
||||||
log_error("Metadata does not support mirroring.");
|
segtype_is_raid(lp->segtype)) &&
|
||||||
|
!(vg->fid->fmt->features & FMT_SEGMENTS)) {
|
||||||
|
log_error("Metadata does not support %s.",
|
||||||
|
segtype_is_raid(lp->segtype) ? "RAID" : "mirroring");
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -3632,9 +3882,12 @@ int lv_create_single(struct volume_group *vg,
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (lp->mirrors > 1 && !activation()) {
|
if ((segtype_is_mirrored(lp->segtype) ||
|
||||||
log_error("Can't create mirror without using "
|
segtype_is_raid(lp->segtype)) && !activation()) {
|
||||||
"device-mapper kernel driver.");
|
log_error("Can't create %s without using "
|
||||||
|
"device-mapper kernel driver.",
|
||||||
|
segtype_is_raid(lp->segtype) ? lp->segtype->name :
|
||||||
|
"mirror");
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -3654,18 +3907,15 @@ int lv_create_single(struct volume_group *vg,
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (lp->mirrors > 1) {
|
if (segtype_is_mirrored(lp->segtype) || segtype_is_raid(lp->segtype)) {
|
||||||
init_mirror_in_sync(lp->nosync);
|
init_mirror_in_sync(lp->nosync);
|
||||||
|
|
||||||
if (lp->nosync) {
|
if (lp->nosync) {
|
||||||
log_warn("WARNING: New mirror won't be synchronised. "
|
log_warn("WARNING: New %s won't be synchronised. "
|
||||||
"Don't read what you didn't write!");
|
"Don't read what you didn't write!",
|
||||||
|
lp->segtype->name);
|
||||||
status |= LV_NOTSYNCED;
|
status |= LV_NOTSYNCED;
|
||||||
}
|
}
|
||||||
|
|
||||||
lp->segtype = get_segtype_from_string(cmd, "mirror");
|
|
||||||
if (!lp->segtype)
|
|
||||||
return_0;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!(lv = lv_create_empty(lp->lv_name ? lp->lv_name : "lvol%d", NULL,
|
if (!(lv = lv_create_empty(lp->lv_name ? lp->lv_name : "lvol%d", NULL,
|
||||||
@ -3688,15 +3938,18 @@ int lv_create_single(struct volume_group *vg,
|
|||||||
if (!dm_list_empty(&lp->tags))
|
if (!dm_list_empty(&lp->tags))
|
||||||
dm_list_splice(&lv->tags, &lp->tags);
|
dm_list_splice(&lv->tags, &lp->tags);
|
||||||
|
|
||||||
if (!lv_extend(lv, lp->segtype, lp->stripes, lp->stripe_size,
|
lp->region_size = adjusted_mirror_region_size(vg->extent_size,
|
||||||
lp->mirrors,
|
lp->extents,
|
||||||
adjusted_mirror_region_size(vg->extent_size,
|
lp->region_size);
|
||||||
lp->extents,
|
|
||||||
lp->region_size),
|
if (!lv_extend(lv, lp->segtype,
|
||||||
|
lp->stripes, lp->stripe_size,
|
||||||
|
lp->mirrors, lp->region_size,
|
||||||
lp->extents, lp->pvh, lp->alloc))
|
lp->extents, lp->pvh, lp->alloc))
|
||||||
return_0;
|
return_0;
|
||||||
|
|
||||||
if ((lp->mirrors > 1) && lp->log_count) {
|
if (lp->log_count &&
|
||||||
|
!seg_is_raid(first_seg(lv)) && seg_is_mirrored(first_seg(lv))) {
|
||||||
if (!add_mirror_log(cmd, lv, lp->log_count,
|
if (!add_mirror_log(cmd, lv, lp->log_count,
|
||||||
first_seg(lv)->region_size,
|
first_seg(lv)->region_size,
|
||||||
lp->pvh, lp->alloc)) {
|
lp->pvh, lp->alloc)) {
|
||||||
|
@ -94,18 +94,22 @@ int check_lv_segments(struct logical_volume *lv, int complete_vg)
|
|||||||
inc_error_count;
|
inc_error_count;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (complete_vg && seg->log_lv) {
|
if (complete_vg && seg->log_lv &&
|
||||||
if (!seg_is_mirrored(seg)) {
|
!seg_is_mirrored(seg) && !(seg->status & RAID_IMAGE)) {
|
||||||
log_error("LV %s: segment %u has log LV but "
|
log_error("LV %s: segment %u log LV %s is not a "
|
||||||
"is not mirrored",
|
"mirror log or a RAID image",
|
||||||
lv->name, seg_count);
|
lv->name, seg_count, seg->log_lv->name);
|
||||||
inc_error_count;
|
inc_error_count;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Check mirror log - which is attached to the mirrored seg
|
||||||
|
*/
|
||||||
|
if (complete_vg && seg->log_lv && seg_is_mirrored(seg)) {
|
||||||
if (!(seg->log_lv->status & MIRROR_LOG)) {
|
if (!(seg->log_lv->status & MIRROR_LOG)) {
|
||||||
log_error("LV %s: segment %u log LV %s is not "
|
log_error("LV %s: segment %u log LV %s is not "
|
||||||
"a mirror log",
|
"a mirror log",
|
||||||
lv->name, seg_count, seg->log_lv->name);
|
lv->name, seg_count, seg->log_lv->name);
|
||||||
inc_error_count;
|
inc_error_count;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -113,7 +117,7 @@ int check_lv_segments(struct logical_volume *lv, int complete_vg)
|
|||||||
find_mirror_seg(seg2) != seg) {
|
find_mirror_seg(seg2) != seg) {
|
||||||
log_error("LV %s: segment %u log LV does not "
|
log_error("LV %s: segment %u log LV does not "
|
||||||
"point back to mirror segment",
|
"point back to mirror segment",
|
||||||
lv->name, seg_count);
|
lv->name, seg_count);
|
||||||
inc_error_count;
|
inc_error_count;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -189,6 +193,7 @@ int check_lv_segments(struct logical_volume *lv, int complete_vg)
|
|||||||
dm_list_iterate_items(sl, &seg_lv(seg, s)->segs_using_this_lv)
|
dm_list_iterate_items(sl, &seg_lv(seg, s)->segs_using_this_lv)
|
||||||
if (sl->seg == seg)
|
if (sl->seg == seg)
|
||||||
seg_found++;
|
seg_found++;
|
||||||
|
|
||||||
if (!seg_found) {
|
if (!seg_found) {
|
||||||
log_error("LV %s segment %d uses LV %s,"
|
log_error("LV %s segment %d uses LV %s,"
|
||||||
" but missing ptr from %s to %s",
|
" but missing ptr from %s to %s",
|
||||||
@ -205,7 +210,8 @@ int check_lv_segments(struct logical_volume *lv, int complete_vg)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (complete_vg && seg_is_mirrored(seg) &&
|
if (complete_vg &&
|
||||||
|
seg_is_mirrored(seg) && !seg_is_raid(seg) &&
|
||||||
seg_type(seg, s) == AREA_LV &&
|
seg_type(seg, s) == AREA_LV &&
|
||||||
seg_lv(seg, s)->le_count != seg->area_len) {
|
seg_lv(seg, s)->le_count != seg->area_len) {
|
||||||
log_error("LV %s: mirrored LV segment %u has "
|
log_error("LV %s: mirrored LV segment %u has "
|
||||||
@ -227,6 +233,8 @@ int check_lv_segments(struct logical_volume *lv, int complete_vg)
|
|||||||
continue;
|
continue;
|
||||||
if (lv == seg_lv(seg, s))
|
if (lv == seg_lv(seg, s))
|
||||||
seg_found++;
|
seg_found++;
|
||||||
|
if (seg_is_raid(seg) && (lv == seg_metalv(seg, s)))
|
||||||
|
seg_found++;
|
||||||
}
|
}
|
||||||
if (seg_is_replicator_dev(seg)) {
|
if (seg_is_replicator_dev(seg)) {
|
||||||
dm_list_iterate_items(rsite, &seg->replicator->rsites) {
|
dm_list_iterate_items(rsite, &seg->replicator->rsites) {
|
||||||
|
@ -46,6 +46,14 @@
|
|||||||
#define EXPORTED_VG 0x00000002U /* VG PV */
|
#define EXPORTED_VG 0x00000002U /* VG PV */
|
||||||
#define RESIZEABLE_VG 0x00000004U /* VG */
|
#define RESIZEABLE_VG 0x00000004U /* VG */
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Since the RAID flags are LV (and seg) only and the above three
|
||||||
|
* are VG/PV only, these flags are reused.
|
||||||
|
*/
|
||||||
|
#define RAID 0x00000001U /* LV */
|
||||||
|
#define RAID_META 0x00000002U /* LV */
|
||||||
|
#define RAID_IMAGE 0x00000004U /* LV */
|
||||||
|
|
||||||
/* May any free extents on this PV be used or must they be left free? */
|
/* May any free extents on this PV be used or must they be left free? */
|
||||||
#define ALLOCATABLE_PV 0x00000008U /* PV */
|
#define ALLOCATABLE_PV 0x00000008U /* PV */
|
||||||
|
|
||||||
@ -293,7 +301,7 @@ struct lv_segment {
|
|||||||
uint64_t status;
|
uint64_t status;
|
||||||
|
|
||||||
/* FIXME Fields depend on segment type */
|
/* FIXME Fields depend on segment type */
|
||||||
uint32_t stripe_size;
|
uint32_t stripe_size; /* For stripe and RAID - in sectors */
|
||||||
uint32_t area_count;
|
uint32_t area_count;
|
||||||
uint32_t area_len;
|
uint32_t area_len;
|
||||||
uint32_t chunk_size; /* For snapshots - in sectors */
|
uint32_t chunk_size; /* For snapshots - in sectors */
|
||||||
@ -309,6 +317,7 @@ struct lv_segment {
|
|||||||
struct dm_list tags;
|
struct dm_list tags;
|
||||||
|
|
||||||
struct lv_segment_area *areas;
|
struct lv_segment_area *areas;
|
||||||
|
struct lv_segment_area *meta_areas; /* For RAID */
|
||||||
|
|
||||||
struct logical_volume *replicator;/* For replicator-devs - link to replicator LV */
|
struct logical_volume *replicator;/* For replicator-devs - link to replicator LV */
|
||||||
struct logical_volume *rlog_lv; /* For replicators */
|
struct logical_volume *rlog_lv; /* For replicators */
|
||||||
@ -320,6 +329,7 @@ struct lv_segment {
|
|||||||
#define seg_type(seg, s) (seg)->areas[(s)].type
|
#define seg_type(seg, s) (seg)->areas[(s)].type
|
||||||
#define seg_pv(seg, s) (seg)->areas[(s)].u.pv.pvseg->pv
|
#define seg_pv(seg, s) (seg)->areas[(s)].u.pv.pvseg->pv
|
||||||
#define seg_lv(seg, s) (seg)->areas[(s)].u.lv.lv
|
#define seg_lv(seg, s) (seg)->areas[(s)].u.lv.lv
|
||||||
|
#define seg_metalv(seg, s) (seg)->meta_areas[(s)].u.lv.lv
|
||||||
|
|
||||||
struct pe_range {
|
struct pe_range {
|
||||||
struct dm_list list;
|
struct dm_list list;
|
||||||
|
@ -233,6 +233,7 @@ int mdas_empty_or_ignored(struct dm_list *mdas);
|
|||||||
#define seg_dev(seg, s) (seg)->areas[(s)].u.pv.pvseg->pv->dev
|
#define seg_dev(seg, s) (seg)->areas[(s)].u.pv.pvseg->pv->dev
|
||||||
#define seg_pe(seg, s) (seg)->areas[(s)].u.pv.pvseg->pe
|
#define seg_pe(seg, s) (seg)->areas[(s)].u.pv.pvseg->pe
|
||||||
#define seg_le(seg, s) (seg)->areas[(s)].u.lv.le
|
#define seg_le(seg, s) (seg)->areas[(s)].u.lv.le
|
||||||
|
#define seg_metale(seg, s) (seg)->meta_areas[(s)].u.lv.le
|
||||||
|
|
||||||
struct name_list {
|
struct name_list {
|
||||||
struct dm_list list;
|
struct dm_list list;
|
||||||
|
@ -38,6 +38,7 @@ struct dev_manager;
|
|||||||
#define SEG_MONITORED 0x00000080U
|
#define SEG_MONITORED 0x00000080U
|
||||||
#define SEG_REPLICATOR 0x00000100U
|
#define SEG_REPLICATOR 0x00000100U
|
||||||
#define SEG_REPLICATOR_DEV 0x00000200U
|
#define SEG_REPLICATOR_DEV 0x00000200U
|
||||||
|
#define SEG_RAID 0x00000400U
|
||||||
#define SEG_UNKNOWN 0x80000000U
|
#define SEG_UNKNOWN 0x80000000U
|
||||||
|
|
||||||
#define seg_is_mirrored(seg) ((seg)->segtype->flags & SEG_AREAS_MIRRORED ? 1 : 0)
|
#define seg_is_mirrored(seg) ((seg)->segtype->flags & SEG_AREAS_MIRRORED ? 1 : 0)
|
||||||
@ -46,6 +47,7 @@ struct dev_manager;
|
|||||||
#define seg_is_striped(seg) ((seg)->segtype->flags & SEG_AREAS_STRIPED ? 1 : 0)
|
#define seg_is_striped(seg) ((seg)->segtype->flags & SEG_AREAS_STRIPED ? 1 : 0)
|
||||||
#define seg_is_snapshot(seg) ((seg)->segtype->flags & SEG_SNAPSHOT ? 1 : 0)
|
#define seg_is_snapshot(seg) ((seg)->segtype->flags & SEG_SNAPSHOT ? 1 : 0)
|
||||||
#define seg_is_virtual(seg) ((seg)->segtype->flags & SEG_VIRTUAL ? 1 : 0)
|
#define seg_is_virtual(seg) ((seg)->segtype->flags & SEG_VIRTUAL ? 1 : 0)
|
||||||
|
#define seg_is_raid(seg) ((seg)->segtype->flags & SEG_RAID ? 1 : 0)
|
||||||
#define seg_can_split(seg) ((seg)->segtype->flags & SEG_CAN_SPLIT ? 1 : 0)
|
#define seg_can_split(seg) ((seg)->segtype->flags & SEG_CAN_SPLIT ? 1 : 0)
|
||||||
#define seg_cannot_be_zeroed(seg) ((seg)->segtype->flags & SEG_CANNOT_BE_ZEROED ? 1 : 0)
|
#define seg_cannot_be_zeroed(seg) ((seg)->segtype->flags & SEG_CANNOT_BE_ZEROED ? 1 : 0)
|
||||||
#define seg_monitored(seg) ((seg)->segtype->flags & SEG_MONITORED ? 1 : 0)
|
#define seg_monitored(seg) ((seg)->segtype->flags & SEG_MONITORED ? 1 : 0)
|
||||||
@ -53,14 +55,19 @@ struct dev_manager;
|
|||||||
|
|
||||||
#define segtype_is_striped(segtype) ((segtype)->flags & SEG_AREAS_STRIPED ? 1 : 0)
|
#define segtype_is_striped(segtype) ((segtype)->flags & SEG_AREAS_STRIPED ? 1 : 0)
|
||||||
#define segtype_is_mirrored(segtype) ((segtype)->flags & SEG_AREAS_MIRRORED ? 1 : 0)
|
#define segtype_is_mirrored(segtype) ((segtype)->flags & SEG_AREAS_MIRRORED ? 1 : 0)
|
||||||
|
#define segtype_is_raid(segtype) ((segtype)->flags & SEG_RAID ? 1 : 0)
|
||||||
#define segtype_is_virtual(segtype) ((segtype)->flags & SEG_VIRTUAL ? 1 : 0)
|
#define segtype_is_virtual(segtype) ((segtype)->flags & SEG_VIRTUAL ? 1 : 0)
|
||||||
|
|
||||||
struct segment_type {
|
struct segment_type {
|
||||||
struct dm_list list; /* Internal */
|
struct dm_list list; /* Internal */
|
||||||
struct cmd_context *cmd; /* lvm_register_segtype() sets this. */
|
struct cmd_context *cmd; /* lvm_register_segtype() sets this. */
|
||||||
|
|
||||||
uint32_t flags;
|
uint32_t flags;
|
||||||
|
uint32_t parity_devs; /* Parity drives required by segtype */
|
||||||
|
|
||||||
struct segtype_handler *ops;
|
struct segtype_handler *ops;
|
||||||
const char *name;
|
const char *name;
|
||||||
|
|
||||||
void *library; /* lvm_register_segtype() sets this. */
|
void *library; /* lvm_register_segtype() sets this. */
|
||||||
void *private; /* For the segtype handler to use. */
|
void *private; /* For the segtype handler to use. */
|
||||||
};
|
};
|
||||||
@ -117,7 +124,21 @@ struct segment_type *init_striped_segtype(struct cmd_context *cmd);
|
|||||||
struct segment_type *init_zero_segtype(struct cmd_context *cmd);
|
struct segment_type *init_zero_segtype(struct cmd_context *cmd);
|
||||||
struct segment_type *init_error_segtype(struct cmd_context *cmd);
|
struct segment_type *init_error_segtype(struct cmd_context *cmd);
|
||||||
struct segment_type *init_free_segtype(struct cmd_context *cmd);
|
struct segment_type *init_free_segtype(struct cmd_context *cmd);
|
||||||
struct segment_type *init_unknown_segtype(struct cmd_context *cmd, const char *name);
|
struct segment_type *init_unknown_segtype(struct cmd_context *cmd,
|
||||||
|
const char *name);
|
||||||
|
#ifdef RAID_INTERNAL
|
||||||
|
struct segment_type *init_raid1_segtype(struct cmd_context *cmd);
|
||||||
|
struct segment_type *init_raid4_segtype(struct cmd_context *cmd);
|
||||||
|
struct segment_type *init_raid5_segtype(struct cmd_context *cmd);
|
||||||
|
struct segment_type *init_raid5_la_segtype(struct cmd_context *cmd);
|
||||||
|
struct segment_type *init_raid5_ra_segtype(struct cmd_context *cmd);
|
||||||
|
struct segment_type *init_raid5_ls_segtype(struct cmd_context *cmd);
|
||||||
|
struct segment_type *init_raid5_rs_segtype(struct cmd_context *cmd);
|
||||||
|
struct segment_type *init_raid6_segtype(struct cmd_context *cmd);
|
||||||
|
struct segment_type *init_raid6_zr_segtype(struct cmd_context *cmd);
|
||||||
|
struct segment_type *init_raid6_nr_segtype(struct cmd_context *cmd);
|
||||||
|
struct segment_type *init_raid6_nc_segtype(struct cmd_context *cmd);
|
||||||
|
#endif
|
||||||
|
|
||||||
#ifdef REPLICATOR_INTERNAL
|
#ifdef REPLICATOR_INTERNAL
|
||||||
int init_replicator_segtype(struct segtype_library *seglib);
|
int init_replicator_segtype(struct segtype_library *seglib);
|
||||||
|
@ -493,6 +493,9 @@
|
|||||||
/* Define to 1 to include built-in support for GFS pool metadata. */
|
/* Define to 1 to include built-in support for GFS pool metadata. */
|
||||||
#undef POOL_INTERNAL
|
#undef POOL_INTERNAL
|
||||||
|
|
||||||
|
/* Define to 1 to include built-in support for raid. */
|
||||||
|
#undef RAID_INTERNAL
|
||||||
|
|
||||||
/* Define to 1 to include the LVM readline shell. */
|
/* Define to 1 to include the LVM readline shell. */
|
||||||
#undef READLINE_SUPPORT
|
#undef READLINE_SUPPORT
|
||||||
|
|
||||||
|
@ -361,6 +361,18 @@ int apply_lvname_restrictions(const char *name)
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (strstr(name, "_rimage")) {
|
||||||
|
log_error("Names including \"_rimage\" are reserved. "
|
||||||
|
"Please choose a different LV name.");
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (strstr(name, "_rmeta")) {
|
||||||
|
log_error("Names including \"_rmeta\" are reserved. "
|
||||||
|
"Please choose a different LV name.");
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
if (strstr(name, "_vorigin")) {
|
if (strstr(name, "_vorigin")) {
|
||||||
log_error("Names including \"_vorigin\" are reserved. "
|
log_error("Names including \"_vorigin\" are reserved. "
|
||||||
"Please choose a different LV name.");
|
"Please choose a different LV name.");
|
||||||
|
352
lib/raid/raid.c
Normal file
352
lib/raid/raid.c
Normal file
@ -0,0 +1,352 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (C) 2011 Red Hat, Inc. All rights reserved.
|
||||||
|
*
|
||||||
|
* This file is part of LVM2.
|
||||||
|
*
|
||||||
|
* This copyrighted material is made available to anyone wishing to use,
|
||||||
|
* modify, copy, or redistribute it subject to the terms and conditions
|
||||||
|
* of the GNU Lesser General Public License v.2.1.
|
||||||
|
*
|
||||||
|
* You should have received a copy of the GNU Lesser General Public License
|
||||||
|
* along with this program; if not, write to the Free Software Foundation,
|
||||||
|
* Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include "lib.h"
|
||||||
|
#include "toolcontext.h"
|
||||||
|
#include "segtype.h"
|
||||||
|
#include "display.h"
|
||||||
|
#include "text_export.h"
|
||||||
|
#include "text_import.h"
|
||||||
|
#include "config.h"
|
||||||
|
#include "str_list.h"
|
||||||
|
#include "targets.h"
|
||||||
|
#include "lvm-string.h"
|
||||||
|
#include "activate.h"
|
||||||
|
#include "metadata.h"
|
||||||
|
#include "lv_alloc.h"
|
||||||
|
|
||||||
|
static const char *_raid_name(const struct lv_segment *seg)
|
||||||
|
{
|
||||||
|
return seg->segtype->name;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int _raid_text_import_area_count(const struct config_node *sn,
|
||||||
|
uint32_t *area_count)
|
||||||
|
{
|
||||||
|
if (!get_config_uint32(sn, "device_count", area_count)) {
|
||||||
|
log_error("Couldn't read 'device_count' for "
|
||||||
|
"segment '%s'.", config_parent_name(sn));
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int
|
||||||
|
_raid_text_import_areas(struct lv_segment *seg, const struct config_node *sn,
|
||||||
|
const struct config_node *cn)
|
||||||
|
{
|
||||||
|
unsigned int s;
|
||||||
|
const struct config_value *cv;
|
||||||
|
struct logical_volume *lv1;
|
||||||
|
const char *seg_name = config_parent_name(sn);
|
||||||
|
|
||||||
|
if (!seg->area_count) {
|
||||||
|
log_error("No areas found for segment %s", seg_name);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
for (cv = cn->v, s = 0; cv && s < seg->area_count; s++, cv = cv->next) {
|
||||||
|
if (cv->type != CFG_STRING) {
|
||||||
|
log_error("Bad volume name in areas array for segment %s.", seg_name);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!cv->next) {
|
||||||
|
log_error("Missing data device in areas array for segment %s.", seg_name);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Metadata device comes first */
|
||||||
|
if (!(lv1 = find_lv(seg->lv->vg, cv->v.str))) {
|
||||||
|
log_error("Couldn't find volume '%s' for segment '%s'.",
|
||||||
|
cv->v.str ? : "NULL", seg_name);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
if (!set_lv_segment_area_lv(seg, s, lv1, 0, RAID_META))
|
||||||
|
return_0;
|
||||||
|
|
||||||
|
/* Data device comes second */
|
||||||
|
cv = cv->next;
|
||||||
|
if (!(lv1 = find_lv(seg->lv->vg, cv->v.str))) {
|
||||||
|
log_error("Couldn't find volume '%s' for segment '%s'.",
|
||||||
|
cv->v.str ? : "NULL", seg_name);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
if (!set_lv_segment_area_lv(seg, s, lv1, 0, RAID_IMAGE))
|
||||||
|
return_0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Check we read the correct number of RAID data/meta pairs.
|
||||||
|
*/
|
||||||
|
if (cv || (s < seg->area_count)) {
|
||||||
|
log_error("Incorrect number of areas in area array "
|
||||||
|
"for segment '%s'.", seg_name);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int
|
||||||
|
_raid_text_import(struct lv_segment *seg, const struct config_node *sn,
|
||||||
|
struct dm_hash_table *pv_hash)
|
||||||
|
{
|
||||||
|
const struct config_node *cn;
|
||||||
|
|
||||||
|
if (find_config_node(sn, "region_size")) {
|
||||||
|
if (!get_config_uint32(sn, "region_size", &seg->region_size)) {
|
||||||
|
log_error("Couldn't read 'region_size' for "
|
||||||
|
"segment %s of logical volume %s.",
|
||||||
|
config_parent_name(sn), seg->lv->name);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (find_config_node(sn, "stripe_size")) {
|
||||||
|
if (!get_config_uint32(sn, "stripe_size", &seg->stripe_size)) {
|
||||||
|
log_error("Couldn't read 'stripe_size' for "
|
||||||
|
"segment %s of logical volume %s.",
|
||||||
|
config_parent_name(sn), seg->lv->name);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (!(cn = find_config_node(sn, "raids"))) {
|
||||||
|
log_error("Couldn't find RAID array for "
|
||||||
|
"segment %s of logical volume %s.",
|
||||||
|
config_parent_name(sn), seg->lv->name);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!_raid_text_import_areas(seg, sn, cn)) {
|
||||||
|
log_error("Failed to import RAID images");
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
seg->status |= RAID;
|
||||||
|
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int
|
||||||
|
_raid_text_export(const struct lv_segment *seg, struct formatter *f)
|
||||||
|
{
|
||||||
|
outf(f, "device_count = %u", seg->area_count);
|
||||||
|
if (seg->region_size)
|
||||||
|
outf(f, "region_size = %" PRIu32, seg->region_size);
|
||||||
|
if (seg->stripe_size)
|
||||||
|
outf(f, "stripe_size = %" PRIu32, seg->stripe_size);
|
||||||
|
|
||||||
|
return out_areas(f, seg, "raid");
|
||||||
|
}
|
||||||
|
|
||||||
|
static int
|
||||||
|
_raid_add_target_line(struct dev_manager *dm __attribute__((unused)),
|
||||||
|
struct dm_pool *mem __attribute__((unused)),
|
||||||
|
struct cmd_context *cmd __attribute__((unused)),
|
||||||
|
void **target_state __attribute__((unused)),
|
||||||
|
struct lv_segment *seg,
|
||||||
|
const struct lv_activate_opts *laopts __attribute__((unused)),
|
||||||
|
struct dm_tree_node *node, uint64_t len,
|
||||||
|
uint32_t *pvmove_mirror_count __attribute__((unused)))
|
||||||
|
{
|
||||||
|
if (!seg->area_count) {
|
||||||
|
log_error(INTERNAL_ERROR "_raid_add_target_line called "
|
||||||
|
"with no areas for %s.", seg->lv->name);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!seg->region_size) {
|
||||||
|
log_error("Missing region size for mirror segment.");
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!dm_tree_node_add_raid_target(node, len, _raid_name(seg),
|
||||||
|
seg->region_size, seg->stripe_size,
|
||||||
|
0, 0))
|
||||||
|
return_0;
|
||||||
|
|
||||||
|
return add_areas_line(dm, seg, node, 0u, seg->area_count);
|
||||||
|
}
|
||||||
|
|
||||||
|
static int _raid_target_status_compatible(const char *type)
|
||||||
|
{
|
||||||
|
return (strstr(type, "raid") != NULL);
|
||||||
|
}
|
||||||
|
|
||||||
|
static int _raid_target_percent(void **target_state,
|
||||||
|
percent_t *percent,
|
||||||
|
struct dm_pool *mem,
|
||||||
|
struct cmd_context *cmd,
|
||||||
|
struct lv_segment *seg, char *params,
|
||||||
|
uint64_t *total_numerator,
|
||||||
|
uint64_t *total_denominator)
|
||||||
|
{
|
||||||
|
int i;
|
||||||
|
uint64_t numerator, denominator;
|
||||||
|
char *pos = params;
|
||||||
|
/*
|
||||||
|
* Status line:
|
||||||
|
* <raid_type> <#devs> <status_chars> <synced>/<total>
|
||||||
|
* Example:
|
||||||
|
* raid1 2 AA 1024000/1024000
|
||||||
|
*/
|
||||||
|
for (i = 0; i < 3; i++) {
|
||||||
|
pos = strstr(pos, " ");
|
||||||
|
if (pos)
|
||||||
|
pos++;
|
||||||
|
else
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
if (!pos || (sscanf(pos, "%" PRIu64 "/%" PRIu64 "%n",
|
||||||
|
&numerator, &denominator, &i) != 2)) {
|
||||||
|
log_error("Failed to parse %s status fraction: %s",
|
||||||
|
seg->segtype->name, params);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
*total_numerator += numerator;
|
||||||
|
*total_denominator += denominator;
|
||||||
|
|
||||||
|
if (seg)
|
||||||
|
seg->extents_copied = seg->area_len * numerator / denominator;
|
||||||
|
|
||||||
|
*percent = make_percent(numerator, denominator);
|
||||||
|
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static int
|
||||||
|
_raid_target_present(struct cmd_context *cmd,
|
||||||
|
const struct lv_segment *seg __attribute__((unused)),
|
||||||
|
unsigned *attributes __attribute__((unused)))
|
||||||
|
{
|
||||||
|
static int _raid_checked = 0;
|
||||||
|
static int _raid_present = 0;
|
||||||
|
|
||||||
|
if (!_raid_checked)
|
||||||
|
_raid_present = target_present(cmd, "raid", 1);
|
||||||
|
|
||||||
|
_raid_checked = 1;
|
||||||
|
|
||||||
|
return _raid_present;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int
|
||||||
|
_raid_modules_needed(struct dm_pool *mem,
|
||||||
|
const struct lv_segment *seg __attribute__((unused)),
|
||||||
|
struct dm_list *modules)
|
||||||
|
{
|
||||||
|
if (!str_list_add(mem, modules, "raid")) {
|
||||||
|
log_error("raid module string list allocation failed");
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
static void _raid_destroy(struct segment_type *segtype)
|
||||||
|
{
|
||||||
|
dm_free((void *) segtype);
|
||||||
|
}
|
||||||
|
|
||||||
|
static struct segtype_handler _raid_ops = {
|
||||||
|
.name = _raid_name,
|
||||||
|
.text_import_area_count = _raid_text_import_area_count,
|
||||||
|
.text_import = _raid_text_import,
|
||||||
|
.text_export = _raid_text_export,
|
||||||
|
.add_target_line = _raid_add_target_line,
|
||||||
|
.target_status_compatible = _raid_target_status_compatible,
|
||||||
|
.target_percent = _raid_target_percent,
|
||||||
|
.target_present = _raid_target_present,
|
||||||
|
.modules_needed = _raid_modules_needed,
|
||||||
|
.destroy = _raid_destroy,
|
||||||
|
};
|
||||||
|
|
||||||
|
static struct segment_type *init_raid_segtype(struct cmd_context *cmd,
|
||||||
|
const char *raid_type)
|
||||||
|
{
|
||||||
|
struct segment_type *segtype = dm_malloc(sizeof(*segtype));
|
||||||
|
|
||||||
|
if (!segtype)
|
||||||
|
return_NULL;
|
||||||
|
|
||||||
|
segtype->cmd = cmd;
|
||||||
|
|
||||||
|
segtype->flags = SEG_RAID;
|
||||||
|
segtype->parity_devs = strstr(raid_type, "raid6") ? 2 : 1;
|
||||||
|
|
||||||
|
segtype->ops = &_raid_ops;
|
||||||
|
segtype->name = raid_type;
|
||||||
|
|
||||||
|
segtype->private = NULL;
|
||||||
|
|
||||||
|
log_very_verbose("Initialised segtype: %s", segtype->name);
|
||||||
|
|
||||||
|
return segtype;
|
||||||
|
}
|
||||||
|
|
||||||
|
struct segment_type *init_raid1_segtype(struct cmd_context *cmd)
|
||||||
|
{
|
||||||
|
struct segment_type *segtype;
|
||||||
|
|
||||||
|
segtype = init_raid_segtype(cmd, "raid1");
|
||||||
|
if (!segtype)
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
segtype->flags |= SEG_AREAS_MIRRORED;
|
||||||
|
segtype->parity_devs = 0;
|
||||||
|
|
||||||
|
return segtype;
|
||||||
|
}
|
||||||
|
struct segment_type *init_raid4_segtype(struct cmd_context *cmd)
|
||||||
|
{
|
||||||
|
return init_raid_segtype(cmd, "raid4");
|
||||||
|
}
|
||||||
|
struct segment_type *init_raid5_segtype(struct cmd_context *cmd)
|
||||||
|
{
|
||||||
|
return init_raid_segtype(cmd, "raid5");
|
||||||
|
}
|
||||||
|
struct segment_type *init_raid5_la_segtype(struct cmd_context *cmd)
|
||||||
|
{
|
||||||
|
return init_raid_segtype(cmd, "raid5_la");
|
||||||
|
}
|
||||||
|
struct segment_type *init_raid5_ra_segtype(struct cmd_context *cmd)
|
||||||
|
{
|
||||||
|
return init_raid_segtype(cmd, "raid5_ra");
|
||||||
|
}
|
||||||
|
struct segment_type *init_raid5_ls_segtype(struct cmd_context *cmd)
|
||||||
|
{
|
||||||
|
return init_raid_segtype(cmd, "raid5_ls");
|
||||||
|
}
|
||||||
|
struct segment_type *init_raid5_rs_segtype(struct cmd_context *cmd)
|
||||||
|
{
|
||||||
|
return init_raid_segtype(cmd, "raid5_rs");
|
||||||
|
}
|
||||||
|
struct segment_type *init_raid6_segtype(struct cmd_context *cmd)
|
||||||
|
{
|
||||||
|
return init_raid_segtype(cmd, "raid6");
|
||||||
|
}
|
||||||
|
struct segment_type *init_raid6_zr_segtype(struct cmd_context *cmd)
|
||||||
|
{
|
||||||
|
return init_raid_segtype(cmd, "raid6_zr");
|
||||||
|
}
|
||||||
|
struct segment_type *init_raid6_nr_segtype(struct cmd_context *cmd)
|
||||||
|
{
|
||||||
|
return init_raid_segtype(cmd, "raid6_nr");
|
||||||
|
}
|
||||||
|
struct segment_type *init_raid6_nc_segtype(struct cmd_context *cmd)
|
||||||
|
{
|
||||||
|
return init_raid_segtype(cmd, "raid6_nc");
|
||||||
|
}
|
@ -467,6 +467,14 @@ int dm_tree_node_add_mirror_target_log(struct dm_tree_node *node,
|
|||||||
unsigned area_count,
|
unsigned area_count,
|
||||||
uint32_t flags);
|
uint32_t flags);
|
||||||
|
|
||||||
|
int dm_tree_node_add_raid_target(struct dm_tree_node *node,
|
||||||
|
uint64_t size,
|
||||||
|
const char *raid_type,
|
||||||
|
uint32_t region_size,
|
||||||
|
uint32_t stripe_size,
|
||||||
|
uint64_t reserved1,
|
||||||
|
uint64_t reserved2);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Replicator operation mode
|
* Replicator operation mode
|
||||||
* Note: API for Replicator is not yet stable
|
* Note: API for Replicator is not yet stable
|
||||||
|
@ -42,6 +42,16 @@ enum {
|
|||||||
SEG_SNAPSHOT_MERGE,
|
SEG_SNAPSHOT_MERGE,
|
||||||
SEG_STRIPED,
|
SEG_STRIPED,
|
||||||
SEG_ZERO,
|
SEG_ZERO,
|
||||||
|
SEG_RAID1,
|
||||||
|
SEG_RAID4,
|
||||||
|
SEG_RAID5_LA,
|
||||||
|
SEG_RAID5_RA,
|
||||||
|
SEG_RAID5_LS,
|
||||||
|
SEG_RAID5_RS,
|
||||||
|
SEG_RAID6_ZR,
|
||||||
|
SEG_RAID6_NR,
|
||||||
|
SEG_RAID6_NC,
|
||||||
|
SEG_LAST,
|
||||||
};
|
};
|
||||||
|
|
||||||
/* FIXME Add crypt and multipath support */
|
/* FIXME Add crypt and multipath support */
|
||||||
@ -61,6 +71,18 @@ struct {
|
|||||||
{ SEG_SNAPSHOT_MERGE, "snapshot-merge" },
|
{ SEG_SNAPSHOT_MERGE, "snapshot-merge" },
|
||||||
{ SEG_STRIPED, "striped" },
|
{ SEG_STRIPED, "striped" },
|
||||||
{ SEG_ZERO, "zero"},
|
{ SEG_ZERO, "zero"},
|
||||||
|
{ SEG_RAID1, "raid1"},
|
||||||
|
{ SEG_RAID4, "raid4"},
|
||||||
|
{ SEG_RAID5_LA, "raid5_la"},
|
||||||
|
{ SEG_RAID5_RA, "raid5_ra"},
|
||||||
|
{ SEG_RAID5_LS, "raid5_ls"},
|
||||||
|
{ SEG_RAID5_RS, "raid5_rs"},
|
||||||
|
{ SEG_RAID6_ZR, "raid6_zr"},
|
||||||
|
{ SEG_RAID6_NR, "raid6_nr"},
|
||||||
|
{ SEG_RAID6_NC, "raid6_nc"},
|
||||||
|
{ SEG_RAID5_LS, "raid5"}, /* same as "raid5_ls" (default for MD also) */
|
||||||
|
{ SEG_RAID6_ZR, "raid6"}, /* same as "raid6_zr" */
|
||||||
|
{ SEG_LAST, NULL },
|
||||||
};
|
};
|
||||||
|
|
||||||
/* Some segment types have a list of areas of other devices attached */
|
/* Some segment types have a list of areas of other devices attached */
|
||||||
@ -100,7 +122,7 @@ struct load_segment {
|
|||||||
unsigned area_count; /* Linear + Striped + Mirrored + Crypt + Replicator */
|
unsigned area_count; /* Linear + Striped + Mirrored + Crypt + Replicator */
|
||||||
struct dm_list areas; /* Linear + Striped + Mirrored + Crypt + Replicator */
|
struct dm_list areas; /* Linear + Striped + Mirrored + Crypt + Replicator */
|
||||||
|
|
||||||
uint32_t stripe_size; /* Striped */
|
uint32_t stripe_size; /* Striped + raid */
|
||||||
|
|
||||||
int persistent; /* Snapshot */
|
int persistent; /* Snapshot */
|
||||||
uint32_t chunk_size; /* Snapshot */
|
uint32_t chunk_size; /* Snapshot */
|
||||||
@ -109,7 +131,7 @@ struct load_segment {
|
|||||||
struct dm_tree_node *merge; /* Snapshot */
|
struct dm_tree_node *merge; /* Snapshot */
|
||||||
|
|
||||||
struct dm_tree_node *log; /* Mirror + Replicator */
|
struct dm_tree_node *log; /* Mirror + Replicator */
|
||||||
uint32_t region_size; /* Mirror */
|
uint32_t region_size; /* Mirror + raid */
|
||||||
unsigned clustered; /* Mirror */
|
unsigned clustered; /* Mirror */
|
||||||
unsigned mirror_area_count; /* Mirror */
|
unsigned mirror_area_count; /* Mirror */
|
||||||
uint32_t flags; /* Mirror log */
|
uint32_t flags; /* Mirror log */
|
||||||
@ -1499,6 +1521,17 @@ static int _emit_areas_line(struct dm_task *dmt __attribute__((unused)),
|
|||||||
EMIT_PARAMS(*pos, "%s", synctype);
|
EMIT_PARAMS(*pos, "%s", synctype);
|
||||||
}
|
}
|
||||||
break;
|
break;
|
||||||
|
case SEG_RAID1:
|
||||||
|
case SEG_RAID4:
|
||||||
|
case SEG_RAID5_LA:
|
||||||
|
case SEG_RAID5_RA:
|
||||||
|
case SEG_RAID5_LS:
|
||||||
|
case SEG_RAID5_RS:
|
||||||
|
case SEG_RAID6_ZR:
|
||||||
|
case SEG_RAID6_NR:
|
||||||
|
case SEG_RAID6_NC:
|
||||||
|
EMIT_PARAMS(*pos, " %s", devbuf);
|
||||||
|
break;
|
||||||
default:
|
default:
|
||||||
EMIT_PARAMS(*pos, "%s%s %" PRIu64, first_time ? "" : " ",
|
EMIT_PARAMS(*pos, "%s%s %" PRIu64, first_time ? "" : " ",
|
||||||
devbuf, area->offset);
|
devbuf, area->offset);
|
||||||
@ -1676,6 +1709,43 @@ static int _mirror_emit_segment_line(struct dm_task *dmt, uint32_t major,
|
|||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static int _raid_emit_segment_line(struct dm_task *dmt, uint32_t major,
|
||||||
|
uint32_t minor, struct load_segment *seg,
|
||||||
|
uint64_t *seg_start, char *params,
|
||||||
|
size_t paramsize)
|
||||||
|
{
|
||||||
|
int param_count = 1; /* mandatory 'chunk size'/'stripe size' arg */
|
||||||
|
int pos = 0;
|
||||||
|
|
||||||
|
if ((seg->flags & DM_NOSYNC) || (seg->flags & DM_FORCESYNC))
|
||||||
|
param_count++;
|
||||||
|
|
||||||
|
if (seg->region_size)
|
||||||
|
param_count += 2;
|
||||||
|
|
||||||
|
if ((seg->type == SEG_RAID1) && seg->stripe_size)
|
||||||
|
log_error("WARNING: Ignoring RAID1 stripe size");
|
||||||
|
|
||||||
|
EMIT_PARAMS(pos, "%s %d %u", dm_segtypes[seg->type].target,
|
||||||
|
param_count, seg->stripe_size);
|
||||||
|
|
||||||
|
if (seg->flags & DM_NOSYNC)
|
||||||
|
EMIT_PARAMS(pos, " nosync");
|
||||||
|
else if (seg->flags & DM_FORCESYNC)
|
||||||
|
EMIT_PARAMS(pos, " sync");
|
||||||
|
|
||||||
|
if (seg->region_size)
|
||||||
|
EMIT_PARAMS(pos, " region_size %u", seg->region_size);
|
||||||
|
|
||||||
|
/* Print number of metadata/data device pairs */
|
||||||
|
EMIT_PARAMS(pos, " %u", seg->area_count/2);
|
||||||
|
|
||||||
|
if (_emit_areas_line(dmt, seg, params, paramsize, &pos) <= 0)
|
||||||
|
return_0;
|
||||||
|
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
static int _emit_segment_line(struct dm_task *dmt, uint32_t major,
|
static int _emit_segment_line(struct dm_task *dmt, uint32_t major,
|
||||||
uint32_t minor, struct load_segment *seg,
|
uint32_t minor, struct load_segment *seg,
|
||||||
uint64_t *seg_start, char *params,
|
uint64_t *seg_start, char *params,
|
||||||
@ -1683,6 +1753,7 @@ static int _emit_segment_line(struct dm_task *dmt, uint32_t major,
|
|||||||
{
|
{
|
||||||
int pos = 0;
|
int pos = 0;
|
||||||
int r;
|
int r;
|
||||||
|
int target_type_is_raid = 0;
|
||||||
char originbuf[DM_FORMAT_DEV_BUFSIZE], cowbuf[DM_FORMAT_DEV_BUFSIZE];
|
char originbuf[DM_FORMAT_DEV_BUFSIZE], cowbuf[DM_FORMAT_DEV_BUFSIZE];
|
||||||
|
|
||||||
switch(seg->type) {
|
switch(seg->type) {
|
||||||
@ -1736,6 +1807,22 @@ static int _emit_segment_line(struct dm_task *dmt, uint32_t major,
|
|||||||
seg->iv_offset != DM_CRYPT_IV_DEFAULT ?
|
seg->iv_offset != DM_CRYPT_IV_DEFAULT ?
|
||||||
seg->iv_offset : *seg_start);
|
seg->iv_offset : *seg_start);
|
||||||
break;
|
break;
|
||||||
|
case SEG_RAID1:
|
||||||
|
case SEG_RAID4:
|
||||||
|
case SEG_RAID5_LA:
|
||||||
|
case SEG_RAID5_RA:
|
||||||
|
case SEG_RAID5_LS:
|
||||||
|
case SEG_RAID5_RS:
|
||||||
|
case SEG_RAID6_ZR:
|
||||||
|
case SEG_RAID6_NR:
|
||||||
|
case SEG_RAID6_NC:
|
||||||
|
target_type_is_raid = 1;
|
||||||
|
r = _raid_emit_segment_line(dmt, major, minor, seg, seg_start,
|
||||||
|
params, paramsize);
|
||||||
|
if (!r)
|
||||||
|
return_0;
|
||||||
|
|
||||||
|
break;
|
||||||
}
|
}
|
||||||
|
|
||||||
switch(seg->type) {
|
switch(seg->type) {
|
||||||
@ -1767,7 +1854,9 @@ static int _emit_segment_line(struct dm_task *dmt, uint32_t major,
|
|||||||
" %" PRIu64 " %s %s", major, minor,
|
" %" PRIu64 " %s %s", major, minor,
|
||||||
*seg_start, seg->size, dm_segtypes[seg->type].target, params);
|
*seg_start, seg->size, dm_segtypes[seg->type].target, params);
|
||||||
|
|
||||||
if (!dm_task_add_target(dmt, *seg_start, seg->size, dm_segtypes[seg->type].target, params))
|
if (!dm_task_add_target(dmt, *seg_start, seg->size,
|
||||||
|
target_type_is_raid ? "raid" :
|
||||||
|
dm_segtypes[seg->type].target, params))
|
||||||
return_0;
|
return_0;
|
||||||
|
|
||||||
*seg_start += seg->size;
|
*seg_start += seg->size;
|
||||||
@ -2250,6 +2339,30 @@ int dm_tree_node_add_mirror_target(struct dm_tree_node *node,
|
|||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
int dm_tree_node_add_raid_target(struct dm_tree_node *node,
|
||||||
|
uint64_t size,
|
||||||
|
const char *raid_type,
|
||||||
|
uint32_t region_size,
|
||||||
|
uint32_t stripe_size,
|
||||||
|
uint64_t reserved1,
|
||||||
|
uint64_t reserved2)
|
||||||
|
{
|
||||||
|
int i;
|
||||||
|
struct load_segment *seg = NULL;
|
||||||
|
|
||||||
|
for (i = 0; dm_segtypes[i].target && !seg; i++)
|
||||||
|
if (!strcmp(raid_type, dm_segtypes[i].target))
|
||||||
|
if (!(seg = _add_segment(node,
|
||||||
|
dm_segtypes[i].type, size)))
|
||||||
|
return_0;
|
||||||
|
|
||||||
|
seg->region_size = region_size;
|
||||||
|
seg->stripe_size = stripe_size;
|
||||||
|
seg->area_count = 0;
|
||||||
|
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
int dm_tree_node_add_replicator_target(struct dm_tree_node *node,
|
int dm_tree_node_add_replicator_target(struct dm_tree_node *node,
|
||||||
uint64_t size,
|
uint64_t size,
|
||||||
const char *rlog_uuid,
|
const char *rlog_uuid,
|
||||||
|
@ -199,11 +199,11 @@ of space.
|
|||||||
.TP
|
.TP
|
||||||
.I \-\-type SegmentType
|
.I \-\-type SegmentType
|
||||||
Create a logical volume that uses the specified segment type
|
Create a logical volume that uses the specified segment type
|
||||||
(e.g. "mirror", "snapshot", "striped"). Especially useful when no
|
(e.g. "raid5", "mirror", "snapshot"). Many segment types have a
|
||||||
existing commandline switch alias enables the use of the desired type
|
|
||||||
(e.g. "error" or "zero" types). Many segment types already have a
|
|
||||||
commandline switch alias that will enable their use (-s is an alias for
|
commandline switch alias that will enable their use (-s is an alias for
|
||||||
--type snapshot).
|
--type snapshot). However, this argument must be used when no existing
|
||||||
|
commandline switch alias is available for the desired type, as is the case
|
||||||
|
with "error", "zero", "raid4", "raid5", or "raid6".
|
||||||
.TP
|
.TP
|
||||||
.I \-\-virtualsize VirtualSize
|
.I \-\-virtualsize VirtualSize
|
||||||
Create a sparse device of the given size (in MB by default) using a snapshot.
|
Create a sparse device of the given size (in MB by default) using a snapshot.
|
||||||
@ -258,7 +258,12 @@ under 100MB of actual data on it.
|
|||||||
.br
|
.br
|
||||||
creates a linear logical volume "vg00/lvol1" using physical extents
|
creates a linear logical volume "vg00/lvol1" using physical extents
|
||||||
/dev/sda:0-7 and /dev/sdb:0-7 for allocation of extents.
|
/dev/sda:0-7 and /dev/sdb:0-7 for allocation of extents.
|
||||||
|
.br
|
||||||
|
|
||||||
|
"lvcreate --type raid5 -L 5G -i 3 -I 64 -n my_lv vg00"
|
||||||
|
.br
|
||||||
|
creates a 5GiB RAID5 logical volume "vg00/my_lv", with 3 stripes (plus
|
||||||
|
a parity drive for a total of 4 devices) and a stripesize of 64kiB.
|
||||||
|
|
||||||
.SH SEE ALSO
|
.SH SEE ALSO
|
||||||
.BR lvm (8),
|
.BR lvm (8),
|
||||||
|
@ -320,6 +320,50 @@ static int _read_mirror_params(struct lvcreate_params *lp,
|
|||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static int _read_raid_params(struct lvcreate_params *lp,
|
||||||
|
struct cmd_context *cmd)
|
||||||
|
{
|
||||||
|
if (!segtype_is_raid(lp->segtype))
|
||||||
|
return 1;
|
||||||
|
|
||||||
|
if (arg_count(cmd, corelog_ARG) ||
|
||||||
|
arg_count(cmd, mirrorlog_ARG)) {
|
||||||
|
log_error("Log options not applicable to %s segtype",
|
||||||
|
lp->segtype->name);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* get_stripe_params is called before _read_raid_params
|
||||||
|
* and already sets:
|
||||||
|
* lp->stripes
|
||||||
|
* lp->stripe_size
|
||||||
|
*
|
||||||
|
* For RAID 4/5/6, these values must be set.
|
||||||
|
*/
|
||||||
|
if (!segtype_is_mirrored(lp->segtype) && (lp->stripes < 2)) {
|
||||||
|
log_error("Number of stripes to %s not specified",
|
||||||
|
lp->segtype->name);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* _read_mirror_params is called before _read_raid_params
|
||||||
|
* and already sets:
|
||||||
|
* lp->nosync
|
||||||
|
* lp->region_size
|
||||||
|
*
|
||||||
|
* But let's ensure that programmers don't reorder
|
||||||
|
* that by checking and warning if they aren't set.
|
||||||
|
*/
|
||||||
|
if (!lp->region_size) {
|
||||||
|
log_error("Programmer error: lp->region_size not set.");
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
static int _lvcreate_params(struct lvcreate_params *lp,
|
static int _lvcreate_params(struct lvcreate_params *lp,
|
||||||
struct lvcreate_cmdline_params *lcp,
|
struct lvcreate_cmdline_params *lcp,
|
||||||
struct cmd_context *cmd,
|
struct cmd_context *cmd,
|
||||||
@ -328,6 +372,7 @@ static int _lvcreate_params(struct lvcreate_params *lp,
|
|||||||
int contiguous;
|
int contiguous;
|
||||||
unsigned pagesize;
|
unsigned pagesize;
|
||||||
struct arg_value_group_list *current_group;
|
struct arg_value_group_list *current_group;
|
||||||
|
const char *segtype_str;
|
||||||
const char *tag;
|
const char *tag;
|
||||||
|
|
||||||
memset(lp, 0, sizeof(*lp));
|
memset(lp, 0, sizeof(*lp));
|
||||||
@ -337,7 +382,11 @@ static int _lvcreate_params(struct lvcreate_params *lp,
|
|||||||
/*
|
/*
|
||||||
* Check selected options are compatible and determine segtype
|
* Check selected options are compatible and determine segtype
|
||||||
*/
|
*/
|
||||||
lp->segtype = get_segtype_from_string(cmd, arg_str_value(cmd, type_ARG, "striped"));
|
segtype_str = "striped";
|
||||||
|
if (arg_count(cmd, mirrors_ARG))
|
||||||
|
segtype_str = find_config_tree_str(cmd, "activation/mirror_segtype_default", DEFAULT_MIRROR_SEGTYPE);
|
||||||
|
|
||||||
|
lp->segtype = get_segtype_from_string(cmd, arg_str_value(cmd, type_ARG, segtype_str));
|
||||||
|
|
||||||
if (arg_count(cmd, snapshot_ARG) || seg_is_snapshot(lp) ||
|
if (arg_count(cmd, snapshot_ARG) || seg_is_snapshot(lp) ||
|
||||||
arg_count(cmd, virtualsize_ARG))
|
arg_count(cmd, virtualsize_ARG))
|
||||||
@ -345,7 +394,7 @@ static int _lvcreate_params(struct lvcreate_params *lp,
|
|||||||
|
|
||||||
lp->mirrors = 1;
|
lp->mirrors = 1;
|
||||||
|
|
||||||
/* Default to 2 mirrored areas if --type mirror */
|
/* Default to 2 mirrored areas if '--type mirror|raid1' */
|
||||||
if (segtype_is_mirrored(lp->segtype))
|
if (segtype_is_mirrored(lp->segtype))
|
||||||
lp->mirrors = 2;
|
lp->mirrors = 2;
|
||||||
|
|
||||||
@ -386,15 +435,12 @@ static int _lvcreate_params(struct lvcreate_params *lp,
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (lp->mirrors > 1) {
|
if (segtype_is_mirrored(lp->segtype) || segtype_is_raid(lp->segtype)) {
|
||||||
if (lp->snapshot) {
|
if (lp->snapshot) {
|
||||||
log_error("mirrors and snapshots are currently "
|
log_error("mirrors and snapshots are currently "
|
||||||
"incompatible");
|
"incompatible");
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!(lp->segtype = get_segtype_from_string(cmd, "striped")))
|
|
||||||
return_0;
|
|
||||||
} else {
|
} else {
|
||||||
if (arg_count(cmd, corelog_ARG)) {
|
if (arg_count(cmd, corelog_ARG)) {
|
||||||
log_error("--corelog is only available with mirrors");
|
log_error("--corelog is only available with mirrors");
|
||||||
@ -426,7 +472,8 @@ static int _lvcreate_params(struct lvcreate_params *lp,
|
|||||||
if (!_lvcreate_name_params(lp, cmd, &argc, &argv) ||
|
if (!_lvcreate_name_params(lp, cmd, &argc, &argv) ||
|
||||||
!_read_size_params(lp, lcp, cmd) ||
|
!_read_size_params(lp, lcp, cmd) ||
|
||||||
!get_stripe_params(cmd, &lp->stripes, &lp->stripe_size) ||
|
!get_stripe_params(cmd, &lp->stripes, &lp->stripe_size) ||
|
||||||
!_read_mirror_params(lp, cmd))
|
!_read_mirror_params(lp, cmd) ||
|
||||||
|
!_read_raid_params(lp, cmd))
|
||||||
return_0;
|
return_0;
|
||||||
|
|
||||||
lp->activate = arg_uint_value(cmd, available_ARG, CHANGE_AY);
|
lp->activate = arg_uint_value(cmd, available_ARG, CHANGE_AY);
|
||||||
|
Loading…
Reference in New Issue
Block a user