mirror of
git://sourceware.org/git/lvm2.git
synced 2024-12-22 17:35:59 +03:00
01b06cb71b
There's a race when asking lvmpolld about progress_status and actually reading the progress info from kernel: Even with lvmpolld being used we read status info from LVM2 command issued by a user (client side from lvmpolld perspective). The whole cycle may look like following: 1) set up an operation that requires polling (i.e. pvmove /dev/sda) 2) notify lvmpolld about such operation (lvmpolld_poll_init()) 3) in case 1) was not called with --background it would continue with: 4) Ask lvmpolld about progress status. it may respond with one of: a) in_progress b) not_found c) finished d) any low level error 5) provided the answer was 4a) try to read progress info from polling LV (i.e. vg00/pvmove1). Repeat steps 4) and 5) until the answer is != 4a). And now we got into racy configuration: lvmpolld answered with in_progress but it may be the that in_between 4) and 5) the operation has already finished and polling LV is already gone or there's nothing to ask for. Up to now, 5) would report warning and it could print such warning many times if --interval was set to 0. We don't want to scary users by warnings in such situation so let's just print these messages in verbose mode. Error messages due to error while reading kernel status info (on existing, active and locked LV) remained the same.
634 lines
17 KiB
C
634 lines
17 KiB
C
/*
|
|
* Copyright (C) 2003-2004 Sistina Software, Inc. All rights reserved.
|
|
* Copyright (C) 2004-2015 Red Hat, Inc. All rights reserved.
|
|
*
|
|
* This file is part of LVM2.
|
|
*
|
|
* This copyrighted material is made available to anyone wishing to use,
|
|
* modify, copy, or redistribute it subject to the terms and conditions
|
|
* of the GNU Lesser General Public License v.2.1.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public License
|
|
* along with this program; if not, write to the Free Software Foundation,
|
|
* Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
*/
|
|
|
|
#include <time.h>
|
|
|
|
#include "tools.h"
|
|
#include "polldaemon.h"
|
|
#include "lvm2cmdline.h"
|
|
#include "lvmpolld-client.h"
|
|
|
|
#define WAIT_AT_LEAST_NANOSECS 100000
|
|
|
|
progress_t poll_mirror_progress(struct cmd_context *cmd,
|
|
struct logical_volume *lv, const char *name,
|
|
struct daemon_parms *parms)
|
|
{
|
|
dm_percent_t segment_percent = DM_PERCENT_0, overall_percent = DM_PERCENT_0;
|
|
uint32_t event_nr = 0;
|
|
|
|
if (!lv_is_mirrored(lv) ||
|
|
!lv_mirror_percent(cmd, lv, !parms->interval, &segment_percent,
|
|
&event_nr) ||
|
|
(segment_percent == DM_PERCENT_INVALID)) {
|
|
log_error("ABORTING: Mirror percentage check failed.");
|
|
return PROGRESS_CHECK_FAILED;
|
|
}
|
|
|
|
overall_percent = copy_percent(lv);
|
|
if (parms->progress_display)
|
|
log_print_unless_silent("%s: %s: %.1f%%", name, parms->progress_title,
|
|
dm_percent_to_float(overall_percent));
|
|
else
|
|
log_verbose("%s: %s: %.1f%%", name, parms->progress_title,
|
|
dm_percent_to_float(overall_percent));
|
|
|
|
if (segment_percent != DM_PERCENT_100)
|
|
return PROGRESS_UNFINISHED;
|
|
|
|
if (overall_percent == DM_PERCENT_100)
|
|
return PROGRESS_FINISHED_ALL;
|
|
|
|
return PROGRESS_FINISHED_SEGMENT;
|
|
}
|
|
|
|
static int _check_lv_status(struct cmd_context *cmd,
|
|
struct volume_group *vg,
|
|
struct logical_volume *lv,
|
|
const char *name, struct daemon_parms *parms,
|
|
int *finished)
|
|
{
|
|
struct dm_list *lvs_changed;
|
|
progress_t progress;
|
|
|
|
/* By default, caller should not retry */
|
|
*finished = 1;
|
|
|
|
if (parms->aborting) {
|
|
if (!(lvs_changed = lvs_using_lv(cmd, vg, lv))) {
|
|
log_error("Failed to generate list of copied LVs: "
|
|
"can't abort.");
|
|
return 0;
|
|
}
|
|
if (!parms->poll_fns->finish_copy(cmd, vg, lv, lvs_changed))
|
|
return_0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
progress = parms->poll_fns->poll_progress(cmd, lv, name, parms);
|
|
if (progress == PROGRESS_CHECK_FAILED)
|
|
return_0;
|
|
|
|
if (progress == PROGRESS_UNFINISHED) {
|
|
/* The only case the caller *should* try again later */
|
|
*finished = 0;
|
|
return 1;
|
|
}
|
|
|
|
if (!(lvs_changed = lvs_using_lv(cmd, vg, lv))) {
|
|
log_error("ABORTING: Failed to generate list of copied LVs");
|
|
return 0;
|
|
}
|
|
|
|
/* Finished? Or progress to next segment? */
|
|
if (progress == PROGRESS_FINISHED_ALL) {
|
|
if (!parms->poll_fns->finish_copy(cmd, vg, lv, lvs_changed))
|
|
return_0;
|
|
} else {
|
|
if (parms->poll_fns->update_metadata &&
|
|
!parms->poll_fns->update_metadata(cmd, vg, lv, lvs_changed, 0)) {
|
|
log_error("ABORTING: Segment progression failed.");
|
|
parms->poll_fns->finish_copy(cmd, vg, lv, lvs_changed);
|
|
return 0;
|
|
}
|
|
*finished = 0; /* Another segment */
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
static void _nanosleep(unsigned secs, unsigned allow_zero_time)
|
|
{
|
|
struct timespec wtime = {
|
|
.tv_sec = secs,
|
|
};
|
|
|
|
if (!secs && !allow_zero_time)
|
|
wtime.tv_nsec = WAIT_AT_LEAST_NANOSECS;
|
|
|
|
while (!nanosleep(&wtime, &wtime) && errno == EINTR) {}
|
|
}
|
|
|
|
static void _sleep_and_rescan_devices(struct daemon_parms *parms)
|
|
{
|
|
if (parms->interval && !parms->aborting) {
|
|
dev_close_all();
|
|
_nanosleep(parms->interval, 1);
|
|
/* Devices might have changed while we slept */
|
|
init_full_scan_done(0);
|
|
}
|
|
}
|
|
|
|
int wait_for_single_lv(struct cmd_context *cmd, struct poll_operation_id *id,
|
|
struct daemon_parms *parms)
|
|
{
|
|
struct volume_group *vg;
|
|
struct logical_volume *lv;
|
|
int finished = 0;
|
|
|
|
/* Poll for completion */
|
|
while (!finished) {
|
|
if (parms->wait_before_testing)
|
|
_sleep_and_rescan_devices(parms);
|
|
|
|
/* Locks the (possibly renamed) VG again */
|
|
vg = vg_read(cmd, id->vg_name, NULL, READ_FOR_UPDATE);
|
|
if (vg_read_error(vg)) {
|
|
release_vg(vg);
|
|
log_error("ABORTING: Can't reread VG for %s.", id->display_name);
|
|
/* What more could we do here? */
|
|
return 0;
|
|
}
|
|
|
|
lv = find_lv(vg, id->lv_name);
|
|
|
|
if (lv && id->uuid && strcmp(id->uuid, (char *)&lv->lvid))
|
|
lv = NULL;
|
|
if (lv && parms->lv_type && !(lv->status & parms->lv_type))
|
|
lv = NULL;
|
|
|
|
if (!lv) {
|
|
if (parms->lv_type == PVMOVE)
|
|
log_print_unless_silent("%s: No pvmove in progress - already finished or aborted.",
|
|
id->display_name);
|
|
else
|
|
log_print_unless_silent("Can't find LV in %s for %s.",
|
|
vg->name, id->display_name);
|
|
|
|
unlock_and_release_vg(cmd, vg, vg->name);
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* If the LV is not active locally, the kernel cannot be
|
|
* queried for its status. We must exit in this case.
|
|
*/
|
|
if (!lv_is_active_locally(lv)) {
|
|
log_print_unless_silent("%s: Interrupted: No longer active.", id->display_name);
|
|
unlock_and_release_vg(cmd, vg, vg->name);
|
|
return 1;
|
|
}
|
|
|
|
if (!_check_lv_status(cmd, vg, lv, id->display_name, parms, &finished)) {
|
|
unlock_and_release_vg(cmd, vg, vg->name);
|
|
return_0;
|
|
}
|
|
|
|
unlock_and_release_vg(cmd, vg, vg->name);
|
|
|
|
/*
|
|
* FIXME Sleeping after testing, while preferred, also works around
|
|
* unreliable "finished" state checking in _percent_run. If the
|
|
* above _check_lv_status is deferred until after the first sleep it
|
|
* may be that a polldaemon will run without ever completing.
|
|
*
|
|
* This happens when one snapshot-merge polldaemon is racing with
|
|
* another (polling the same LV). The first to see the LV status
|
|
* reach the "finished" state will alter the LV that the other
|
|
* polldaemon(s) are polling. These other polldaemon(s) can then
|
|
* continue polling an LV that doesn't have a "status".
|
|
*/
|
|
if (!parms->wait_before_testing && !finished)
|
|
_sleep_and_rescan_devices(parms);
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
struct poll_id_list {
|
|
struct dm_list list;
|
|
struct poll_operation_id *id;
|
|
};
|
|
|
|
static struct poll_operation_id *copy_poll_operation_id(struct dm_pool *mem,
|
|
const struct poll_operation_id *id)
|
|
{
|
|
struct poll_operation_id *copy;
|
|
|
|
if (!id)
|
|
return_NULL;
|
|
|
|
copy = (struct poll_operation_id *) dm_pool_alloc(mem, sizeof(struct poll_operation_id));
|
|
if (!copy) {
|
|
log_error("Poll operation ID allocation failed.");
|
|
return NULL;
|
|
}
|
|
|
|
copy->display_name = id->display_name ? dm_pool_strdup(mem, id->display_name) : NULL;
|
|
copy->lv_name = id->lv_name ? dm_pool_strdup(mem, id->lv_name) : NULL;
|
|
copy->vg_name = id->vg_name ? dm_pool_strdup(mem, id->vg_name) : NULL;
|
|
copy->uuid = id->uuid ? dm_pool_strdup(mem, id->uuid) : NULL;
|
|
|
|
if (!copy->display_name || !copy->lv_name || !copy->vg_name || !copy->uuid) {
|
|
log_error("Failed to copy one or more poll_operation_id members.");
|
|
return NULL;
|
|
}
|
|
|
|
return copy;
|
|
}
|
|
|
|
static struct poll_id_list* poll_id_list_create(struct dm_pool *mem,
|
|
const struct poll_operation_id *id)
|
|
{
|
|
struct poll_id_list *idl = (struct poll_id_list *) dm_pool_alloc(mem, sizeof(struct poll_id_list));
|
|
|
|
if (!idl) {
|
|
log_error("Poll ID list allocation failed.");
|
|
return NULL;
|
|
}
|
|
|
|
if (!(idl->id = copy_poll_operation_id(mem, id))) {
|
|
dm_pool_free(mem, idl);
|
|
return NULL;
|
|
}
|
|
|
|
return idl;
|
|
}
|
|
|
|
static int _poll_vg(struct cmd_context *cmd, const char *vgname,
|
|
struct volume_group *vg, struct processing_handle *handle)
|
|
{
|
|
struct daemon_parms *parms;
|
|
struct lv_list *lvl;
|
|
struct dm_list idls;
|
|
struct poll_id_list *idl;
|
|
struct poll_operation_id id;
|
|
struct logical_volume *lv;
|
|
int finished;
|
|
|
|
if (!handle || !(parms = (struct daemon_parms *) handle->custom_handle)) {
|
|
log_error(INTERNAL_ERROR "Handle is undefined.");
|
|
return ECMD_FAILED;
|
|
}
|
|
|
|
dm_list_init(&idls);
|
|
|
|
/*
|
|
* first iterate all LVs in a VG and collect LVs suitable
|
|
* for polling (or an abort) which takes place below
|
|
*/
|
|
dm_list_iterate_items(lvl, &vg->lvs) {
|
|
lv = lvl->lv;
|
|
if (!(lv->status & parms->lv_type))
|
|
continue;
|
|
id.display_name = parms->poll_fns->get_copy_name_from_lv(lv);
|
|
if (!id.display_name && !parms->aborting)
|
|
continue;
|
|
|
|
if (!id.display_name) {
|
|
log_error("Device name for LV %s not found in metadata. "
|
|
"(unfinished pvmove mirror removal?)", display_lvname(lv));
|
|
goto err;
|
|
}
|
|
|
|
/* FIXME Need to do the activation from _set_up_pvmove here
|
|
* if it's not running and we're not aborting. */
|
|
if (!lv_is_active(lv)) {
|
|
log_print_unless_silent("%s: Skipping inactive LV. Try lvchange or vgchange.", id.display_name);
|
|
continue;
|
|
}
|
|
|
|
id.lv_name = lv->name;
|
|
id.vg_name = vg->name;
|
|
id.uuid = lv->lvid.s;
|
|
|
|
idl = poll_id_list_create(cmd->mem, &id);
|
|
if (!idl) {
|
|
log_error("Failed to create poll_id_list.");
|
|
goto err;
|
|
}
|
|
|
|
dm_list_add(&idls, &idl->list);
|
|
}
|
|
|
|
/* perform the poll operation on LVs collected in previous cycle */
|
|
dm_list_iterate_items(idl, &idls) {
|
|
if (!(lv = find_lv(vg, idl->id->lv_name)))
|
|
continue;
|
|
if (idl->id->uuid && strcmp(idl->id->uuid, (char *)&lv->lvid))
|
|
continue;
|
|
if (parms->lv_type && !(lv->status & parms->lv_type))
|
|
continue;
|
|
if (_check_lv_status(cmd, vg, lv, idl->id->display_name, parms, &finished) && !finished)
|
|
parms->outstanding_count++;
|
|
}
|
|
|
|
err:
|
|
if (!dm_list_empty(&idls))
|
|
dm_pool_free(cmd->mem, dm_list_item(dm_list_first(&idls), struct poll_id_list));
|
|
|
|
return ECMD_PROCESSED;
|
|
}
|
|
|
|
static void _poll_for_all_vgs(struct cmd_context *cmd,
|
|
struct processing_handle *handle)
|
|
{
|
|
struct daemon_parms *parms = (struct daemon_parms *) handle->custom_handle;
|
|
|
|
while (1) {
|
|
parms->outstanding_count = 0;
|
|
process_each_vg(cmd, 0, NULL, READ_FOR_UPDATE, handle, _poll_vg);
|
|
if (!parms->outstanding_count)
|
|
break;
|
|
if (parms->interval)
|
|
dev_close_all();
|
|
_nanosleep(parms->interval, 1);
|
|
}
|
|
}
|
|
|
|
#ifdef LVMPOLLD_SUPPORT
|
|
typedef struct {
|
|
struct daemon_parms *parms;
|
|
struct dm_list idls;
|
|
} lvmpolld_parms_t;
|
|
|
|
static int report_progress(struct cmd_context *cmd, struct poll_operation_id *id,
|
|
struct daemon_parms *parms)
|
|
{
|
|
struct volume_group *vg;
|
|
struct logical_volume *lv;
|
|
|
|
vg = vg_read(cmd, id->vg_name, NULL, 0);
|
|
if (vg_read_error(vg)) {
|
|
release_vg(vg);
|
|
log_error("Can't reread VG for %s", id->display_name);
|
|
return 0;
|
|
}
|
|
|
|
lv = find_lv(vg, id->lv_name);
|
|
|
|
if (lv && id->uuid && strcmp(id->uuid, (char *)&lv->lvid))
|
|
lv = NULL;
|
|
if (lv && parms->lv_type && !(lv->status & parms->lv_type))
|
|
lv = NULL;
|
|
|
|
if (!lv) {
|
|
if (parms->lv_type == PVMOVE)
|
|
log_verbose("%s: No pvmove in progress - already finished or aborted.",
|
|
id->display_name);
|
|
else
|
|
log_verbose("Can't find LV in %s for %s. Already finished or removed.",
|
|
vg->name, id->display_name);
|
|
goto out;
|
|
}
|
|
|
|
if (!lv_is_active_locally(lv)) {
|
|
log_verbose("%s: Interrupted: No longer active.", id->display_name);
|
|
goto out;
|
|
}
|
|
|
|
if (parms->poll_fns->poll_progress(cmd, lv, id->display_name, parms) == PROGRESS_CHECK_FAILED) {
|
|
unlock_and_release_vg(cmd, vg, vg->name);
|
|
return_0;
|
|
}
|
|
|
|
out:
|
|
unlock_and_release_vg(cmd, vg, vg->name);
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int _lvmpolld_init_poll_vg(struct cmd_context *cmd, const char *vgname,
|
|
struct volume_group *vg, struct processing_handle *handle)
|
|
{
|
|
int r;
|
|
struct lv_list *lvl;
|
|
struct logical_volume *lv;
|
|
struct poll_id_list *idl;
|
|
struct poll_operation_id id;
|
|
lvmpolld_parms_t *lpdp = (lvmpolld_parms_t *) handle->custom_handle;
|
|
|
|
dm_list_iterate_items(lvl, &vg->lvs) {
|
|
lv = lvl->lv;
|
|
if (!(lv->status & lpdp->parms->lv_type))
|
|
continue;
|
|
|
|
id.display_name = lpdp->parms->poll_fns->get_copy_name_from_lv(lv);
|
|
if (!id.display_name && !lpdp->parms->aborting)
|
|
continue;
|
|
|
|
if (!lv->lvid.s) {
|
|
log_print_unless_silent("Missing LV uuid within: %s/%s", id.vg_name, id.lv_name);
|
|
continue;
|
|
}
|
|
|
|
id.vg_name = lv->vg->name;
|
|
id.lv_name = lv->name;
|
|
id.uuid = lv->lvid.s;
|
|
|
|
r = lvmpolld_poll_init(cmd, &id, lpdp->parms);
|
|
|
|
if (r && !lpdp->parms->background) {
|
|
if (!(idl = poll_id_list_create(cmd->mem, &id)))
|
|
return ECMD_FAILED;
|
|
|
|
dm_list_add(&lpdp->idls, &idl->list);
|
|
}
|
|
}
|
|
|
|
return ECMD_PROCESSED;
|
|
}
|
|
|
|
static void _lvmpolld_poll_for_all_vgs(struct cmd_context *cmd,
|
|
struct daemon_parms *parms,
|
|
struct processing_handle *handle)
|
|
{
|
|
int r;
|
|
struct dm_list *first;
|
|
struct poll_id_list *idl, *tlv;
|
|
unsigned finished;
|
|
lvmpolld_parms_t lpdp = {
|
|
.parms = parms
|
|
};
|
|
|
|
dm_list_init(&lpdp.idls);
|
|
|
|
handle->custom_handle = &lpdp;
|
|
|
|
process_each_vg(cmd, 0, NULL, 0, handle, _lvmpolld_init_poll_vg);
|
|
|
|
first = dm_list_first(&lpdp.idls);
|
|
|
|
while (!dm_list_empty(&lpdp.idls)) {
|
|
dm_list_iterate_items_safe(idl, tlv, &lpdp.idls) {
|
|
r = lvmpolld_request_info(idl->id, lpdp.parms,
|
|
&finished);
|
|
if (!r || finished)
|
|
dm_list_del(&idl->list);
|
|
else if (!parms->aborting)
|
|
report_progress(cmd, idl->id, lpdp.parms);
|
|
}
|
|
|
|
if (lpdp.parms->interval)
|
|
dev_close_all();
|
|
|
|
_nanosleep(lpdp.parms->interval, 0);
|
|
}
|
|
|
|
if (first)
|
|
dm_pool_free(cmd->mem, dm_list_item(first, struct poll_id_list));
|
|
}
|
|
|
|
static int _lvmpoll_daemon(struct cmd_context *cmd, struct poll_operation_id *id,
|
|
struct daemon_parms *parms)
|
|
{
|
|
int r;
|
|
struct processing_handle *handle = NULL;
|
|
unsigned finished = 0;
|
|
|
|
if (parms->aborting)
|
|
parms->interval = 0;
|
|
|
|
if (id) {
|
|
r = lvmpolld_poll_init(cmd, id, parms);
|
|
if (r && !parms->background) {
|
|
while (1) {
|
|
if (!(r = lvmpolld_request_info(id, parms, &finished)) ||
|
|
finished ||
|
|
(!parms->aborting && !(r = report_progress(cmd, id, parms))))
|
|
break;
|
|
|
|
if (parms->interval)
|
|
dev_close_all();
|
|
|
|
_nanosleep(parms->interval, 0);
|
|
}
|
|
}
|
|
|
|
return r ? ECMD_PROCESSED : ECMD_FAILED;
|
|
} else {
|
|
/* process all in-flight operations */
|
|
if (!(handle = init_processing_handle(cmd))) {
|
|
log_error("Failed to initialize processing handle.");
|
|
return ECMD_FAILED;
|
|
} else {
|
|
_lvmpolld_poll_for_all_vgs(cmd, parms, handle);
|
|
destroy_processing_handle(cmd, handle);
|
|
return ECMD_PROCESSED;
|
|
}
|
|
}
|
|
}
|
|
#else
|
|
# define _lvmpoll_daemon(cmd, id, parms) (ECMD_FAILED)
|
|
#endif /* LVMPOLLD_SUPPORT */
|
|
|
|
/*
|
|
* Only allow *one* return from poll_daemon() (the parent).
|
|
* If there is a child it must exit (ignoring the memory leak messages).
|
|
* - 'background' is advisory so a child polldaemon may not be used even
|
|
* if it was requested.
|
|
*/
|
|
static int _poll_daemon(struct cmd_context *cmd, struct poll_operation_id *id,
|
|
struct daemon_parms *parms)
|
|
{
|
|
struct processing_handle *handle = NULL;
|
|
int daemon_mode = 0;
|
|
int ret = ECMD_PROCESSED;
|
|
|
|
if (parms->background) {
|
|
daemon_mode = become_daemon(cmd, 0);
|
|
if (daemon_mode == 0)
|
|
return ECMD_PROCESSED; /* Parent */
|
|
else if (daemon_mode == 1)
|
|
parms->progress_display = 0; /* Child */
|
|
/* FIXME Use wait_event (i.e. interval = 0) and */
|
|
/* fork one daemon per copy? */
|
|
}
|
|
|
|
/*
|
|
* Process one specific task or all incomplete tasks?
|
|
*/
|
|
if (id) {
|
|
if (!wait_for_single_lv(cmd, id, parms)) {
|
|
stack;
|
|
ret = ECMD_FAILED;
|
|
}
|
|
} else {
|
|
if (!parms->interval)
|
|
parms->interval = find_config_tree_int(cmd, activation_polling_interval_CFG, NULL);
|
|
if (!(handle = init_processing_handle(cmd))) {
|
|
log_error("Failed to initialize processing handle.");
|
|
ret = ECMD_FAILED;
|
|
} else {
|
|
handle->custom_handle = parms;
|
|
_poll_for_all_vgs(cmd, handle);
|
|
}
|
|
}
|
|
|
|
if (parms->background && daemon_mode == 1) {
|
|
destroy_processing_handle(cmd, handle);
|
|
/*
|
|
* child was successfully forked:
|
|
* background polldaemon must not return to the caller
|
|
* because it will redundantly continue performing the
|
|
* caller's task (that the parent already performed)
|
|
*/
|
|
/* FIXME Attempt proper cleanup */
|
|
_exit(lvm_return_code(ret));
|
|
}
|
|
|
|
destroy_processing_handle(cmd, handle);
|
|
return ret;
|
|
}
|
|
|
|
static int _daemon_parms_init(struct cmd_context *cmd, struct daemon_parms *parms,
|
|
unsigned background, struct poll_functions *poll_fns,
|
|
const char *progress_title, uint64_t lv_type)
|
|
{
|
|
sign_t interval_sign;
|
|
|
|
parms->aborting = arg_is_set(cmd, abort_ARG);
|
|
parms->background = background;
|
|
interval_sign = arg_sign_value(cmd, interval_ARG, SIGN_NONE);
|
|
if (interval_sign == SIGN_MINUS) {
|
|
log_error("Argument to --interval cannot be negative.");
|
|
return 0;
|
|
}
|
|
parms->interval = arg_uint_value(cmd, interval_ARG,
|
|
find_config_tree_int(cmd, activation_polling_interval_CFG, NULL));
|
|
parms->wait_before_testing = (interval_sign == SIGN_PLUS);
|
|
parms->progress_title = progress_title;
|
|
parms->lv_type = lv_type;
|
|
parms->poll_fns = poll_fns;
|
|
|
|
if (parms->interval && !parms->aborting)
|
|
log_verbose("Checking progress %s waiting every %u seconds.",
|
|
(parms->wait_before_testing ? "after" : "before"),
|
|
parms->interval);
|
|
|
|
parms->progress_display = parms->interval ? 1 : 0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
int poll_daemon(struct cmd_context *cmd, unsigned background,
|
|
uint64_t lv_type, struct poll_functions *poll_fns,
|
|
const char *progress_title, struct poll_operation_id *id)
|
|
{
|
|
struct daemon_parms parms;
|
|
|
|
if (!_daemon_parms_init(cmd, &parms, background, poll_fns, progress_title, lv_type))
|
|
return_EINVALID_CMD_LINE;
|
|
|
|
if (lvmpolld_use())
|
|
return _lvmpoll_daemon(cmd, id, &parms);
|
|
else {
|
|
/* classical polling allows only PMVOVE or 0 values */
|
|
parms.lv_type &= PVMOVE;
|
|
return _poll_daemon(cmd, id, &parms);
|
|
}
|
|
}
|