/* * Copyright (C) 2001-2004 Sistina Software, Inc. All rights reserved. * Copyright (C) 2004-2013 Red Hat, Inc. All rights reserved. * * This file is part of LVM2. * * This copyrighted material is made available to anyone wishing to use, * modify, copy, or redistribute it subject to the terms and conditions * of the GNU Lesser General Public License v.2.1. * * You should have received a copy of the GNU Lesser General Public License * along with this program; if not, write to the Free Software Foundation, * Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA */ #include "lib.h" #include "metadata.h" #include "display.h" #include "activate.h" #include "toolcontext.h" #include "segtype.h" #include "str_list.h" #include "lvmlockd.h" #include #include static struct utsname _utsname; static int _utsinit = 0; static char *_format_pvsegs(struct dm_pool *mem, const struct lv_segment *seg, int range_format, int metadata_areas_only) { unsigned int s; const char *name = NULL; uint32_t extent = 0; char extent_str[32]; if (!dm_pool_begin_object(mem, 256)) { log_error("dm_pool_begin_object failed"); return NULL; } if (metadata_areas_only && (!seg_is_raid(seg) || lv_is_raid_metadata(seg->lv) || lv_is_raid_image(seg->lv))) goto out; for (s = 0; s < seg->area_count; s++) { switch (metadata_areas_only ? seg_metatype(seg, s) : seg_type(seg, s)) { case AREA_LV: name = metadata_areas_only ? seg_metalv(seg, s)->name : seg_lv(seg, s)->name; extent = metadata_areas_only ? seg_le(seg, s) : 0; break; case AREA_PV: /* Raid metadata never uses PVs directly */ if (metadata_areas_only) continue; name = dev_name(seg_dev(seg, s)); extent = seg_pe(seg, s); break; case AREA_UNASSIGNED: name = "unassigned"; extent = 0; break; default: log_error(INTERNAL_ERROR "Unknown area segtype."); return NULL; } if (!dm_pool_grow_object(mem, name, strlen(name))) { log_error("dm_pool_grow_object failed"); return NULL; } if (dm_snprintf(extent_str, sizeof(extent_str), "%s%" PRIu32 "%s", range_format ? ":" : "(", extent, range_format ? "-" : ")") < 0) { log_error("Extent number dm_snprintf failed"); return NULL; } if (!dm_pool_grow_object(mem, extent_str, strlen(extent_str))) { log_error("dm_pool_grow_object failed"); return NULL; } if (range_format) { if (dm_snprintf(extent_str, sizeof(extent_str), FMTu32, metadata_areas_only ? extent + seg_metalv(seg, s)->le_count - 1 : extent + seg->area_len - 1) < 0) { log_error("Extent number dm_snprintf failed"); return NULL; } if (!dm_pool_grow_object(mem, extent_str, strlen(extent_str))) { log_error("dm_pool_grow_object failed"); return NULL; } } if ((s != seg->area_count - 1) && !dm_pool_grow_object(mem, range_format ? " " : ",", 1)) { log_error("dm_pool_grow_object failed"); return NULL; } } out: if (!dm_pool_grow_object(mem, "\0", 1)) { log_error("dm_pool_grow_object failed"); return NULL; } return dm_pool_end_object(mem); } char *lvseg_devices(struct dm_pool *mem, const struct lv_segment *seg) { return _format_pvsegs(mem, seg, 0, 0); } char *lvseg_metadata_devices(struct dm_pool *mem, const struct lv_segment *seg) { return _format_pvsegs(mem, seg, 0, 1); } char *lvseg_seg_pe_ranges(struct dm_pool *mem, const struct lv_segment *seg) { return _format_pvsegs(mem, seg, 1, 0); } char *lvseg_seg_metadata_le_ranges(struct dm_pool *mem, const struct lv_segment *seg) { return _format_pvsegs(mem, seg, 1, 1); } char *lvseg_tags_dup(const struct lv_segment *seg) { return tags_format_and_copy(seg->lv->vg->vgmem, &seg->tags); } char *lvseg_segtype_dup(struct dm_pool *mem, const struct lv_segment *seg) { return dm_pool_strdup(mem, lvseg_name(seg)); } char *lvseg_discards_dup(struct dm_pool *mem, const struct lv_segment *seg) { return dm_pool_strdup(mem, get_pool_discards_name(seg->discards)); } char *lvseg_cachemode_dup(struct dm_pool *mem, const struct lv_segment *seg) { const char *name = get_cache_mode_name(seg); if (!name) return_NULL; return dm_pool_strdup(mem, name); } #ifdef DMEVENTD # include "libdevmapper-event.h" #endif char *lvseg_monitor_dup(struct dm_pool *mem, const struct lv_segment *seg) { const char *s = ""; #ifdef DMEVENTD struct lvinfo info; int pending = 0, monitored; struct lv_segment *segm = (struct lv_segment *) seg; if (lv_is_cow(seg->lv) && !lv_is_merging_cow(seg->lv)) segm = first_seg(seg->lv->snapshot->lv); // log_debug("Query LV:%s mon:%s segm:%s tgtm:%p segmon:%d statusm:%d", seg->lv->name, segm->lv->name, segm->segtype->name, segm->segtype->ops->target_monitored, seg_monitored(segm), (int)(segm->status & PVMOVE)); if ((dmeventd_monitor_mode() != 1) || !segm->segtype->ops || !segm->segtype->ops->target_monitored) /* Nothing to do, monitoring not supported */; else if (lv_is_cow_covering_origin(seg->lv)) /* Nothing to do, snapshot already covers origin */; else if (!seg_monitored(segm) || (segm->status & PVMOVE)) s = "not monitored"; else if (lv_info(seg->lv->vg->cmd, seg->lv, 1, &info, 0, 0) && info.exists) { monitored = segm->segtype->ops->target_monitored(segm, &pending); if (pending) s = "pending"; else s = (monitored) ? "monitored" : "not monitored"; } // else log_debug("Not active"); #endif return dm_pool_strdup(mem, s); } uint64_t lvseg_chunksize(const struct lv_segment *seg) { uint64_t size; if (lv_is_cow(seg->lv)) size = (uint64_t) find_snapshot(seg->lv)->chunk_size; else if (seg_is_pool(seg)) size = (uint64_t) seg->chunk_size; else if (seg_is_cache(seg)) return lvseg_chunksize(first_seg(seg->pool_lv)); else size = UINT64_C(0); return size; } const char *lvseg_name(const struct lv_segment *seg) { /* Support even segtypes without 'ops' */ if (seg->segtype->ops && seg->segtype->ops->name) return seg->segtype->ops->name(seg); return seg->segtype->name; } uint64_t lvseg_start(const struct lv_segment *seg) { return (uint64_t) seg->le * seg->lv->vg->extent_size; } uint64_t lvseg_size(const struct lv_segment *seg) { return (uint64_t) seg->len * seg->lv->vg->extent_size; } uint32_t lv_kernel_read_ahead(const struct logical_volume *lv) { struct lvinfo info; if (!lv_info(lv->vg->cmd, lv, 0, &info, 0, 1) || !info.exists) return UINT32_MAX; return info.read_ahead; } static char *_do_lv_origin_dup(struct dm_pool *mem, const struct logical_volume *lv, int uuid) { struct logical_volume *origin; if (lv_is_cow(lv)) origin = origin_from_cow(lv); else if (lv_is_cache(lv) && first_seg(lv)->origin) origin = first_seg(lv)->origin; else if (lv_is_thin_volume(lv) && first_seg(lv)->origin) origin = first_seg(lv)->origin; else if (lv_is_thin_volume(lv) && first_seg(lv)->external_lv) origin = first_seg(lv)->external_lv; else return NULL; if (uuid) return lv_uuid_dup(mem, origin); else return lv_name_dup(mem, origin); } char *lv_origin_dup(struct dm_pool *mem, const struct logical_volume *lv) { return _do_lv_origin_dup(mem, lv, 0); } char *lv_origin_uuid_dup(struct dm_pool *mem, const struct logical_volume *lv) { return _do_lv_origin_dup(mem, lv, 1); } char *lv_name_dup(struct dm_pool *mem, const struct logical_volume *lv) { return dm_pool_strdup(mem, lv->name); } char *lv_fullname_dup(struct dm_pool *mem, const struct logical_volume *lv) { char lvfullname[NAME_LEN * 2 + 2]; if (dm_snprintf(lvfullname, sizeof(lvfullname), "%s/%s", lv->vg->name, lv->name) < 0) { log_error("lvfullname snprintf failed"); return NULL; } return dm_pool_strdup(mem, lvfullname); } struct logical_volume *lv_parent(const struct logical_volume *lv) { struct logical_volume *parent_lv = NULL; if (lv_is_visible(lv)) ; else if (lv_is_mirror_image(lv) || lv_is_mirror_log(lv)) parent_lv = get_only_segment_using_this_lv(lv)->lv; else if (lv_is_raid_image(lv) || lv_is_raid_metadata(lv)) parent_lv = get_only_segment_using_this_lv(lv)->lv; else if (lv_is_cache_pool_data(lv) || lv_is_cache_pool_metadata(lv)) parent_lv = get_only_segment_using_this_lv(lv)->lv; else if (lv_is_thin_pool_data(lv) || lv_is_thin_pool_metadata(lv)) parent_lv = get_only_segment_using_this_lv(lv)->lv; return parent_lv; } char *lv_parent_dup(struct dm_pool *mem, const struct logical_volume *lv) { struct logical_volume *parent_lv = lv_parent(lv); return dm_pool_strdup(mem, parent_lv ? parent_lv->name : ""); } char *lv_modules_dup(struct dm_pool *mem, const struct logical_volume *lv) { struct dm_list *modules; if (!(modules = str_list_create(mem))) { log_error("modules str_list allocation failed"); return NULL; } if (!list_lv_modules(mem, lv, modules)) return_NULL; return tags_format_and_copy(mem, modules); } static char *_do_lv_mirror_log_dup(struct dm_pool *mem, const struct logical_volume *lv, int uuid) { struct lv_segment *seg; dm_list_iterate_items(seg, &lv->segments) { if (seg_is_mirrored(seg) && seg->log_lv) { if (uuid) return lv_uuid_dup(mem, seg->log_lv); else return lv_name_dup(mem, seg->log_lv); } } return NULL; } char *lv_mirror_log_dup(struct dm_pool *mem, const struct logical_volume *lv) { return _do_lv_mirror_log_dup(mem, lv, 0); } char *lv_mirror_log_uuid_dup(struct dm_pool *mem, const struct logical_volume *lv) { return _do_lv_mirror_log_dup(mem, lv, 1); } static char *_do_lv_pool_lv_dup(struct dm_pool *mem, const struct logical_volume *lv, int uuid) { struct lv_segment *seg; dm_list_iterate_items(seg, &lv->segments) { if (seg->pool_lv && (seg_is_thin_volume(seg) || seg_is_cache(seg))) { if (uuid) return lv_uuid_dup(mem, seg->pool_lv); else return lv_name_dup(mem, seg->pool_lv); } } return NULL; } char *lv_pool_lv_dup(struct dm_pool *mem, const struct logical_volume *lv) { return _do_lv_pool_lv_dup(mem, lv, 0); } char *lv_pool_lv_uuid_dup(struct dm_pool *mem, const struct logical_volume *lv) { return _do_lv_pool_lv_dup(mem, lv, 1); } static char *_do_lv_data_lv_dup(struct dm_pool *mem, const struct logical_volume *lv, int uuid) { struct lv_segment *seg = (lv_is_thin_pool(lv) || lv_is_cache_pool(lv)) ? first_seg(lv) : NULL; if (seg) { if (uuid) return lv_uuid_dup(mem, seg_lv(seg, 0)); else return lv_name_dup(mem, seg_lv(seg, 0)); } return NULL; } char *lv_data_lv_dup(struct dm_pool *mem, const struct logical_volume *lv) { return _do_lv_data_lv_dup(mem, lv, 0); } char *lv_data_lv_uuid_dup(struct dm_pool *mem, const struct logical_volume *lv) { return _do_lv_data_lv_dup(mem, lv, 1); } static char *_do_lv_metadata_lv_dup(struct dm_pool *mem, const struct logical_volume *lv, int uuid) { struct lv_segment *seg = (lv_is_thin_pool(lv) || lv_is_cache_pool(lv)) ? first_seg(lv) : NULL; if (seg) { if (uuid) return lv_uuid_dup(mem, seg->metadata_lv); else return lv_name_dup(mem, seg->metadata_lv); } return NULL; } char *lv_metadata_lv_dup(struct dm_pool *mem, const struct logical_volume *lv) { return _do_lv_metadata_lv_dup(mem, lv, 0); } char *lv_metadata_lv_uuid_dup(struct dm_pool *mem, const struct logical_volume *lv) { return _do_lv_metadata_lv_dup(mem, lv, 1); } const char *lv_layer(const struct logical_volume *lv) { if (lv_is_thin_pool(lv)) return "tpool"; else if (lv_is_origin(lv) || lv_is_external_origin(lv)) return "real"; return NULL; } int lv_kernel_minor(const struct logical_volume *lv) { struct lvinfo info; if (lv_info(lv->vg->cmd, lv, 0, &info, 0, 0) && info.exists) return info.minor; return -1; } int lv_kernel_major(const struct logical_volume *lv) { struct lvinfo info; if (lv_info(lv->vg->cmd, lv, 0, &info, 0, 0) && info.exists) return info.major; return -1; } static char *_do_lv_convert_lv_dup(struct dm_pool *mem, const struct logical_volume *lv, int uuid) { struct lv_segment *seg; if (lv_is_converting(lv) || lv_is_mirrored(lv)) { seg = first_seg(lv); /* Temporary mirror is always area_num == 0 */ if (seg_type(seg, 0) == AREA_LV && is_temporary_mirror_layer(seg_lv(seg, 0))) { if (uuid) return lv_uuid_dup(mem, seg_lv(seg, 0)); else return lv_name_dup(mem, seg_lv(seg, 0)); } } return NULL; } char *lv_convert_lv_dup(struct dm_pool *mem, const struct logical_volume *lv) { return _do_lv_convert_lv_dup(mem, lv, 0); } char *lv_convert_lv_uuid_dup(struct dm_pool *mem, const struct logical_volume *lv) { return _do_lv_convert_lv_dup(mem, lv, 1); } static char *_do_lv_move_pv_dup(struct dm_pool *mem, const struct logical_volume *lv, int uuid) { struct logical_volume *mimage0_lv; struct lv_segment *seg; struct pv_segment *pvseg; dm_list_iterate_items(seg, &lv->segments) { if (seg->status & PVMOVE) { if (seg_type(seg, 0) == AREA_LV) { /* atomic pvmove */ mimage0_lv = seg_lv(seg, 0); if (!lv_is_mirror_image(mimage0_lv)) { log_error(INTERNAL_ERROR "Bad pvmove structure"); return NULL; } pvseg = seg_pvseg(first_seg(mimage0_lv), 0); } else /* Segment pvmove */ pvseg = seg_pvseg(seg, 0); if (uuid) return pv_uuid_dup(mem, pvseg->pv); else return pv_name_dup(mem, pvseg->pv); } } return NULL; } char *lv_move_pv_dup(struct dm_pool *mem, const struct logical_volume *lv) { return _do_lv_move_pv_dup(mem, lv, 0); } char *lv_move_pv_uuid_dup(struct dm_pool *mem, const struct logical_volume *lv) { return _do_lv_move_pv_dup(mem, lv, 1); } uint64_t lv_origin_size(const struct logical_volume *lv) { struct lv_segment *seg; if (lv_is_cow(lv)) return (uint64_t) find_snapshot(lv)->len * lv->vg->extent_size; if (lv_is_thin_volume(lv) && (seg = first_seg(lv)) && seg->external_lv) return seg->external_lv->size; if (lv_is_origin(lv)) return lv->size; return 0; } uint64_t lv_metadata_size(const struct logical_volume *lv) { struct lv_segment *seg = (lv_is_thin_pool(lv) || lv_is_cache_pool(lv)) ? first_seg(lv) : NULL; return seg ? seg->metadata_lv->size : 0; } char *lv_path_dup(struct dm_pool *mem, const struct logical_volume *lv) { char *repstr; size_t len; /* Only for visible devices that get a link from /dev/vg */ if (!*lv->vg->name || !lv_is_visible(lv) || lv_is_thin_pool(lv)) return dm_pool_strdup(mem, ""); len = strlen(lv->vg->cmd->dev_dir) + strlen(lv->vg->name) + strlen(lv->name) + 2; if (!(repstr = dm_pool_zalloc(mem, len))) { log_error("dm_pool_alloc failed"); return NULL; } if (dm_snprintf(repstr, len, "%s%s/%s", lv->vg->cmd->dev_dir, lv->vg->name, lv->name) < 0) { log_error("lvpath snprintf failed"); return NULL; } return repstr; } char *lv_dmpath_dup(struct dm_pool *mem, const struct logical_volume *lv) { char *name; char *repstr; size_t len; if (!*lv->vg->name) return dm_pool_strdup(mem, ""); if (!(name = dm_build_dm_name(mem, lv->vg->name, lv->name, NULL))) { log_error("dm_build_dm_name failed"); return NULL; } len = strlen(dm_dir()) + strlen(name) + 2; if (!(repstr = dm_pool_zalloc(mem, len))) { log_error("dm_pool_alloc failed"); return NULL; } if (dm_snprintf(repstr, len, "%s/%s", dm_dir(), name) < 0) { log_error("lv_dmpath snprintf failed"); return NULL; } return repstr; } char *lv_uuid_dup(struct dm_pool *mem, const struct logical_volume *lv) { return id_format_and_copy(mem ? mem : lv->vg->vgmem, &lv->lvid.id[1]); } char *lv_tags_dup(const struct logical_volume *lv) { return tags_format_and_copy(lv->vg->vgmem, &lv->tags); } uint64_t lv_size(const struct logical_volume *lv) { return lv->size; } int lv_mirror_image_in_sync(const struct logical_volume *lv) { dm_percent_t percent; struct lv_segment *seg = first_seg(lv); struct lv_segment *mirror_seg; if (!(lv->status & MIRROR_IMAGE) || !seg || !(mirror_seg = find_mirror_seg(seg))) { log_error(INTERNAL_ERROR "Cannot find mirror segment."); return 0; } if (!lv_mirror_percent(lv->vg->cmd, mirror_seg->lv, 0, &percent, NULL)) return_0; return (percent == DM_PERCENT_100) ? 1 : 0; } int lv_raid_image_in_sync(const struct logical_volume *lv) { unsigned s; dm_percent_t percent; char *raid_health; struct lv_segment *seg, *raid_seg = NULL; /* * If the LV is not active locally, * it doesn't make sense to check status */ if (!lv_is_active_locally(lv)) return 0; /* Assume not in-sync */ if (!lv_is_raid_image(lv)) { log_error(INTERNAL_ERROR "%s is not a RAID image", lv->name); return 0; } if ((seg = first_seg(lv))) raid_seg = get_only_segment_using_this_lv(seg->lv); if (!raid_seg) { log_error("Failed to find RAID segment for %s", lv->name); return 0; } if (!seg_is_raid(raid_seg)) { log_error("%s on %s is not a RAID segment", raid_seg->lv->name, lv->name); return 0; } if (!lv_raid_percent(raid_seg->lv, &percent)) return_0; if (percent == DM_PERCENT_100) return 1; /* Find out which sub-LV this is. */ for (s = 0; s < raid_seg->area_count; s++) if (seg_lv(raid_seg, s) == lv) break; if (s == raid_seg->area_count) { log_error(INTERNAL_ERROR "sub-LV %s was not found in raid segment", lv->name); return 0; } if (!lv_raid_dev_health(raid_seg->lv, &raid_health)) return_0; if (raid_health[s] == 'A') return 1; return 0; } /* * _lv_raid_healthy * @lv: A RAID_IMAGE, RAID_META, or RAID logical volume. * * Returns: 1 if healthy, 0 if device is not health */ int lv_raid_healthy(const struct logical_volume *lv) { unsigned s; char *raid_health; struct lv_segment *seg, *raid_seg = NULL; /* * If the LV is not active locally, * it doesn't make sense to check status */ if (!lv_is_active_locally(lv)) return 1; /* assume healthy */ if (!lv_is_raid_type(lv)) { log_error(INTERNAL_ERROR "%s is not of RAID type", lv->name); return 0; } if (lv_is_raid(lv)) raid_seg = first_seg(lv); else if ((seg = first_seg(lv))) raid_seg = get_only_segment_using_this_lv(seg->lv); if (!raid_seg) { log_error("Failed to find RAID segment for %s", lv->name); return 0; } if (!seg_is_raid(raid_seg)) { log_error("%s on %s is not a RAID segment", raid_seg->lv->name, lv->name); return 0; } if (!lv_raid_dev_health(raid_seg->lv, &raid_health)) return_0; if (lv_is_raid(lv)) { if (strchr(raid_health, 'D')) return 0; else return 1; } /* Find out which sub-LV this is. */ for (s = 0; s < raid_seg->area_count; s++) if ((lv_is_raid_image(lv) && (seg_lv(raid_seg, s) == lv)) || (lv_is_raid_metadata(lv) && (seg_metalv(raid_seg,s) == lv))) break; if (s == raid_seg->area_count) { log_error(INTERNAL_ERROR "sub-LV %s was not found in raid segment", lv->name); return 0; } if (raid_health[s] == 'D') return 0; return 1; } char *lv_attr_dup_with_info_and_seg_status(struct dm_pool *mem, const struct lv_with_info_and_seg_status *lvdm) { dm_percent_t snap_percent; const struct logical_volume *lv = lvdm->lv; struct lv_segment *seg; char *repstr; if (!(repstr = dm_pool_zalloc(mem, 11))) { log_error("dm_pool_alloc failed"); return 0; } /* Blank if this is a "free space" LV. */ if (!*lv->name) goto out; if (lv_is_pvmove(lv)) repstr[0] = 'p'; else if (lv->status & CONVERTING) repstr[0] = 'c'; /* Origin takes precedence over mirror and thin volume */ else if (lv_is_origin(lv) || lv_is_external_origin(lv)) repstr[0] = (lv_is_merging_origin(lv)) ? 'O' : 'o'; else if (lv_is_pool_metadata(lv) || lv_is_pool_metadata_spare(lv) || lv_is_raid_metadata(lv)) repstr[0] = 'e'; else if (lv_is_cache_type(lv)) repstr[0] = 'C'; else if (lv_is_raid(lv)) repstr[0] = (lv->status & LV_NOTSYNCED) ? 'R' : 'r'; else if (lv_is_mirror(lv)) repstr[0] = (lv->status & LV_NOTSYNCED) ? 'M' : 'm'; else if (lv_is_thin_volume(lv)) repstr[0] = lv_is_merging_origin(lv) ? 'O' : (lv_is_merging_thin_snapshot(lv) ? 'S' : 'V'); else if (lv_is_virtual(lv)) repstr[0] = 'v'; else if (lv_is_thin_pool(lv)) repstr[0] = 't'; else if (lv_is_thin_pool_data(lv)) repstr[0] = 'T'; else if (lv_is_mirror_image(lv)) repstr[0] = (lv_mirror_image_in_sync(lv)) ? 'i' : 'I'; else if (lv_is_raid_image(lv)) /* * Visible RAID_IMAGES are sub-LVs that have been exposed for * top-level use by being split from the RAID array with * '--splitmirrors 1 --trackchanges'. They always report 'I'. */ repstr[0] = (!lv_is_visible(lv) && lv_raid_image_in_sync(lv)) ? 'i' : 'I'; else if (lv_is_mirror_log(lv)) repstr[0] = 'l'; else if (lv_is_cow(lv)) repstr[0] = (lv_is_merging_cow(lv)) ? 'S' : 's'; else if (lv_is_cache_origin(lv)) repstr[0] = 'o'; else repstr[0] = '-'; if (lv_is_pvmove(lv)) repstr[1] = '-'; else if (lv->status & LVM_WRITE) repstr[1] = 'w'; else if (lv->status & LVM_READ) repstr[1] = 'r'; else repstr[1] = '-'; repstr[2] = alloc_policy_char(lv->alloc); if (lv_is_locked(lv)) repstr[2] = toupper(repstr[2]); repstr[3] = (lv->status & FIXED_MINOR) ? 'm' : '-'; if (!activation() || !lvdm->info_ok) { repstr[4] = 'X'; /* Unknown */ repstr[5] = 'X'; /* Unknown */ } else if (lvdm->info.exists) { if (lvdm->info.suspended) repstr[4] = 's'; /* Suspended */ else if (lvdm->info.live_table) repstr[4] = 'a'; /* Active */ else if (lvdm->info.inactive_table) repstr[4] = 'i'; /* Inactive with table */ else repstr[4] = 'd'; /* Inactive without table */ /* Snapshot dropped? */ if (lvdm->info.live_table && lv_is_cow(lv)) { if (!lv_snapshot_percent(lv, &snap_percent) || snap_percent == DM_PERCENT_INVALID) { if (lvdm->info.suspended) repstr[4] = 'S'; /* Susp Inv snapshot */ else repstr[4] = 'I'; /* Invalid snapshot */ } else if (snap_percent == LVM_PERCENT_MERGE_FAILED) { if (lvdm->info.suspended) repstr[4] = 'M'; /* Susp snapshot merge failed */ else repstr[4] = 'm'; /* snapshot merge failed */ } } /* * 'R' indicates read-only activation of a device that * does not have metadata flagging it as read-only. */ if (repstr[1] != 'r' && lvdm->info.read_only) repstr[1] = 'R'; repstr[5] = (lvdm->info.open_count) ? 'o' : '-'; } else { repstr[4] = '-'; repstr[5] = '-'; } if (lv_is_thin_pool(lv) || lv_is_thin_volume(lv)) repstr[6] = 't'; else if (lv_is_cache_pool(lv) || lv_is_cache(lv) || lv_is_cache_origin(lv)) repstr[6] = 'C'; else if (lv_is_raid_type(lv)) repstr[6] = 'r'; else if (lv_is_mirror_type(lv) || lv_is_pvmove(lv)) repstr[6] = 'm'; else if (lv_is_cow(lv) || lv_is_origin(lv)) repstr[6] = 's'; else if (lv_has_unknown_segments(lv)) repstr[6] = 'u'; else if (lv_is_virtual(lv)) repstr[6] = 'v'; else repstr[6] = '-'; if (((lv_is_thin_volume(lv) && (seg = first_seg(lv)) && seg->pool_lv && (seg = first_seg(seg->pool_lv))) || (lv_is_thin_pool(lv) && (seg = first_seg(lv)))) && seg->zero_new_blocks) repstr[7] = 'z'; else repstr[7] = '-'; repstr[8] = '-'; if (lv->status & PARTIAL_LV) repstr[8] = 'p'; else if (lv_is_raid_type(lv)) { uint64_t n; if (!activation()) repstr[8] = 'X'; /* Unknown */ else if (!lv_raid_healthy(lv)) repstr[8] = 'r'; /* RAID needs 'r'efresh */ else if (lv_is_raid(lv)) { if (lv_raid_mismatch_count(lv, &n) && n) repstr[8] = 'm'; /* RAID has 'm'ismatches */ } else if (lv->status & LV_WRITEMOSTLY) repstr[8] = 'w'; /* sub-LV has 'w'ritemostly */ } else if (lv_is_thin_pool(lv) && (lvdm->seg_status.type != SEG_STATUS_NONE)) { if (lvdm->seg_status.type == SEG_STATUS_UNKNOWN) repstr[8] = 'X'; /* Unknown */ else if (lvdm->seg_status.thin_pool->fail) repstr[8] = 'F'; else if (lvdm->seg_status.thin_pool->out_of_data_space) repstr[8] = 'D'; else if (lvdm->seg_status.thin_pool->read_only) repstr[8] = 'M'; } if (lv->status & LV_ACTIVATION_SKIP) repstr[9] = 'k'; else repstr[9] = '-'; out: return repstr; } /* backward compatible internal API for lvm2api, TODO improve it */ char *lv_attr_dup(struct dm_pool *mem, const struct logical_volume *lv) { char *ret = NULL; struct lv_with_info_and_seg_status status = { .seg_status.type = SEG_STATUS_NONE, .lv = lv }; if (!(status.seg_status.mem = dm_pool_create("reporter_pool", 1024))) return_0; if (!(status.info_ok = lv_info_with_seg_status(lv->vg->cmd, lv, first_seg(lv), 1, &status, 1, 1))) goto_bad; ret = lv_attr_dup_with_info_and_seg_status(mem, &status); bad: dm_pool_destroy(status.seg_status.mem); return ret; } int lv_set_creation(struct logical_volume *lv, const char *hostname, uint64_t timestamp) { const char *hn; if (!hostname) { if (!_utsinit) { if (uname(&_utsname)) { log_error("uname failed: %s", strerror(errno)); memset(&_utsname, 0, sizeof(_utsname)); } _utsinit = 1; } hostname = _utsname.nodename; } if (!(hn = dm_hash_lookup(lv->vg->hostnames, hostname))) { if (!(hn = dm_pool_strdup(lv->vg->vgmem, hostname))) { log_error("Failed to duplicate hostname"); return 0; } if (!dm_hash_insert(lv->vg->hostnames, hostname, (void*)hn)) return_0; } lv->hostname = hn; lv->timestamp = timestamp ? : (uint64_t) time(NULL); return 1; } char *lv_time_dup(struct dm_pool *mem, const struct logical_volume *lv, int iso_mode) { char buffer[4096]; struct tm *local_tm; time_t ts = (time_t)lv->timestamp; const char *format = iso_mode ? DEFAULT_TIME_FORMAT : lv->vg->cmd->time_format; if (!ts || !(local_tm = localtime(&ts)) || !strftime(buffer, sizeof(buffer), format, local_tm)) buffer[0] = 0; return dm_pool_strdup(mem, buffer); } char *lv_host_dup(struct dm_pool *mem, const struct logical_volume *lv) { return dm_pool_strdup(mem, lv->hostname ? : ""); } static int _lv_is_exclusive(struct logical_volume *lv) { struct lv_segment *seg; /* Some seg types require exclusive activation */ /* FIXME Scan recursively */ dm_list_iterate_items(seg, &lv->segments) if (seg_only_exclusive(seg)) return 1; /* Origin has no seg type require exlusiveness */ return lv_is_origin(lv); } int lv_active_change(struct cmd_context *cmd, struct logical_volume *lv, enum activation_change activate, int needs_exclusive) { const char *ay_with_mode = NULL; if (activate == CHANGE_ASY) ay_with_mode = "sh"; if (activate == CHANGE_AEY) ay_with_mode = "ex"; if (is_change_activating(activate) && !lockd_lv(cmd, lv, ay_with_mode, LDLV_PERSISTENT)) { log_error("Failed to lock logical volume %s/%s", lv->vg->name, lv->name); return 0; } switch (activate) { case CHANGE_AN: deactivate: log_verbose("Deactivating logical volume \"%s\"", lv->name); if (!deactivate_lv(cmd, lv)) return_0; break; case CHANGE_ALN: if (vg_is_clustered(lv->vg) && (needs_exclusive || _lv_is_exclusive(lv))) { if (!lv_is_active_locally(lv)) { log_error("Cannot deactivate remotely exclusive device locally."); return 0; } /* Unlock whole exclusive activation */ goto deactivate; } log_verbose("Deactivating logical volume \"%s\" locally.", lv->name); if (!deactivate_lv_local(cmd, lv)) return_0; break; case CHANGE_ALY: case CHANGE_AAY: if (needs_exclusive || _lv_is_exclusive(lv)) { log_verbose("Activating logical volume \"%s\" exclusively locally.", lv->name); if (!activate_lv_excl_local(cmd, lv)) return_0; } else { log_verbose("Activating logical volume \"%s\" locally.", lv->name); if (!activate_lv_local(cmd, lv)) return_0; } break; case CHANGE_AEY: exclusive: log_verbose("Activating logical volume \"%s\" exclusively.", lv->name); if (!activate_lv_excl(cmd, lv)) return_0; break; case CHANGE_ASY: case CHANGE_AY: default: if (needs_exclusive || _lv_is_exclusive(lv)) goto exclusive; log_verbose("Activating logical volume \"%s\".", lv->name); if (!activate_lv(cmd, lv)) return_0; } if (!is_change_activating(activate) && !lockd_lv(cmd, lv, "un", LDLV_PERSISTENT)) log_error("Failed to unlock logical volume %s/%s", lv->vg->name, lv->name); return 1; } char *lv_active_dup(struct dm_pool *mem, const struct logical_volume *lv) { const char *s; if (!activation()) { s = "unknown"; goto out; } if (vg_is_clustered(lv->vg)) { //const struct logical_volume *lvo = lv; lv = lv_lock_holder(lv); //log_debug("Holder for %s => %s.", lvo->name, lv->name); } if (!lv_is_active(lv)) s = ""; /* not active */ else if (!vg_is_clustered(lv->vg)) s = "active"; else if (lv_is_active_exclusive(lv)) /* exclusive cluster activation */ s = lv_is_active_exclusive_locally(lv) ? "local exclusive" : "remote exclusive"; else /* locally active */ s = lv_is_active_but_not_locally(lv) ? "remotely" : "locally"; out: return dm_pool_strdup(mem, s); } char *lv_profile_dup(struct dm_pool *mem, const struct logical_volume *lv) { const char *profile_name = lv->profile ? lv->profile->name : ""; return dm_pool_strdup(mem, profile_name); } char *lv_lock_args_dup(struct dm_pool *mem, const struct logical_volume *lv) { const char *lock_args = lv->lock_args ? lv->lock_args : ""; return dm_pool_strdup(mem, lock_args); } /* For given LV find recursively the LV which holds lock for it */ const struct logical_volume *lv_lock_holder(const struct logical_volume *lv) { const struct seg_list *sl; if (lv_is_cow(lv)) return lv_lock_holder(origin_from_cow(lv)); if (lv_is_thin_pool(lv)) /* Find any active LV from the pool */ dm_list_iterate_items(sl, &lv->segs_using_this_lv) if (lv_is_active(sl->seg->lv)) { log_debug("Thin volume \"%s\" is active.", sl->seg->lv->name); return sl->seg->lv; } /* RAID changes visibility of splitted LVs but references them still as leg/meta */ if ((lv_is_raid_image(lv) || lv_is_raid_metadata(lv)) && lv_is_visible(lv)) return lv; /* For other types, by default look for the first user */ dm_list_iterate_items(sl, &lv->segs_using_this_lv) { /* FIXME: complete this exception list */ if (lv_is_thin_volume(lv) && lv_is_thin_volume(sl->seg->lv) && first_seg(lv)->pool_lv == sl->seg->pool_lv) continue; /* Skip thin snaphost */ if (lv_is_external_origin(lv) && lv_is_thin_volume(sl->seg->lv)) continue; /* Skip external origin */ if (lv_is_pending_delete(sl->seg->lv)) continue; /* Skip deleted LVs */ return lv_lock_holder(sl->seg->lv); } return lv; } struct profile *lv_config_profile(const struct logical_volume *lv) { return lv->profile ? : lv->vg->profile; }