2014-01-23 12:56:17 +04:00
/*
* Copyright ( C ) 2013 - 2014 Red Hat , Inc . All rights reserved .
*
* This file is part of LVM2 .
*
* This copyrighted material is made available to anyone wishing to use ,
* modify , copy , or redistribute it subject to the terms and conditions
* of the GNU Lesser General Public License v .2 .1 .
*
* You should have received a copy of the GNU Lesser General Public License
* along with this program ; if not , write to the Free Software Foundation ,
2016-01-21 13:49:46 +03:00
* Inc . , 51 Franklin Street , Fifth Floor , Boston , MA 02110 - 1301 USA
2014-01-23 12:56:17 +04:00
*/
/*
* This file holds common pool functions .
*/
2018-05-14 12:30:20 +03:00
# include "lib/misc/lib.h"
# include "lib/activate/activate.h"
# include "lib/locking/locking.h"
# include "lib/metadata/metadata.h"
# include "lib/metadata/segtype.h"
# include "lib/metadata/lv_alloc.h"
# include "lib/config/defaults.h"
# include "lib/device/dev-type.h"
# include "lib/display/display.h"
# include "lib/commands/toolcontext.h"
2016-03-01 17:20:09 +03:00
# include <stddef.h>
2014-01-23 12:56:17 +04:00
int attach_pool_metadata_lv ( struct lv_segment * pool_seg ,
struct logical_volume * metadata_lv )
{
2014-05-23 16:24:28 +04:00
if ( ! seg_is_pool ( pool_seg ) ) {
2014-01-23 12:56:17 +04:00
log_error ( INTERNAL_ERROR
" Unable to attach pool metadata LV to %s segtype. " ,
2014-10-20 20:40:39 +04:00
lvseg_name ( pool_seg ) ) ;
2014-01-23 12:56:17 +04:00
return 0 ;
}
pool_seg - > metadata_lv = metadata_lv ;
2014-01-28 22:25:07 +04:00
metadata_lv - > status | = seg_is_thin_pool ( pool_seg ) ?
THIN_POOL_METADATA : CACHE_POOL_METADATA ;
2014-01-23 12:56:17 +04:00
lv_set_hidden ( metadata_lv ) ;
return add_seg_to_segs_using_this_lv ( metadata_lv , pool_seg ) ;
}
2014-07-08 17:19:47 +04:00
int detach_pool_metadata_lv ( struct lv_segment * pool_seg , struct logical_volume * * metadata_lv )
{
struct logical_volume * lv = pool_seg - > metadata_lv ;
if ( ! lv | |
! lv_is_pool_metadata ( lv ) | |
! remove_seg_from_segs_using_this_lv ( lv , pool_seg ) ) {
log_error ( INTERNAL_ERROR " Logical volume %s is not valid pool. " ,
display_lvname ( pool_seg - > lv ) ) ;
return 0 ;
}
lv_set_visible ( lv ) ;
lv - > status & = ~ ( THIN_POOL_METADATA | CACHE_POOL_METADATA ) ;
* metadata_lv = lv ;
pool_seg - > metadata_lv = NULL ;
return 1 ;
}
2014-01-23 12:56:17 +04:00
int attach_pool_data_lv ( struct lv_segment * pool_seg ,
struct logical_volume * pool_data_lv )
{
2014-05-23 16:24:28 +04:00
if ( ! seg_is_pool ( pool_seg ) ) {
2014-01-23 12:56:17 +04:00
log_error ( INTERNAL_ERROR
2014-01-23 09:04:27 +04:00
" Unable to attach pool data LV to %s segtype. " ,
2014-10-20 20:40:39 +04:00
lvseg_name ( pool_seg ) ) ;
2014-01-23 12:56:17 +04:00
return 0 ;
}
if ( ! set_lv_segment_area_lv ( pool_seg , 0 , pool_data_lv ,
2014-01-28 22:25:07 +04:00
0 , seg_is_thin_pool ( pool_seg ) ?
THIN_POOL_DATA : CACHE_POOL_DATA ) )
2014-01-23 12:56:17 +04:00
return_0 ;
2014-01-28 22:25:07 +04:00
pool_seg - > lv - > status | = seg_is_thin_pool ( pool_seg ) ?
THIN_POOL : CACHE_POOL ;
2014-01-23 12:56:17 +04:00
lv_set_hidden ( pool_data_lv ) ;
return 1 ;
}
int attach_pool_lv ( struct lv_segment * seg ,
struct logical_volume * pool_lv ,
struct logical_volume * origin ,
2016-03-01 17:21:36 +03:00
struct generic_logical_volume * indirect_origin ,
2014-01-23 12:56:17 +04:00
struct logical_volume * merge_lv )
{
2016-03-01 17:21:36 +03:00
struct glv_list * glvl ;
2014-01-28 22:25:07 +04:00
if ( ! seg_is_thin_volume ( seg ) & & ! seg_is_cache ( seg ) ) {
2014-01-23 12:56:17 +04:00
log_error ( INTERNAL_ERROR " Unable to attach pool to %s/%s "
2014-01-28 22:25:07 +04:00
" that is not cache or thin volume. " ,
2014-01-23 12:56:17 +04:00
pool_lv - > vg - > name , seg - > lv - > name ) ;
return 0 ;
}
seg - > pool_lv = pool_lv ;
seg - > origin = origin ;
2014-01-28 22:25:07 +04:00
seg - > lv - > status | = seg_is_cache ( seg ) ? CACHE : THIN_VOLUME ;
2014-01-23 12:56:17 +04:00
2019-10-16 22:22:20 +03:00
if ( seg_is_cache ( seg ) ) {
lv_set_hidden ( pool_lv ) ; /* Used cache-pool/cachevol is hidden */
if ( lv_is_cache_vol ( pool_lv ) )
/*
* This flag is added to the segtype name so that old versions of lvm
* ( if they happen to be used with new metadata with a cache LV using a
* cachevol ) will report an error when they see the unknown
* cache + CACHE_USES_CACHEVOL segment type . Otherwise the old version
* would expect to find a cache pool and fail .
*/
seg - > lv - > status | = LV_CACHE_USES_CACHEVOL ;
}
2014-11-09 23:04:33 +03:00
2014-01-23 12:56:17 +04:00
if ( origin & & ! add_seg_to_segs_using_this_lv ( origin , seg ) )
return_0 ;
2016-03-01 17:21:36 +03:00
if ( indirect_origin ) {
if ( ! ( glvl = get_or_create_glvl ( seg - > lv - > vg - > vgmem , seg - > lv , NULL ) ) )
return_0 ;
seg - > indirect_origin = indirect_origin ;
if ( indirect_origin - > is_historical )
dm_list_add ( & indirect_origin - > historical - > indirect_glvs , & glvl - > list ) ;
else
dm_list_add ( & indirect_origin - > live - > indirect_glvs , & glvl - > list ) ;
}
2014-01-23 12:56:17 +04:00
if ( ! add_seg_to_segs_using_this_lv ( pool_lv , seg ) )
return_0 ;
if ( merge_lv ) {
if ( origin ! = merge_lv ) {
if ( ! add_seg_to_segs_using_this_lv ( merge_lv , seg ) )
return_0 ;
}
init_snapshot_merge ( seg , merge_lv ) ;
}
return 1 ;
}
2016-03-01 17:20:09 +03:00
static struct glv_list * _init_historical_glvl ( struct dm_pool * mem , struct lv_segment * seg )
{
struct glv_list * glvl ;
struct historical_logical_volume * hlv ;
if ( ! ( glvl = dm_pool_zalloc ( mem , sizeof ( struct glv_list ) ) ) )
goto_bad ;
if ( ! ( glvl - > glv = dm_pool_zalloc ( mem , sizeof ( struct generic_logical_volume ) ) ) )
goto_bad ;
if ( ! ( hlv = dm_pool_zalloc ( mem , sizeof ( struct historical_logical_volume ) ) ) )
goto_bad ;
hlv - > lvid = seg - > lv - > lvid ;
hlv - > name = seg - > lv - > name ;
hlv - > vg = seg - > lv - > vg ;
hlv - > timestamp = seg - > lv - > timestamp ;
toollib: do not process just created historical LV
When executing process_each_lv_in_vg, we process live LVs first and
after that, we process any historical LVs. In case we have just removed
an LV, which also means we have just made it "historical" and so it
appears as fresh item in vg->historical_lvs list, we have to skip it
when we get to processing historical LVs inside the same process_each_lv_in_vg
call.
The simplest approach here, without introducing another LV list, is to
simply mark such historical LVs as "fresh" directly in struct
historical_logical_volume when we have just removed the original LV
and created the historical LV for it. Then, we just need to check the
flag when processing historical LVs and skip it if it is "fresh".
When we read historical LVs out of metadata, they are marked as
"not fresh" and so they can be processed as usual.
This was mainly an issue in conjuction with -S|--select use:
# lvmconfig --type diff
metadata {
record_lvs_history=1
}
(In this example, a thin pool with lvol1 thin LV and lvol2 and lvol3 snapshots.)
# lvs -H vg -o name,pool_lv,full_ancestors,full_descendants
LV Pool FAncestors FDescendants
lvol1 pool lvol2,lvol3
lvol2 pool lvol1 lvol3
lvol3 pool lvol2,lvol1
pool
# lvremove -S 'name=lvol2'
Logical volume "lvol2" successfully removed.
Historical logical volume "lvol2" successfully removed.
...here, the historical LV lvol2 should not have been removed because
we have just removed its original non-historical lvol2 and the fresh
historical lvol2 must not be included in the same processing spree.
2022-10-12 15:41:58 +03:00
hlv - > fresh = 1 ;
2016-03-01 17:20:09 +03:00
dm_list_init ( & hlv - > indirect_glvs ) ;
glvl - > glv - > is_historical = 1 ;
glvl - > glv - > historical = hlv ;
return glvl ;
bad :
log_error ( " Initialization of historical LV representation for removed logical "
2021-03-09 15:13:42 +03:00
" volume %s failed. " , display_lvname ( seg - > lv ) ) ;
2016-03-01 17:20:09 +03:00
if ( glvl )
dm_pool_free ( mem , glvl ) ;
return NULL ;
}
static struct generic_logical_volume * _create_historical_glv ( struct lv_segment * seg_to_remove )
{
struct dm_pool * mem = seg_to_remove - > lv - > vg - > vgmem ;
struct generic_logical_volume * historical_glv , * origin_glv = NULL ;
struct glv_list * historical_glvl ;
int origin_glv_created = 0 ;
if ( ! ( historical_glvl = _init_historical_glvl ( mem , seg_to_remove ) ) )
goto_bad ;
historical_glv = historical_glvl - > glv ;
if ( seg_to_remove - > origin ) {
if ( ! ( origin_glv = get_or_create_glv ( mem , seg_to_remove - > origin , & origin_glv_created ) ) )
goto_bad ;
if ( ! add_glv_to_indirect_glvs ( mem , origin_glv , historical_glv ) )
goto_bad ;
} else if ( seg_to_remove - > indirect_origin ) {
origin_glv = seg_to_remove - > indirect_origin ;
if ( ! remove_glv_from_indirect_glvs ( origin_glv , seg_to_remove - > lv - > this_glv ) )
goto_bad ;
if ( ! add_glv_to_indirect_glvs ( mem , origin_glv , historical_glv ) )
goto_bad ;
}
dm_list_add ( & seg_to_remove - > lv - > vg - > historical_lvs , & historical_glvl - > list ) ;
return historical_glvl - > glv ;
bad :
log_error ( " Failed to create historical LV representation for removed logical "
2021-03-09 15:13:42 +03:00
" volume %s. " , display_lvname ( seg_to_remove - > lv ) ) ;
2016-03-01 17:20:09 +03:00
if ( origin_glv_created )
seg_to_remove - > origin - > this_glv = NULL ;
if ( historical_glvl )
dm_pool_free ( mem , historical_glvl ) ;
return NULL ;
}
static int _set_up_historical_lv ( struct lv_segment * seg_to_remove ,
struct generic_logical_volume * * previous_glv )
{
struct generic_logical_volume * glv = NULL ;
2016-03-01 17:25:28 +03:00
if ( seg_to_remove - > lv - > vg - > cmd - > record_historical_lvs ) {
if ( seg_to_remove - > origin | | seg_to_remove - > indirect_origin | |
dm_list_size ( & seg_to_remove - > lv - > segs_using_this_lv ) | |
dm_list_size ( & seg_to_remove - > lv - > indirect_glvs ) ) {
if ( ! ( glv = _create_historical_glv ( seg_to_remove ) ) )
return_0 ;
}
} else {
if ( seg_to_remove - > indirect_origin & &
! remove_glv_from_indirect_glvs ( seg_to_remove - > indirect_origin ,
seg_to_remove - > lv - > this_glv ) )
2016-03-01 17:20:09 +03:00
return_0 ;
}
* previous_glv = glv ;
return 1 ;
}
2014-01-23 12:56:17 +04:00
int detach_pool_lv ( struct lv_segment * seg )
{
2016-03-01 17:20:09 +03:00
struct generic_logical_volume * previous_glv = NULL , * glv , * user_glv ;
struct glv_list * user_glvl , * tglvl ;
2014-01-23 12:56:17 +04:00
struct lv_thin_message * tmsg , * tmp ;
struct seg_list * sl , * tsl ;
int no_update = 0 ;
if ( ! seg - > pool_lv ) {
log_error ( INTERNAL_ERROR
" No pool associated with %s LV, %s. " ,
2021-03-09 15:13:42 +03:00
lvseg_name ( seg ) , display_lvname ( seg - > lv ) ) ;
2014-01-23 12:56:17 +04:00
return 0 ;
}
2014-01-28 22:25:07 +04:00
if ( seg_is_cache ( seg ) ) {
if ( ! remove_seg_from_segs_using_this_lv ( seg - > pool_lv , seg ) )
return_0 ;
2014-04-01 20:02:36 +04:00
seg - > lv - > status & = ~ CACHE ;
2019-10-14 23:32:13 +03:00
seg - > lv - > status & = ~ LV_CACHE_USES_CACHEVOL ;
2014-11-09 23:04:33 +03:00
lv_set_visible ( seg - > pool_lv ) ;
2019-10-14 11:45:10 +03:00
seg - > pool_lv - > status & = ~ LV_CACHE_VOL ;
2014-01-28 22:25:07 +04:00
seg - > pool_lv = NULL ;
return 1 ;
}
2014-01-23 12:56:17 +04:00
if ( ! lv_is_thin_pool ( seg - > pool_lv ) ) {
2021-03-09 15:13:42 +03:00
log_error ( INTERNAL_ERROR " Cannot detach pool from LV %s. " ,
display_lvname ( seg - > lv ) ) ;
2014-01-23 12:56:17 +04:00
return 0 ;
}
/* Drop any message referencing removed segment */
dm_list_iterate_items_safe ( tmsg , tmp , & ( first_seg ( seg - > pool_lv ) - > thin_messages ) ) {
switch ( tmsg - > type ) {
case DM_THIN_MESSAGE_CREATE_SNAP :
case DM_THIN_MESSAGE_CREATE_THIN :
if ( tmsg - > u . lv = = seg - > lv ) {
log_debug_metadata ( " Discarding message for LV %s. " ,
2021-03-09 15:13:42 +03:00
display_lvname ( tmsg - > u . lv ) ) ;
2014-01-23 12:56:17 +04:00
dm_list_del ( & tmsg - > list ) ;
no_update = 1 ; /* Replacing existing */
}
break ;
case DM_THIN_MESSAGE_DELETE :
if ( tmsg - > u . delete_id = = seg - > device_id ) {
log_error ( INTERNAL_ERROR " Trying to delete %u again. " ,
tmsg - > u . delete_id ) ;
return 0 ;
}
break ;
default :
log_error ( INTERNAL_ERROR " Unsupported message type %u. " , tmsg - > type ) ;
break ;
}
}
2016-03-01 17:20:09 +03:00
if ( ! _set_up_historical_lv ( seg , & previous_glv ) )
return_0 ;
2014-01-23 12:56:17 +04:00
if ( ! detach_thin_external_origin ( seg ) )
return_0 ;
2020-09-25 23:42:53 +03:00
if ( seg - > device_id & & /* Only thins with device_id > 0 can be deleted */
2022-08-24 16:02:07 +03:00
! attach_thin_pool_message ( first_seg ( seg - > pool_lv ) ,
DM_THIN_MESSAGE_DELETE ,
NULL , seg - > device_id , no_update ) )
2014-01-23 12:56:17 +04:00
return_0 ;
if ( ! remove_seg_from_segs_using_this_lv ( seg - > pool_lv , seg ) )
return_0 ;
if ( seg - > origin & &
! remove_seg_from_segs_using_this_lv ( seg - > origin , seg ) )
return_0 ;
/* If thin origin, remove it from related thin snapshots */
/*
* TODO : map removal of origin as snapshot lvconvert - - merge ?
* i . e . rename thin snapshot to origin thin origin
*/
dm_list_iterate_items_safe ( sl , tsl , & seg - > lv - > segs_using_this_lv ) {
if ( ! seg_is_thin_volume ( sl - > seg ) | |
( seg - > lv ! = sl - > seg - > origin ) )
continue ;
2016-03-01 17:20:09 +03:00
if ( previous_glv ) {
if ( ! ( user_glv = get_or_create_glv ( seg - > lv - > vg - > vgmem , sl - > seg - > lv , NULL ) ) )
return_0 ;
if ( ! add_glv_to_indirect_glvs ( seg - > lv - > vg - > vgmem , previous_glv , user_glv ) )
return_0 ;
}
2014-01-23 12:56:17 +04:00
if ( ! remove_seg_from_segs_using_this_lv ( seg - > lv , sl - > seg ) )
return_0 ;
/* Thin snapshot is now regular thin volume */
sl - > seg - > origin = NULL ;
}
2016-03-01 17:20:09 +03:00
dm_list_iterate_items_safe ( user_glvl , tglvl , & seg - > lv - > indirect_glvs ) {
user_glv = user_glvl - > glv ;
if ( ! ( glv = get_or_create_glv ( seg - > lv - > vg - > vgmem , seg - > lv , NULL ) ) )
return_0 ;
if ( ! remove_glv_from_indirect_glvs ( glv , user_glv ) )
return_0 ;
if ( previous_glv ) {
if ( ! add_glv_to_indirect_glvs ( seg - > lv - > vg - > vgmem , previous_glv , user_glv ) )
return_0 ;
}
}
2014-01-23 12:56:17 +04:00
seg - > lv - > status & = ~ THIN_VOLUME ;
seg - > pool_lv = NULL ;
seg - > origin = NULL ;
2016-03-01 17:20:09 +03:00
seg - > indirect_origin = NULL ;
2014-01-23 12:56:17 +04:00
return 1 ;
}
struct lv_segment * find_pool_seg ( const struct lv_segment * seg )
{
2014-11-11 15:31:25 +03:00
struct lv_segment * pool_seg = NULL ;
struct seg_list * sl ;
2014-01-23 12:56:17 +04:00
2014-11-11 15:31:25 +03:00
dm_list_iterate_items ( sl , & seg - > lv - > segs_using_this_lv ) {
/* Needs to be he only item in list */
if ( lv_is_pending_delete ( sl - > seg - > lv ) )
continue ;
if ( pool_seg ) {
log_error ( " %s is referenced by more then one segments (%s, %s). " ,
display_lvname ( seg - > lv ) , display_lvname ( pool_seg - > lv ) ,
display_lvname ( sl - > seg - > lv ) ) ;
return NULL ; /* More then one segment */
}
pool_seg = sl - > seg ;
}
2014-01-23 12:56:17 +04:00
2014-11-12 12:17:17 +03:00
if ( ! pool_seg ) {
log_error ( " Pool segment not found for %s. " , display_lvname ( seg - > lv ) ) ;
return NULL ;
}
2014-05-23 16:24:28 +04:00
if ( ( lv_is_thin_type ( seg - > lv ) & & ! seg_is_pool ( pool_seg ) ) ) {
2014-01-28 22:25:07 +04:00
log_error ( " %s on %s is not a %s pool segment " ,
2021-03-09 15:13:42 +03:00
display_lvname ( pool_seg - > lv ) , display_lvname ( seg - > lv ) ,
2014-01-28 22:25:07 +04:00
lv_is_thin_type ( seg - > lv ) ? " thin " : " cache " ) ;
2014-01-23 12:56:17 +04:00
return NULL ;
}
return pool_seg ;
}
2014-10-06 14:18:57 +04:00
int validate_pool_chunk_size ( struct cmd_context * cmd ,
const struct segment_type * segtype ,
uint32_t chunk_size )
{
2017-03-05 19:41:16 +03:00
if ( segtype_is_cache ( segtype ) | | segtype_is_cache_pool ( segtype ) )
return validate_cache_chunk_size ( cmd , chunk_size ) ;
2014-10-06 14:18:57 +04:00
2017-03-05 19:41:16 +03:00
return validate_thin_pool_chunk_size ( cmd , chunk_size ) ;
2014-10-06 14:18:57 +04:00
}
2014-07-23 00:20:18 +04:00
int recalculate_pool_chunk_size_with_dev_hints ( struct logical_volume * pool_lv ,
2023-07-02 12:21:15 +03:00
struct logical_volume * pool_data_lv ,
2014-07-23 00:20:18 +04:00
int chunk_size_calc_policy )
{
struct lv_segment * seg ;
struct physical_volume * pv ;
struct cmd_context * cmd = pool_lv - > vg - > cmd ;
unsigned long previous_hint = 0 , hint = 0 ;
uint32_t min_chunk_size , max_chunk_size ;
2017-03-09 18:24:28 +03:00
if ( ! chunk_size_calc_policy )
return 1 ; /* Chunk size was specified by user */
2014-07-23 00:20:18 +04:00
if ( lv_is_thin_pool ( pool_lv ) ) {
min_chunk_size = DM_THIN_MIN_DATA_BLOCK_SIZE ;
max_chunk_size = DM_THIN_MAX_DATA_BLOCK_SIZE ;
} else if ( lv_is_cache_pool ( pool_lv ) ) {
min_chunk_size = DM_CACHE_MIN_DATA_BLOCK_SIZE ;
max_chunk_size = DM_CACHE_MAX_DATA_BLOCK_SIZE ;
} else {
log_error ( INTERNAL_ERROR " %s is not a pool logical volume. " , display_lvname ( pool_lv ) ) ;
return 0 ;
}
dm_list_iterate_items ( seg , & pool_data_lv - > segments ) {
2023-10-14 01:27:04 +03:00
switch ( seg - > area_count ? seg_type ( seg , 0 ) : AREA_UNASSIGNED ) {
2014-10-26 21:45:17 +03:00
case AREA_PV :
pv = seg_pv ( seg , 0 ) ;
if ( chunk_size_calc_policy = = THIN_CHUNK_SIZE_CALC_METHOD_PERFORMANCE )
hint = dev_optimal_io_size ( cmd - > dev_types , pv_dev ( pv ) ) ;
else
hint = dev_minimum_io_size ( cmd - > dev_types , pv_dev ( pv ) ) ;
if ( ! hint )
continue ;
if ( previous_hint )
2016-08-18 16:06:13 +03:00
hint = lcm ( previous_hint , hint ) ;
2014-10-26 21:45:17 +03:00
previous_hint = hint ;
break ;
case AREA_LV :
/* FIXME: hint for stacked (raid) LVs - estimate geometry from LV ?? */
default :
break ;
}
2014-07-23 00:20:18 +04:00
}
if ( ! hint )
log_debug_alloc ( " No usable device hint found while recalculating "
" pool chunk size for %s. " , display_lvname ( pool_lv ) ) ;
else if ( ( hint < min_chunk_size ) | | ( hint > max_chunk_size ) )
log_debug_alloc ( " Calculated chunk size %s for pool %s "
" is out of allowed range (%s-%s). " ,
display_size ( cmd , hint ) , display_lvname ( pool_lv ) ,
display_size ( cmd , min_chunk_size ) ,
display_size ( cmd , max_chunk_size ) ) ;
2017-03-09 18:24:28 +03:00
else if ( hint > first_seg ( pool_lv ) - > chunk_size ) {
log_debug_alloc ( " Updating chunk size %s for pool %s to %s. " ,
display_size ( cmd , first_seg ( pool_lv ) - > chunk_size ) ,
display_lvname ( pool_lv ) ,
display_size ( cmd , hint ) ) ;
first_seg ( pool_lv ) - > chunk_size = hint ;
}
2014-07-23 00:20:18 +04:00
return 1 ;
}
2014-01-23 12:56:17 +04:00
int create_pool ( struct logical_volume * pool_lv ,
const struct segment_type * segtype ,
struct alloc_handle * ah , uint32_t stripes , uint32_t stripe_size )
{
const struct segment_type * striped ;
struct logical_volume * meta_lv , * data_lv ;
struct lv_segment * seg ;
char name [ NAME_LEN ] ;
2014-10-24 16:22:25 +04:00
int r ;
2014-01-23 12:56:17 +04:00
if ( pool_lv - > le_count ) {
2014-01-28 22:25:07 +04:00
log_error ( INTERNAL_ERROR " Pool %s already has extents. " ,
2021-03-09 15:13:42 +03:00
display_lvname ( pool_lv ) ) ;
2014-01-23 12:56:17 +04:00
return 0 ;
}
2014-10-24 16:22:25 +04:00
if ( dm_snprintf ( name , sizeof ( name ) , " %s_%s " , pool_lv - > name ,
( segtype_is_cache_pool ( segtype ) ) ?
" cmeta " : " tmeta " ) < 0 ) {
log_error ( " Name of logical volume %s is too long to be a pool name. " ,
display_lvname ( pool_lv ) ) ;
return 0 ;
}
2014-01-23 12:56:17 +04:00
/* LV is not yet a pool, so it's extension from lvcreate */
2015-09-22 21:04:12 +03:00
if ( ! ( striped = get_segtype_from_string ( pool_lv - > vg - > cmd , SEG_TYPE_NAME_STRIPED ) ) )
2014-01-23 12:56:17 +04:00
return_0 ;
2014-10-24 16:22:25 +04:00
if ( activation ( ) & & striped - > ops - > target_present & &
! striped - > ops - > target_present ( pool_lv - > vg - > cmd , NULL , NULL ) ) {
2014-01-23 12:56:17 +04:00
log_error ( " %s: Required device-mapper target(s) not "
2014-10-24 16:22:25 +04:00
" detected in your kernel. " , striped - > name ) ;
2014-01-23 12:56:17 +04:00
return 0 ;
}
/* Metadata segment */
if ( ! lv_add_segment ( ah , stripes , 1 , pool_lv , striped , 1 , 0 , 0 ) )
return_0 ;
if ( ! activation ( ) )
log_warn ( " WARNING: Pool %s is created without initialization. " ,
2021-03-09 15:13:42 +03:00
display_lvname ( pool_lv ) ) ;
2014-10-20 17:03:24 +04:00
else if ( ! test_mode ( ) ) {
2014-01-23 12:56:17 +04:00
if ( ! vg_write ( pool_lv - > vg ) | | ! vg_commit ( pool_lv - > vg ) )
return_0 ;
/*
* If killed here , only the VISIBLE striped pool LV is left
* and user could easily remove it .
*
* FIXME : implement lazy clearing when activation is disabled
*/
2014-01-24 15:28:35 +04:00
/*
* pool_lv is a new LV so the VG lock protects us
* Pass in LV_TEMPORARY flag , since device is activated purely for wipe
* and later it is either deactivated ( in cluster )
* or directly converted to invisible device via suspend / resume
*/
pool_lv - > status | = LV_TEMPORARY ;
2018-06-05 21:21:28 +03:00
if ( ! activate_lv ( pool_lv - > vg - > cmd , pool_lv ) ) {
2014-10-24 16:22:25 +04:00
log_error ( " Aborting. Failed to activate pool metadata %s. " ,
display_lvname ( pool_lv ) ) ;
2014-01-23 12:56:17 +04:00
goto bad ;
}
2020-06-24 13:11:21 +03:00
/* Clear pool metadata device. */
if ( ! ( r = wipe_lv ( pool_lv , ( struct wipe_params ) { . is_metadata = 1 } ) ) ) {
2014-10-24 16:22:25 +04:00
log_error ( " Aborting. Failed to wipe pool metadata %s. " ,
display_lvname ( pool_lv ) ) ;
}
2014-01-24 16:13:37 +04:00
pool_lv - > status & = ~ LV_TEMPORARY ;
2014-03-12 02:07:41 +04:00
/* Deactivates cleared metadata LV */
2018-01-12 15:34:13 +03:00
if ( ! deactivate_lv ( pool_lv - > vg - > cmd , pool_lv ) ) {
2014-10-24 16:22:25 +04:00
log_error ( " Aborting. Could not deactivate pool metadata %s. " ,
display_lvname ( pool_lv ) ) ;
return 0 ;
}
if ( ! r )
goto bad ;
2014-01-23 12:56:17 +04:00
}
if ( ! ( meta_lv = lv_create_empty ( name , NULL , LVM_READ | LVM_WRITE ,
ALLOC_INHERIT , pool_lv - > vg ) ) )
goto_bad ;
if ( ! move_lv_segments ( meta_lv , pool_lv , 0 , 0 ) )
goto_bad ;
/* Pool data segment */
if ( ! lv_add_segment ( ah , 0 , stripes , pool_lv , striped , stripe_size , 0 , 0 ) )
goto_bad ;
2024-05-03 22:12:45 +03:00
if ( pool_lv - > vg - > cmd - > lvcreate_vcp & & ! convert_vdo_lv ( pool_lv , pool_lv - > vg - > cmd - > lvcreate_vcp ) ) {
/* Conversion to VDO commits metadata,
* try to deactivate pool LV , and remove metadata LV */
if ( ! deactivate_lv ( pool_lv - > vg - > cmd , pool_lv ) )
log_error ( " Failedto deactivate pool volume %s. " ,
display_lvname ( pool_lv ) ) ;
if ( ! lv_remove ( meta_lv ) | |
! vg_write ( meta_lv - > vg ) | | ! vg_commit ( meta_lv - > vg ) )
log_error ( " Manual intervention may be required to "
" remove abandoned LV(s) before retrying. " ) ;
2023-12-14 16:20:19 +03:00
goto_bad ;
2024-05-03 22:12:45 +03:00
}
2023-12-14 16:20:19 +03:00
2014-01-23 12:56:17 +04:00
if ( ! ( data_lv = insert_layer_for_lv ( pool_lv - > vg - > cmd , pool_lv ,
pool_lv - > status ,
2014-01-28 22:25:07 +04:00
( segtype_is_cache_pool ( segtype ) ) ?
" _cdata " : " _tdata " ) ) )
2014-01-23 12:56:17 +04:00
goto_bad ;
seg = first_seg ( pool_lv ) ;
/* Drop reference as attach_pool_data_lv() takes it again */
if ( ! remove_seg_from_segs_using_this_lv ( data_lv , seg ) )
goto_bad ;
2014-01-28 22:25:07 +04:00
seg - > segtype = segtype ; /* Set as thin_pool or cache_pool segment */
2014-01-23 12:56:17 +04:00
if ( ! attach_pool_data_lv ( seg , data_lv ) )
goto_bad ;
if ( ! attach_pool_metadata_lv ( seg , meta_lv ) )
goto_bad ;
return 1 ;
bad :
if ( activation ( ) ) {
2014-10-24 16:22:25 +04:00
/* Without activation there was no intermediate commit */
2014-01-23 12:56:17 +04:00
if ( ! lv_remove ( pool_lv ) | |
! vg_write ( pool_lv - > vg ) | | ! vg_commit ( pool_lv - > vg ) )
log_error ( " Manual intervention may be required to "
" remove abandoned LV(s) before retrying. " ) ;
}
return 0 ;
}
2014-07-08 17:19:47 +04:00
struct logical_volume * alloc_pool_metadata ( struct logical_volume * pool_lv ,
2023-06-29 14:53:43 +03:00
uint32_t read_ahead ,
2014-07-08 17:19:47 +04:00
uint32_t stripes , uint32_t stripe_size ,
2014-10-30 15:04:06 +03:00
uint32_t extents , alloc_policy_t alloc ,
2014-07-08 17:19:47 +04:00
struct dm_list * pvh )
{
struct logical_volume * metadata_lv ;
/* FIXME: Make lvm2api usable */
struct lvcreate_params lvc = {
. activate = CHANGE_ALY ,
. alloc = alloc ,
2014-10-30 15:04:06 +03:00
. extents = extents ,
2014-07-08 17:19:47 +04:00
. major = - 1 ,
. minor = - 1 ,
. permission = LVM_READ | LVM_WRITE ,
. pvh = pvh ,
. read_ahead = read_ahead ,
. stripe_size = stripe_size ,
. stripes = stripes ,
2014-10-30 15:04:41 +03:00
. tags = DM_LIST_HEAD_INIT ( lvc . tags ) ,
2014-10-24 16:22:25 +04:00
. temporary = 1 ,
2014-10-30 15:04:41 +03:00
. zero = 1 ,
2020-06-24 13:11:21 +03:00
. is_metadata = 1 ,
2023-08-17 20:59:36 +03:00
. lv_name = " pool_metadata%d " ,
2014-07-08 17:19:47 +04:00
} ;
2015-09-22 21:04:12 +03:00
if ( ! ( lvc . segtype = get_segtype_from_string ( pool_lv - > vg - > cmd , SEG_TYPE_NAME_STRIPED ) ) )
2014-07-08 17:19:47 +04:00
return_0 ;
/* FIXME: allocate properly space for metadata_lv */
if ( ! ( metadata_lv = lv_create_single ( pool_lv - > vg , & lvc ) ) )
return_0 ;
return metadata_lv ;
}
2023-12-08 16:10:39 +03:00
int add_metadata_to_pool ( struct lv_segment * pool_seg ,
struct logical_volume * metadata_lv )
{
struct cmd_context * cmd = metadata_lv - > vg - > cmd ;
char name [ NAME_LEN ] ; /* generated sub lv name */
if ( ! deactivate_lv ( cmd , metadata_lv ) ) {
log_error ( " Aborting. Failed to deactivate metadata lv. "
" Manual intervention required. " ) ;
return 0 ;
}
if ( ( dm_snprintf ( name , sizeof ( name ) , " %s%s " , pool_seg - > lv - > name ,
seg_is_thin_pool ( pool_seg ) ? " _tmeta " : " _cmeta " ) < 0 ) ) {
log_error ( " Failed to create internal lv names, %s name is too long. " ,
pool_seg - > lv - > name ) ;
return 0 ;
}
/* Rename LVs to the pool _[ct]meta LV naming scheme. */
if ( ( strcmp ( metadata_lv - > name , name ) ! = 0 ) & &
! lv_rename_update ( cmd , metadata_lv , name , 0 ) )
return_0 ;
if ( ! attach_pool_metadata_lv ( pool_seg , metadata_lv ) )
return_0 ;
return 1 ;
}
2014-07-08 17:19:47 +04:00
static struct logical_volume * _alloc_pool_metadata_spare ( struct volume_group * vg ,
uint32_t extents ,
struct dm_list * pvh )
{
struct logical_volume * lv ;
/* FIXME: Make lvm2api usable */
struct lvcreate_params lp = {
. activate = CHANGE_ALY ,
. alloc = ALLOC_INHERIT ,
. extents = extents ,
. major = - 1 ,
. minor = - 1 ,
. permission = LVM_READ | LVM_WRITE ,
. pvh = pvh ? : & vg - > pvs ,
. read_ahead = DM_READ_AHEAD_AUTO ,
. stripes = 1 ,
2014-10-30 15:04:41 +03:00
. tags = DM_LIST_HEAD_INIT ( lp . tags ) ,
2014-07-08 17:19:47 +04:00
. temporary = 1 ,
2014-10-30 15:04:41 +03:00
. zero = 1 ,
2020-06-24 13:11:21 +03:00
. is_metadata = 1 ,
2014-07-08 17:19:47 +04:00
} ;
2015-09-22 21:04:12 +03:00
if ( ! ( lp . segtype = get_segtype_from_string ( vg - > cmd , SEG_TYPE_NAME_STRIPED ) ) )
2014-07-08 17:19:47 +04:00
return_0 ;
/* FIXME: Maybe using silent mode ? */
2014-10-20 17:02:59 +04:00
log_verbose ( " Preparing pool metadata spare volume for Volume group %s. " , vg - > name ) ;
2014-07-08 17:19:47 +04:00
if ( ! ( lv = lv_create_single ( vg , & lp ) ) )
return_0 ;
/* Spare LV should not be active */
2018-01-12 15:34:13 +03:00
if ( ! deactivate_lv ( vg - > cmd , lv ) ) {
2014-07-08 17:19:47 +04:00
log_error ( " Unable to deactivate pool metadata spare LV. "
" Manual intervention required. " ) ;
return 0 ;
}
if ( ! vg_set_pool_metadata_spare ( lv ) )
return_0 ;
return lv ;
}
/*
* Create / resize pool metadata spare LV
* Caller does vg_write ( ) , vg_commit ( ) with pool creation
* extents is 0 , max size is determined
*/
int handle_pool_metadata_spare ( struct volume_group * vg , uint32_t extents ,
struct dm_list * pvh , int poolmetadataspare )
{
2021-01-30 01:20:18 +03:00
/* Max usable size of any spare volume is currently 16GiB rouned to extent size */
const uint64_t MAX_SIZE = ( UINT64_C ( 2 * 16 ) * 1024 * 1024 + vg - > extent_size - 1 ) / vg - > extent_size ;
2014-07-08 17:19:47 +04:00
struct logical_volume * lv = vg - > pool_metadata_spare_lv ;
uint32_t seg_mirrors ;
struct lv_segment * seg ;
const struct lv_list * lvl ;
if ( ! extents )
/* Find maximal size of metadata LV */
dm_list_iterate_items ( lvl , & vg - > lvs )
if ( lv_is_pool_metadata ( lvl - > lv ) & &
2021-01-30 01:20:18 +03:00
( lvl - > lv - > le_count > extents ) ) {
2014-07-08 17:19:47 +04:00
extents = lvl - > lv - > le_count ;
2021-01-30 01:20:18 +03:00
if ( extents > = MAX_SIZE )
break ;
}
2014-07-08 17:19:47 +04:00
if ( ! poolmetadataspare ) {
/* TODO: Not showing when lvm.conf would define 'n' ? */
if ( DEFAULT_POOL_METADATA_SPARE & & extents )
/* Warn if there would be any user */
log_warn ( " WARNING: recovery of pools without pool "
" metadata spare LV is not automated. " ) ;
return 1 ;
}
2021-07-20 23:33:05 +03:00
if ( ! extents ) {
/* pmspare is not needed */
if ( lv ) {
log_debug_metadata ( " Dropping unused pool metadata spare LV %s. " ,
display_lvname ( lv ) ) ;
if ( ! lv_remove_single ( vg - > cmd , lv , DONT_PROMPT , 0 ) )
return_0 ;
}
return 1 ;
}
2021-01-30 01:20:18 +03:00
if ( extents > MAX_SIZE )
extents = MAX_SIZE ;
2014-07-08 17:19:47 +04:00
if ( ! lv ) {
2022-09-10 00:07:07 +03:00
log_debug ( " Adding new pool metadata spare %u extents. " , extents ) ;
2014-07-08 17:19:47 +04:00
if ( ! _alloc_pool_metadata_spare ( vg , extents , pvh ) )
return_0 ;
return 1 ;
}
seg = last_seg ( lv ) ;
seg_mirrors = lv_mirror_count ( lv ) ;
2022-09-10 00:07:07 +03:00
log_debug ( " Extending pool metadata spare from %u to %u extents. " ,
lv - > le_count , extents ) ;
2014-07-08 17:19:47 +04:00
/* Check spare LV is big enough and preserve segtype */
if ( ( lv - > le_count < extents ) & & seg & &
2024-05-27 14:24:06 +03:00
/* coverity[format_string_injection] lv name is already validated */
2014-07-08 17:19:47 +04:00
! lv_extend ( lv , seg - > segtype ,
seg - > area_count / seg_mirrors ,
seg - > stripe_size ,
seg_mirrors ,
seg - > region_size ,
2014-10-26 10:13:59 +03:00
extents - lv - > le_count ,
2014-07-08 17:19:47 +04:00
pvh , lv - > alloc , 0 ) )
return_0 ;
return 1 ;
}
2021-01-12 19:59:29 +03:00
int update_pool_metadata_min_max ( struct cmd_context * cmd ,
uint32_t extent_size ,
uint64_t min_metadata_size , /* required min */
uint64_t max_metadata_size , /* writable max */
uint64_t * metadata_size , /* current calculated */
struct logical_volume * metadata_lv , /* name of converted LV or NULL */
uint32_t * metadata_extents ) /* resulting extent count */
{
max_metadata_size = dm_round_up ( max_metadata_size , extent_size ) ;
min_metadata_size = dm_round_up ( min_metadata_size , extent_size ) ;
if ( * metadata_size > max_metadata_size ) {
if ( metadata_lv ) {
log_print_unless_silent ( " Size %s of pool metadata volume %s is bigger then maximum usable size %s. " ,
display_size ( cmd , * metadata_size ) ,
display_lvname ( metadata_lv ) ,
display_size ( cmd , max_metadata_size ) ) ;
} else {
if ( * metadata_extents )
log_print_unless_silent ( " Reducing pool metadata size %s to maximum usable size %s. " ,
display_size ( cmd , * metadata_size ) ,
display_size ( cmd , max_metadata_size ) ) ;
* metadata_size = max_metadata_size ;
}
} else if ( * metadata_size < min_metadata_size ) {
if ( metadata_lv ) {
log_error ( " Can't use volume %s with size %s as pool metadata. Minimal required size is %s. " ,
display_lvname ( metadata_lv ) ,
display_size ( cmd , * metadata_size ) ,
display_size ( cmd , min_metadata_size ) ) ;
return 0 ;
} else {
if ( * metadata_extents )
log_print_unless_silent ( " Extending pool metadata size %s to required minimal size %s. " ,
display_size ( cmd , * metadata_size ) ,
display_size ( cmd , min_metadata_size ) ) ;
* metadata_size = min_metadata_size ;
}
}
if ( ! ( * metadata_extents = extents_from_size ( cmd , * metadata_size , extent_size ) ) )
return_0 ;
return 1 ;
}
2014-07-08 17:19:47 +04:00
int vg_set_pool_metadata_spare ( struct logical_volume * lv )
{
char new_name [ NAME_LEN ] ;
struct volume_group * vg = lv - > vg ;
if ( vg - > pool_metadata_spare_lv ) {
if ( vg - > pool_metadata_spare_lv = = lv )
return 1 ;
if ( ! vg_remove_pool_metadata_spare ( vg ) )
return_0 ;
}
if ( dm_snprintf ( new_name , sizeof ( new_name ) , " %s_pmspare " , lv - > name ) < 0 ) {
log_error ( " Can't create pool metadata spare. Name of pool LV "
" %s is too long. " , lv - > name ) ;
return 0 ;
}
2014-10-20 17:02:59 +04:00
log_verbose ( " Renaming %s as pool metadata spare volume %s. " , lv - > name , new_name ) ;
2014-07-08 17:19:47 +04:00
if ( ! lv_rename_update ( vg - > cmd , lv , new_name , 0 ) )
return_0 ;
lv_set_hidden ( lv ) ;
lv - > status | = POOL_METADATA_SPARE ;
vg - > pool_metadata_spare_lv = lv ;
return 1 ;
}
int vg_remove_pool_metadata_spare ( struct volume_group * vg )
{
char new_name [ NAME_LEN ] ;
char * c ;
struct logical_volume * lv = vg - > pool_metadata_spare_lv ;
if ( ! ( lv - > status & POOL_METADATA_SPARE ) ) {
log_error ( INTERNAL_ERROR " LV %s is not pool metadata spare. " ,
2015-11-18 11:14:24 +03:00
display_lvname ( lv ) ) ;
2014-07-08 17:19:47 +04:00
return 0 ;
}
vg - > pool_metadata_spare_lv = NULL ;
lv - > status & = ~ POOL_METADATA_SPARE ;
lv_set_visible ( lv ) ;
/* Cut off suffix _pmspare */
2024-04-04 19:42:24 +03:00
if ( ! _dm_strncpy ( new_name , lv - > name , sizeof ( new_name ) ) | |
2015-11-18 11:14:24 +03:00
! ( c = strchr ( new_name , ' _ ' ) ) ) {
2014-07-08 17:19:47 +04:00
log_error ( INTERNAL_ERROR " LV %s has no suffix for pool metadata spare. " ,
2015-11-18 11:14:24 +03:00
display_lvname ( lv ) ) ;
2014-07-08 17:19:47 +04:00
return 0 ;
}
* c = 0 ;
/* If the name is in use, generate new lvol%d */
2016-03-01 17:31:48 +03:00
if ( lv_name_is_used_in_vg ( vg , new_name , NULL ) & &
2014-07-08 17:19:47 +04:00
! generate_lv_name ( vg , " lvol%d " , new_name , sizeof ( new_name ) ) ) {
log_error ( " Failed to generate unique name for "
" pool metadata spare logical volume. " ) ;
return 0 ;
}
log_print_unless_silent ( " Renaming existing pool metadata spare "
2015-11-18 11:14:24 +03:00
" logical volume \" %s \" to \" %s/%s \" . " ,
2019-01-24 16:12:42 +03:00
display_lvname ( lv ) , vg - > name , new_name ) ;
2014-07-08 17:19:47 +04:00
if ( ! lv_rename_update ( vg - > cmd , lv , new_name , 0 ) )
return_0 ;
/* To display default warning */
( void ) handle_pool_metadata_spare ( vg , 0 , 0 , 0 ) ;
return 1 ;
}