2002-02-25 15:02:33 +03:00
/*
2008-01-30 17:00:02 +03:00
* Copyright ( C ) 2002 - 2004 Sistina Software , Inc . All rights reserved .
2018-02-27 16:13:00 +03:00
* Copyright ( C ) 2004 - 2018 Red Hat , Inc . All rights reserved .
2002-02-25 15:02:33 +03:00
*
2004-03-30 23:35:44 +04:00
* This file is part of LVM2 .
*
* This copyrighted material is made available to anyone wishing to use ,
* modify , copy , or redistribute it subject to the terms and conditions
2007-08-21 00:55:30 +04:00
* of the GNU Lesser General Public License v .2 .1 .
2004-03-30 23:35:44 +04:00
*
2007-08-21 00:55:30 +04:00
* You should have received a copy of the GNU Lesser General Public License
2004-03-30 23:35:44 +04:00
* along with this program ; if not , write to the Free Software Foundation ,
2016-01-21 13:49:46 +03:00
* Inc . , 51 Franklin Street , Fifth Floor , Boston , MA 02110 - 1301 USA
2002-02-25 15:02:33 +03:00
*/
2018-05-14 12:30:20 +03:00
# include "lib/misc/lib.h"
2002-02-25 15:02:33 +03:00
# include "dev_manager.h"
2018-05-14 12:30:20 +03:00
# include "lib/misc/lvm-string.h"
2002-02-26 14:49:17 +03:00
# include "fs.h"
2018-05-14 12:30:20 +03:00
# include "lib/config/defaults.h"
# include "lib/metadata/segtype.h"
# include "lib/display/display.h"
# include "lib/commands/toolcontext.h"
# include "lib/activate/targets.h"
# include "lib/config/config.h"
# include "lib/activate/activate.h"
# include "lib/misc/lvm-exec.h"
# include "lib/datastruct/str_list.h"
2021-04-06 15:47:11 +03:00
# include "lib/misc/lvm-signal.h"
2002-02-25 15:02:33 +03:00
2002-02-26 17:44:13 +03:00
# include <limits.h>
2002-03-07 20:37:38 +03:00
# include <dirent.h>
2002-02-25 15:02:33 +03:00
2003-08-20 16:53:57 +04:00
# define MAX_TARGET_PARAMSIZE 50000
2013-10-08 15:27:21 +04:00
# define LVM_UDEV_NOSCAN_FLAG DM_SUBSYSTEM_UDEV_FLAG0
2018-03-09 16:49:38 +03:00
# define CRYPT_TEMP "CRYPT-TEMP"
2019-09-13 14:22:07 +03:00
# define CRYPT_SUBDEV "CRYPT-SUBDEV"
2018-03-09 16:49:38 +03:00
# define STRATIS "stratis-"
2003-08-20 16:53:57 +04:00
2002-03-16 01:59:12 +03:00
typedef enum {
2005-11-09 01:52:26 +03:00
PRELOAD ,
2002-03-16 01:59:12 +03:00
ACTIVATE ,
2002-03-27 21:17:43 +03:00
DEACTIVATE ,
2002-03-16 01:59:12 +03:00
SUSPEND ,
2006-08-09 01:20:00 +04:00
SUSPEND_WITH_LOCKFS ,
2005-11-09 01:52:26 +03:00
CLEAN
2002-03-27 21:17:43 +03:00
} action_t ;
2002-03-16 01:59:12 +03:00
2014-07-31 00:55:11 +04:00
/* This list must match lib/misc/lvm-string.c:build_dm_uuid(). */
2024-05-03 13:43:14 +03:00
static const char * const _uuid_suffix_list [ ] = { " pool " , " cdata " , " cmeta " , " cvol " , " tdata " , " tmeta " , " vdata " , " vpool " , " imeta " , NULL } ;
2014-07-31 00:55:11 +04:00
2014-11-13 12:08:40 +03:00
struct dlid_list {
struct dm_list list ;
const char * dlid ;
const struct logical_volume * lv ;
} ;
2002-02-25 15:02:33 +03:00
struct dev_manager {
2005-10-17 03:03:59 +04:00
struct dm_pool * mem ;
2002-02-25 15:02:33 +03:00
2004-05-05 01:25:57 +04:00
struct cmd_context * cmd ;
void * target_state ;
2003-05-06 16:00:29 +04:00
uint32_t pvmove_mirror_count ;
2009-05-20 13:52:37 +04:00
int flush_required ;
2013-07-09 14:34:49 +04:00
int activation ; /* building activation tree */
thin: move pool messaging from resume to suspend
Existing messaging intarface for thin-pool has a few 'weak' points:
* Message were posted with each 'resume' operation, thus not allowing
activation of thin-pool with the existing state.
* Acceleration skipped suspend step has not worked in cluster,
since clvmd resumes only nodes which are suspended (have proper lock
state).
* Resume may fail and code is not really designed to 'fail' in this
phase (generic rule here is resume DOES NOT fail unless something serious
is wrong and lvm2 tool usually doesn't handle recovery path in this case.)
* Full thin-pool suspend happened, when taken a thin-volume snapshot.
With this patch the new method relocates message passing into suspend
state.
This has a few drawbacks with current API, but overal it performs
better and gives are more posibilities to deal with errors.
Patch introduces a new logic for 'origin-only' suspend of thin-pool and
this also relates to thin-volume when taking snapshot.
When suspend_origin_only operation is invoked on a pool with
queued messages then only those messages are posted to thin-pool and
actual suspend of thin pool and data and metadata volume is skipped.
This makes taking a snapshot of thin-volume lighter operation and
avoids blocking of other unrelated active thin volumes.
Also fail now happens in 'suspend' state where the 'Fail' is more expected
and it is better handled through error paths.
Activation of thin-pool is now not sending any message and leaves upto a tool
to decided later how to finish unfinished double-commit transaction.
Problem which needs some API improvements relates to the lvm2 tree
construction. For the suspend tree we do not add target table line
into the tree, but only a device is inserted into a tree.
Current mechanism to attach messages for thin-pool requires the libdm
to know about thin-pool target, so lvm2 currently takes assumption, node
is really a thin-pool and fills in the table line for this node (which
should be ensured by the PRELOAD phase, but it's a misuse of internal API)
we would possibly need to be able to attach message to 'any' node.
Other thing to notice - current messaging interface in thin-pool
target requires to suspend thin volume origin first and then send
a create message, but this could not have any 'nice' solution on lvm2
side and IMHO we should introduce something like 'create_after_resume'
message.
Patch also changes the moment, where lvm2 transaction id is increased.
Now it happens only after successful finish of kernel transaction id
change. This change was needed to handle properly activation of pool,
which is in the middle of unfinished transaction, and also this corrects
usage of thin-pool by external apps like Docker.
2015-07-01 14:31:37 +03:00
int suspend ; /* building suspend tree */
2014-11-13 12:08:40 +03:00
unsigned track_pending_delete ;
2011-06-11 04:03:06 +04:00
unsigned track_pvmove_deps ;
2003-04-25 02:09:13 +04:00
2015-11-25 18:04:27 +03:00
const char * vg_name ;
2002-02-25 15:02:33 +03:00
} ;
2005-11-09 01:52:26 +03:00
struct lv_layer {
2014-09-22 17:50:07 +04:00
const struct logical_volume * lv ;
2005-11-09 01:52:26 +03:00
const char * old_name ;
2018-02-28 19:22:09 +03:00
int visible_component ;
2005-11-09 01:52:26 +03:00
} ;
2002-03-07 19:48:46 +03:00
2018-03-08 12:27:04 +03:00
int read_only_lv ( const struct logical_volume * lv , const struct lv_activate_opts * laopts , const char * layer )
2005-11-09 01:52:26 +03:00
{
2018-03-08 12:27:04 +03:00
if ( layer & & lv_is_cow ( lv ) )
return 0 ; /* Keep snapshot's COW volume writable */
2018-03-13 00:21:13 +03:00
if ( lv_is_raid_image ( lv ) | | lv_is_raid_metadata ( lv ) )
return 0 ; /* Keep RAID SubLvs writable */
2019-09-14 02:08:14 +03:00
if ( ! layer ) {
2020-03-11 14:57:44 +03:00
if ( lv_is_thin_pool ( lv ) | | lv_is_vdo_pool ( lv ) )
2019-09-14 02:08:14 +03:00
return 1 ;
}
2014-02-13 17:42:29 +04:00
return ( laopts - > read_only | | ! ( lv - > status & LVM_WRITE ) ) ;
2005-11-09 01:52:26 +03:00
}
2002-02-25 17:46:57 +03:00
/*
* Low level device - layer operations .
2016-12-01 16:56:32 +03:00
*
* Unless task is DM_DEVICE_TARGET_MSG , also calls dm_task_run ( )
2002-02-25 17:46:57 +03:00
*/
2016-12-01 16:56:32 +03:00
static struct dm_task * _setup_task_run ( int task , struct dm_info * info ,
const char * name , const char * uuid ,
uint32_t * event_nr ,
uint32_t major , uint32_t minor ,
int with_open_count ,
int with_flush ,
int query_inactive )
2002-02-25 17:46:57 +03:00
{
2022-02-16 02:33:25 +03:00
char vsn [ 80 ] ;
unsigned maj , min ;
2002-02-25 17:46:57 +03:00
struct dm_task * dmt ;
2008-01-30 16:19:47 +03:00
if ( ! ( dmt = dm_task_create ( task ) ) )
return_NULL ;
2002-02-25 17:46:57 +03:00
2011-01-05 17:03:36 +03:00
if ( name & & ! dm_task_set_name ( dmt , name ) )
goto_out ;
2002-03-25 21:54:59 +03:00
2011-01-05 17:03:36 +03:00
if ( uuid & & * uuid & & ! dm_task_set_uuid ( dmt , uuid ) )
goto_out ;
2002-03-25 21:54:59 +03:00
2011-01-05 17:03:36 +03:00
if ( event_nr & & ! dm_task_set_event_nr ( dmt , * event_nr ) )
goto_out ;
2003-04-30 19:26:25 +04:00
2011-01-05 17:03:36 +03:00
if ( major & & ! dm_task_set_major_minor ( dmt , major , minor , 1 ) )
goto_out ;
2008-12-19 18:23:03 +03:00
2011-07-01 18:09:19 +04:00
if ( activation_checks ( ) & & ! dm_task_enable_checks ( dmt ) )
goto_out ;
2014-11-02 19:22:32 +03:00
2016-12-01 16:56:32 +03:00
if ( query_inactive & & ! dm_task_query_inactive_table ( dmt ) ) {
log_error ( " Failed to set query_inactive_table. " ) ;
goto out ;
}
2014-11-02 19:22:32 +03:00
if ( ! with_open_count & & ! dm_task_no_open_count ( dmt ) )
log_warn ( " WARNING: Failed to disable open_count. " ) ;
2016-05-27 14:45:47 +03:00
if ( ! with_flush & & ! dm_task_no_flush ( dmt ) )
log_warn ( " WARNING: Failed to set no_flush. " ) ;
2021-12-15 13:24:31 +03:00
switch ( task ) {
case DM_DEVICE_TARGET_MSG :
2016-12-01 16:56:32 +03:00
return dmt ; /* TARGET_MSG needs more local tweaking before task_run() */
2021-12-15 13:24:31 +03:00
case DM_DEVICE_LIST :
2022-02-16 02:33:25 +03:00
/* Use 'newuuid' only with DM version that supports it */
if ( driver_version ( vsn , sizeof ( vsn ) ) & &
( sscanf ( vsn , " %u.%u " , & maj , & min ) = = 2 ) & &
( maj = = 4 ? min > = 19 : maj > 4 ) & &
! dm_task_set_newuuid ( dmt , " " ) ) // new uuid has no meaning here
2021-12-15 13:24:31 +03:00
log_warn ( " WARNING: Failed to query uuid with LIST. " ) ;
break ;
default :
break ;
}
2016-12-01 16:56:32 +03:00
if ( ! dm_task_run ( dmt ) )
goto_out ;
if ( info & & ! dm_task_get_info ( dmt , info ) )
goto_out ;
2002-02-25 17:46:57 +03:00
return dmt ;
2016-12-01 16:56:32 +03:00
out :
2011-01-05 17:03:36 +03:00
dm_task_destroy ( dmt ) ;
2016-12-01 16:53:35 +03:00
2011-01-05 17:03:36 +03:00
return NULL ;
2002-02-25 17:46:57 +03:00
}
2024-04-05 15:00:11 +03:00
/* Read info from DM VDO 'stats' message */
static int _vdo_pool_message_stats ( struct dm_pool * mem ,
const struct logical_volume * lv ,
struct lv_status_vdo * status )
{
const char * response ;
const char * dlid ;
struct dm_task * dmt = NULL ;
int r = 0 ;
unsigned i ;
const char * p ;
struct vdo_msg_elem {
const char * name ;
uint64_t * val ;
} const vme [ ] = { /* list of properties lvm2 wants to parse */
{ " dataBlocksUsed " , & status - > data_blocks_used } ,
{ " logicalBlocksUsed " , & status - > logical_blocks_used }
} ;
for ( i = 0 ; i < DM_ARRAY_SIZE ( vme ) ; + + i )
* vme [ i ] . val = ULLONG_MAX ;
if ( ! ( dlid = build_dm_uuid ( mem , lv , lv_layer ( lv ) ) ) )
return_0 ;
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_TARGET_MSG , NULL , NULL , dlid , 0 , 0 , 0 , 0 , 0 , 0 ) ) )
return_0 ;
if ( ! dm_task_set_message ( dmt , " stats " ) )
goto_out ;
if ( ! dm_task_run ( dmt ) )
goto_out ;
log_debug_activation ( " Checking VDO pool stats message for LV %s. " ,
display_lvname ( lv ) ) ;
if ( ( response = dm_task_get_message_response ( dmt ) ) ) {
for ( i = 0 ; i < DM_ARRAY_SIZE ( vme ) ; + + i ) {
errno = 0 ;
if ( ! ( p = strstr ( response , vme [ i ] . name ) ) | |
! ( p = strchr ( p , ' : ' ) ) | |
( ( * vme [ i ] . val = strtoul ( p + 1 , NULL , 10 ) ) = = ULLONG_MAX ) | | errno ) {
log_debug ( " Cannot parse %s in VDO DM stats message. " , vme [ i ] . name ) ;
* vme [ i ] . val = ULLONG_MAX ;
goto out ;
}
if ( * vme [ i ] . val ! = ULLONG_MAX )
log_debug ( " VDO property %s = " FMTu64 , vme [ i ] . name , * vme [ i ] . val ) ;
}
}
r = 1 ;
out :
dm_task_destroy ( dmt ) ;
return r ;
}
2014-11-04 17:00:32 +03:00
static int _get_segment_status_from_target_params ( const char * target_name ,
const char * params ,
2021-09-09 15:59:38 +03:00
const struct dm_info * dminfo ,
2014-11-04 17:00:32 +03:00
struct lv_seg_status * seg_status )
{
2016-12-05 12:20:42 +03:00
const struct lv_segment * seg = seg_status - > seg ;
const struct segment_type * segtype = seg - > segtype ;
2014-11-04 17:00:32 +03:00
2016-12-02 15:57:52 +03:00
seg_status - > type = SEG_STATUS_UNKNOWN ; /* Parsing failed */
2016-12-05 12:20:42 +03:00
/* Switch to snapshot segtype status logic for merging origin */
/* This is 'dynamic' decision, both states are valid */
if ( lv_is_merging_origin ( seg - > lv ) ) {
if ( ! strcmp ( target_name , TARGET_NAME_SNAPSHOT_ORIGIN ) ) {
seg_status - > type = SEG_STATUS_NONE ;
return 1 ; /* Merge has not yet started */
}
if ( ! strcmp ( target_name , TARGET_NAME_SNAPSHOT_MERGE ) & &
! ( segtype = get_segtype_from_string ( seg - > lv - > vg - > cmd , TARGET_NAME_SNAPSHOT ) ) )
return_0 ;
/* Merging, parse 'snapshot' status of merge progress */
}
2016-12-02 15:57:52 +03:00
if ( ! params ) {
log_warn ( " WARNING: Cannot find matching %s segment for %s. " ,
2016-12-05 12:20:42 +03:00
segtype - > name , display_lvname ( seg_status - > seg - > lv ) ) ;
2016-12-02 15:57:52 +03:00
return 0 ;
}
2016-05-25 17:14:46 +03:00
2016-12-05 12:20:42 +03:00
/* Validate target_name segtype from DM table with lvm2 metadata segtype */
2018-06-04 12:05:28 +03:00
if ( ! lv_is_locked ( seg - > lv ) & &
strcmp ( segtype - > name , target_name ) & &
2016-12-05 12:20:42 +03:00
/* If kernel's type isn't an exact match is it compatible? */
( ! segtype - > ops - > target_status_compatible | |
! segtype - > ops - > target_status_compatible ( target_name ) ) ) {
2022-02-15 23:16:10 +03:00
log_warn ( " WARNING: Detected %s segment type does not match expected type %s for %s. " ,
2016-12-05 12:20:42 +03:00
target_name , segtype - > name , display_lvname ( seg_status - > seg - > lv ) ) ;
return 0 ;
2014-11-04 17:00:32 +03:00
}
2016-12-05 12:20:42 +03:00
/* TODO: move into segtype method */
2015-09-24 16:59:07 +03:00
if ( segtype_is_cache ( segtype ) ) {
2015-01-20 15:14:16 +03:00
if ( ! dm_get_status_cache ( seg_status - > mem , params , & ( seg_status - > cache ) ) )
return_0 ;
2015-01-14 14:51:59 +03:00
seg_status - > type = SEG_STATUS_CACHE ;
2015-09-24 16:59:07 +03:00
} else if ( segtype_is_raid ( segtype ) ) {
2015-01-20 15:14:16 +03:00
if ( ! dm_get_status_raid ( seg_status - > mem , params , & seg_status - > raid ) )
2015-01-14 14:51:59 +03:00
return_0 ;
seg_status - > type = SEG_STATUS_RAID ;
2015-09-24 16:59:07 +03:00
} else if ( segtype_is_thin_volume ( segtype ) ) {
2015-01-20 15:14:16 +03:00
if ( ! dm_get_status_thin ( seg_status - > mem , params , & seg_status - > thin ) )
2015-01-14 14:51:59 +03:00
return_0 ;
seg_status - > type = SEG_STATUS_THIN ;
2015-09-24 16:59:07 +03:00
} else if ( segtype_is_thin_pool ( segtype ) ) {
2015-01-20 15:14:16 +03:00
if ( ! dm_get_status_thin_pool ( seg_status - > mem , params , & seg_status - > thin_pool ) )
2015-01-14 14:51:59 +03:00
return_0 ;
seg_status - > type = SEG_STATUS_THIN_POOL ;
2015-09-24 16:59:07 +03:00
} else if ( segtype_is_snapshot ( segtype ) ) {
2015-01-20 15:14:16 +03:00
if ( ! dm_get_status_snapshot ( seg_status - > mem , params , & seg_status - > snapshot ) )
2015-01-14 14:51:59 +03:00
return_0 ;
2014-11-04 17:00:32 +03:00
seg_status - > type = SEG_STATUS_SNAPSHOT ;
2018-07-02 18:20:30 +03:00
} else if ( segtype_is_vdo_pool ( segtype ) ) {
2024-04-05 15:00:11 +03:00
if ( ! _vdo_pool_message_stats ( seg_status - > mem , seg - > lv , & seg_status - > vdo_pool ) )
stack ;
2021-09-09 15:59:38 +03:00
if ( ! parse_vdo_pool_status ( seg_status - > mem , seg - > lv , params , dminfo , & seg_status - > vdo_pool ) )
2018-07-02 18:20:30 +03:00
return_0 ;
seg_status - > type = SEG_STATUS_VDO_POOL ;
2018-08-27 22:53:09 +03:00
} else if ( segtype_is_writecache ( segtype ) ) {
if ( ! dm_get_status_writecache ( seg_status - > mem , params , & ( seg_status - > writecache ) ) )
return_0 ;
seg_status - > type = SEG_STATUS_WRITECACHE ;
2019-11-21 01:07:27 +03:00
} else if ( segtype_is_integrity ( segtype ) ) {
if ( ! dm_get_status_integrity ( seg_status - > mem , params , & ( seg_status - > integrity ) ) )
return_0 ;
seg_status - > type = SEG_STATUS_INTEGRITY ;
2016-12-02 15:57:52 +03:00
} else
2016-12-05 12:20:42 +03:00
/*
* TODO : Add support for other segment types too !
* Status not supported
*/
2016-12-02 15:57:52 +03:00
seg_status - > type = SEG_STATUS_NONE ;
2014-11-04 17:00:32 +03:00
return 1 ;
}
typedef enum {
INFO , /* DM_DEVICE_INFO ioctl */
STATUS , /* DM_DEVICE_STATUS ioctl */
} info_type_t ;
2017-02-24 02:50:00 +03:00
/* Return length of segment depending on type and reshape_len */
static uint32_t _seg_len ( const struct lv_segment * seg )
{
uint32_t reshape_len = seg_is_raid ( seg ) ? ( ( seg - > area_count - seg - > segtype - > parity_devs ) * seg - > reshape_len ) : 0 ;
return seg - > len - reshape_len ;
}
2016-12-18 16:57:16 +03:00
static int _info_run ( const char * dlid , struct dm_info * dminfo ,
uint32_t * read_ahead ,
2014-11-04 17:00:32 +03:00
struct lv_seg_status * seg_status ,
2019-10-31 13:45:28 +03:00
const char * name_check ,
2014-11-04 17:00:32 +03:00
int with_open_count , int with_read_ahead ,
uint32_t major , uint32_t minor )
2002-03-15 00:17:30 +03:00
{
int r = 0 ;
struct dm_task * dmt ;
2003-11-13 17:11:41 +03:00
int dmtask ;
2016-12-18 16:57:16 +03:00
int with_flush ; /* TODO: arg for _info_run */
2014-11-04 17:00:32 +03:00
void * target = NULL ;
2022-10-26 15:38:29 +03:00
uint64_t target_start , target_length , start , extent_size , length , length_crop = 0 ;
2016-12-01 12:37:03 +03:00
char * target_name , * target_params ;
2019-11-01 21:35:28 +03:00
const char * devname ;
2014-11-04 17:00:32 +03:00
2016-12-18 16:57:16 +03:00
if ( seg_status ) {
dmtask = DM_DEVICE_STATUS ;
with_flush = 0 ;
} else {
dmtask = DM_DEVICE_INFO ;
with_flush = 1 ; /* doesn't really matter */
2014-11-04 17:00:32 +03:00
}
2002-03-15 00:17:30 +03:00
2016-12-18 16:57:16 +03:00
if ( ! ( dmt = _setup_task_run ( dmtask , dminfo , NULL , dlid , 0 , major , minor ,
with_open_count , with_flush , 0 ) ) )
2008-01-30 16:19:47 +03:00
return_0 ;
2002-03-15 00:17:30 +03:00
2019-10-31 13:45:28 +03:00
if ( name_check & & dminfo - > exists & &
2019-11-01 21:35:28 +03:00
( devname = dm_task_get_name ( dmt ) ) & &
( strcmp ( name_check , devname ) ! = 0 ) )
2019-10-31 13:45:28 +03:00
dminfo - > exists = 0 ; /* mismatching name -> device does not exist */
2019-11-01 23:01:51 +03:00
if ( with_read_ahead & & read_ahead & & dminfo - > exists ) {
2007-12-03 21:00:38 +03:00
if ( ! dm_task_get_read_ahead ( dmt , read_ahead ) )
goto_out ;
} else if ( read_ahead )
2007-11-29 18:04:12 +03:00
* read_ahead = DM_READ_AHEAD_NONE ;
2007-11-12 23:51:54 +03:00
2016-12-01 12:37:03 +03:00
/* Query status only for active device */
2016-12-18 16:57:16 +03:00
if ( seg_status & & dminfo - > exists ) {
2022-10-26 15:38:29 +03:00
extent_size = length = seg_status - > seg - > lv - > vg - > extent_size ;
start = extent_size * seg_status - > seg - > le ;
2017-02-24 02:50:00 +03:00
length * = _seg_len ( seg_status - > seg ) ;
2016-12-01 12:37:03 +03:00
2017-04-08 20:38:17 +03:00
/* Uses max DM_THIN_MAX_METADATA_SIZE sectors for metadata device */
if ( lv_is_thin_pool_metadata ( seg_status - > seg - > lv ) & &
( length > DM_THIN_MAX_METADATA_SIZE ) )
2021-01-12 19:59:29 +03:00
length_crop = DM_THIN_MAX_METADATA_SIZE ;
2017-04-08 20:38:17 +03:00
2018-06-29 12:11:14 +03:00
/* Uses virtual size with headers for VDO pool device */
if ( lv_is_vdo_pool ( seg_status - > seg - > lv ) )
length = get_vdo_pool_virtual_size ( seg_status - > seg ) ;
2019-11-21 01:07:27 +03:00
if ( lv_is_integrity ( seg_status - > seg - > lv ) )
length = seg_status - > seg - > integrity_data_sectors ;
2014-11-04 17:00:32 +03:00
do {
target = dm_get_next_target ( dmt , target , & target_start ,
& target_length , & target_name , & target_params ) ;
2016-12-01 12:37:03 +03:00
2021-01-12 19:59:29 +03:00
if ( ( start = = target_start ) & &
( ( length = = target_length ) | |
2022-10-26 15:38:29 +03:00
( ( lv_is_vdo_pool ( seg_status - > seg - > lv ) ) & & /* should fit within extent size */
( length < target_length ) & & ( ( length + extent_size ) > target_length ) ) | |
2021-01-12 19:59:29 +03:00
( length_crop & & ( length_crop = = target_length ) ) ) )
2016-12-01 12:37:03 +03:00
break ; /* Keep target_params when matching segment is found */
target_params = NULL ; /* Marking this target_params unusable */
2014-11-04 17:00:32 +03:00
} while ( target ) ;
2017-06-27 00:34:38 +03:00
if ( ! target_name | |
2021-09-09 15:59:38 +03:00
! _get_segment_status_from_target_params ( target_name , target_params , dminfo , seg_status ) )
2016-12-01 12:37:03 +03:00
stack ;
2014-11-04 17:00:32 +03:00
}
2002-03-15 00:17:30 +03:00
r = 1 ;
out :
dm_task_destroy ( dmt ) ;
2016-12-01 16:53:35 +03:00
2002-03-15 00:17:30 +03:00
return r ;
}
2002-02-25 17:46:57 +03:00
mirror: Avoid reading from mirrors that have failed devices
Addresses: rhbz855398 (Allow VGs to be built on cluster mirrors),
and other issues.
The LVM code attempts to avoid reading labels from devices that are
suspended to try to avoid situations that may cause the commands to
block indefinitely. When scanning devices, 'ignore_suspended_devices'
can be set so the code (lib/activate/dev_manager.c:device_is_usable())
checks any DM devices it finds and avoids them if they are suspended.
The mirror target has an additional mechanism that can cause I/O to
be blocked. If a device in a mirror fails, all I/O will be blocked
by the kernel until a new table (a linear target or a mirror with
replacement devices) is loaded. The mirror indicates that this condition
has happened by marking a 'D' for the faulty device in its status
output. This condition must also be checked by 'device_is_usable()' to
avoid the possibility of blocking LVM commands indefinitely due to an
attempt to read the blocked mirror for labels.
Until now, mirrors were avoided if the 'ignore_suspended_devices'
condition was set. This check seemed to suggest, "if we are concerned
about suspended devices, then let's ignore mirrors altogether just
in case". This is insufficient and doesn't solve any problems. All
devices that are suspended are already avoided if
'ignore_suspended_devices' is set; and if a mirror is blocking because
of an error condition, it will block the LVM command regardless of the
setting of that variable.
Rather than avoiding mirrors whenever 'ignore_suspended_devices' is
set, this patch causes mirrors to be avoided whenever they are blocking
due to an error. (As mentioned above, the case where a DM device is
suspended is already covered.) This solves a number of issues that weren't
handled before. For example, pvcreate (or any command that does a
pv_read or vg_read, which eventually call device_is_usable()) will be
protected from blocked mirrors regardless of how
'ignore_suspended_devices' is set. Additionally, a mirror that is
neither suspended nor blocking is /allowed/ to be read regardless
of how 'ignore_suspended_devices' is set. (The latter point being the
source of the fix for rhbz855398.)
2012-10-24 08:10:33 +04:00
/*
* ignore_blocked_mirror_devices
* @ dev
* @ start
* @ length
* @ mirror_status_str
*
* When a DM ' mirror ' target is created with ' block_on_error ' or
* ' handle_errors ' , it will block I / O if there is a device failure
* until the mirror is reconfigured . Thus , LVM should never attempt
* to read labels from a mirror that has a failed device . ( LVM
* commands are issued to repair mirrors ; and if LVM is blocked
* attempting to read a mirror , a circular dependency would be created . )
*
* This function is a slimmed - down version of lib / mirror / mirrored . c :
2012-10-25 09:42:45 +04:00
* _mirrored_transient_status ( ) .
mirror: Avoid reading from mirrors that have failed devices
Addresses: rhbz855398 (Allow VGs to be built on cluster mirrors),
and other issues.
The LVM code attempts to avoid reading labels from devices that are
suspended to try to avoid situations that may cause the commands to
block indefinitely. When scanning devices, 'ignore_suspended_devices'
can be set so the code (lib/activate/dev_manager.c:device_is_usable())
checks any DM devices it finds and avoids them if they are suspended.
The mirror target has an additional mechanism that can cause I/O to
be blocked. If a device in a mirror fails, all I/O will be blocked
by the kernel until a new table (a linear target or a mirror with
replacement devices) is loaded. The mirror indicates that this condition
has happened by marking a 'D' for the faulty device in its status
output. This condition must also be checked by 'device_is_usable()' to
avoid the possibility of blocking LVM commands indefinitely due to an
attempt to read the blocked mirror for labels.
Until now, mirrors were avoided if the 'ignore_suspended_devices'
condition was set. This check seemed to suggest, "if we are concerned
about suspended devices, then let's ignore mirrors altogether just
in case". This is insufficient and doesn't solve any problems. All
devices that are suspended are already avoided if
'ignore_suspended_devices' is set; and if a mirror is blocking because
of an error condition, it will block the LVM command regardless of the
setting of that variable.
Rather than avoiding mirrors whenever 'ignore_suspended_devices' is
set, this patch causes mirrors to be avoided whenever they are blocking
due to an error. (As mentioned above, the case where a DM device is
suspended is already covered.) This solves a number of issues that weren't
handled before. For example, pvcreate (or any command that does a
pv_read or vg_read, which eventually call device_is_usable()) will be
protected from blocked mirrors regardless of how
'ignore_suspended_devices' is set. Additionally, a mirror that is
neither suspended nor blocking is /allowed/ to be read regardless
of how 'ignore_suspended_devices' is set. (The latter point being the
source of the fix for rhbz855398.)
2012-10-24 08:10:33 +04:00
*
* If a failed device is detected in the status string , then it must be
* determined if ' block_on_error ' or ' handle_errors ' was used when
* creating the mirror . This info can only be determined from the mirror
* table . The ' dev ' , ' start ' , ' length ' trio allow us to correlate the
* ' mirror_status_str ' with the correct device table in order to check
* for blocking .
*
* Returns : 1 if mirror should be ignored , 0 if safe to use
*/
2021-12-15 13:26:21 +03:00
static int _ignore_blocked_mirror_devices ( struct cmd_context * cmd ,
struct device * dev ,
mirror: Avoid reading from mirrors that have failed devices
Addresses: rhbz855398 (Allow VGs to be built on cluster mirrors),
and other issues.
The LVM code attempts to avoid reading labels from devices that are
suspended to try to avoid situations that may cause the commands to
block indefinitely. When scanning devices, 'ignore_suspended_devices'
can be set so the code (lib/activate/dev_manager.c:device_is_usable())
checks any DM devices it finds and avoids them if they are suspended.
The mirror target has an additional mechanism that can cause I/O to
be blocked. If a device in a mirror fails, all I/O will be blocked
by the kernel until a new table (a linear target or a mirror with
replacement devices) is loaded. The mirror indicates that this condition
has happened by marking a 'D' for the faulty device in its status
output. This condition must also be checked by 'device_is_usable()' to
avoid the possibility of blocking LVM commands indefinitely due to an
attempt to read the blocked mirror for labels.
Until now, mirrors were avoided if the 'ignore_suspended_devices'
condition was set. This check seemed to suggest, "if we are concerned
about suspended devices, then let's ignore mirrors altogether just
in case". This is insufficient and doesn't solve any problems. All
devices that are suspended are already avoided if
'ignore_suspended_devices' is set; and if a mirror is blocking because
of an error condition, it will block the LVM command regardless of the
setting of that variable.
Rather than avoiding mirrors whenever 'ignore_suspended_devices' is
set, this patch causes mirrors to be avoided whenever they are blocking
due to an error. (As mentioned above, the case where a DM device is
suspended is already covered.) This solves a number of issues that weren't
handled before. For example, pvcreate (or any command that does a
pv_read or vg_read, which eventually call device_is_usable()) will be
protected from blocked mirrors regardless of how
'ignore_suspended_devices' is set. Additionally, a mirror that is
neither suspended nor blocking is /allowed/ to be read regardless
of how 'ignore_suspended_devices' is set. (The latter point being the
source of the fix for rhbz855398.)
2012-10-24 08:10:33 +04:00
uint64_t start , uint64_t length ,
char * mirror_status_str )
{
2015-11-30 23:17:25 +03:00
struct dm_pool * mem ;
struct dm_status_mirror * sm ;
mirror: Avoid reading from mirrors that have failed devices
Addresses: rhbz855398 (Allow VGs to be built on cluster mirrors),
and other issues.
The LVM code attempts to avoid reading labels from devices that are
suspended to try to avoid situations that may cause the commands to
block indefinitely. When scanning devices, 'ignore_suspended_devices'
can be set so the code (lib/activate/dev_manager.c:device_is_usable())
checks any DM devices it finds and avoids them if they are suspended.
The mirror target has an additional mechanism that can cause I/O to
be blocked. If a device in a mirror fails, all I/O will be blocked
by the kernel until a new table (a linear target or a mirror with
replacement devices) is loaded. The mirror indicates that this condition
has happened by marking a 'D' for the faulty device in its status
output. This condition must also be checked by 'device_is_usable()' to
avoid the possibility of blocking LVM commands indefinitely due to an
attempt to read the blocked mirror for labels.
Until now, mirrors were avoided if the 'ignore_suspended_devices'
condition was set. This check seemed to suggest, "if we are concerned
about suspended devices, then let's ignore mirrors altogether just
in case". This is insufficient and doesn't solve any problems. All
devices that are suspended are already avoided if
'ignore_suspended_devices' is set; and if a mirror is blocking because
of an error condition, it will block the LVM command regardless of the
setting of that variable.
Rather than avoiding mirrors whenever 'ignore_suspended_devices' is
set, this patch causes mirrors to be avoided whenever they are blocking
due to an error. (As mentioned above, the case where a DM device is
suspended is already covered.) This solves a number of issues that weren't
handled before. For example, pvcreate (or any command that does a
pv_read or vg_read, which eventually call device_is_usable()) will be
protected from blocked mirrors regardless of how
'ignore_suspended_devices' is set. Additionally, a mirror that is
neither suspended nor blocking is /allowed/ to be read regardless
of how 'ignore_suspended_devices' is set. (The latter point being the
source of the fix for rhbz855398.)
2012-10-24 08:10:33 +04:00
unsigned i , check_for_blocking = 0 ;
uint64_t s , l ;
2013-04-12 20:30:04 +04:00
char * p , * params , * target_type = NULL ;
mirror: Avoid reading from mirrors that have failed devices
Addresses: rhbz855398 (Allow VGs to be built on cluster mirrors),
and other issues.
The LVM code attempts to avoid reading labels from devices that are
suspended to try to avoid situations that may cause the commands to
block indefinitely. When scanning devices, 'ignore_suspended_devices'
can be set so the code (lib/activate/dev_manager.c:device_is_usable())
checks any DM devices it finds and avoids them if they are suspended.
The mirror target has an additional mechanism that can cause I/O to
be blocked. If a device in a mirror fails, all I/O will be blocked
by the kernel until a new table (a linear target or a mirror with
replacement devices) is loaded. The mirror indicates that this condition
has happened by marking a 'D' for the faulty device in its status
output. This condition must also be checked by 'device_is_usable()' to
avoid the possibility of blocking LVM commands indefinitely due to an
attempt to read the blocked mirror for labels.
Until now, mirrors were avoided if the 'ignore_suspended_devices'
condition was set. This check seemed to suggest, "if we are concerned
about suspended devices, then let's ignore mirrors altogether just
in case". This is insufficient and doesn't solve any problems. All
devices that are suspended are already avoided if
'ignore_suspended_devices' is set; and if a mirror is blocking because
of an error condition, it will block the LVM command regardless of the
setting of that variable.
Rather than avoiding mirrors whenever 'ignore_suspended_devices' is
set, this patch causes mirrors to be avoided whenever they are blocking
due to an error. (As mentioned above, the case where a DM device is
suspended is already covered.) This solves a number of issues that weren't
handled before. For example, pvcreate (or any command that does a
pv_read or vg_read, which eventually call device_is_usable()) will be
protected from blocked mirrors regardless of how
'ignore_suspended_devices' is set. Additionally, a mirror that is
neither suspended nor blocking is /allowed/ to be read regardless
of how 'ignore_suspended_devices' is set. (The latter point being the
source of the fix for rhbz855398.)
2012-10-24 08:10:33 +04:00
void * next = NULL ;
2012-12-07 02:37:21 +04:00
struct dm_task * dmt = NULL ;
int r = 0 ;
2024-07-02 23:07:43 +03:00
char fake_dev_name [ 16 ] ;
2024-08-22 15:45:35 +03:00
struct device fake_dev = { . fd = 0 } ;
2024-07-02 23:07:43 +03:00
struct dm_str_list * alias ;
mirror: Avoid reading from mirrors that have failed devices
Addresses: rhbz855398 (Allow VGs to be built on cluster mirrors),
and other issues.
The LVM code attempts to avoid reading labels from devices that are
suspended to try to avoid situations that may cause the commands to
block indefinitely. When scanning devices, 'ignore_suspended_devices'
can be set so the code (lib/activate/dev_manager.c:device_is_usable())
checks any DM devices it finds and avoids them if they are suspended.
The mirror target has an additional mechanism that can cause I/O to
be blocked. If a device in a mirror fails, all I/O will be blocked
by the kernel until a new table (a linear target or a mirror with
replacement devices) is loaded. The mirror indicates that this condition
has happened by marking a 'D' for the faulty device in its status
output. This condition must also be checked by 'device_is_usable()' to
avoid the possibility of blocking LVM commands indefinitely due to an
attempt to read the blocked mirror for labels.
Until now, mirrors were avoided if the 'ignore_suspended_devices'
condition was set. This check seemed to suggest, "if we are concerned
about suspended devices, then let's ignore mirrors altogether just
in case". This is insufficient and doesn't solve any problems. All
devices that are suspended are already avoided if
'ignore_suspended_devices' is set; and if a mirror is blocking because
of an error condition, it will block the LVM command regardless of the
setting of that variable.
Rather than avoiding mirrors whenever 'ignore_suspended_devices' is
set, this patch causes mirrors to be avoided whenever they are blocking
due to an error. (As mentioned above, the case where a DM device is
suspended is already covered.) This solves a number of issues that weren't
handled before. For example, pvcreate (or any command that does a
pv_read or vg_read, which eventually call device_is_usable()) will be
protected from blocked mirrors regardless of how
'ignore_suspended_devices' is set. Additionally, a mirror that is
neither suspended nor blocking is /allowed/ to be read regardless
of how 'ignore_suspended_devices' is set. (The latter point being the
source of the fix for rhbz855398.)
2012-10-24 08:10:33 +04:00
2015-11-30 23:17:25 +03:00
if ( ! ( mem = dm_pool_create ( " blocked_mirrors " , 128 ) ) )
2012-12-07 02:37:21 +04:00
return_0 ;
mirror: Avoid reading from mirrors that have failed devices
Addresses: rhbz855398 (Allow VGs to be built on cluster mirrors),
and other issues.
The LVM code attempts to avoid reading labels from devices that are
suspended to try to avoid situations that may cause the commands to
block indefinitely. When scanning devices, 'ignore_suspended_devices'
can be set so the code (lib/activate/dev_manager.c:device_is_usable())
checks any DM devices it finds and avoids them if they are suspended.
The mirror target has an additional mechanism that can cause I/O to
be blocked. If a device in a mirror fails, all I/O will be blocked
by the kernel until a new table (a linear target or a mirror with
replacement devices) is loaded. The mirror indicates that this condition
has happened by marking a 'D' for the faulty device in its status
output. This condition must also be checked by 'device_is_usable()' to
avoid the possibility of blocking LVM commands indefinitely due to an
attempt to read the blocked mirror for labels.
Until now, mirrors were avoided if the 'ignore_suspended_devices'
condition was set. This check seemed to suggest, "if we are concerned
about suspended devices, then let's ignore mirrors altogether just
in case". This is insufficient and doesn't solve any problems. All
devices that are suspended are already avoided if
'ignore_suspended_devices' is set; and if a mirror is blocking because
of an error condition, it will block the LVM command regardless of the
setting of that variable.
Rather than avoiding mirrors whenever 'ignore_suspended_devices' is
set, this patch causes mirrors to be avoided whenever they are blocking
due to an error. (As mentioned above, the case where a DM device is
suspended is already covered.) This solves a number of issues that weren't
handled before. For example, pvcreate (or any command that does a
pv_read or vg_read, which eventually call device_is_usable()) will be
protected from blocked mirrors regardless of how
'ignore_suspended_devices' is set. Additionally, a mirror that is
neither suspended nor blocking is /allowed/ to be read regardless
of how 'ignore_suspended_devices' is set. (The latter point being the
source of the fix for rhbz855398.)
2012-10-24 08:10:33 +04:00
2015-11-30 23:17:25 +03:00
if ( ! dm_get_status_mirror ( mem , mirror_status_str , & sm ) )
goto_out ;
for ( i = 0 ; i < sm - > dev_count ; + + i )
if ( sm - > devs [ i ] . health ! = DM_STATUS_MIRROR_ALIVE ) {
2016-12-01 16:53:35 +03:00
log_debug_activation ( " %s: Mirror image %d marked as failed. " ,
2013-01-08 02:30:29 +04:00
dev_name ( dev ) , i ) ;
mirror: Avoid reading from mirrors that have failed devices
Addresses: rhbz855398 (Allow VGs to be built on cluster mirrors),
and other issues.
The LVM code attempts to avoid reading labels from devices that are
suspended to try to avoid situations that may cause the commands to
block indefinitely. When scanning devices, 'ignore_suspended_devices'
can be set so the code (lib/activate/dev_manager.c:device_is_usable())
checks any DM devices it finds and avoids them if they are suspended.
The mirror target has an additional mechanism that can cause I/O to
be blocked. If a device in a mirror fails, all I/O will be blocked
by the kernel until a new table (a linear target or a mirror with
replacement devices) is loaded. The mirror indicates that this condition
has happened by marking a 'D' for the faulty device in its status
output. This condition must also be checked by 'device_is_usable()' to
avoid the possibility of blocking LVM commands indefinitely due to an
attempt to read the blocked mirror for labels.
Until now, mirrors were avoided if the 'ignore_suspended_devices'
condition was set. This check seemed to suggest, "if we are concerned
about suspended devices, then let's ignore mirrors altogether just
in case". This is insufficient and doesn't solve any problems. All
devices that are suspended are already avoided if
'ignore_suspended_devices' is set; and if a mirror is blocking because
of an error condition, it will block the LVM command regardless of the
setting of that variable.
Rather than avoiding mirrors whenever 'ignore_suspended_devices' is
set, this patch causes mirrors to be avoided whenever they are blocking
due to an error. (As mentioned above, the case where a DM device is
suspended is already covered.) This solves a number of issues that weren't
handled before. For example, pvcreate (or any command that does a
pv_read or vg_read, which eventually call device_is_usable()) will be
protected from blocked mirrors regardless of how
'ignore_suspended_devices' is set. Additionally, a mirror that is
neither suspended nor blocking is /allowed/ to be read regardless
of how 'ignore_suspended_devices' is set. (The latter point being the
source of the fix for rhbz855398.)
2012-10-24 08:10:33 +04:00
check_for_blocking = 1 ;
}
2015-11-30 23:17:25 +03:00
if ( ! check_for_blocking & & sm - > log_count ) {
if ( sm - > logs [ 0 ] . health ! = DM_STATUS_MIRROR_ALIVE ) {
2016-12-01 16:53:35 +03:00
log_debug_activation ( " %s: Mirror log device marked as failed. " ,
2013-01-08 02:30:29 +04:00
dev_name ( dev ) ) ;
2012-10-25 09:42:45 +04:00
check_for_blocking = 1 ;
} else {
2024-07-02 23:07:43 +03:00
dev_init ( & fake_dev ) ;
if ( dm_snprintf ( fake_dev_name , sizeof ( fake_dev_name ) , " %u:%u " ,
2015-11-30 23:17:25 +03:00
sm - > logs [ 0 ] . major , sm - > logs [ 0 ] . minor ) < 0 )
2012-10-25 09:42:45 +04:00
goto_out ;
2024-07-02 23:07:43 +03:00
if ( ! ( alias = dm_pool_zalloc ( mem , sizeof ( * alias ) ) ) )
goto_out ;
if ( ! ( alias - > str = dm_pool_strdup ( mem , fake_dev_name ) ) )
2012-10-25 09:42:45 +04:00
goto_out ;
2024-07-02 23:07:43 +03:00
dm_list_add ( & fake_dev . aliases , & alias - > list ) ;
fake_dev . flags = DEV_REGULAR ;
fake_dev . dev = MKDEV ( sm - > logs [ 0 ] . major , sm - > logs [ 0 ] . minor ) ;
2012-10-25 09:42:45 +04:00
2024-07-02 23:07:43 +03:00
if ( dm_device_is_usable ( cmd , & fake_dev , ( struct dev_usable_check_params )
2024-06-28 21:56:12 +03:00
{ . check_empty = 1 ,
. check_blocked = 1 ,
. check_suspended = ignore_suspended_devices ( ) ,
. check_error_target = 1 ,
. check_reserved = 0 } , NULL ) )
2015-12-05 00:02:24 +03:00
goto out ; /* safe to use */
stack ;
2012-10-25 09:42:45 +04:00
}
}
2012-12-07 02:37:21 +04:00
if ( ! check_for_blocking ) {
r = 1 ;
goto out ;
}
mirror: Avoid reading from mirrors that have failed devices
Addresses: rhbz855398 (Allow VGs to be built on cluster mirrors),
and other issues.
The LVM code attempts to avoid reading labels from devices that are
suspended to try to avoid situations that may cause the commands to
block indefinitely. When scanning devices, 'ignore_suspended_devices'
can be set so the code (lib/activate/dev_manager.c:device_is_usable())
checks any DM devices it finds and avoids them if they are suspended.
The mirror target has an additional mechanism that can cause I/O to
be blocked. If a device in a mirror fails, all I/O will be blocked
by the kernel until a new table (a linear target or a mirror with
replacement devices) is loaded. The mirror indicates that this condition
has happened by marking a 'D' for the faulty device in its status
output. This condition must also be checked by 'device_is_usable()' to
avoid the possibility of blocking LVM commands indefinitely due to an
attempt to read the blocked mirror for labels.
Until now, mirrors were avoided if the 'ignore_suspended_devices'
condition was set. This check seemed to suggest, "if we are concerned
about suspended devices, then let's ignore mirrors altogether just
in case". This is insufficient and doesn't solve any problems. All
devices that are suspended are already avoided if
'ignore_suspended_devices' is set; and if a mirror is blocking because
of an error condition, it will block the LVM command regardless of the
setting of that variable.
Rather than avoiding mirrors whenever 'ignore_suspended_devices' is
set, this patch causes mirrors to be avoided whenever they are blocking
due to an error. (As mentioned above, the case where a DM device is
suspended is already covered.) This solves a number of issues that weren't
handled before. For example, pvcreate (or any command that does a
pv_read or vg_read, which eventually call device_is_usable()) will be
protected from blocked mirrors regardless of how
'ignore_suspended_devices' is set. Additionally, a mirror that is
neither suspended nor blocking is /allowed/ to be read regardless
of how 'ignore_suspended_devices' is set. (The latter point being the
source of the fix for rhbz855398.)
2012-10-24 08:10:33 +04:00
/*
* We avoid another system call if we can , but if a device is
* dead , we have no choice but to look up the table too .
*/
2016-12-01 16:56:32 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_TABLE , NULL , NULL , NULL , NULL ,
MAJOR ( dev - > dev ) , MINOR ( dev - > dev ) , 0 , 1 , 0 ) ) )
mirror: Avoid reading from mirrors that have failed devices
Addresses: rhbz855398 (Allow VGs to be built on cluster mirrors),
and other issues.
The LVM code attempts to avoid reading labels from devices that are
suspended to try to avoid situations that may cause the commands to
block indefinitely. When scanning devices, 'ignore_suspended_devices'
can be set so the code (lib/activate/dev_manager.c:device_is_usable())
checks any DM devices it finds and avoids them if they are suspended.
The mirror target has an additional mechanism that can cause I/O to
be blocked. If a device in a mirror fails, all I/O will be blocked
by the kernel until a new table (a linear target or a mirror with
replacement devices) is loaded. The mirror indicates that this condition
has happened by marking a 'D' for the faulty device in its status
output. This condition must also be checked by 'device_is_usable()' to
avoid the possibility of blocking LVM commands indefinitely due to an
attempt to read the blocked mirror for labels.
Until now, mirrors were avoided if the 'ignore_suspended_devices'
condition was set. This check seemed to suggest, "if we are concerned
about suspended devices, then let's ignore mirrors altogether just
in case". This is insufficient and doesn't solve any problems. All
devices that are suspended are already avoided if
'ignore_suspended_devices' is set; and if a mirror is blocking because
of an error condition, it will block the LVM command regardless of the
setting of that variable.
Rather than avoiding mirrors whenever 'ignore_suspended_devices' is
set, this patch causes mirrors to be avoided whenever they are blocking
due to an error. (As mentioned above, the case where a DM device is
suspended is already covered.) This solves a number of issues that weren't
handled before. For example, pvcreate (or any command that does a
pv_read or vg_read, which eventually call device_is_usable()) will be
protected from blocked mirrors regardless of how
'ignore_suspended_devices' is set. Additionally, a mirror that is
neither suspended nor blocking is /allowed/ to be read regardless
of how 'ignore_suspended_devices' is set. (The latter point being the
source of the fix for rhbz855398.)
2012-10-24 08:10:33 +04:00
goto_out ;
do {
next = dm_get_next_target ( dmt , next , & s , & l ,
& target_type , & params ) ;
2016-02-23 22:16:06 +03:00
if ( ( s = = start ) & & ( l = = length ) & &
target_type & & params ) {
2016-03-22 20:46:15 +03:00
if ( strcmp ( target_type , TARGET_NAME_MIRROR ) )
mirror: Avoid reading from mirrors that have failed devices
Addresses: rhbz855398 (Allow VGs to be built on cluster mirrors),
and other issues.
The LVM code attempts to avoid reading labels from devices that are
suspended to try to avoid situations that may cause the commands to
block indefinitely. When scanning devices, 'ignore_suspended_devices'
can be set so the code (lib/activate/dev_manager.c:device_is_usable())
checks any DM devices it finds and avoids them if they are suspended.
The mirror target has an additional mechanism that can cause I/O to
be blocked. If a device in a mirror fails, all I/O will be blocked
by the kernel until a new table (a linear target or a mirror with
replacement devices) is loaded. The mirror indicates that this condition
has happened by marking a 'D' for the faulty device in its status
output. This condition must also be checked by 'device_is_usable()' to
avoid the possibility of blocking LVM commands indefinitely due to an
attempt to read the blocked mirror for labels.
Until now, mirrors were avoided if the 'ignore_suspended_devices'
condition was set. This check seemed to suggest, "if we are concerned
about suspended devices, then let's ignore mirrors altogether just
in case". This is insufficient and doesn't solve any problems. All
devices that are suspended are already avoided if
'ignore_suspended_devices' is set; and if a mirror is blocking because
of an error condition, it will block the LVM command regardless of the
setting of that variable.
Rather than avoiding mirrors whenever 'ignore_suspended_devices' is
set, this patch causes mirrors to be avoided whenever they are blocking
due to an error. (As mentioned above, the case where a DM device is
suspended is already covered.) This solves a number of issues that weren't
handled before. For example, pvcreate (or any command that does a
pv_read or vg_read, which eventually call device_is_usable()) will be
protected from blocked mirrors regardless of how
'ignore_suspended_devices' is set. Additionally, a mirror that is
neither suspended nor blocking is /allowed/ to be read regardless
of how 'ignore_suspended_devices' is set. (The latter point being the
source of the fix for rhbz855398.)
2012-10-24 08:10:33 +04:00
goto_out ;
2013-04-12 20:30:04 +04:00
if ( ( ( p = strstr ( params , " block_on_error " ) ) & &
( p [ 15 ] = = ' \0 ' | | p [ 15 ] = = ' ' ) ) | |
( ( p = strstr ( params , " handle_errors " ) ) & &
( p [ 14 ] = = ' \0 ' | | p [ 14 ] = = ' ' ) ) ) {
2016-12-01 16:53:35 +03:00
log_debug_activation ( " %s: I/O blocked to mirror device. " ,
2013-01-08 02:30:29 +04:00
dev_name ( dev ) ) ;
2012-12-07 02:37:21 +04:00
goto out ;
mirror: Avoid reading from mirrors that have failed devices
Addresses: rhbz855398 (Allow VGs to be built on cluster mirrors),
and other issues.
The LVM code attempts to avoid reading labels from devices that are
suspended to try to avoid situations that may cause the commands to
block indefinitely. When scanning devices, 'ignore_suspended_devices'
can be set so the code (lib/activate/dev_manager.c:device_is_usable())
checks any DM devices it finds and avoids them if they are suspended.
The mirror target has an additional mechanism that can cause I/O to
be blocked. If a device in a mirror fails, all I/O will be blocked
by the kernel until a new table (a linear target or a mirror with
replacement devices) is loaded. The mirror indicates that this condition
has happened by marking a 'D' for the faulty device in its status
output. This condition must also be checked by 'device_is_usable()' to
avoid the possibility of blocking LVM commands indefinitely due to an
attempt to read the blocked mirror for labels.
Until now, mirrors were avoided if the 'ignore_suspended_devices'
condition was set. This check seemed to suggest, "if we are concerned
about suspended devices, then let's ignore mirrors altogether just
in case". This is insufficient and doesn't solve any problems. All
devices that are suspended are already avoided if
'ignore_suspended_devices' is set; and if a mirror is blocking because
of an error condition, it will block the LVM command regardless of the
setting of that variable.
Rather than avoiding mirrors whenever 'ignore_suspended_devices' is
set, this patch causes mirrors to be avoided whenever they are blocking
due to an error. (As mentioned above, the case where a DM device is
suspended is already covered.) This solves a number of issues that weren't
handled before. For example, pvcreate (or any command that does a
pv_read or vg_read, which eventually call device_is_usable()) will be
protected from blocked mirrors regardless of how
'ignore_suspended_devices' is set. Additionally, a mirror that is
neither suspended nor blocking is /allowed/ to be read regardless
of how 'ignore_suspended_devices' is set. (The latter point being the
source of the fix for rhbz855398.)
2012-10-24 08:10:33 +04:00
}
}
} while ( next ) ;
2012-12-07 02:37:21 +04:00
r = 1 ;
mirror: Avoid reading from mirrors that have failed devices
Addresses: rhbz855398 (Allow VGs to be built on cluster mirrors),
and other issues.
The LVM code attempts to avoid reading labels from devices that are
suspended to try to avoid situations that may cause the commands to
block indefinitely. When scanning devices, 'ignore_suspended_devices'
can be set so the code (lib/activate/dev_manager.c:device_is_usable())
checks any DM devices it finds and avoids them if they are suspended.
The mirror target has an additional mechanism that can cause I/O to
be blocked. If a device in a mirror fails, all I/O will be blocked
by the kernel until a new table (a linear target or a mirror with
replacement devices) is loaded. The mirror indicates that this condition
has happened by marking a 'D' for the faulty device in its status
output. This condition must also be checked by 'device_is_usable()' to
avoid the possibility of blocking LVM commands indefinitely due to an
attempt to read the blocked mirror for labels.
Until now, mirrors were avoided if the 'ignore_suspended_devices'
condition was set. This check seemed to suggest, "if we are concerned
about suspended devices, then let's ignore mirrors altogether just
in case". This is insufficient and doesn't solve any problems. All
devices that are suspended are already avoided if
'ignore_suspended_devices' is set; and if a mirror is blocking because
of an error condition, it will block the LVM command regardless of the
setting of that variable.
Rather than avoiding mirrors whenever 'ignore_suspended_devices' is
set, this patch causes mirrors to be avoided whenever they are blocking
due to an error. (As mentioned above, the case where a DM device is
suspended is already covered.) This solves a number of issues that weren't
handled before. For example, pvcreate (or any command that does a
pv_read or vg_read, which eventually call device_is_usable()) will be
protected from blocked mirrors regardless of how
'ignore_suspended_devices' is set. Additionally, a mirror that is
neither suspended nor blocking is /allowed/ to be read regardless
of how 'ignore_suspended_devices' is set. (The latter point being the
source of the fix for rhbz855398.)
2012-10-24 08:10:33 +04:00
out :
2012-12-07 02:37:21 +04:00
if ( dmt )
dm_task_destroy ( dmt ) ;
2015-11-30 23:17:25 +03:00
dm_pool_destroy ( mem ) ;
2012-12-07 02:37:21 +04:00
return r ;
mirror: Avoid reading from mirrors that have failed devices
Addresses: rhbz855398 (Allow VGs to be built on cluster mirrors),
and other issues.
The LVM code attempts to avoid reading labels from devices that are
suspended to try to avoid situations that may cause the commands to
block indefinitely. When scanning devices, 'ignore_suspended_devices'
can be set so the code (lib/activate/dev_manager.c:device_is_usable())
checks any DM devices it finds and avoids them if they are suspended.
The mirror target has an additional mechanism that can cause I/O to
be blocked. If a device in a mirror fails, all I/O will be blocked
by the kernel until a new table (a linear target or a mirror with
replacement devices) is loaded. The mirror indicates that this condition
has happened by marking a 'D' for the faulty device in its status
output. This condition must also be checked by 'device_is_usable()' to
avoid the possibility of blocking LVM commands indefinitely due to an
attempt to read the blocked mirror for labels.
Until now, mirrors were avoided if the 'ignore_suspended_devices'
condition was set. This check seemed to suggest, "if we are concerned
about suspended devices, then let's ignore mirrors altogether just
in case". This is insufficient and doesn't solve any problems. All
devices that are suspended are already avoided if
'ignore_suspended_devices' is set; and if a mirror is blocking because
of an error condition, it will block the LVM command regardless of the
setting of that variable.
Rather than avoiding mirrors whenever 'ignore_suspended_devices' is
set, this patch causes mirrors to be avoided whenever they are blocking
due to an error. (As mentioned above, the case where a DM device is
suspended is already covered.) This solves a number of issues that weren't
handled before. For example, pvcreate (or any command that does a
pv_read or vg_read, which eventually call device_is_usable()) will be
protected from blocked mirrors regardless of how
'ignore_suspended_devices' is set. Additionally, a mirror that is
neither suspended nor blocking is /allowed/ to be read regardless
of how 'ignore_suspended_devices' is set. (The latter point being the
source of the fix for rhbz855398.)
2012-10-24 08:10:33 +04:00
}
2015-06-17 14:37:53 +03:00
static int _device_is_suspended ( int major , int minor )
{
struct dm_task * dmt ;
struct dm_info info ;
2016-12-01 16:56:32 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_INFO , & info ,
NULL , NULL , NULL ,
major , minor , 0 , 0 , 0 ) ) )
2016-04-21 21:19:24 +03:00
return_0 ;
2015-06-17 14:37:53 +03:00
dm_task_destroy ( dmt ) ;
2016-12-01 16:56:32 +03:00
return ( info . exists & & info . suspended ) ;
2015-06-17 14:37:53 +03:00
}
static int _ignore_suspended_snapshot_component ( struct device * dev )
{
struct dm_task * dmt ;
void * next = NULL ;
char * params , * target_type = NULL ;
uint64_t start , length ;
int major1 , minor1 , major2 , minor2 ;
int r = 0 ;
2016-12-01 16:56:32 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_TABLE , NULL ,
NULL , NULL , NULL ,
MAJOR ( dev - > dev ) , MINOR ( dev - > dev ) , 0 , 1 , 0 ) ) )
2016-04-21 21:19:24 +03:00
return_0 ;
2015-06-17 14:37:53 +03:00
do {
next = dm_get_next_target ( dmt , next , & start , & length , & target_type , & params ) ;
2018-03-08 15:01:12 +03:00
if ( ! target_type )
continue ;
if ( ! strcmp ( target_type , TARGET_NAME_SNAPSHOT ) ) {
2015-11-09 11:27:18 +03:00
if ( ! params | | sscanf ( params , " %d:%d %d:%d " , & major1 , & minor1 , & major2 , & minor2 ) ! = 4 ) {
2024-05-25 23:36:18 +03:00
log_warn ( " WARNING: Incorrect snapshot table found for %u:%u. " ,
MAJOR ( dev - > dev ) , MINOR ( dev - > dev ) ) ;
2016-11-25 15:46:06 +03:00
goto out ;
2015-06-17 14:37:53 +03:00
}
2015-06-17 15:12:18 +03:00
r = r | | _device_is_suspended ( major1 , minor1 ) | | _device_is_suspended ( major2 , minor2 ) ;
2016-03-22 20:46:15 +03:00
} else if ( ! strcmp ( target_type , TARGET_NAME_SNAPSHOT_ORIGIN ) ) {
2015-11-09 11:27:18 +03:00
if ( ! params | | sscanf ( params , " %d:%d " , & major1 , & minor1 ) ! = 2 ) {
2024-05-25 23:36:18 +03:00
log_warn ( " WARNING: Incorrect snapshot-origin table found for %u:%u. " ,
MAJOR ( dev - > dev ) , MINOR ( dev - > dev ) ) ;
2016-11-25 15:46:06 +03:00
goto out ;
2015-06-17 14:37:53 +03:00
}
2015-06-17 15:12:18 +03:00
r = r | | _device_is_suspended ( major1 , minor1 ) ;
2015-06-17 14:37:53 +03:00
}
} while ( next ) ;
out :
dm_task_destroy ( dmt ) ;
2016-12-01 16:53:35 +03:00
2015-06-17 14:37:53 +03:00
return r ;
}
2015-09-02 17:13:31 +03:00
static int _ignore_unusable_thins ( struct device * dev )
{
/* TODO make function for thin testing */
struct dm_pool * mem ;
struct dm_status_thin_pool * status ;
struct dm_task * dmt = NULL ;
void * next = NULL ;
uint64_t start , length ;
char * target_type = NULL ;
char * params ;
int minor , major ;
int r = 0 ;
if ( ! ( mem = dm_pool_create ( " unusable_thins " , 128 ) ) )
return_0 ;
2016-12-01 16:56:32 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_TABLE , NULL , NULL , NULL , NULL ,
MAJOR ( dev - > dev ) , MINOR ( dev - > dev ) , 0 , 1 , 0 ) ) )
2015-09-02 17:13:31 +03:00
goto_out ;
2016-04-08 20:28:35 +03:00
2015-09-02 17:13:31 +03:00
dm_get_next_target ( dmt , next , & start , & length , & target_type , & params ) ;
2016-04-08 20:28:35 +03:00
if ( ! params | | sscanf ( params , " %d:%d " , & major , & minor ) ! = 2 ) {
2024-05-25 23:36:18 +03:00
log_warn ( " WARNING: Cannot get thin-pool major:minor for thin device %u:%u. " ,
MAJOR ( dev - > dev ) , MINOR ( dev - > dev ) ) ;
2015-09-02 17:13:31 +03:00
goto out ;
}
dm_task_destroy ( dmt ) ;
2016-12-01 16:56:32 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_STATUS , NULL , NULL , NULL , NULL ,
major , minor , 0 , 0 , 0 ) ) )
2015-09-02 17:13:31 +03:00
goto_out ;
2016-04-08 20:28:35 +03:00
2015-09-02 17:13:31 +03:00
dm_get_next_target ( dmt , next , & start , & length , & target_type , & params ) ;
if ( ! dm_get_status_thin_pool ( mem , params , & status ) )
2015-11-08 19:04:03 +03:00
goto_out ;
2015-09-02 17:13:31 +03:00
if ( status - > read_only | | status - > out_of_data_space ) {
log_warn ( " WARNING: %s: Thin's thin-pool needs inspection. " ,
dev_name ( dev ) ) ;
goto out ;
}
r = 1 ;
out :
if ( dmt )
dm_task_destroy ( dmt ) ;
dm_pool_destroy ( mem ) ;
return r ;
}
2018-03-08 12:32:15 +03:00
static int _ignore_invalid_snapshot ( const char * params )
{
struct dm_status_snapshot * s ;
struct dm_pool * mem ;
2018-03-08 15:18:39 +03:00
int r = 0 ;
2018-03-08 12:32:15 +03:00
if ( ! ( mem = dm_pool_create ( " invalid snapshots " , 128 ) ) )
return_0 ;
if ( ! dm_get_status_snapshot ( mem , params , & s ) )
2018-03-08 15:18:39 +03:00
stack ;
else
r = s - > invalid ;
dm_pool_destroy ( mem ) ;
return r ;
}
static int _ignore_frozen_raid ( struct device * dev , const char * params )
{
struct dm_status_raid * s ;
struct dm_pool * mem ;
int r = 0 ;
if ( ! ( mem = dm_pool_create ( " frozen raid " , 128 ) ) )
2018-03-08 12:32:15 +03:00
return_0 ;
2018-03-08 15:18:39 +03:00
if ( ! dm_get_status_raid ( mem , params , & s ) )
stack ;
else if ( s - > sync_action & & ! strcmp ( s - > sync_action , " frozen " ) ) {
2024-05-25 23:36:18 +03:00
log_warn ( " WARNING: %s frozen raid device (%u:%u) needs inspection. " ,
dev_name ( dev ) , MAJOR ( dev - > dev ) , MINOR ( dev - > dev ) ) ;
2018-03-08 15:18:39 +03:00
r = 1 ;
}
2018-03-08 12:32:15 +03:00
dm_pool_destroy ( mem ) ;
return r ;
}
2021-12-15 13:26:21 +03:00
static int _is_usable_uuid ( const struct device * dev , const char * name , const char * uuid , int check_reserved , int check_lv , int * is_lv )
{
char * vgname , * lvname , * layer ;
char vg_name [ NAME_LEN ] ;
if ( ! check_reserved & & ! check_lv )
return 1 ;
if ( ! strncmp ( uuid , UUID_PREFIX , sizeof ( UUID_PREFIX ) - 1 ) ) { /* with LVM- prefix */
if ( check_reserved ) {
/* Check internal lvm devices */
if ( strlen ( uuid ) > ( sizeof ( UUID_PREFIX ) + 2 * ID_LEN ) ) { /* 68 with suffix */
log_debug_activation ( " %s: Reserved uuid %s on internal LV device %s not usable. " ,
dev_name ( dev ) , uuid , name ) ;
return 0 ;
}
/* Recognize some older reserved LVs just from the LV name (snapshot, pvmove...) */
vgname = vg_name ;
2024-04-04 19:42:24 +03:00
if ( ! _dm_strncpy ( vg_name , name , sizeof ( vg_name ) ) | |
2021-12-15 13:26:21 +03:00
! dm_split_lvm_name ( NULL , NULL , & vgname , & lvname , & layer ) )
return_0 ;
/* FIXME: fails to handle dev aliases i.e. /dev/dm-5, replace with UUID suffix */
if ( lvname & & ( is_reserved_lvname ( lvname ) | | * layer ) ) {
log_debug_activation ( " %s: Reserved internal LV device %s/%s%s%s not usable. " ,
dev_name ( dev ) , vgname , lvname , * layer ? " - " : " " , layer ) ;
return 0 ;
}
}
if ( check_lv ) {
/* Skip LVs */
if ( is_lv )
* is_lv = 1 ;
return 0 ;
}
}
if ( check_reserved & &
( ! strncmp ( uuid , CRYPT_TEMP , sizeof ( CRYPT_TEMP ) - 1 ) | |
! strncmp ( uuid , CRYPT_SUBDEV , sizeof ( CRYPT_SUBDEV ) - 1 ) | |
! strncmp ( uuid , STRATIS , sizeof ( STRATIS ) - 1 ) ) ) {
/* Skip private crypto devices */
log_debug_activation ( " %s: Reserved uuid %s on %s device %s not usable. " ,
dev_name ( dev ) , uuid ,
uuid [ 0 ] = = ' C ' ? " crypto " : " stratis " ,
name ) ;
return 0 ;
}
return 1 ;
}
2013-08-08 02:42:26 +04:00
/*
2024-06-28 21:56:12 +03:00
* dm_device_is_usable
2013-08-08 02:42:26 +04:00
* @ dev
* @ check_lv_names
*
* A device is considered not usable if it is :
* 1 ) An empty device ( no targets )
* 2 ) A blocked mirror ( i . e . a mirror with a failure and block_on_error set )
* 3 ) ignore_suspended_devices is set and
* a ) the device is suspended
* b ) it is a snapshot origin
* 4 ) an error target
* 5 ) the LV name is a reserved name .
*
* Returns : 1 if usable , 0 otherwise
*/
2024-06-28 21:56:12 +03:00
int dm_device_is_usable ( struct cmd_context * cmd , struct device * dev , struct dev_usable_check_params check , int * is_lv )
2007-01-26 00:22:30 +03:00
{
struct dm_task * dmt ;
struct dm_info info ;
2010-05-13 22:38:38 +04:00
const char * name , * uuid ;
2024-06-07 16:57:42 +03:00
const struct dm_active_device * dm_dev ;
2008-01-30 17:00:02 +03:00
uint64_t start , length ;
char * target_type = NULL ;
2021-12-15 13:26:21 +03:00
char * params ;
2007-01-26 02:03:48 +03:00
void * next = NULL ;
2021-03-18 19:17:21 +03:00
int only_error_or_zero_target = 1 ;
2007-01-26 00:22:30 +03:00
int r = 0 ;
2024-07-08 23:32:41 +03:00
if ( dm_devs_cache_use ( ) & &
2024-06-07 16:57:42 +03:00
/* With cache we can avoid status calls for unusable UUIDs */
2024-07-01 19:50:59 +03:00
( dm_dev = dm_devs_cache_get_by_devno ( cmd , dev - > dev ) ) & &
2024-06-07 16:57:42 +03:00
! _is_usable_uuid ( dev , dm_dev - > name , dm_dev - > uuid , check . check_reserved , check . check_lv , is_lv ) )
return 0 ;
2016-12-01 16:56:32 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_STATUS , & info , NULL , NULL , NULL ,
MAJOR ( dev - > dev ) , MINOR ( dev - > dev ) , 0 , 0 , 0 ) ) )
2016-04-21 21:19:24 +03:00
return_0 ;
2016-04-08 17:00:03 +03:00
2010-08-09 18:05:16 +04:00
if ( ! info . exists )
2007-01-26 00:22:30 +03:00
goto out ;
2007-01-26 02:03:48 +03:00
name = dm_task_get_name ( dmt ) ;
2010-05-13 22:38:38 +04:00
uuid = dm_task_get_uuid ( dmt ) ;
2007-01-26 02:03:48 +03:00
2014-09-23 14:47:11 +04:00
if ( check . check_empty & & ! info . target_count ) {
2013-01-08 02:30:29 +04:00
log_debug_activation ( " %s: Empty device %s not usable. " , dev_name ( dev ) , name ) ;
2010-10-25 14:37:34 +04:00
goto out ;
}
2014-09-23 14:47:11 +04:00
if ( check . check_suspended & & info . suspended ) {
2013-01-08 02:30:29 +04:00
log_debug_activation ( " %s: Suspended device %s not usable. " , dev_name ( dev ) , name ) ;
2010-08-09 18:05:16 +04:00
goto out ;
}
2021-12-15 13:26:21 +03:00
if ( uuid & &
! _is_usable_uuid ( dev , name , uuid , check . check_reserved , check . check_lv , is_lv ) )
goto out ;
2014-03-12 22:38:34 +04:00
mirror: Avoid reading from mirrors that have failed devices
Addresses: rhbz855398 (Allow VGs to be built on cluster mirrors),
and other issues.
The LVM code attempts to avoid reading labels from devices that are
suspended to try to avoid situations that may cause the commands to
block indefinitely. When scanning devices, 'ignore_suspended_devices'
can be set so the code (lib/activate/dev_manager.c:device_is_usable())
checks any DM devices it finds and avoids them if they are suspended.
The mirror target has an additional mechanism that can cause I/O to
be blocked. If a device in a mirror fails, all I/O will be blocked
by the kernel until a new table (a linear target or a mirror with
replacement devices) is loaded. The mirror indicates that this condition
has happened by marking a 'D' for the faulty device in its status
output. This condition must also be checked by 'device_is_usable()' to
avoid the possibility of blocking LVM commands indefinitely due to an
attempt to read the blocked mirror for labels.
Until now, mirrors were avoided if the 'ignore_suspended_devices'
condition was set. This check seemed to suggest, "if we are concerned
about suspended devices, then let's ignore mirrors altogether just
in case". This is insufficient and doesn't solve any problems. All
devices that are suspended are already avoided if
'ignore_suspended_devices' is set; and if a mirror is blocking because
of an error condition, it will block the LVM command regardless of the
setting of that variable.
Rather than avoiding mirrors whenever 'ignore_suspended_devices' is
set, this patch causes mirrors to be avoided whenever they are blocking
due to an error. (As mentioned above, the case where a DM device is
suspended is already covered.) This solves a number of issues that weren't
handled before. For example, pvcreate (or any command that does a
pv_read or vg_read, which eventually call device_is_usable()) will be
protected from blocked mirrors regardless of how
'ignore_suspended_devices' is set. Additionally, a mirror that is
neither suspended nor blocking is /allowed/ to be read regardless
of how 'ignore_suspended_devices' is set. (The latter point being the
source of the fix for rhbz855398.)
2012-10-24 08:10:33 +04:00
/* FIXME Also check for mpath no paths */
2008-01-30 17:00:02 +03:00
do {
next = dm_get_next_target ( dmt , next , & start , & length ,
& target_type , & params ) ;
mirror: Avoid reading from mirrors that have failed devices
Addresses: rhbz855398 (Allow VGs to be built on cluster mirrors),
and other issues.
The LVM code attempts to avoid reading labels from devices that are
suspended to try to avoid situations that may cause the commands to
block indefinitely. When scanning devices, 'ignore_suspended_devices'
can be set so the code (lib/activate/dev_manager.c:device_is_usable())
checks any DM devices it finds and avoids them if they are suspended.
The mirror target has an additional mechanism that can cause I/O to
be blocked. If a device in a mirror fails, all I/O will be blocked
by the kernel until a new table (a linear target or a mirror with
replacement devices) is loaded. The mirror indicates that this condition
has happened by marking a 'D' for the faulty device in its status
output. This condition must also be checked by 'device_is_usable()' to
avoid the possibility of blocking LVM commands indefinitely due to an
attempt to read the blocked mirror for labels.
Until now, mirrors were avoided if the 'ignore_suspended_devices'
condition was set. This check seemed to suggest, "if we are concerned
about suspended devices, then let's ignore mirrors altogether just
in case". This is insufficient and doesn't solve any problems. All
devices that are suspended are already avoided if
'ignore_suspended_devices' is set; and if a mirror is blocking because
of an error condition, it will block the LVM command regardless of the
setting of that variable.
Rather than avoiding mirrors whenever 'ignore_suspended_devices' is
set, this patch causes mirrors to be avoided whenever they are blocking
due to an error. (As mentioned above, the case where a DM device is
suspended is already covered.) This solves a number of issues that weren't
handled before. For example, pvcreate (or any command that does a
pv_read or vg_read, which eventually call device_is_usable()) will be
protected from blocked mirrors regardless of how
'ignore_suspended_devices' is set. Additionally, a mirror that is
neither suspended nor blocking is /allowed/ to be read regardless
of how 'ignore_suspended_devices' is set. (The latter point being the
source of the fix for rhbz855398.)
2012-10-24 08:10:33 +04:00
2018-03-08 15:01:12 +03:00
if ( ! target_type )
continue ;
if ( check . check_blocked & & ! strcmp ( target_type , TARGET_NAME_MIRROR ) ) {
Mirror: Fix hangs and lock-ups caused by attempting label reads of mirrors
There is a problem with the way mirrors have been designed to handle
failures that is resulting in stuck LVM processes and hung I/O. When
mirrors encounter a write failure, they block I/O and notify userspace
to reconfigure the mirror to remove failed devices. This process is
open to a couple races:
1) Any LVM process other than the one that is meant to deal with the
mirror failure can attempt to read the mirror, fail, and block other
LVM commands (including the repair command) from proceeding due to
holding a lock on the volume group.
2) If there are multiple mirrors that suffer a failure in the same
volume group, a repair can block while attempting to read the LVM
label from one mirror while trying to repair the other.
Mitigation of these races has been attempted by disallowing label reading
of mirrors that are either suspended or are indicated as blocking by
the kernel. While this has closed the window of opportunity for hitting
the above problems considerably, it hasn't closed it completely. This is
because it is still possible to start an LVM command, read the status of
the mirror as healthy, and then perform the read for the label at the
moment after a the failure is discovered by the kernel.
I can see two solutions to this problem:
1) Allow users to configure whether mirrors can be candidates for LVM
labels (i.e. whether PVs can be created on mirror LVs). If the user
chooses to allow label scanning of mirror LVs, it will be at the expense
of a possible hang in I/O or LVM processes.
2) Instrument a way to allow asynchronous label reading - allowing
blocked label reads to be ignored while continuing to process the LVM
command. This would action would allow LVM commands to continue even
though they would have otherwise blocked trying to read a mirror. They
can then release their lock and allow a repair command to commence. In
the event of #2 above, the repair command already in progress can continue
and repair the failed mirror.
This patch brings solution #1. If solution #2 is developed later on, the
configuration option created in #1 can be negated - allowing mirrors to
be scanned for labels by default once again.
2013-10-23 04:14:33 +04:00
if ( ignore_lvm_mirrors ( ) ) {
log_debug_activation ( " %s: Scanning mirror devices is disabled. " , dev_name ( dev ) ) ;
goto out ;
}
2021-12-15 13:26:21 +03:00
if ( ! _ignore_blocked_mirror_devices ( cmd , dev , start ,
Mirror: Fix hangs and lock-ups caused by attempting label reads of mirrors
There is a problem with the way mirrors have been designed to handle
failures that is resulting in stuck LVM processes and hung I/O. When
mirrors encounter a write failure, they block I/O and notify userspace
to reconfigure the mirror to remove failed devices. This process is
open to a couple races:
1) Any LVM process other than the one that is meant to deal with the
mirror failure can attempt to read the mirror, fail, and block other
LVM commands (including the repair command) from proceeding due to
holding a lock on the volume group.
2) If there are multiple mirrors that suffer a failure in the same
volume group, a repair can block while attempting to read the LVM
label from one mirror while trying to repair the other.
Mitigation of these races has been attempted by disallowing label reading
of mirrors that are either suspended or are indicated as blocking by
the kernel. While this has closed the window of opportunity for hitting
the above problems considerably, it hasn't closed it completely. This is
because it is still possible to start an LVM command, read the status of
the mirror as healthy, and then perform the read for the label at the
moment after a the failure is discovered by the kernel.
I can see two solutions to this problem:
1) Allow users to configure whether mirrors can be candidates for LVM
labels (i.e. whether PVs can be created on mirror LVs). If the user
chooses to allow label scanning of mirror LVs, it will be at the expense
of a possible hang in I/O or LVM processes.
2) Instrument a way to allow asynchronous label reading - allowing
blocked label reads to be ignored while continuing to process the LVM
command. This would action would allow LVM commands to continue even
though they would have otherwise blocked trying to read a mirror. They
can then release their lock and allow a repair command to commence. In
the event of #2 above, the repair command already in progress can continue
and repair the failed mirror.
This patch brings solution #1. If solution #2 is developed later on, the
configuration option created in #1 can be negated - allowing mirrors to
be scanned for labels by default once again.
2013-10-23 04:14:33 +04:00
length , params ) ) {
log_debug_activation ( " %s: Mirror device %s not usable. " ,
dev_name ( dev ) , name ) ;
goto out ;
}
2010-05-13 22:38:38 +04:00
}
2010-08-26 18:21:50 +04:00
/*
2015-01-09 13:24:16 +03:00
* FIXME : Snapshot origin could be sitting on top of a mirror
* which could be blocking I / O . We should add a check for the
* stack here and see if there ' s blocked mirror underneath .
* Currently , mirrors used as origin or snapshot is not
* supported anymore and in general using mirrors in a stack
* is disabled by default ( with a warning that if enabled ,
* it could cause various deadlocks ) .
2015-06-17 14:37:53 +03:00
* Similar situation can happen with RAID devices where
* a RAID device can be snapshotted .
* If one of the RAID legs are down and we ' re doing
* lvconvert - - repair , there ' s a time period in which
* snapshot components are ( besides other devs ) suspended .
* See also https : //bugzilla.redhat.com/show_bug.cgi?id=1219222
* for an example where this causes problems .
*
* This is a quick check for now , but replace it with more
* robust and better check that would check the stack
2024-08-30 00:05:41 +03:00
* correctly , not just snapshots but any combination possible
2015-06-17 14:37:53 +03:00
* in a stack - use proper dm tree to check this instead .
2010-08-26 18:21:50 +04:00
*/
2018-03-08 15:01:12 +03:00
if ( check . check_suspended & &
2016-03-22 20:46:15 +03:00
( ! strcmp ( target_type , TARGET_NAME_SNAPSHOT ) | | ! strcmp ( target_type , TARGET_NAME_SNAPSHOT_ORIGIN ) ) & &
2015-06-17 14:37:53 +03:00
_ignore_suspended_snapshot_component ( dev ) ) {
log_debug_activation ( " %s: %s device %s not usable. " , dev_name ( dev ) , target_type , name ) ;
2010-08-26 18:21:50 +04:00
goto out ;
2015-06-17 14:37:53 +03:00
}
2010-10-24 21:36:58 +04:00
2018-03-08 15:01:12 +03:00
if ( ! strcmp ( target_type , TARGET_NAME_SNAPSHOT ) & &
2018-03-08 12:32:15 +03:00
_ignore_invalid_snapshot ( params ) ) {
log_debug_activation ( " %s: Invalid %s device %s not usable. " , dev_name ( dev ) , target_type , name ) ;
goto out ;
2018-03-08 15:18:39 +03:00
}
if ( ! strncmp ( target_type , TARGET_NAME_RAID , 4 ) & & _ignore_frozen_raid ( dev , params ) ) {
log_debug_activation ( " %s: Frozen %s device %s not usable. " ,
dev_name ( dev ) , target_type , name ) ;
goto out ;
2018-03-08 12:32:15 +03:00
}
2015-09-02 17:13:31 +03:00
/* TODO: extend check struct ? */
2018-03-08 15:01:12 +03:00
if ( ! strcmp ( target_type , TARGET_NAME_THIN ) & &
2015-09-02 17:13:31 +03:00
! _ignore_unusable_thins ( dev ) ) {
log_debug_activation ( " %s: %s device %s not usable. " , dev_name ( dev ) , target_type , name ) ;
goto out ;
}
2021-03-18 19:17:21 +03:00
if ( only_error_or_zero_target & &
strcmp ( target_type , TARGET_NAME_ERROR ) & &
strcmp ( target_type , TARGET_NAME_ZERO ) )
only_error_or_zero_target = 0 ;
2008-01-30 17:00:02 +03:00
} while ( next ) ;
2007-01-26 00:22:30 +03:00
2021-03-18 19:17:21 +03:00
/* Skip devices consisting entirely of error or zero targets. */
2010-10-25 14:37:34 +04:00
/* FIXME Deal with device stacked above error targets? */
2021-03-18 19:17:21 +03:00
if ( check . check_error_target & & only_error_or_zero_target ) {
2013-01-08 02:30:29 +04:00
log_debug_activation ( " %s: Error device %s not usable. " ,
dev_name ( dev ) , name ) ;
2010-10-25 14:37:34 +04:00
goto out ;
}
2007-01-26 00:22:30 +03:00
/* FIXME Also check dependencies? */
r = 1 ;
out :
dm_task_destroy ( dmt ) ;
return r ;
}
2024-05-23 02:36:50 +03:00
/* Read UUID from a given DM device into buf_uuid */
2024-06-28 21:35:12 +03:00
int devno_dm_uuid ( struct cmd_context * cmd , int major , int minor ,
char * uuid_buf , size_t uuid_buf_size )
2024-05-23 02:36:50 +03:00
{
struct dm_task * dmt ;
struct dm_info info ;
2024-06-03 00:10:05 +03:00
const struct dm_active_device * dm_dev ;
2024-05-23 02:36:50 +03:00
const char * uuid ;
int r = 0 ;
2024-08-01 19:30:10 +03:00
if ( major ! = cmd - > dev_types - > device_mapper_major )
return 0 ;
2024-07-08 23:32:41 +03:00
if ( dm_devs_cache_use ( ) ) {
2024-07-01 19:50:59 +03:00
if ( ( dm_dev = dm_devs_cache_get_by_devno ( cmd , MKDEV ( major , minor ) ) ) ) {
2024-06-03 00:10:05 +03:00
dm_strncpy ( uuid_buf , dm_dev - > uuid , uuid_buf_size ) ;
2024-05-23 02:36:50 +03:00
return 1 ;
}
uuid_buf [ 0 ] = 0 ;
return 0 ;
}
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_INFO , & info , NULL , NULL , NULL ,
major , minor , 0 , 0 , 0 ) ) )
return_0 ;
if ( info . exists & & ( uuid = dm_task_get_uuid ( dmt ) ) )
r = dm_strncpy ( uuid_buf , uuid , uuid_buf_size ) ;
dm_task_destroy ( dmt ) ;
return r ;
}
2024-06-28 21:35:12 +03:00
int dev_dm_uuid ( struct cmd_context * cmd , struct device * dev ,
char * uuid_buf , size_t uuid_buf_size )
{
return devno_dm_uuid ( cmd , MAJOR ( dev - > dev ) , MINOR ( dev - > dev ) ,
uuid_buf , uuid_buf_size ) ;
}
2016-04-08 18:27:12 +03:00
/*
* If active LVs were activated by a version of LVM2 before 2.02 .00 we must
* perform additional checks to find them because they do not have the LVM -
* prefix on their dm uuids .
* As of 2.02 .150 , we ' ve chosen to disable this compatibility arbitrarily if
* we ' re running kernel version 3 or above .
*/
# define MIN_KERNEL_MAJOR 3
static int _original_uuid_format_check_required ( struct cmd_context * cmd )
2016-04-07 22:37:26 +03:00
{
2016-04-08 18:27:12 +03:00
static int _kernel_major = 0 ;
if ( ! _kernel_major ) {
2016-04-08 20:34:56 +03:00
if ( ( sscanf ( cmd - > kernel_vsn , " %d " , & _kernel_major ) = = 1 ) & &
( _kernel_major > = MIN_KERNEL_MAJOR ) )
2016-04-08 18:27:12 +03:00
log_debug_activation ( " Skipping checks for old devices without " UUID_PREFIX
" dm uuid prefix (kernel vsn %d >= %d). " , _kernel_major , MIN_KERNEL_MAJOR ) ;
2016-04-08 20:34:56 +03:00
else
_kernel_major = - 1 ;
2016-04-07 22:37:26 +03:00
}
2016-04-08 20:34:56 +03:00
return ( _kernel_major = = - 1 ) ;
2016-04-07 22:37:26 +03:00
}
2016-12-18 16:57:16 +03:00
static int _info ( struct cmd_context * cmd ,
const char * name , const char * dlid ,
2019-10-31 13:45:28 +03:00
int with_open_count , int with_read_ahead , int with_name_check ,
2014-11-04 17:00:32 +03:00
struct dm_info * dminfo , uint32_t * read_ahead ,
struct lv_seg_status * seg_status )
2002-03-25 21:54:59 +03:00
{
2014-07-31 00:55:11 +04:00
char old_style_dlid [ sizeof ( UUID_PREFIX ) + 2 * ID_LEN ] ;
const char * suffix , * suffix_position ;
2019-10-31 13:45:28 +03:00
const char * name_check = ( with_name_check ) ? name : NULL ;
2014-07-31 00:55:11 +04:00
unsigned i = 0 ;
2010-01-26 10:58:23 +03:00
2016-12-18 16:57:16 +03:00
log_debug_activation ( " Getting device info for %s [%s]. " , name , dlid ) ;
2014-07-31 00:55:11 +04:00
/* Check for dlid */
2019-10-31 13:45:28 +03:00
if ( ! _info_run ( dlid , dminfo , read_ahead , seg_status , name_check ,
2016-12-18 16:57:16 +03:00
with_open_count , with_read_ahead , 0 , 0 ) )
return_0 ;
if ( dminfo - > exists )
2010-02-24 23:00:56 +03:00
return 1 ;
2014-07-31 00:55:11 +04:00
/* Check for original version of dlid before the suffixes got added in 2.02.106 */
2021-09-24 18:40:35 +03:00
if ( ( suffix_position = strrchr ( dlid , ' - ' ) ) ) {
2024-05-03 13:43:14 +03:00
while ( ( suffix = _uuid_suffix_list [ i + + ] ) ) {
2014-07-31 00:55:11 +04:00
if ( strcmp ( suffix_position + 1 , suffix ) )
continue ;
2024-04-04 18:03:30 +03:00
dm_strncpy ( old_style_dlid , dlid , sizeof ( old_style_dlid ) ) ;
2016-12-18 16:57:16 +03:00
if ( ! _info_run ( old_style_dlid , dminfo , read_ahead , seg_status ,
2019-10-31 13:45:28 +03:00
name_check , with_open_count , with_read_ahead ,
0 , 0 ) )
2016-12-18 16:57:16 +03:00
return_0 ;
if ( dminfo - > exists )
2014-07-31 00:55:11 +04:00
return 1 ;
}
}
2016-04-08 18:27:12 +03:00
/* Must we still check for the pre-2.02.00 dm uuid format? */
if ( ! _original_uuid_format_check_required ( cmd ) )
2016-12-18 16:57:16 +03:00
return 1 ;
2016-04-07 22:37:26 +03:00
2014-07-31 00:55:11 +04:00
/* Check for dlid before UUID_PREFIX was added */
2016-12-18 16:57:16 +03:00
if ( ! _info_run ( dlid + sizeof ( UUID_PREFIX ) - 1 , dminfo , read_ahead , seg_status ,
2019-10-31 13:45:28 +03:00
name_check , with_open_count , with_read_ahead , 0 , 0 ) )
2016-12-18 16:57:16 +03:00
return_0 ;
2002-03-25 21:54:59 +03:00
2016-12-18 16:57:16 +03:00
return 1 ;
2002-03-25 21:54:59 +03:00
}
2018-08-17 23:45:52 +03:00
int dev_manager_remove_dm_major_minor ( uint32_t major , uint32_t minor )
{
struct dm_task * dmt ;
int r = 0 ;
log_verbose ( " Removing dm dev %u:%u " , major , minor ) ;
if ( ! ( dmt = dm_task_create ( DM_DEVICE_REMOVE ) ) )
return_0 ;
if ( ! dm_task_set_major ( dmt , major ) | | ! dm_task_set_minor ( dmt , minor ) ) {
log_error ( " Failed to set device number for remove %u:%u " , major , minor ) ;
goto out ;
}
r = dm_task_run ( dmt ) ;
out :
dm_task_destroy ( dmt ) ;
return r ;
}
2008-12-19 18:23:03 +03:00
static int _info_by_dev ( uint32_t major , uint32_t minor , struct dm_info * info )
{
2019-10-31 13:45:28 +03:00
return _info_run ( NULL , info , NULL , NULL , NULL , 0 , 0 , major , minor ) ;
2008-12-19 18:23:03 +03:00
}
2019-03-13 14:59:59 +03:00
int dev_manager_check_prefix_dm_major_minor ( uint32_t major , uint32_t minor , const char * prefix )
{
struct dm_task * dmt ;
const char * uuid ;
int r = 1 ;
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_INFO , NULL , NULL , NULL , 0 , major , minor , 0 , 0 , 0 ) ) )
return_0 ;
if ( ! ( uuid = dm_task_get_uuid ( dmt ) ) | | strncasecmp ( uuid , prefix , strlen ( prefix ) ) )
r = 0 ;
dm_task_destroy ( dmt ) ;
return r ;
}
2024-06-28 21:35:12 +03:00
/*
* Get a list of active dm devices from the kernel .
* The ' devs ' list contains a struct dm_active_device .
*/
int dev_manager_get_dm_active_devices ( const char * prefix , struct dm_list * * devs , unsigned * devs_features )
2021-12-15 13:24:31 +03:00
{
struct dm_task * dmt ;
int r = 1 ;
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_LIST , NULL , NULL , NULL , 0 , 0 , 0 , 0 , 0 , 0 ) ) )
return_0 ;
if ( ! dm_task_get_device_list ( dmt , devs , devs_features ) ) {
r = 0 ;
goto_out ;
}
out :
dm_task_destroy ( dmt ) ;
return r ;
}
2016-12-18 16:57:16 +03:00
int dev_manager_info ( struct cmd_context * cmd ,
const struct logical_volume * lv , const char * layer ,
2019-10-31 13:45:28 +03:00
int with_open_count , int with_read_ahead , int with_name_check ,
2014-11-04 17:00:32 +03:00
struct dm_info * dminfo , uint32_t * read_ahead ,
struct lv_seg_status * seg_status )
2005-10-17 22:00:02 +04:00
{
2024-06-21 12:08:30 +03:00
char old_style_dlid [ sizeof ( UUID_PREFIX ) + 2 * ID_LEN ] ;
2010-08-03 17:13:01 +04:00
char * dlid , * name ;
2016-12-01 16:59:02 +03:00
int r = 0 ;
2010-02-24 23:00:56 +03:00
2016-04-08 18:27:12 +03:00
if ( ! ( name = dm_build_dm_name ( cmd - > mem , lv - > vg - > name , lv - > name , layer ) ) )
2015-11-25 17:45:19 +03:00
return_0 ;
2005-10-19 17:59:18 +04:00
2016-12-01 16:59:02 +03:00
if ( ! ( dlid = build_dm_uuid ( cmd - > mem , lv , layer ) ) )
2015-11-25 17:45:19 +03:00
goto_out ;
2005-10-19 17:59:18 +04:00
2024-06-21 12:08:30 +03:00
dm_strncpy ( old_style_dlid , dlid , sizeof ( old_style_dlid ) ) ;
2024-07-08 23:32:41 +03:00
if ( dm_devs_cache_use ( ) & &
2024-07-01 19:50:59 +03:00
! dm_devs_cache_get_by_uuid ( cmd , dlid ) & &
! dm_devs_cache_get_by_uuid ( cmd , old_style_dlid ) ) {
2021-12-15 13:45:22 +03:00
log_debug ( " Cached as inactive %s. " , name ) ;
if ( dminfo )
memset ( dminfo , 0 , sizeof ( * dminfo ) ) ;
r = 1 ;
goto out ;
}
2019-10-31 13:45:28 +03:00
if ( ! ( r = _info ( cmd , name , dlid ,
with_open_count , with_read_ahead , with_name_check ,
2016-12-18 16:57:16 +03:00
dminfo , read_ahead , seg_status ) ) )
stack ;
2013-09-20 00:23:43 +04:00
out :
2016-04-08 18:27:12 +03:00
dm_pool_free ( cmd - > mem , name ) ;
2013-09-20 00:23:43 +04:00
2010-02-24 23:00:56 +03:00
return r ;
2005-10-17 22:00:02 +04:00
}
2024-03-21 19:23:33 +03:00
static struct dm_tree_node * _cached_dm_tree_node ( struct dm_pool * mem ,
struct dm_tree * dtree ,
const struct logical_volume * lv ,
const char * layer )
{
struct dm_tree_node * dnode ;
char * dlid ;
if ( ! ( dlid = build_dm_uuid ( mem , lv , layer ) ) )
return_NULL ;
dnode = dm_tree_find_node_by_uuid ( dtree , dlid ) ;
dm_pool_free ( mem , dlid ) ;
return dnode ;
}
2014-11-02 22:59:57 +03:00
static const struct dm_info * _cached_dm_info ( struct dm_pool * mem ,
struct dm_tree * dtree ,
const struct logical_volume * lv ,
const char * layer )
2010-01-22 18:40:31 +03:00
{
2013-07-14 17:08:26 +04:00
const struct dm_tree_node * dnode ;
const struct dm_info * dinfo = NULL ;
2010-01-22 18:40:31 +03:00
2024-03-21 19:23:33 +03:00
if ( ! ( dnode = _cached_dm_tree_node ( mem , dtree , lv , layer ) ) )
return NULL ;
2010-01-22 18:40:31 +03:00
if ( ! ( dinfo = dm_tree_node_get_info ( dnode ) ) ) {
2016-04-17 15:44:13 +03:00
log_warn ( " WARNING: Cannot get info from tree node for %s. " ,
display_lvname ( lv ) ) ;
2024-03-21 19:23:33 +03:00
return NULL ;
2010-01-22 18:40:31 +03:00
}
if ( ! dinfo - > exists )
2013-07-14 17:08:26 +04:00
dinfo = NULL ;
2010-01-22 18:40:31 +03:00
return dinfo ;
}
2014-09-22 17:50:07 +04:00
int lv_has_target_type ( struct dm_pool * mem , const struct logical_volume * lv ,
2010-04-23 06:57:39 +04:00
const char * layer , const char * target_type )
2010-01-13 04:54:34 +03:00
{
int r = 0 ;
char * dlid ;
struct dm_task * dmt ;
struct dm_info info ;
void * next = NULL ;
uint64_t start , length ;
char * type = NULL ;
char * params = NULL ;
2014-03-11 20:13:47 +04:00
if ( ! ( dlid = build_dm_uuid ( mem , lv , layer ) ) )
2010-01-13 04:54:34 +03:00
return_0 ;
2016-12-01 16:56:32 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_STATUS , & info , NULL , dlid , 0 , 0 , 0 , 0 , 0 , 0 ) ) )
2011-11-18 23:42:03 +04:00
goto_bad ;
2010-01-13 04:54:34 +03:00
2016-12-01 16:56:32 +03:00
if ( ! info . exists )
2010-01-13 04:54:34 +03:00
goto_out ;
2016-04-15 04:21:27 +03:00
/* If there is a preloaded table, use that in preference. */
if ( info . inactive_table ) {
dm_task_destroy ( dmt ) ;
2016-12-01 16:56:32 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_STATUS , & info , NULL , dlid , 0 , 0 , 0 , 0 , 0 , 1 ) ) )
2016-04-15 04:21:27 +03:00
goto_bad ;
2016-12-01 16:56:32 +03:00
if ( ! info . exists | | ! info . inactive_table )
2016-04-15 04:21:27 +03:00
goto_out ;
}
2010-01-13 04:54:34 +03:00
do {
next = dm_get_next_target ( dmt , next , & start , & length ,
& type , & params ) ;
2016-04-15 04:21:27 +03:00
if ( type & & ! strncmp ( type , target_type , strlen ( target_type ) ) ) {
r = 1 ;
2010-01-13 04:54:34 +03:00
break ;
}
} while ( next ) ;
2011-11-18 23:42:03 +04:00
out :
2010-01-13 04:54:34 +03:00
dm_task_destroy ( dmt ) ;
2011-11-18 23:42:03 +04:00
bad :
dm_pool_free ( mem , dlid ) ;
2010-01-13 04:54:34 +03:00
return r ;
}
2023-03-06 16:52:28 +03:00
static int _lv_has_thin_device_id ( struct dm_pool * mem , const struct logical_volume * lv ,
2016-04-17 15:51:48 +03:00
const char * layer , unsigned device_id )
{
char * dlid ;
struct dm_task * dmt ;
struct dm_info info ;
void * next = NULL ;
uint64_t start , length ;
char * type = NULL ;
char * params = NULL ;
unsigned id = ~ 0 ;
if ( ! ( dlid = build_dm_uuid ( mem , lv , layer ) ) )
return_0 ;
2016-12-01 16:56:32 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_TABLE , & info , NULL , dlid , 0 , 0 , 0 , 0 , 1 , 0 ) ) )
2016-04-17 15:51:48 +03:00
goto_bad ;
2016-12-01 16:56:32 +03:00
if ( ! info . exists )
2016-04-17 15:51:48 +03:00
goto_out ;
/* If there is a preloaded table, use that in preference. */
if ( info . inactive_table ) {
dm_task_destroy ( dmt ) ;
2016-12-01 16:56:32 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_TABLE , & info , NULL , dlid , 0 , 0 , 0 , 0 , 1 , 1 ) ) )
2016-04-17 15:51:48 +03:00
goto_bad ;
2016-12-01 16:56:32 +03:00
if ( ! info . exists | | ! info . inactive_table )
2016-04-17 15:51:48 +03:00
goto_out ;
}
( void ) dm_get_next_target ( dmt , next , & start , & length , & type , & params ) ;
if ( ! type | | strcmp ( type , TARGET_NAME_THIN ) )
goto_out ;
if ( ! params | | sscanf ( params , " %*u:%*u %u " , & id ) ! = 1 )
goto_out ;
log_debug_activation ( " %soaded thin volume %s with id %u is %smatching id %u. " ,
info . inactive_table ? " Prel " : " L " ,
display_lvname ( lv ) , id ,
( device_id ! = id ) ? " not " : " " , device_id ) ;
out :
dm_task_destroy ( dmt ) ;
bad :
dm_pool_free ( mem , dlid ) ;
return ( device_id = = id ) ;
}
2016-12-01 16:53:35 +03:00
int add_linear_area_to_dtree ( struct dm_tree_node * node , uint64_t size , uint32_t extent_size ,
int use_linear_target , const char * vgname , const char * lvname )
2011-11-29 00:37:51 +04:00
{
uint32_t page_size ;
/*
* Use striped or linear target ?
*/
if ( ! use_linear_target ) {
page_size = lvm_getpagesize ( ) > > SECTOR_SHIFT ;
/*
* We ' ll use the extent size as the stripe size .
* Extent size and page size are always powers of 2.
* The striped target requires that the stripe size is
* divisible by the page size .
*/
if ( extent_size > = page_size ) {
/* Use striped target */
if ( ! dm_tree_node_add_striped_target ( node , size , extent_size ) )
return_0 ;
return 1 ;
2016-12-01 16:53:35 +03:00
}
/* Some exotic cases are unsupported by striped. */
log_warn ( " WARNING: Using linear target for %s/%s: Striped requires extent size "
" ( " FMTu32 " sectors) >= page size ( " FMTu32 " ). " ,
vgname , lvname , extent_size , page_size ) ;
2011-11-29 00:37:51 +04:00
}
/*
* Use linear target .
*/
if ( ! dm_tree_node_add_linear_target ( node , size ) )
return_0 ;
return 1 ;
}
2014-06-09 14:08:27 +04:00
static dm_percent_range_t _combine_percent ( dm_percent_t a , dm_percent_t b ,
uint32_t numerator , uint32_t denominator )
2009-10-01 04:35:29 +04:00
{
2014-06-09 14:08:27 +04:00
if ( a = = LVM_PERCENT_MERGE_FAILED | | b = = LVM_PERCENT_MERGE_FAILED )
return LVM_PERCENT_MERGE_FAILED ;
2012-01-21 02:02:04 +04:00
2014-06-09 14:08:27 +04:00
if ( a = = DM_PERCENT_INVALID | | b = = DM_PERCENT_INVALID )
return DM_PERCENT_INVALID ;
2009-10-01 04:35:29 +04:00
2014-06-09 14:08:27 +04:00
if ( a = = DM_PERCENT_100 & & b = = DM_PERCENT_100 )
return DM_PERCENT_100 ;
2009-10-01 04:35:29 +04:00
2014-06-09 14:08:27 +04:00
if ( a = = DM_PERCENT_0 & & b = = DM_PERCENT_0 )
return DM_PERCENT_0 ;
2009-10-01 04:35:29 +04:00
2014-06-09 14:08:27 +04:00
return ( dm_percent_range_t ) dm_make_percent ( numerator , denominator ) ;
2009-10-01 04:35:29 +04:00
}
2003-05-06 16:00:29 +04:00
static int _percent_run ( struct dev_manager * dm , const char * name ,
2005-10-19 17:59:18 +04:00
const char * dlid ,
2003-05-06 16:00:29 +04:00
const char * target_type , int wait ,
2014-06-09 14:08:27 +04:00
const struct logical_volume * lv , dm_percent_t * overall_percent ,
2021-04-06 15:47:11 +03:00
uint32_t * event_nr , int fail_if_percent_unsupported ,
int * interrupted )
2003-04-25 02:09:13 +04:00
{
int r = 0 ;
struct dm_task * dmt ;
2003-04-30 19:26:25 +04:00
struct dm_info info ;
2003-04-25 02:09:13 +04:00
void * next = NULL ;
uint64_t start , length ;
char * type = NULL ;
char * params = NULL ;
2011-12-21 16:59:22 +04:00
const struct dm_list * segh = lv ? & lv - > segments : NULL ;
2003-05-06 16:00:29 +04:00
struct lv_segment * seg = NULL ;
2009-10-01 04:35:29 +04:00
int first_time = 1 ;
2014-06-09 14:08:27 +04:00
dm_percent_t percent = DM_PERCENT_INVALID ;
2003-04-25 02:09:13 +04:00
uint64_t total_numerator = 0 , total_denominator = 0 ;
2016-03-30 11:16:41 +03:00
struct segment_type * segtype ;
2003-04-25 02:09:13 +04:00
2012-06-20 14:05:00 +04:00
* overall_percent = percent ;
2003-04-25 02:09:13 +04:00
2016-03-30 11:16:41 +03:00
if ( ! ( segtype = get_segtype_from_string ( dm - > cmd , target_type ) ) )
return_0 ;
2021-04-06 15:47:11 +03:00
if ( wait )
sigint_allow ( ) ;
2016-12-01 16:56:32 +03:00
if ( ! ( dmt = _setup_task_run ( wait ? DM_DEVICE_WAITEVENT : DM_DEVICE_STATUS , & info ,
name , dlid , event_nr , 0 , 0 , 0 , 0 , 0 ) ) )
2021-04-06 15:47:11 +03:00
goto_bad ;
2003-04-25 02:09:13 +04:00
2016-12-01 16:56:32 +03:00
if ( ! info . exists )
2005-11-09 01:52:26 +03:00
goto_out ;
2003-04-30 19:26:25 +04:00
if ( event_nr )
* event_nr = info . event_nr ;
2003-04-25 02:09:13 +04:00
do {
next = dm_get_next_target ( dmt , next , & start , & length , & type ,
& params ) ;
2003-05-06 16:00:29 +04:00
if ( lv ) {
2008-11-04 01:14:30 +03:00
if ( ! ( segh = dm_list_next ( & lv - > segments , segh ) ) ) {
2003-05-06 16:00:29 +04:00
log_error ( " Number of segments in active LV %s "
2015-11-25 18:06:31 +03:00
" does not match metadata. " ,
display_lvname ( lv ) ) ;
2003-05-06 16:00:29 +04:00
goto out ;
}
2008-11-04 01:14:30 +03:00
seg = dm_list_item ( segh , struct lv_segment ) ;
2003-05-06 16:00:29 +04:00
}
2003-04-25 02:09:13 +04:00
2010-01-15 19:35:26 +03:00
if ( ! type | | ! params )
2003-04-25 02:09:13 +04:00
continue ;
2010-01-15 19:35:26 +03:00
if ( strcmp ( type , target_type ) ) {
/* If kernel's type isn't an exact match is it compatible? */
if ( ! segtype - > ops - > target_status_compatible | |
! segtype - > ops - > target_status_compatible ( type ) )
continue ;
}
2010-12-08 22:26:35 +03:00
if ( ! segtype - > ops - > target_percent )
continue ;
if ( ! segtype - > ops - > target_percent ( & dm - > target_state ,
2010-11-30 14:53:31 +03:00
& percent , dm - > mem ,
2006-05-17 00:53:13 +04:00
dm - > cmd , seg , params ,
2004-05-05 01:25:57 +04:00
& total_numerator ,
2008-07-15 04:25:52 +04:00
& total_denominator ) )
2005-11-09 01:52:26 +03:00
goto_out ;
2004-05-05 01:25:57 +04:00
2009-10-01 04:35:29 +04:00
if ( first_time ) {
2010-11-30 14:53:31 +03:00
* overall_percent = percent ;
2009-10-01 04:35:29 +04:00
first_time = 0 ;
} else
2010-11-30 14:53:31 +03:00
* overall_percent =
_combine_percent ( * overall_percent , percent ,
total_numerator , total_denominator ) ;
2003-04-25 02:09:13 +04:00
} while ( next ) ;
2010-12-20 17:04:43 +03:00
if ( lv & & dm_list_next ( & lv - > segments , segh ) ) {
2003-05-06 16:00:29 +04:00
log_error ( " Number of segments in active LV %s does not "
2015-11-25 18:06:31 +03:00
" match metadata. " , display_lvname ( lv ) ) ;
2003-05-06 16:00:29 +04:00
goto out ;
}
2010-11-30 14:53:31 +03:00
if ( first_time ) {
/* above ->target_percent() was not executed! */
/* FIXME why return PERCENT_100 et. al. in this case? */
2014-06-09 14:08:27 +04:00
* overall_percent = DM_PERCENT_100 ;
2010-11-30 14:53:31 +03:00
if ( fail_if_percent_unsupported )
goto_out ;
2009-10-01 04:35:29 +04:00
}
2003-04-25 02:09:13 +04:00
2017-06-24 17:22:36 +03:00
log_debug_activation ( " LV percent: %s " ,
display_percent ( dm - > cmd , * overall_percent ) ) ;
2003-04-25 02:09:13 +04:00
r = 1 ;
2021-04-06 15:47:11 +03:00
out :
2003-04-25 02:09:13 +04:00
dm_task_destroy ( dmt ) ;
2021-04-06 15:47:11 +03:00
bad :
if ( wait ) {
sigint_restore ( ) ;
if ( sigint_caught ( ) ) {
* interrupted = 1 ;
return_0 ;
}
}
2003-04-25 02:09:13 +04:00
return r ;
}
2005-10-19 17:59:18 +04:00
static int _percent ( struct dev_manager * dm , const char * name , const char * dlid ,
2003-05-06 16:00:29 +04:00
const char * target_type , int wait ,
2014-06-09 14:08:27 +04:00
const struct logical_volume * lv , dm_percent_t * percent ,
2010-11-30 14:53:31 +03:00
uint32_t * event_nr , int fail_if_percent_unsupported )
2003-04-25 02:09:13 +04:00
{
2021-04-06 15:47:11 +03:00
int interrupted = 0 ;
2005-10-26 19:00:51 +04:00
if ( dlid & & * dlid ) {
if ( _percent_run ( dm , NULL , dlid , target_type , wait , lv , percent ,
2021-04-06 15:47:11 +03:00
event_nr , fail_if_percent_unsupported , & interrupted ) )
2005-10-26 19:00:51 +04:00
return 1 ;
2017-07-19 17:16:12 +03:00
2021-04-06 15:47:11 +03:00
if ( ! interrupted & &
_original_uuid_format_check_required ( dm - > cmd ) & &
2017-07-19 17:16:12 +03:00
_percent_run ( dm , NULL , dlid + sizeof ( UUID_PREFIX ) - 1 ,
target_type , wait , lv , percent ,
2021-04-06 15:47:11 +03:00
event_nr , fail_if_percent_unsupported , & interrupted ) )
2005-10-26 19:00:51 +04:00
return 1 ;
}
2003-04-25 02:09:13 +04:00
2021-04-06 15:47:11 +03:00
if ( ! interrupted & & name & &
_percent_run ( dm , name , NULL , target_type , wait , lv , percent ,
event_nr , fail_if_percent_unsupported , & interrupted ) )
2003-04-25 02:09:13 +04:00
return 1 ;
2013-07-17 01:35:56 +04:00
return_0 ;
2003-04-25 02:09:13 +04:00
}
2010-08-17 05:51:12 +04:00
/* FIXME Merge with the percent function */
2014-09-22 17:50:07 +04:00
int dev_manager_transient ( struct dev_manager * dm , const struct logical_volume * lv )
2010-05-24 19:32:20 +04:00
{
int r = 0 ;
struct dm_task * dmt ;
struct dm_info info ;
void * next = NULL ;
uint64_t start , length ;
char * type = NULL ;
char * params = NULL ;
char * dlid = NULL ;
2013-02-02 03:16:36 +04:00
const char * layer = lv_layer ( lv ) ;
2010-05-24 19:32:20 +04:00
const struct dm_list * segh = & lv - > segments ;
struct lv_segment * seg = NULL ;
2014-03-11 20:13:47 +04:00
if ( ! ( dlid = build_dm_uuid ( dm - > mem , lv , layer ) ) )
2010-05-24 19:32:20 +04:00
return_0 ;
2016-12-01 16:56:32 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_STATUS , & info , NULL , dlid , NULL , 0 , 0 , 0 , 0 , 0 ) ) )
2010-05-24 19:32:20 +04:00
return_0 ;
2016-12-01 16:56:32 +03:00
if ( ! info . exists )
2010-05-24 19:32:20 +04:00
goto_out ;
do {
next = dm_get_next_target ( dmt , next , & start , & length , & type ,
& params ) ;
2010-12-01 01:28:06 +03:00
if ( ! ( segh = dm_list_next ( & lv - > segments , segh ) ) ) {
log_error ( " Number of segments in active LV %s "
2015-11-25 18:06:31 +03:00
" does not match metadata. " , display_lvname ( lv ) ) ;
2010-12-01 01:28:06 +03:00
goto out ;
2010-05-24 19:32:20 +04:00
}
2010-12-01 01:28:06 +03:00
seg = dm_list_item ( segh , struct lv_segment ) ;
2010-05-24 19:32:20 +04:00
if ( ! type | | ! params )
continue ;
2012-06-21 14:43:31 +04:00
if ( ! seg ) {
log_error ( INTERNAL_ERROR " Segment is not selected. " ) ;
goto out ;
}
2010-05-24 19:32:20 +04:00
if ( seg - > segtype - > ops - > check_transient_status & &
2015-11-30 23:24:31 +03:00
! seg - > segtype - > ops - > check_transient_status ( dm - > mem , seg , params ) )
2010-05-24 19:32:20 +04:00
goto_out ;
} while ( next ) ;
2010-12-20 17:04:43 +03:00
if ( dm_list_next ( & lv - > segments , segh ) ) {
2010-05-24 19:32:20 +04:00
log_error ( " Number of segments in active LV %s does not "
2015-11-25 18:06:31 +03:00
" match metadata. " , display_lvname ( lv ) ) ;
2010-05-24 19:32:20 +04:00
goto out ;
}
r = 1 ;
out :
dm_task_destroy ( dmt ) ;
return r ;
}
2005-11-09 01:52:26 +03:00
/*
* dev_manager implementation .
*/
struct dev_manager * dev_manager_create ( struct cmd_context * cmd ,
2011-06-11 04:03:06 +04:00
const char * vg_name ,
unsigned track_pvmove_deps )
2002-03-19 02:25:50 +03:00
{
2005-11-09 01:52:26 +03:00
struct dm_pool * mem ;
struct dev_manager * dm ;
2002-03-19 02:25:50 +03:00
2008-01-30 16:19:47 +03:00
if ( ! ( mem = dm_pool_create ( " dev_manager " , 16 * 1024 ) ) )
return_NULL ;
2002-03-19 02:25:50 +03:00
2009-05-07 16:01:21 +04:00
if ( ! ( dm = dm_pool_zalloc ( mem , sizeof ( * dm ) ) ) )
2007-04-26 20:44:59 +04:00
goto_bad ;
2002-03-19 02:25:50 +03:00
2005-11-09 01:52:26 +03:00
dm - > cmd = cmd ;
dm - > mem = mem ;
2015-11-25 18:04:27 +03:00
dm - > vg_name = vg_name ;
2004-05-13 00:40:34 +04:00
2011-06-11 04:03:06 +04:00
/*
* When we manipulate ( normally suspend / resume ) the PVMOVE
* device directly , there ' s no need to touch the LVs above .
*/
dm - > track_pvmove_deps = track_pvmove_deps ;
2005-11-09 01:52:26 +03:00
dm - > target_state = NULL ;
2005-01-13 01:58:21 +03:00
2009-08-04 19:36:13 +04:00
dm_udev_set_sync_support ( cmd - > current_settings . udev_sync ) ;
2005-11-09 01:52:26 +03:00
return dm ;
2004-05-13 00:40:34 +04:00
2005-11-09 01:52:26 +03:00
bad :
dm_pool_destroy ( mem ) ;
2016-12-01 16:53:35 +03:00
2005-11-09 01:52:26 +03:00
return NULL ;
2004-05-13 00:40:34 +04:00
}
2005-11-09 01:52:26 +03:00
void dev_manager_destroy ( struct dev_manager * dm )
2004-05-13 00:40:34 +04:00
{
2005-11-09 01:52:26 +03:00
dm_pool_destroy ( dm - > mem ) ;
}
2004-05-13 00:40:34 +04:00
2006-05-16 20:48:31 +04:00
void dev_manager_release ( void )
{
dm_lib_release ( ) ;
}
2005-11-09 01:52:26 +03:00
void dev_manager_exit ( void )
{
dm_lib_exit ( ) ;
2004-05-13 00:40:34 +04:00
}
2005-11-09 01:52:26 +03:00
int dev_manager_snapshot_percent ( struct dev_manager * dm ,
2006-04-06 18:06:27 +04:00
const struct logical_volume * lv ,
2014-06-09 14:08:27 +04:00
dm_percent_t * percent )
2004-05-13 00:40:34 +04:00
{
2012-01-21 01:56:01 +04:00
const struct logical_volume * snap_lv ;
2005-11-09 01:52:26 +03:00
char * name ;
const char * dlid ;
2010-02-10 17:38:24 +03:00
int fail_if_percent_unsupported = 0 ;
if ( lv_is_merging_origin ( lv ) ) {
/*
* Set ' fail_if_percent_unsupported ' , otherwise passing
* unsupported LV types to _percent will lead to a default
* successful return with percent_range as PERCENT_100 .
* - For a merging origin , this will result in a polldaemon
* that runs infinitely ( because completion is PERCENT_0 )
* - We unfortunately don ' t yet _know_ if a snapshot - merge
* target is active ( activation is deferred if dev is open ) ;
* so we can ' t short - circuit origin devices based purely on
* existing LVM LV attributes .
*/
fail_if_percent_unsupported = 1 ;
}
2005-11-09 01:52:26 +03:00
2012-01-21 01:56:01 +04:00
if ( lv_is_merging_cow ( lv ) ) {
/* must check percent of origin for a merging snapshot */
snap_lv = origin_from_cow ( lv ) ;
} else
snap_lv = lv ;
2005-11-09 01:52:26 +03:00
/*
* Build a name for the top layer .
*/
2012-01-21 01:56:01 +04:00
if ( ! ( name = dm_build_dm_name ( dm - > mem , snap_lv - > vg - > name , snap_lv - > name , NULL ) ) )
2005-11-09 01:52:26 +03:00
return_0 ;
2014-03-11 20:13:47 +04:00
if ( ! ( dlid = build_dm_uuid ( dm - > mem , snap_lv , NULL ) ) )
2005-11-09 01:52:26 +03:00
return_0 ;
2004-05-13 00:40:34 +04:00
2005-11-09 01:52:26 +03:00
/*
* Try and get some info on this device .
*/
2016-03-22 20:46:15 +03:00
if ( ! _percent ( dm , name , dlid , TARGET_NAME_SNAPSHOT , 0 , NULL , percent ,
2013-07-17 01:35:56 +04:00
NULL , fail_if_percent_unsupported ) )
2008-01-30 16:19:47 +03:00
return_0 ;
2004-05-13 00:40:34 +04:00
2005-11-09 01:52:26 +03:00
/* If the snapshot isn't available, percent will be -1 */
2004-05-13 00:40:34 +04:00
return 1 ;
}
2005-11-09 01:52:26 +03:00
/* FIXME Merge with snapshot_percent, auto-detecting target type */
/* FIXME Cope with more than one target */
int dev_manager_mirror_percent ( struct dev_manager * dm ,
2010-01-16 01:58:25 +03:00
const struct logical_volume * lv , int wait ,
2014-06-09 14:08:27 +04:00
dm_percent_t * percent , uint32_t * event_nr )
2002-02-25 17:46:57 +03:00
{
2005-11-09 01:52:26 +03:00
char * name ;
const char * dlid ;
2011-08-03 02:07:20 +04:00
const char * target_type = first_seg ( lv ) - > segtype - > name ;
2013-02-02 03:16:36 +04:00
const char * layer = lv_layer ( lv ) ;
2002-02-25 17:46:57 +03:00
2002-03-26 16:41:37 +03:00
/*
2005-11-09 01:52:26 +03:00
* Build a name for the top layer .
2002-03-26 16:41:37 +03:00
*/
2011-08-30 18:55:15 +04:00
if ( ! ( name = dm_build_dm_name ( dm - > mem , lv - > vg - > name , lv - > name , layer ) ) )
2005-11-09 01:52:26 +03:00
return_0 ;
2002-03-08 13:41:48 +03:00
2015-11-25 17:45:19 +03:00
if ( ! ( dlid = build_dm_uuid ( dm - > mem , lv , layer ) ) )
return_0 ;
2002-03-15 00:17:30 +03:00
2016-12-01 16:53:35 +03:00
log_debug_activation ( " Getting device %s status percentage for %s. " ,
2013-01-08 02:30:29 +04:00
target_type , name ) ;
2016-12-01 16:53:35 +03:00
if ( ! _percent ( dm , name , dlid , target_type , wait , lv , percent , event_nr , 0 ) )
2008-01-30 16:19:47 +03:00
return_0 ;
2002-03-15 00:17:30 +03:00
2005-11-09 01:52:26 +03:00
return 1 ;
}
2003-07-05 02:34:56 +04:00
2013-02-01 21:31:47 +04:00
int dev_manager_raid_status ( struct dev_manager * dm ,
const struct logical_volume * lv ,
2021-03-17 11:50:09 +03:00
struct lv_status_raid * * status , int * exists )
2013-02-01 21:31:47 +04:00
{
int r = 0 ;
const char * dlid ;
struct dm_task * dmt ;
struct dm_info info ;
uint64_t start , length ;
char * type = NULL ;
char * params = NULL ;
2013-02-05 03:10:16 +04:00
const char * layer = lv_layer ( lv ) ;
2021-03-17 11:50:09 +03:00
struct dm_status_raid * sr ;
* exists = - 1 ;
if ( ! ( * status = dm_pool_zalloc ( dm - > mem , sizeof ( struct lv_status_cache ) ) ) )
return_0 ;
2013-02-01 21:31:47 +04:00
2014-03-11 20:13:47 +04:00
if ( ! ( dlid = build_dm_uuid ( dm - > mem , lv , layer ) ) )
2013-02-01 21:31:47 +04:00
return_0 ;
2016-12-01 16:56:32 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_STATUS , & info , NULL , dlid , 0 , 0 , 0 , 0 , 0 , 0 ) ) )
2013-02-01 21:31:47 +04:00
return_0 ;
2021-03-17 11:50:09 +03:00
if ( ! ( * exists = info . exists ) )
goto out ;
log_debug_activation ( " Checking raid status for volume %s. " ,
display_lvname ( lv ) ) ;
2013-02-01 21:31:47 +04:00
dm_get_next_target ( dmt , NULL , & start , & length , & type , & params ) ;
2016-12-01 16:53:35 +03:00
if ( ! type | | strcmp ( type , TARGET_NAME_RAID ) ) {
log_error ( " Expected %s segment type but got %s instead. " ,
TARGET_NAME_RAID , type ? type : " NULL " ) ;
2013-04-09 00:04:08 +04:00
goto out ;
}
2013-06-15 04:28:54 +04:00
/* FIXME Check there's only one target */
2021-03-17 11:50:09 +03:00
if ( ! dm_get_status_raid ( dm - > mem , params , & sr ) )
2013-02-01 21:31:47 +04:00
goto_out ;
2021-03-17 11:50:09 +03:00
( * status ) - > mem = dm - > mem ; /* User has to destroy this mem pool later */
( * status ) - > raid = sr ;
( * status ) - > in_sync = dm_make_percent ( sr - > insync_regions , sr - > total_regions ) ;
2013-02-01 21:31:47 +04:00
r = 1 ;
out :
dm_task_destroy ( dmt ) ;
return r ;
}
2013-04-12 00:33:59 +04:00
int dev_manager_raid_message ( struct dev_manager * dm ,
const struct logical_volume * lv ,
const char * msg )
{
int r = 0 ;
const char * dlid ;
struct dm_task * dmt ;
const char * layer = lv_layer ( lv ) ;
2014-09-16 00:33:53 +04:00
if ( ! lv_is_raid ( lv ) ) {
2016-12-01 16:53:35 +03:00
log_error ( INTERNAL_ERROR " %s is not a RAID logical volume. " ,
2015-11-25 18:06:31 +03:00
display_lvname ( lv ) ) ;
2013-04-12 00:33:59 +04:00
return 0 ;
}
2017-03-04 00:10:21 +03:00
/* These are the supported RAID messages for dm-raid v1.9.0 */
2015-11-10 23:32:46 +03:00
if ( strcmp ( msg , " idle " ) & &
strcmp ( msg , " frozen " ) & &
strcmp ( msg , " resync " ) & &
strcmp ( msg , " recover " ) & &
strcmp ( msg , " check " ) & &
2017-03-04 00:10:21 +03:00
strcmp ( msg , " repair " ) ) {
2016-12-01 16:53:35 +03:00
log_error ( INTERNAL_ERROR " Unknown RAID message: %s. " , msg ) ;
2013-04-12 00:33:59 +04:00
return 0 ;
}
2014-03-11 20:13:47 +04:00
if ( ! ( dlid = build_dm_uuid ( dm - > mem , lv , layer ) ) )
2013-04-12 00:33:59 +04:00
return_0 ;
2016-12-01 16:56:32 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_TARGET_MSG , NULL , NULL , dlid , 0 , 0 , 0 , 0 , 1 , 0 ) ) )
2013-04-12 00:33:59 +04:00
return_0 ;
if ( ! dm_task_set_message ( dmt , msg ) )
goto_out ;
if ( ! dm_task_run ( dmt ) )
goto_out ;
r = 1 ;
out :
dm_task_destroy ( dmt ) ;
return r ;
}
2018-08-27 22:53:09 +03:00
int dev_manager_writecache_message ( struct dev_manager * dm ,
const struct logical_volume * lv ,
const char * msg )
{
int r = 0 ;
const char * dlid ;
struct dm_task * dmt ;
const char * layer = lv_layer ( lv ) ;
if ( ! lv_is_writecache ( lv ) ) {
log_error ( INTERNAL_ERROR " %s is not a writecache logical volume. " ,
display_lvname ( lv ) ) ;
return 0 ;
}
if ( ! ( dlid = build_dm_uuid ( dm - > mem , lv , layer ) ) )
return_0 ;
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_TARGET_MSG , NULL , NULL , dlid , 0 , 0 , 0 , 0 , 1 , 0 ) ) )
return_0 ;
if ( ! dm_task_set_message ( dmt , msg ) )
goto_out ;
if ( ! dm_task_run ( dmt ) )
goto_out ;
r = 1 ;
out :
dm_task_destroy ( dmt ) ;
return r ;
}
2014-01-28 22:24:51 +04:00
int dev_manager_cache_status ( struct dev_manager * dm ,
const struct logical_volume * lv ,
2021-03-17 13:20:23 +03:00
struct lv_status_cache * * status , int * exists )
2014-01-28 22:24:51 +04:00
{
int r = 0 ;
const char * dlid ;
struct dm_task * dmt ;
struct dm_info info ;
uint64_t start , length ;
char * type = NULL ;
char * params = NULL ;
2014-11-03 14:52:29 +03:00
struct dm_status_cache * c ;
2014-01-28 22:24:51 +04:00
2021-03-17 13:20:23 +03:00
* exists = - 1 ;
2014-11-03 14:52:29 +03:00
if ( ! ( dlid = build_dm_uuid ( dm - > mem , lv , lv_layer ( lv ) ) ) )
return_0 ;
2016-12-01 16:56:32 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_STATUS , & info , NULL , dlid , 0 , 0 , 0 , 0 , 0 , 0 ) ) )
2014-01-28 22:24:51 +04:00
return_0 ;
2021-03-17 13:20:23 +03:00
if ( ! ( * exists = info . exists ) )
goto out ;
log_debug_activation ( " Checking status for cache volume %s. " ,
display_lvname ( lv ) ) ;
2014-01-28 22:24:51 +04:00
dm_get_next_target ( dmt , NULL , & start , & length , & type , & params ) ;
2016-03-22 20:46:15 +03:00
if ( ! type | | strcmp ( type , TARGET_NAME_CACHE ) ) {
2016-12-01 16:53:35 +03:00
log_error ( " Expected %s segment type but got %s instead. " ,
TARGET_NAME_CACHE , type ? type : " NULL " ) ;
2014-01-28 22:24:51 +04:00
goto out ;
}
2014-11-03 14:52:29 +03:00
/*
* FIXME :
* - > target_percent ( ) API is able to transfer only a single value .
* Needs to be able to pass whole structure .
*/
2016-03-30 11:16:09 +03:00
if ( ! dm_get_status_cache ( dm - > mem , params , & c ) )
2014-01-28 22:24:51 +04:00
goto_out ;
2020-09-27 02:11:47 +03:00
if ( ! ( * status = dm_pool_zalloc ( dm - > mem , sizeof ( struct lv_status_cache ) ) ) )
goto_out ;
2014-11-06 22:36:53 +03:00
( * status ) - > mem = dm - > mem ; /* User has to destroy this mem pool later */
2020-09-27 02:11:47 +03:00
( * status ) - > cache = c ;
2016-03-10 19:56:43 +03:00
if ( c - > fail | | c - > error ) {
( * status ) - > data_usage =
( * status ) - > metadata_usage =
( * status ) - > dirty_usage = DM_PERCENT_INVALID ;
} else {
( * status ) - > data_usage = dm_make_percent ( c - > used_blocks ,
c - > total_blocks ) ;
( * status ) - > metadata_usage = dm_make_percent ( c - > metadata_used_blocks ,
c - > metadata_total_blocks ) ;
2016-09-09 22:26:38 +03:00
( * status ) - > dirty_usage = ( c - > used_blocks ) ?
2016-09-09 21:45:37 +03:00
dm_make_percent ( c - > dirty_blocks ,
c - > used_blocks ) : DM_PERCENT_0 ;
2016-03-10 19:56:43 +03:00
}
2014-01-28 22:24:51 +04:00
r = 1 ;
out :
dm_task_destroy ( dmt ) ;
return r ;
}
2012-01-25 12:48:42 +04:00
int dev_manager_thin_pool_status ( struct dev_manager * dm ,
2020-09-27 02:11:47 +03:00
const struct logical_volume * lv , int flush ,
2021-03-17 13:20:23 +03:00
struct lv_status_thin_pool * * status , int * exists )
2012-01-25 12:48:42 +04:00
{
2020-09-27 02:11:47 +03:00
struct dm_status_thin_pool * dm_status ;
2012-01-25 12:48:42 +04:00
const char * dlid ;
struct dm_task * dmt ;
struct dm_info info ;
uint64_t start , length ;
char * type = NULL ;
char * params = NULL ;
int r = 0 ;
2021-03-17 13:20:23 +03:00
* exists = - 1 ;
if ( ! ( * status = dm_pool_zalloc ( dm - > mem , sizeof ( struct lv_status_thin_pool ) ) ) )
return_0 ;
2012-01-25 12:48:42 +04:00
/* Build dlid for the thin pool layer */
2014-03-11 20:13:47 +04:00
if ( ! ( dlid = build_dm_uuid ( dm - > mem , lv , lv_layer ( lv ) ) ) )
2012-01-25 12:48:42 +04:00
return_0 ;
2016-12-01 16:56:32 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_STATUS , & info , NULL , dlid , 0 , 0 , 0 , 0 , flush , 0 ) ) )
2012-01-25 12:48:42 +04:00
return_0 ;
2021-03-17 13:20:23 +03:00
if ( ! ( * exists = info . exists ) )
goto out ;
log_debug_activation ( " Checking thin pool status for LV %s. " ,
display_lvname ( lv ) ) ;
2012-01-25 12:48:42 +04:00
dm_get_next_target ( dmt , NULL , & start , & length , & type , & params ) ;
2020-09-27 02:11:47 +03:00
if ( ! type | | strcmp ( type , TARGET_NAME_THIN_POOL ) ) {
log_error ( " Expected %s segment type but got %s instead. " ,
TARGET_NAME_THIN_POOL , type ? type : " NULL " ) ;
goto out ;
}
2013-06-15 04:28:54 +04:00
2020-09-27 02:11:47 +03:00
if ( ! dm_get_status_thin_pool ( dm - > mem , params , & dm_status ) )
2012-01-25 12:48:42 +04:00
goto_out ;
2020-09-27 02:11:47 +03:00
( * status ) - > mem = dm - > mem ;
( * status ) - > thin_pool = dm_status ;
if ( dm_status - > fail | | dm_status - > error ) {
( * status ) - > data_usage =
( * status ) - > metadata_usage = DM_PERCENT_INVALID ;
} else {
( * status ) - > data_usage = dm_make_percent ( dm_status - > used_data_blocks ,
dm_status - > total_data_blocks ) ;
( * status ) - > metadata_usage = dm_make_percent ( dm_status - > used_metadata_blocks ,
dm_status - > total_metadata_blocks ) ;
}
2012-01-25 12:48:42 +04:00
r = 1 ;
out :
dm_task_destroy ( dmt ) ;
return r ;
}
2020-09-27 02:11:47 +03:00
int dev_manager_thin_status ( struct dev_manager * dm ,
const struct logical_volume * lv , int flush ,
2021-03-17 13:20:23 +03:00
struct lv_status_thin * * status , int * exists )
2011-12-21 17:09:33 +04:00
{
2020-09-27 02:11:47 +03:00
struct dm_status_thin * dm_status ;
2011-12-21 17:09:33 +04:00
const char * dlid ;
2020-09-27 02:11:47 +03:00
struct dm_task * dmt ;
struct dm_info info ;
uint64_t start , length ;
char * type = NULL ;
char * params = NULL ;
uint64_t csize ;
int r = 0 ;
2011-12-21 17:09:33 +04:00
2021-03-17 13:20:23 +03:00
* exists = - 1 ;
if ( ! ( * status = dm_pool_zalloc ( dm - > mem , sizeof ( struct lv_status_thin ) ) ) )
return_0 ;
2020-09-27 02:11:47 +03:00
if ( ! ( dlid = build_dm_uuid ( dm - > mem , lv , lv_layer ( lv ) ) ) )
2011-12-21 17:09:33 +04:00
return_0 ;
2020-09-27 02:11:47 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_STATUS , & info , NULL , dlid , 0 , 0 , 0 , 0 , flush , 0 ) ) )
2011-12-21 17:09:33 +04:00
return_0 ;
2021-03-17 13:20:23 +03:00
if ( ! ( * exists = info . exists ) )
goto out ;
log_debug_activation ( " Checking thin status for LV %s. " ,
display_lvname ( lv ) ) ;
2011-12-21 17:09:33 +04:00
2020-09-27 02:11:47 +03:00
dm_get_next_target ( dmt , NULL , & start , & length , & type , & params ) ;
2011-12-21 17:09:33 +04:00
2020-09-27 02:11:47 +03:00
if ( ! type | | strcmp ( type , TARGET_NAME_THIN ) ) {
log_error ( " Expected %s segment type but got %s instead. " ,
TARGET_NAME_THIN , type ? type : " NULL " ) ;
goto out ;
}
2012-01-19 19:27:54 +04:00
2020-09-27 02:11:47 +03:00
if ( ! dm_get_status_thin ( dm - > mem , params , & dm_status ) )
goto_out ;
2012-01-19 19:27:54 +04:00
2020-09-27 02:11:47 +03:00
( * status ) - > mem = dm - > mem ;
( * status ) - > thin = dm_status ;
if ( dm_status - > fail )
( * status ) - > usage = DM_PERCENT_INVALID ;
else {
/* Pool allocates whole chunk so round-up to nearest one */
csize = first_seg ( first_seg ( lv ) - > pool_lv ) - > chunk_size ;
csize = ( ( lv - > size + csize - 1 ) / csize ) * csize ;
if ( dm_status - > mapped_sectors > csize ) {
log_warn ( " WARNING: LV %s maps %s while the size is only %s. " ,
display_lvname ( lv ) ,
display_size ( dm - > cmd , dm_status - > mapped_sectors ) ,
display_size ( dm - > cmd , csize ) ) ;
/* Don't show nonsense numbers like i.e. 1000% full */
dm_status - > mapped_sectors = csize ;
}
( * status ) - > usage = dm_make_percent ( dm_status - > mapped_sectors , csize ) ;
}
2016-12-01 16:53:35 +03:00
2020-09-27 02:11:47 +03:00
r = 1 ;
out :
dm_task_destroy ( dmt ) ;
2012-01-19 19:27:54 +04:00
2020-09-27 02:11:47 +03:00
return r ;
2012-01-19 19:27:54 +04:00
}
2019-10-26 00:31:08 +03:00
/*
* Explore state of running DM table to obtain currently used deviceId
*/
2013-12-04 16:57:27 +04:00
int dev_manager_thin_device_id ( struct dev_manager * dm ,
const struct logical_volume * lv ,
2021-03-17 13:20:23 +03:00
uint32_t * device_id , int * exists )
2013-12-04 16:57:27 +04:00
{
const char * dlid ;
struct dm_task * dmt ;
struct dm_info info ;
uint64_t start , length ;
char * params , * target_type = NULL ;
2019-10-26 00:31:08 +03:00
const char * layer = lv_layer ( lv ) ;
2013-12-04 16:57:27 +04:00
int r = 0 ;
2021-03-17 13:20:23 +03:00
* exists = - 1 ;
2019-10-26 00:31:08 +03:00
if ( lv_is_merging_origin ( lv ) & & ! lv_info ( lv - > vg - > cmd , lv , 1 , NULL , 0 , 0 ) )
/* If the merge has already happened, that table
* can already be using correct LV without - real layer */
layer = NULL ;
2013-12-04 16:57:27 +04:00
/* Build dlid for the thin layer */
2019-10-26 00:31:08 +03:00
if ( ! ( dlid = build_dm_uuid ( dm - > mem , lv , layer ) ) )
2013-12-04 16:57:27 +04:00
return_0 ;
2016-12-01 16:56:32 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_TABLE , & info , NULL , dlid , 0 , 0 , 0 , 0 , 1 , 0 ) ) )
2013-12-04 16:57:27 +04:00
return_0 ;
2021-03-17 13:20:23 +03:00
if ( ! ( * exists = info . exists ) )
goto out ;
log_debug_activation ( " Checking device id for LV %s. " ,
display_lvname ( lv ) ) ;
2013-12-04 16:57:27 +04:00
if ( dm_get_next_target ( dmt , NULL , & start , & length ,
& target_type , & params ) ) {
2015-11-25 18:06:31 +03:00
log_error ( " More then one table line found for %s. " ,
display_lvname ( lv ) ) ;
2013-12-04 16:57:27 +04:00
goto out ;
}
2016-03-22 20:46:15 +03:00
if ( ! target_type | | strcmp ( target_type , TARGET_NAME_THIN ) ) {
2015-11-25 18:06:31 +03:00
log_error ( " Unexpected target type %s found for thin %s. " ,
target_type , display_lvname ( lv ) ) ;
2013-12-04 16:57:27 +04:00
goto out ;
}
2015-11-09 11:27:18 +03:00
if ( ! params | | sscanf ( params , " %*u:%*u %u " , device_id ) ! = 1 ) {
2015-11-25 18:06:31 +03:00
log_error ( " Cannot parse table like parameters %s for %s. " ,
params , display_lvname ( lv ) ) ;
2013-12-04 16:57:27 +04:00
goto out ;
}
r = 1 ;
out :
dm_task_destroy ( dmt ) ;
return r ;
}
2018-06-29 12:15:54 +03:00
int dev_manager_vdo_pool_status ( struct dev_manager * dm ,
2021-03-17 13:20:23 +03:00
const struct logical_volume * lv , int flush ,
struct lv_status_vdo * * status , int * exists )
2018-06-29 12:15:54 +03:00
{
const char * dlid ;
struct dm_info info ;
uint64_t start , length ;
struct dm_task * dmt = NULL ;
char * type = NULL ;
char * params = NULL ;
int r = 0 ;
2021-03-17 13:20:23 +03:00
* exists = - 1 ;
if ( ! ( * status = dm_pool_zalloc ( dm - > mem , sizeof ( struct lv_status_vdo ) ) ) )
return_0 ;
2018-06-29 12:15:54 +03:00
if ( ! ( dlid = build_dm_uuid ( dm - > mem , lv , lv_layer ( lv ) ) ) )
return_0 ;
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_STATUS , & info , NULL , dlid , 0 , 0 , 0 , 0 , flush , 0 ) ) )
return_0 ;
2021-03-17 13:20:23 +03:00
if ( ! ( * exists = info . exists ) )
goto out ;
log_debug_activation ( " Checking VDO pool status for LV %s. " ,
display_lvname ( lv ) ) ;
2018-06-29 12:15:54 +03:00
if ( dm_get_next_target ( dmt , NULL , & start , & length , & type , & params ) ) {
log_error ( " More then one table line found for %s. " ,
display_lvname ( lv ) ) ;
goto out ;
}
if ( ! type | | strcmp ( type , TARGET_NAME_VDO ) ) {
log_error ( " Expected %s segment type but got %s instead. " ,
TARGET_NAME_VDO , type ? type : " NULL " ) ;
goto out ;
}
2024-04-05 15:00:11 +03:00
if ( ! _vdo_pool_message_stats ( dm - > mem , lv , * status ) )
2024-02-19 16:13:20 +03:00
stack ;
2021-09-09 15:59:38 +03:00
if ( ! parse_vdo_pool_status ( dm - > mem , lv , params , & info , * status ) )
2019-01-21 02:48:05 +03:00
goto_out ;
2018-06-29 12:15:54 +03:00
2021-03-17 13:20:23 +03:00
( * status ) - > mem = dm - > mem ;
2018-06-29 12:15:54 +03:00
r = 1 ;
out :
dm_task_destroy ( dmt ) ;
return r ;
}
2023-01-15 23:24:28 +03:00
int dev_manager_vdo_pool_size_config ( struct dev_manager * dm ,
const struct logical_volume * lv ,
struct vdo_pool_size_config * cfg )
{
const char * dlid ;
struct dm_info info ;
uint64_t start , length ;
struct dm_task * dmt = NULL ;
char * type = NULL ;
char * params = NULL ;
int r = 0 ;
unsigned version = 0 ;
memset ( cfg , 0 , sizeof ( * cfg ) ) ;
if ( ! ( dlid = build_dm_uuid ( dm - > mem , lv , lv_layer ( lv ) ) ) )
return_0 ;
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_TABLE , & info , NULL , dlid , 0 , 0 , 0 , 0 , 0 , 0 ) ) )
return_0 ;
if ( ! info . exists )
goto inactive ; /* VDO device is not active, should not happen here... */
log_debug_activation ( " Checking VDO pool table line for LV %s. " ,
display_lvname ( lv ) ) ;
if ( dm_get_next_target ( dmt , NULL , & start , & length , & type , & params ) ) {
log_error ( " More then one table line found for %s. " ,
display_lvname ( lv ) ) ;
goto out ;
}
if ( ! type | | strcmp ( type , TARGET_NAME_VDO ) ) {
log_error ( " Expected %s segment type but got %s instead. " ,
TARGET_NAME_VDO , type ? type : " NULL " ) ;
goto out ;
}
if ( sscanf ( params , " V%u %*s " FMTu64 " %*u " FMTu32 ,
& version , & cfg - > physical_size , & cfg - > block_map_cache_size_mb ) ! = 3 ) {
log_error ( " Failed to parse VDO parameters %s for LV %s. " ,
params , display_lvname ( lv ) ) ;
goto out ;
}
switch ( version ) {
case 2 : break ;
case 4 : break ;
default : log_warn ( " WARNING: Unknown VDO table line version %u. " , version ) ;
}
cfg - > virtual_size = length ;
cfg - > physical_size * = 8 ; // From 4K unit to 512B
cfg - > block_map_cache_size_mb / = 256 ; // From 4K unit to MiB
cfg - > index_memory_size_mb = first_seg ( lv ) - > vdo_params . index_memory_size_mb ; // Preserved
inactive :
r = 1 ;
out :
dm_task_destroy ( dmt ) ;
return r ;
}
2013-12-04 16:57:27 +04:00
2005-11-09 01:52:26 +03:00
/*************************/
/* NEW CODE STARTS HERE */
/*************************/
2010-02-24 23:00:56 +03:00
static int _dev_manager_lv_mknodes ( const struct logical_volume * lv )
2005-11-09 01:52:26 +03:00
{
char * name ;
2011-08-30 18:55:15 +04:00
if ( ! ( name = dm_build_dm_name ( lv - > vg - > cmd - > mem , lv - > vg - > name ,
2015-11-25 18:06:31 +03:00
lv - > name , NULL ) ) )
2005-11-09 01:52:26 +03:00
return_0 ;
return fs_add_lv ( lv , name ) ;
2002-03-18 16:09:27 +03:00
}
2010-02-24 23:00:56 +03:00
static int _dev_manager_lv_rmnodes ( const struct logical_volume * lv )
2003-04-25 02:09:13 +04:00
{
2005-11-09 01:52:26 +03:00
return fs_del_lv ( lv ) ;
}
2016-06-30 16:54:27 +03:00
static int _lv_has_mknode ( const struct logical_volume * lv )
{
return ( lv_is_visible ( lv ) & &
( ! lv_is_thin_pool ( lv ) | | lv_is_new_thin_pool ( lv ) ) ) ;
}
2010-02-24 23:00:56 +03:00
int dev_manager_mknodes ( const struct logical_volume * lv )
{
struct dm_info dminfo ;
2016-12-02 15:46:47 +03:00
struct dm_task * dmt ;
2010-08-03 17:13:01 +04:00
char * name ;
2010-02-24 23:00:56 +03:00
int r = 0 ;
2011-08-30 18:55:15 +04:00
if ( ! ( name = dm_build_dm_name ( lv - > vg - > cmd - > mem , lv - > vg - > name , lv - > name , NULL ) ) )
2010-02-24 23:00:56 +03:00
return_0 ;
2016-12-02 15:46:47 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_MKNODES , & dminfo , name , NULL , 0 , 0 , 0 , 0 , 0 , 0 ) ) )
return_0 ;
if ( dminfo . exists ) {
2018-02-28 19:22:09 +03:00
/* read-only component LV is also made visible */
if ( _lv_has_mknode ( lv ) | | ( dminfo . read_only & & lv_is_component ( lv ) ) )
2016-12-02 15:46:47 +03:00
r = _dev_manager_lv_mknodes ( lv ) ;
2023-07-24 20:03:06 +03:00
else
r = 1 ;
2016-12-02 15:46:47 +03:00
} else
r = _dev_manager_lv_rmnodes ( lv ) ;
dm_task_destroy ( dmt ) ;
2010-02-24 23:00:56 +03:00
return r ;
}
2013-05-13 13:46:24 +04:00
# ifdef UDEV_SYNC_SUPPORT
/*
* Until the DM_UEVENT_GENERATED_FLAG was introduced in kernel patch
* 856 a6f1dbd8940e72755af145ebcd806408ecedd
* some operations could not be performed by udev , requiring our fallback code .
*/
static int _dm_driver_has_stable_udev_support ( void )
{
char vsn [ 80 ] ;
unsigned maj , min , patchlevel ;
return driver_version ( vsn , sizeof ( vsn ) ) & &
( sscanf ( vsn , " %u.%u.%u " , & maj , & min , & patchlevel ) = = 3 ) & &
( maj = = 4 ? min > = 18 : maj > 4 ) ;
}
static int _check_udev_fallback ( struct cmd_context * cmd )
{
struct config_info * settings = & cmd - > current_settings ;
if ( settings - > udev_fallback ! = - 1 )
goto out ;
/*
* Use udev fallback automatically in case udev
* is disabled via DM_DISABLE_UDEV environment
* variable or udev rules are switched off .
*/
settings - > udev_fallback = ! settings - > udev_rules ? 1 :
2013-06-25 14:31:53 +04:00
find_config_tree_bool ( cmd , activation_verify_udev_operations_CFG , NULL ) ;
2013-05-13 13:46:24 +04:00
/* Do not rely fully on udev if the udev support is known to be incomplete. */
if ( ! settings - > udev_fallback & & ! _dm_driver_has_stable_udev_support ( ) ) {
log_very_verbose ( " Kernel driver has incomplete udev support so "
" LVM will check and perform some operations itself. " ) ;
settings - > udev_fallback = 1 ;
}
out :
return settings - > udev_fallback ;
}
# else /* UDEV_SYNC_SUPPORT */
static int _check_udev_fallback ( struct cmd_context * cmd )
{
/* We must use old node/symlink creation code if not compiled with udev support at all! */
return cmd - > current_settings . udev_fallback = 1 ;
}
# endif /* UDEV_SYNC_SUPPORT */
2014-09-22 17:50:07 +04:00
static uint16_t _get_udev_flags ( struct dev_manager * dm , const struct logical_volume * lv ,
2018-02-28 19:22:09 +03:00
const char * layer , int noscan , int temporary ,
int visible_component )
2010-04-23 18:16:32 +04:00
{
uint16_t udev_flags = 0 ;
2011-06-17 18:50:53 +04:00
/*
* Instruct also libdevmapper to disable udev
* fallback in accordance to LVM2 settings .
*/
2013-05-13 13:46:24 +04:00
if ( ! _check_udev_fallback ( dm - > cmd ) )
2011-06-17 18:50:53 +04:00
udev_flags | = DM_UDEV_DISABLE_LIBRARY_FALLBACK ;
2010-04-23 18:16:32 +04:00
/*
* Is this top - level and visible device ?
* If not , create just the / dev / mapper content .
*/
2011-10-29 00:34:45 +04:00
/* FIXME: add target's method for this */
2018-02-28 19:22:09 +03:00
if ( lv_is_new_thin_pool ( lv ) | | visible_component )
2014-11-04 12:33:35 +03:00
/* New thin-pool is regular LV with -tpool UUID suffix. */
udev_flags | = DM_UDEV_DISABLE_DISK_RULES_FLAG |
DM_UDEV_DISABLE_OTHER_RULES_FLAG ;
2019-09-14 02:13:33 +03:00
else if ( layer | | ! lv_is_visible ( lv ) | | lv_is_thin_pool ( lv ) | | lv_is_vdo_pool ( lv ) )
2010-04-23 18:16:32 +04:00
udev_flags | = DM_UDEV_DISABLE_SUBSYSTEM_RULES_FLAG |
DM_UDEV_DISABLE_DISK_RULES_FLAG |
DM_UDEV_DISABLE_OTHER_RULES_FLAG ;
/*
* There ' s no need for other udev rules to touch special LVs with
* reserved names . We don ' t need to populate / dev / disk here either .
* Even if they happen to be visible and top - level .
*/
else if ( is_reserved_lvname ( lv - > name ) )
udev_flags | = DM_UDEV_DISABLE_DISK_RULES_FLAG |
DM_UDEV_DISABLE_OTHER_RULES_FLAG ;
/*
* Snapshots and origins could have the same rule applied that will
* give symlinks exactly the same name ( e . g . a name based on
* filesystem UUID ) . We give preference to origins to make such
* naming deterministic ( e . g . symlinks in / dev / disk / by - uuid ) .
*/
if ( lv_is_cow ( lv ) )
udev_flags | = DM_UDEV_LOW_PRIORITY_FLAG ;
/*
* Finally , add flags to disable / dev / mapper and / dev / < vgname > content
* to be created by udev if it is requested by user ' s configuration .
* This is basically an explicit fallback to old node / symlink creation
* without udev .
*/
if ( ! dm - > cmd - > current_settings . udev_rules )
udev_flags | = DM_UDEV_DISABLE_DM_RULES_FLAG |
DM_UDEV_DISABLE_SUBSYSTEM_RULES_FLAG ;
2013-10-08 15:27:21 +04:00
/*
activation: flag temporary LVs internally
Add LV_TEMPORARY flag for LVs with limited existence during command
execution. Such LVs are temporary in way that they need to be activated,
some action done and then removed immediately. Such LVs are just like
any normal LV - the only difference is that they are removed during
LVM command execution. This is also the case for LVs representing
future pool metadata spare LVs which we need to initialize by using
the usual LV before they are declared as pool metadata spare.
We can optimize some other parts like udev to do a better job if
it knows that the LV is temporary and any processing on it is just
useless.
This flag is orthogonal to LV_NOSCAN flag introduced recently
as LV_NOSCAN flag is primarily used to mark an LV for the scanning
to be avoided before the zeroing of the device happens. The LV_TEMPORARY
flag makes a difference between a full-fledged LV visible in the system
and the LV just used as a temporary overlay for some action that needs to
be done on underlying PVs.
For example: lvcreate --thinpool POOL --zero n -L 1G vg
- first, the usual LV is created to do a clean up for pool metadata
spare. The LV is activated, zeroed, deactivated.
- between "activated" and "zeroed" stage, the LV_NOSCAN flag is used
to avoid any scanning in udev
- betwen "zeroed" and "deactivated" stage, we need to avoid the WATCH
udev rule, but since the LV is just a usual LV, we can't make a
difference. The LV_TEMPORARY internal LV flag helps here. If we
create the LV with this flag, the DM_UDEV_DISABLE_DISK_RULES
and DM_UDEV_DISABLE_OTHER_RULES flag are set (just like as it is
with "invisible" and non-top-level LVs) - udev is directed to
skip WATCH rule use.
- if the LV_TEMPORARY flag was not used, there would normally be
a WATCH event generated once the LV is closed after "zeroed"
stage. This will make problems with immediated deactivation that
follows.
2013-10-23 16:06:39 +04:00
* LVM subsystem specific flags .
2013-10-08 15:27:21 +04:00
*/
activation: flag temporary LVs internally
Add LV_TEMPORARY flag for LVs with limited existence during command
execution. Such LVs are temporary in way that they need to be activated,
some action done and then removed immediately. Such LVs are just like
any normal LV - the only difference is that they are removed during
LVM command execution. This is also the case for LVs representing
future pool metadata spare LVs which we need to initialize by using
the usual LV before they are declared as pool metadata spare.
We can optimize some other parts like udev to do a better job if
it knows that the LV is temporary and any processing on it is just
useless.
This flag is orthogonal to LV_NOSCAN flag introduced recently
as LV_NOSCAN flag is primarily used to mark an LV for the scanning
to be avoided before the zeroing of the device happens. The LV_TEMPORARY
flag makes a difference between a full-fledged LV visible in the system
and the LV just used as a temporary overlay for some action that needs to
be done on underlying PVs.
For example: lvcreate --thinpool POOL --zero n -L 1G vg
- first, the usual LV is created to do a clean up for pool metadata
spare. The LV is activated, zeroed, deactivated.
- between "activated" and "zeroed" stage, the LV_NOSCAN flag is used
to avoid any scanning in udev
- betwen "zeroed" and "deactivated" stage, we need to avoid the WATCH
udev rule, but since the LV is just a usual LV, we can't make a
difference. The LV_TEMPORARY internal LV flag helps here. If we
create the LV with this flag, the DM_UDEV_DISABLE_DISK_RULES
and DM_UDEV_DISABLE_OTHER_RULES flag are set (just like as it is
with "invisible" and non-top-level LVs) - udev is directed to
skip WATCH rule use.
- if the LV_TEMPORARY flag was not used, there would normally be
a WATCH event generated once the LV is closed after "zeroed"
stage. This will make problems with immediated deactivation that
follows.
2013-10-23 16:06:39 +04:00
if ( noscan )
udev_flags | = DM_SUBSYSTEM_UDEV_FLAG0 ;
if ( temporary )
udev_flags | = DM_UDEV_DISABLE_DISK_RULES_FLAG |
DM_UDEV_DISABLE_OTHER_RULES_FLAG ;
2013-10-08 15:27:21 +04:00
2010-04-23 18:16:32 +04:00
return udev_flags ;
}
2017-11-24 15:55:42 +03:00
static int _add_lv_to_dtree ( struct dev_manager * dm , struct dm_tree * dtree ,
const struct logical_volume * lv , int origin_only ) ;
2019-08-20 13:30:25 +03:00
static int _add_new_lv_to_dtree ( struct dev_manager * dm , struct dm_tree * dtree ,
const struct logical_volume * lv ,
struct lv_activate_opts * laopts ,
const char * layer ) ;
/*
* Check for device holders ( ATM used only for removed pvmove targets )
* and add them into dtree structures .
* When ' laopts ! = NULL ' add them as new nodes - which also corrects READ_AHEAD .
2024-08-30 00:05:41 +03:00
* Note : correct table are already explicitly PRELOADED .
2019-08-20 13:30:25 +03:00
*/
2017-11-24 22:59:38 +03:00
static int _check_holder ( struct dev_manager * dm , struct dm_tree * dtree ,
2019-08-20 13:30:25 +03:00
const struct logical_volume * lv ,
struct lv_activate_opts * laopts ,
uint32_t major , const char * d_name )
2017-11-24 15:55:42 +03:00
{
const char * default_uuid_prefix = dm_uuid_prefix ( ) ;
const size_t default_uuid_prefix_len = strlen ( default_uuid_prefix ) ;
2017-11-24 22:59:38 +03:00
const char * name ;
const char * uuid ;
struct dm_info info ;
struct dm_task * dmt ;
2017-11-24 15:55:42 +03:00
struct logical_volume * lv_det ;
2017-11-24 22:59:38 +03:00
union lvid id ;
int dev , r = 0 ;
2017-11-24 15:55:42 +03:00
2017-11-24 22:59:38 +03:00
errno = 0 ;
dev = strtoll ( d_name + 3 , NULL , 10 ) ;
if ( errno ) {
log_error ( " Failed to parse dm device minor number from %s. " , d_name ) ;
2017-11-24 15:55:42 +03:00
return 0 ;
}
2017-11-24 22:59:38 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_INFO , & info , NULL , NULL , NULL ,
major , dev , 0 , 0 , 0 ) ) )
return_0 ;
2017-11-24 15:55:42 +03:00
2017-11-24 22:59:38 +03:00
if ( info . exists ) {
uuid = dm_task_get_uuid ( dmt ) ;
name = dm_task_get_name ( dmt ) ;
2017-11-24 15:55:42 +03:00
2017-11-24 22:59:38 +03:00
log_debug_activation ( " Checking holder of %s %s ( " FMTu32 " : " FMTu32 " ) %s. " ,
display_lvname ( lv ) , uuid , info . major , info . minor ,
name ) ;
2017-11-24 15:55:42 +03:00
2017-11-24 22:59:38 +03:00
/* Skip common uuid prefix */
2017-11-24 15:55:42 +03:00
if ( ! strncmp ( default_uuid_prefix , uuid , default_uuid_prefix_len ) )
uuid + = default_uuid_prefix_len ;
2021-08-03 23:32:33 +03:00
if ( ! memcmp ( uuid , & lv - > vg - > id , ID_LEN ) & &
2017-11-24 15:55:42 +03:00
! dm_tree_find_node_by_uuid ( dtree , uuid ) ) {
2018-11-08 12:02:28 +03:00
/* trims any UUID suffix (i.e. -cow) */
2024-04-04 18:03:30 +03:00
dm_strncpy ( ( char * ) & id , uuid , 2 * sizeof ( struct id ) + 1 ) ;
2017-11-24 22:59:38 +03:00
2017-11-24 15:55:42 +03:00
/* If UUID is not yet in dtree, look for matching LV */
2017-11-24 22:59:38 +03:00
if ( ! ( lv_det = find_lv_in_vg_by_lvid ( lv - > vg , & id ) ) ) {
log_error ( " Cannot find holder with device name %s in VG %s. " ,
name , lv - > vg - > name ) ;
2017-11-24 15:55:42 +03:00
goto out ;
}
if ( lv_is_cow ( lv_det ) )
lv_det = origin_from_cow ( lv_det ) ;
log_debug_activation ( " Found holder %s of %s. " ,
display_lvname ( lv_det ) ,
display_lvname ( lv ) ) ;
2019-08-20 13:30:25 +03:00
if ( ! laopts ) {
if ( ! _add_lv_to_dtree ( dm , dtree , lv_det , 0 ) )
goto_out ;
} else if ( ! _add_new_lv_to_dtree ( dm , dtree , lv_det , laopts , 0 ) )
goto_out ;
2017-11-24 15:55:42 +03:00
}
}
2017-11-24 22:59:38 +03:00
r = 1 ;
out :
dm_task_destroy ( dmt ) ;
return r ;
}
/*
* Add exiting devices which holds given LV device open .
* This is used in case when metadata already do not contain information
* i . e . PVMOVE is being finished and final table is going to be resumed .
*/
static int _add_holders_to_dtree ( struct dev_manager * dm , struct dm_tree * dtree ,
2019-08-20 13:30:25 +03:00
const struct logical_volume * lv ,
struct lv_activate_opts * laopts ,
const struct dm_info * info )
2017-11-24 22:59:38 +03:00
{
const char * sysfs_dir = dm_sysfs_dir ( ) ;
char sysfs_path [ PATH_MAX ] ;
struct dirent * dirent ;
DIR * d ;
int r = 0 ;
/* Sysfs path of holders */
if ( dm_snprintf ( sysfs_path , sizeof ( sysfs_path ) , " %sblock/dm- " FMTu32
" /holders " , sysfs_dir , info - > minor ) < 0 ) {
log_error ( " sysfs_path dm_snprintf failed. " ) ;
return 0 ;
}
if ( ! ( d = opendir ( sysfs_path ) ) ) {
log_sys_error ( " opendir " , sysfs_path ) ;
return 0 ;
}
while ( ( dirent = readdir ( d ) ) )
/* Expects minor is added to 'dm-' prefix */
if ( ! strncmp ( dirent - > d_name , " dm- " , 3 ) & &
2019-08-20 13:30:25 +03:00
! _check_holder ( dm , dtree , lv , laopts , info - > major , dirent - > d_name ) )
2017-11-24 22:59:38 +03:00
goto_out ;
2017-11-24 15:55:42 +03:00
r = 1 ;
out :
if ( closedir ( d ) )
2017-11-24 22:59:38 +03:00
log_sys_debug ( " closedir " , " holders " ) ;
2017-11-24 15:55:42 +03:00
return r ;
}
2005-11-09 16:08:41 +03:00
static int _add_dev_to_dtree ( struct dev_manager * dm , struct dm_tree * dtree ,
2014-09-22 17:50:07 +04:00
const struct logical_volume * lv , const char * layer )
2005-11-09 01:52:26 +03:00
{
char * dlid , * name ;
2008-12-19 18:23:03 +03:00
struct dm_info info , info2 ;
2024-06-03 00:10:05 +03:00
const struct dm_active_device * dm_dev ;
2003-04-25 02:09:13 +04:00
2011-08-30 18:55:15 +04:00
if ( ! ( name = dm_build_dm_name ( dm - > mem , lv - > vg - > name , lv - > name , layer ) ) )
2005-11-09 01:52:26 +03:00
return_0 ;
2003-04-25 02:09:13 +04:00
2019-08-27 13:18:47 +03:00
if ( ! ( dlid = build_dm_uuid ( dm - > track_pending_delete ? dm - > cmd - > pending_delete_mem : dm - > mem , lv , layer ) ) )
2005-11-09 01:52:26 +03:00
return_0 ;
2024-07-08 23:32:41 +03:00
if ( dm_devs_cache_use ( ) ) {
2024-07-01 19:50:59 +03:00
if ( ! ( dm_dev = dm_devs_cache_get_by_uuid ( dm - > cmd , dlid ) ) ) {
2022-02-16 02:33:32 +03:00
log_debug ( " Cached as not present %s. " , name ) ;
return 1 ;
}
info = ( struct dm_info ) {
. exists = 1 ,
2024-06-03 00:10:05 +03:00
. major = MAJOR ( dm_dev - > devno ) ,
. minor = MINOR ( dm_dev - > devno ) ,
2022-02-16 02:33:32 +03:00
} ;
log_debug ( " Cached as present %s %s (%d:%d). " ,
name , dlid , info . major , info . minor ) ;
} else if ( ! _info ( dm - > cmd , name , dlid , 0 , 0 , 0 , & info , NULL , NULL ) )
2016-12-18 16:57:16 +03:00
return_0 ;
2008-12-19 18:23:03 +03:00
/*
* For top level volumes verify that existing device match
* requested major / minor and that major / minor pair is available for use
*/
if ( ! layer & & lv - > major ! = - 1 & & lv - > minor ! = - 1 ) {
2009-02-12 23:42:07 +03:00
/*
* FIXME compare info . major with lv - > major if multiple major support
*/
2016-04-12 12:47:21 +03:00
if ( info . exists & & ( ( int ) info . minor ! = lv - > minor ) ) {
2008-12-19 18:23:03 +03:00
log_error ( " Volume %s (% " PRIu32 " :% " PRIu32 " ) "
" differs from already active device "
2016-12-01 16:53:35 +03:00
" (% " PRIu32 " :% " PRIu32 " ). " ,
2015-11-25 18:06:31 +03:00
display_lvname ( lv ) , lv - > major , lv - > minor ,
info . major , info . minor ) ;
2008-12-19 18:23:03 +03:00
return 0 ;
}
if ( ! info . exists & & _info_by_dev ( lv - > major , lv - > minor , & info2 ) & &
info2 . exists ) {
log_error ( " The requested major:minor pair "
2016-12-01 16:53:35 +03:00
" (% " PRIu32 " :% " PRIu32 " ) is already used. " ,
2008-12-19 18:23:03 +03:00
lv - > major , lv - > minor ) ;
return 0 ;
}
}
2010-04-23 18:16:32 +04:00
if ( info . exists & & ! dm_tree_add_dev_with_udev_flags ( dtree , info . major , info . minor ,
2018-02-28 19:22:09 +03:00
_get_udev_flags ( dm , lv , layer ,
0 , 0 , 0 ) ) ) {
2016-12-01 16:53:35 +03:00
log_error ( " Failed to add device (% " PRIu32 " :% " PRIu32 " ) to dtree. " ,
2005-11-09 01:52:26 +03:00
info . major , info . minor ) ;
return 0 ;
2003-04-25 02:09:13 +04:00
}
2014-11-13 12:08:40 +03:00
if ( info . exists & & dm - > track_pending_delete ) {
2024-03-21 23:01:57 +03:00
log_debug_activation ( " Tracking pending delete for %s%s%s (%s). " ,
display_lvname ( lv ) , layer ? " - " : " " , layer ? : " " , dlid ) ;
2019-08-27 13:18:47 +03:00
if ( ! str_list_add ( dm - > cmd - > pending_delete_mem , & dm - > cmd - > pending_delete , dlid ) )
2014-11-13 12:08:40 +03:00
return_0 ;
}
2017-11-24 15:55:42 +03:00
/*
* Find holders of existing active LV where name starts with ' pvmove ' ,
* but it ' s not anymore PVMOVE LV and also it ' s not PVMOVE _mimage
*/
if ( info . exists & & ! lv_is_pvmove ( lv ) & &
! strchr ( lv - > name , ' _ ' ) & & ! strncmp ( lv - > name , " pvmove " , 6 ) )
2019-08-20 13:30:25 +03:00
if ( ! _add_holders_to_dtree ( dm , dtree , lv , NULL , & info ) )
2017-11-24 15:55:42 +03:00
return_0 ;
2005-11-09 01:52:26 +03:00
return 1 ;
}
2014-07-09 19:24:34 +04:00
struct pool_cb_data {
2012-03-03 01:49:43 +04:00
struct dev_manager * dm ;
2014-07-09 19:24:34 +04:00
const struct logical_volume * pool_lv ;
int skip_zero ; /* to skip zeroed device header (check first 64B) */
int exec ; /* which binary to call */
int opts ;
2018-03-17 23:50:03 +03:00
struct {
unsigned maj ;
unsigned min ;
unsigned patch ;
} version ;
2014-07-09 19:24:34 +04:00
const char * global ;
2012-03-03 01:49:43 +04:00
} ;
2018-03-17 23:50:03 +03:00
/*
* Simple version of check function calling ' tool - V '
*
* Returns 1 if the tool ' s version is equal or better to given .
* Otherwise it returns 0.
*/
static int _check_tool_version ( struct cmd_context * cmd , const char * tool ,
unsigned maj , unsigned min , unsigned patch )
{
const char * argv [ ] = { tool , " -V " , NULL } ;
struct pipe_data pdata ;
FILE * f ;
char buf [ 128 ] = { 0 } ;
char * nl ;
unsigned v_maj , v_min , v_patch ;
int ret = 0 ;
if ( ! ( f = pipe_open ( cmd , argv , 0 , & pdata ) ) ) {
log_warn ( " WARNING: Cannot read output from %s. " , argv [ 0 ] ) ;
} else {
if ( fgets ( buf , sizeof ( buf ) - 1 , f ) & &
( sscanf ( buf , " %u.%u.%u " , & v_maj , & v_min , & v_patch ) = = 3 ) ) {
if ( ( v_maj > maj ) | |
( ( v_maj = = maj ) & &
( ( v_min > min ) | |
( v_min = = min & & v_patch > = patch ) ) ) )
ret = 1 ;
if ( ( nl = strchr ( buf , ' \n ' ) ) )
nl [ 0 ] = 0 ; /* cut newline away */
log_verbose ( " Found version of %s %s is %s then requested %u.%u.%u. " ,
argv [ 0 ] , buf , ret ? " better " : " older " , maj , min , patch ) ;
} else
log_warn ( " WARNING: Cannot parse output '%s' from %s. " , buf , argv [ 0 ] ) ;
( void ) pipe_close ( & pdata ) ;
}
return ret ;
}
2014-07-09 19:24:34 +04:00
static int _pool_callback ( struct dm_tree_node * node ,
dm_node_callback_t type , void * cb_data )
2012-03-03 01:49:43 +04:00
{
2017-11-07 22:54:51 +03:00
int ret , status = 0 , fd ;
2014-07-09 19:24:34 +04:00
const struct pool_cb_data * data = cb_data ;
const struct logical_volume * pool_lv = data - > pool_lv ;
const struct logical_volume * mlv = first_seg ( pool_lv ) - > metadata_lv ;
2020-12-10 02:36:09 +03:00
struct cmd_context * cmd = pool_lv - > vg - > cmd ;
2014-07-09 19:24:34 +04:00
long buf [ 64 / sizeof ( long ) ] ; /* buffer for short disk header (64B) */
2012-03-14 21:12:05 +04:00
int args = 0 ;
2017-12-07 19:49:15 +03:00
char * mpath ;
2023-07-16 18:35:20 +03:00
const char * argv [ DEFAULT_MAX_EXEC_ARGS + 7 ] = { /* Max supported 15 args */
2020-12-10 02:36:09 +03:00
find_config_tree_str_allow_empty ( cmd , data - > exec , NULL )
2017-12-08 15:09:45 +03:00
} ;
2023-07-16 18:35:20 +03:00
if ( ! argv [ 0 ] | | ! * argv [ 0 ] ) /* *_check tool is unconfigured/disabled with "" setting */
2017-12-08 15:09:45 +03:00
return 1 ;
2012-03-03 01:49:43 +04:00
2020-12-10 02:36:09 +03:00
if ( lv_is_cache_vol ( pool_lv ) ) {
if ( ! ( mpath = lv_dmpath_suffix_dup ( data - > dm - > mem , pool_lv , " -cmeta " ) ) ) {
log_error ( " Failed to build device path for checking cachevol metadata %s. " ,
display_lvname ( pool_lv ) ) ;
return 0 ;
}
} else {
if ( ! ( mpath = lv_dmpath_dup ( data - > dm - > mem , mlv ) ) ) {
log_error ( " Failed to build device path for checking pool metadata %s. " ,
display_lvname ( mlv ) ) ;
return 0 ;
}
2017-12-07 19:49:15 +03:00
}
2021-12-15 13:45:22 +03:00
2024-07-08 23:32:41 +03:00
dm_devs_cache_destroy ( ) ;
2021-12-15 13:45:22 +03:00
2020-12-10 02:36:09 +03:00
log_debug ( " Running check command on %s " , mpath ) ;
2017-12-07 19:49:15 +03:00
if ( data - > skip_zero ) {
if ( ( fd = open ( mpath , O_RDONLY ) ) < 0 ) {
log_sys_error ( " open " , mpath ) ;
return 0 ;
}
/* let's assume there is no problem to read 64 bytes */
if ( read ( fd , buf , sizeof ( buf ) ) < ( int ) sizeof ( buf ) ) {
log_sys_error ( " read " , mpath ) ;
if ( close ( fd ) )
log_sys_error ( " close " , mpath ) ;
return 0 ;
}
for ( ret = 0 ; ret < ( int ) DM_ARRAY_SIZE ( buf ) ; + + ret )
if ( buf [ ret ] )
break ;
if ( close ( fd ) )
log_sys_error ( " close " , mpath ) ;
if ( ret = = ( int ) DM_ARRAY_SIZE ( buf ) ) {
log_debug_activation ( " Metadata checking skipped, detected empty disk header on %s. " ,
mpath ) ;
return 1 ;
}
}
2023-07-16 18:35:20 +03:00
if ( ! prepare_exec_args ( cmd , argv , & args , data - > opts ) )
return_0 ;
2012-03-14 21:12:05 +04:00
2017-12-07 19:49:15 +03:00
argv [ + + args ] = mpath ;
2014-07-09 19:24:34 +04:00
2020-12-10 02:36:09 +03:00
if ( ! ( ret = exec_cmd ( cmd , ( const char * const * ) argv ,
2012-03-03 01:49:43 +04:00
& status , 0 ) ) ) {
2018-03-17 23:50:03 +03:00
if ( status = = ENOENT ) {
2018-03-19 14:05:36 +03:00
log_warn ( " WARNING: Check is skipped, please install recommended missing binary %s! " ,
2018-03-17 23:50:03 +03:00
argv [ 0 ] ) ;
return 1 ;
}
if ( ( data - > version . maj | | data - > version . min | | data - > version . patch ) & &
2020-12-10 02:36:09 +03:00
! _check_tool_version ( cmd , argv [ 0 ] ,
2018-03-17 23:50:03 +03:00
data - > version . maj , data - > version . min , data - > version . patch ) ) {
log_warn ( " WARNING: Check is skipped, please upgrade installed version of %s! " ,
argv [ 0 ] ) ;
return 1 ;
}
2012-03-05 18:15:50 +04:00
switch ( type ) {
case DM_NODE_CALLBACK_PRELOADED :
2014-07-09 19:24:34 +04:00
log_err_once ( " Check of pool %s failed (status:%d). "
" Manual repair required! " ,
display_lvname ( pool_lv ) , status ) ;
2012-03-05 18:15:50 +04:00
break ;
default :
2014-07-09 19:24:34 +04:00
log_warn ( " WARNING: Integrity check of metadata for pool "
" %s failed. " , display_lvname ( pool_lv ) ) ;
2012-03-05 18:15:50 +04:00
}
2012-03-03 01:49:43 +04:00
/*
* FIXME : What should we do here ? ?
*
* Maybe mark the node , so it ' s not activating
2014-07-09 19:24:34 +04:00
* as pool but as error / linear and let the
2012-03-03 01:49:43 +04:00
* dm tree resolve the issue .
*/
}
return ret ;
}
2014-07-09 19:24:34 +04:00
static int _pool_register_callback ( struct dev_manager * dm ,
struct dm_tree_node * node ,
const struct logical_volume * lv )
2012-03-03 01:49:43 +04:00
{
2014-07-09 19:24:34 +04:00
struct pool_cb_data * data ;
2012-03-03 01:49:43 +04:00
2014-11-04 11:51:58 +03:00
/* Do not skip metadata of testing even for unused thin pools */
#if 0
2014-07-09 19:24:34 +04:00
/* Skip metadata testing for unused thin pool. */
if ( lv_is_thin_pool ( lv ) & &
( ! first_seg ( lv ) - > transaction_id | |
( ( first_seg ( lv ) - > transaction_id = = 1 ) & &
pool_has_message ( first_seg ( lv ) , NULL , 0 ) ) ) )
2012-03-03 01:49:43 +04:00
return 1 ;
2014-11-04 11:51:58 +03:00
# endif
2024-04-07 14:01:26 +03:00
/* Skip validation of metadata for lvremove and vgremove */
if ( ! dm - > activation & &
( ! strcmp ( dm - > cmd - > name , " lvremove " ) | |
! strcmp ( dm - > cmd - > name , " vgremove " ) ) ) {
log_debug ( " Skipping %s callback registration for command %s. " ,
display_lvname ( lv ) , dm - > cmd - > name ) ;
return 1 ;
}
2012-03-03 01:49:43 +04:00
2014-07-09 19:24:34 +04:00
if ( ! ( data = dm_pool_zalloc ( dm - > mem , sizeof ( * data ) ) ) ) {
2012-03-03 01:49:43 +04:00
log_error ( " Failed to allocated path for callback. " ) ;
return 0 ;
}
data - > dm = dm ;
2014-07-09 19:24:34 +04:00
if ( lv_is_thin_pool ( lv ) ) {
data - > pool_lv = lv ;
data - > skip_zero = 1 ;
data - > exec = global_thin_check_executable_CFG ;
data - > opts = global_thin_check_options_CFG ;
data - > global = " thin " ;
} else if ( lv_is_cache ( lv ) ) { /* cache pool */
data - > pool_lv = first_seg ( lv ) - > pool_lv ;
2016-02-23 14:15:42 +03:00
data - > skip_zero = 1 ; /* cheap read-error detection */
2014-07-09 19:24:34 +04:00
data - > exec = global_cache_check_executable_CFG ;
data - > opts = global_cache_check_options_CFG ;
data - > global = " cache " ;
2018-03-17 23:50:03 +03:00
if ( first_seg ( first_seg ( lv ) - > pool_lv ) - > cache_metadata_format > 1 ) {
data - > version . maj = 0 ;
data - > version . min = 7 ;
}
2014-07-09 19:24:34 +04:00
} else {
log_error ( INTERNAL_ERROR " Registering unsupported pool callback. " ) ;
return 0 ;
}
dm_tree_node_set_callback ( node , _pool_callback , data ) ;
2012-03-03 01:49:43 +04:00
return 1 ;
}
2024-04-04 02:29:30 +03:00
static struct id _get_id_for_meta_or_data ( const struct lv_segment * lvseg , int meta_or_data )
{
/* When ID is provided in form of metadata_id or data_id, otherwise use CVOL ID */
if ( meta_or_data & & lvseg - > metadata_id )
return * lvseg - > metadata_id ;
if ( ! meta_or_data & & lvseg - > data_id )
return * lvseg - > data_id ;
return lvseg - > pool_lv - > lvid . id [ 1 ] ;
}
2019-10-14 11:31:38 +03:00
/* Add special devices _cmeta & _cdata on top of CacheVol to dm tree */
static int _add_cvol_subdev_to_dtree ( struct dev_manager * dm , struct dm_tree * dtree ,
const struct logical_volume * lv , int meta_or_data )
{
const char * layer = meta_or_data ? " cmeta " : " cdata " ;
struct dm_pool * mem = dm - > track_pending_delete ? dm - > cmd - > pending_delete_mem : dm - > mem ;
struct lv_segment * lvseg = first_seg ( lv ) ;
2024-04-04 02:29:30 +03:00
const struct logical_volume * pool_lv = lvseg - > pool_lv ;
2019-10-14 11:31:38 +03:00
struct dm_info info ;
char * name , * dlid ;
2024-04-04 02:29:30 +03:00
union lvid lvid = { { lv - > vg - > id , _get_id_for_meta_or_data ( lvseg , meta_or_data ) } } ;
2019-10-14 11:31:38 +03:00
if ( ! ( dlid = dm_build_dm_uuid ( mem , UUID_PREFIX , ( const char * ) & lvid . s , layer ) ) )
return_0 ;
/* Name is actually not really needed here, but aids debugging... */
if ( ! ( name = dm_build_dm_name ( dm - > mem , lv - > vg - > name , pool_lv - > name , layer ) ) )
return_0 ;
2024-03-21 16:27:13 +03:00
if ( ! _info ( dm - > cmd , name , dlid , 0 , 0 , 0 , & info , NULL , NULL ) )
2019-10-14 11:31:38 +03:00
return_0 ;
if ( info . exists ) {
if ( ! dm_tree_add_dev_with_udev_flags ( dtree , info . major , info . minor ,
_get_udev_flags ( dm , lv , layer , 0 , 0 , 0 ) ) ) {
log_error ( " Failed to add device (% " PRIu32 " :% " PRIu32 " ) to dtree. " , info . major , info . minor ) ;
return 0 ;
}
if ( dm - > track_pending_delete ) {
2024-03-21 23:01:57 +03:00
log_debug_activation ( " Tracking pending delete for %s-%s (%s). " ,
display_lvname ( pool_lv ) , layer , dlid ) ;
2019-10-14 11:31:38 +03:00
if ( ! str_list_add ( mem , & dm - > cmd - > pending_delete , dlid ) )
return_0 ;
}
}
return 1 ;
}
thin: move pool messaging from resume to suspend
Existing messaging intarface for thin-pool has a few 'weak' points:
* Message were posted with each 'resume' operation, thus not allowing
activation of thin-pool with the existing state.
* Acceleration skipped suspend step has not worked in cluster,
since clvmd resumes only nodes which are suspended (have proper lock
state).
* Resume may fail and code is not really designed to 'fail' in this
phase (generic rule here is resume DOES NOT fail unless something serious
is wrong and lvm2 tool usually doesn't handle recovery path in this case.)
* Full thin-pool suspend happened, when taken a thin-volume snapshot.
With this patch the new method relocates message passing into suspend
state.
This has a few drawbacks with current API, but overal it performs
better and gives are more posibilities to deal with errors.
Patch introduces a new logic for 'origin-only' suspend of thin-pool and
this also relates to thin-volume when taking snapshot.
When suspend_origin_only operation is invoked on a pool with
queued messages then only those messages are posted to thin-pool and
actual suspend of thin pool and data and metadata volume is skipped.
This makes taking a snapshot of thin-volume lighter operation and
avoids blocking of other unrelated active thin volumes.
Also fail now happens in 'suspend' state where the 'Fail' is more expected
and it is better handled through error paths.
Activation of thin-pool is now not sending any message and leaves upto a tool
to decided later how to finish unfinished double-commit transaction.
Problem which needs some API improvements relates to the lvm2 tree
construction. For the suspend tree we do not add target table line
into the tree, but only a device is inserted into a tree.
Current mechanism to attach messages for thin-pool requires the libdm
to know about thin-pool target, so lvm2 currently takes assumption, node
is really a thin-pool and fills in the table line for this node (which
should be ensured by the PRELOAD phase, but it's a misuse of internal API)
we would possibly need to be able to attach message to 'any' node.
Other thing to notice - current messaging interface in thin-pool
target requires to suspend thin volume origin first and then send
a create message, but this could not have any 'nice' solution on lvm2
side and IMHO we should introduce something like 'create_after_resume'
message.
Patch also changes the moment, where lvm2 transaction id is increased.
Now it happens only after successful finish of kernel transaction id
change. This change was needed to handle properly activation of pool,
which is in the middle of unfinished transaction, and also this corrects
usage of thin-pool by external apps like Docker.
2015-07-01 14:31:37 +03:00
/* Declaration to resolve suspend tree and message passing for thin-pool */
static int _add_target_to_dtree ( struct dev_manager * dm ,
struct dm_tree_node * dnode ,
struct lv_segment * seg ,
struct lv_activate_opts * laopts ) ;
2005-11-09 01:52:26 +03:00
/*
* Add LV and any known dependencies
*/
2011-08-11 08:18:17 +04:00
static int _add_lv_to_dtree ( struct dev_manager * dm , struct dm_tree * dtree ,
2014-09-22 17:50:07 +04:00
const struct logical_volume * lv , int origin_only )
2005-11-09 01:52:26 +03:00
{
2011-08-11 08:18:17 +04:00
uint32_t s ;
2011-06-11 04:03:06 +04:00
struct seg_list * sl ;
2013-02-21 13:39:47 +04:00
struct dm_list * snh ;
2013-04-19 23:09:28 +04:00
struct lv_segment * seg ;
2014-07-09 19:24:34 +04:00
struct dm_tree_node * node ;
2017-11-10 23:18:55 +03:00
const struct logical_volume * plv ;
2011-06-11 04:03:06 +04:00
2017-11-30 15:26:44 +03:00
if ( lv_is_pvmove ( lv ) & & ( dm - > track_pvmove_deps = = 2 ) )
return 1 ; /* Avoid rechecking of already seen pvmove LV */
2014-04-01 19:53:18 +04:00
if ( lv_is_cache_pool ( lv ) ) {
2014-11-02 21:34:50 +03:00
if ( ! dm_list_empty ( & lv - > segs_using_this_lv ) ) {
if ( ! _add_lv_to_dtree ( dm , dtree , seg_lv ( first_seg ( lv ) , 0 ) , 0 ) )
return_0 ;
if ( ! _add_lv_to_dtree ( dm , dtree , first_seg ( lv ) - > metadata_lv , 0 ) )
return_0 ;
2014-11-11 15:31:25 +03:00
/* Cache pool does not have a real device node */
return 1 ;
}
/* Unused cache pool is activated as metadata */
2014-04-01 19:53:18 +04:00
}
2013-02-02 03:44:07 +04:00
if ( ! origin_only & & ! _add_dev_to_dtree ( dm , dtree , lv , NULL ) )
2005-11-09 01:52:26 +03:00
return_0 ;
/* FIXME Can we avoid doing this every time? */
2013-02-21 13:25:44 +04:00
/* Reused also for lv_is_external_origin(lv) */
2005-11-09 16:08:41 +03:00
if ( ! _add_dev_to_dtree ( dm , dtree , lv , " real " ) )
2005-11-09 01:52:26 +03:00
return_0 ;
2010-08-17 20:25:32 +04:00
if ( ! origin_only & & ! _add_dev_to_dtree ( dm , dtree , lv , " cow " ) )
2005-11-09 01:52:26 +03:00
return_0 ;
2003-04-25 02:09:13 +04:00
2013-02-21 13:39:47 +04:00
if ( origin_only & & lv_is_thin_volume ( lv ) ) {
2013-07-15 13:47:10 +04:00
if ( ! _add_dev_to_dtree ( dm , dtree , lv , lv_layer ( lv ) ) )
2013-02-21 13:39:47 +04:00
return_0 ;
#if 0
/* ? Use origin_only to avoid 'deep' thin pool suspend ? */
2013-07-15 15:05:15 +04:00
/* FIXME Implement dm_tree_node_skip_childrens optimisation */
2014-03-11 20:13:47 +04:00
if ( ! ( uuid = build_dm_uuid ( dm - > mem , lv , lv_layer ( lv ) ) ) )
2013-02-21 13:39:47 +04:00
return_0 ;
2014-07-09 19:24:34 +04:00
if ( ( node = dm_tree_find_node_by_uuid ( dtree , uuid ) ) )
dm_tree_node_skip_childrens ( node , 1 ) ;
2013-02-21 13:39:47 +04:00
# endif
}
2024-02-27 19:23:01 +03:00
if ( dm - > activation & & lv_is_external_origin ( lv ) ) {
2013-07-09 14:34:49 +04:00
/* Find possible users of external origin lv */
dm_list_iterate_items ( sl , & lv - > segs_using_this_lv )
2024-02-27 19:23:01 +03:00
if ( ! _add_dev_to_dtree ( dm , dtree , sl - > seg - > lv , lv_layer ( sl - > seg - > lv ) ) )
2013-07-09 14:34:49 +04:00
return_0 ;
}
2013-02-21 13:39:47 +04:00
if ( lv_is_thin_pool ( lv ) ) {
2013-07-09 14:34:49 +04:00
/*
* For both origin_only and ! origin_only
* skips test for - tpool - real and tpool - cow
*/
2013-02-21 13:39:47 +04:00
if ( ! _add_dev_to_dtree ( dm , dtree , lv , lv_layer ( lv ) ) )
return_0 ;
thin: move pool messaging from resume to suspend
Existing messaging intarface for thin-pool has a few 'weak' points:
* Message were posted with each 'resume' operation, thus not allowing
activation of thin-pool with the existing state.
* Acceleration skipped suspend step has not worked in cluster,
since clvmd resumes only nodes which are suspended (have proper lock
state).
* Resume may fail and code is not really designed to 'fail' in this
phase (generic rule here is resume DOES NOT fail unless something serious
is wrong and lvm2 tool usually doesn't handle recovery path in this case.)
* Full thin-pool suspend happened, when taken a thin-volume snapshot.
With this patch the new method relocates message passing into suspend
state.
This has a few drawbacks with current API, but overal it performs
better and gives are more posibilities to deal with errors.
Patch introduces a new logic for 'origin-only' suspend of thin-pool and
this also relates to thin-volume when taking snapshot.
When suspend_origin_only operation is invoked on a pool with
queued messages then only those messages are posted to thin-pool and
actual suspend of thin pool and data and metadata volume is skipped.
This makes taking a snapshot of thin-volume lighter operation and
avoids blocking of other unrelated active thin volumes.
Also fail now happens in 'suspend' state where the 'Fail' is more expected
and it is better handled through error paths.
Activation of thin-pool is now not sending any message and leaves upto a tool
to decided later how to finish unfinished double-commit transaction.
Problem which needs some API improvements relates to the lvm2 tree
construction. For the suspend tree we do not add target table line
into the tree, but only a device is inserted into a tree.
Current mechanism to attach messages for thin-pool requires the libdm
to know about thin-pool target, so lvm2 currently takes assumption, node
is really a thin-pool and fills in the table line for this node (which
should be ensured by the PRELOAD phase, but it's a misuse of internal API)
we would possibly need to be able to attach message to 'any' node.
Other thing to notice - current messaging interface in thin-pool
target requires to suspend thin volume origin first and then send
a create message, but this could not have any 'nice' solution on lvm2
side and IMHO we should introduce something like 'create_after_resume'
message.
Patch also changes the moment, where lvm2 transaction id is increased.
Now it happens only after successful finish of kernel transaction id
change. This change was needed to handle properly activation of pool,
which is in the middle of unfinished transaction, and also this corrects
usage of thin-pool by external apps like Docker.
2015-07-01 14:31:37 +03:00
/*
* TODO : change API and move this code
* Could be easier to handle this in _add_dev_to_dtree ( )
* and base this according to info . exists ?
*/
2013-07-09 14:34:49 +04:00
if ( ! dm - > activation ) {
2024-03-21 19:23:33 +03:00
if ( ( node = _cached_dm_tree_node ( dm - > mem , dtree , lv , lv_layer ( lv ) ) ) ) {
thin: move pool messaging from resume to suspend
Existing messaging intarface for thin-pool has a few 'weak' points:
* Message were posted with each 'resume' operation, thus not allowing
activation of thin-pool with the existing state.
* Acceleration skipped suspend step has not worked in cluster,
since clvmd resumes only nodes which are suspended (have proper lock
state).
* Resume may fail and code is not really designed to 'fail' in this
phase (generic rule here is resume DOES NOT fail unless something serious
is wrong and lvm2 tool usually doesn't handle recovery path in this case.)
* Full thin-pool suspend happened, when taken a thin-volume snapshot.
With this patch the new method relocates message passing into suspend
state.
This has a few drawbacks with current API, but overal it performs
better and gives are more posibilities to deal with errors.
Patch introduces a new logic for 'origin-only' suspend of thin-pool and
this also relates to thin-volume when taking snapshot.
When suspend_origin_only operation is invoked on a pool with
queued messages then only those messages are posted to thin-pool and
actual suspend of thin pool and data and metadata volume is skipped.
This makes taking a snapshot of thin-volume lighter operation and
avoids blocking of other unrelated active thin volumes.
Also fail now happens in 'suspend' state where the 'Fail' is more expected
and it is better handled through error paths.
Activation of thin-pool is now not sending any message and leaves upto a tool
to decided later how to finish unfinished double-commit transaction.
Problem which needs some API improvements relates to the lvm2 tree
construction. For the suspend tree we do not add target table line
into the tree, but only a device is inserted into a tree.
Current mechanism to attach messages for thin-pool requires the libdm
to know about thin-pool target, so lvm2 currently takes assumption, node
is really a thin-pool and fills in the table line for this node (which
should be ensured by the PRELOAD phase, but it's a misuse of internal API)
we would possibly need to be able to attach message to 'any' node.
Other thing to notice - current messaging interface in thin-pool
target requires to suspend thin volume origin first and then send
a create message, but this could not have any 'nice' solution on lvm2
side and IMHO we should introduce something like 'create_after_resume'
message.
Patch also changes the moment, where lvm2 transaction id is increased.
Now it happens only after successful finish of kernel transaction id
change. This change was needed to handle properly activation of pool,
which is in the middle of unfinished transaction, and also this corrects
usage of thin-pool by external apps like Docker.
2015-07-01 14:31:37 +03:00
if ( origin_only ) {
struct lv_activate_opts laopts = {
. origin_only = 1 ,
. send_messages = 1 /* Node with messages */
} ;
/*
2024-08-30 00:05:41 +03:00
* Add some messages if right node exist in the table only
thin: move pool messaging from resume to suspend
Existing messaging intarface for thin-pool has a few 'weak' points:
* Message were posted with each 'resume' operation, thus not allowing
activation of thin-pool with the existing state.
* Acceleration skipped suspend step has not worked in cluster,
since clvmd resumes only nodes which are suspended (have proper lock
state).
* Resume may fail and code is not really designed to 'fail' in this
phase (generic rule here is resume DOES NOT fail unless something serious
is wrong and lvm2 tool usually doesn't handle recovery path in this case.)
* Full thin-pool suspend happened, when taken a thin-volume snapshot.
With this patch the new method relocates message passing into suspend
state.
This has a few drawbacks with current API, but overal it performs
better and gives are more posibilities to deal with errors.
Patch introduces a new logic for 'origin-only' suspend of thin-pool and
this also relates to thin-volume when taking snapshot.
When suspend_origin_only operation is invoked on a pool with
queued messages then only those messages are posted to thin-pool and
actual suspend of thin pool and data and metadata volume is skipped.
This makes taking a snapshot of thin-volume lighter operation and
avoids blocking of other unrelated active thin volumes.
Also fail now happens in 'suspend' state where the 'Fail' is more expected
and it is better handled through error paths.
Activation of thin-pool is now not sending any message and leaves upto a tool
to decided later how to finish unfinished double-commit transaction.
Problem which needs some API improvements relates to the lvm2 tree
construction. For the suspend tree we do not add target table line
into the tree, but only a device is inserted into a tree.
Current mechanism to attach messages for thin-pool requires the libdm
to know about thin-pool target, so lvm2 currently takes assumption, node
is really a thin-pool and fills in the table line for this node (which
should be ensured by the PRELOAD phase, but it's a misuse of internal API)
we would possibly need to be able to attach message to 'any' node.
Other thing to notice - current messaging interface in thin-pool
target requires to suspend thin volume origin first and then send
a create message, but this could not have any 'nice' solution on lvm2
side and IMHO we should introduce something like 'create_after_resume'
message.
Patch also changes the moment, where lvm2 transaction id is increased.
Now it happens only after successful finish of kernel transaction id
change. This change was needed to handle properly activation of pool,
which is in the middle of unfinished transaction, and also this corrects
usage of thin-pool by external apps like Docker.
2015-07-01 14:31:37 +03:00
* when building SUSPEND tree for origin - only thin - pool .
*
* TODO : Fix call of ' _add_target_to_dtree ( ) ' to add message
* to thin - pool node as we already know the pool node exists
* in the table . Any better / cleaner API way ?
*
* Probably some ' new ' target method to add messages for any node ?
*/
if ( dm - > suspend & &
! dm_list_empty ( & ( first_seg ( lv ) - > thin_messages ) ) & &
! _add_target_to_dtree ( dm , node , first_seg ( lv ) , & laopts ) )
return_0 ;
} else {
/* Setup callback for non-activation partial tree */
/* Activation gets own callback when needed */
/* TODO: extend _cached_dm_info() to return dnode */
if ( ! _pool_register_callback ( dm , node , lv ) )
return_0 ;
}
}
2014-07-09 19:24:34 +04:00
}
}
2019-09-14 02:13:33 +03:00
if ( lv_is_vdo_pool ( lv ) ) {
/*
* For both origin_only and ! origin_only
* skips test for - vpool - real and vpool - cow
*/
if ( ! _add_dev_to_dtree ( dm , dtree , lv , lv_layer ( lv ) ) )
return_0 ;
}
2014-11-10 12:56:43 +03:00
if ( lv_is_cache ( lv ) ) {
2014-11-13 12:08:40 +03:00
if ( ! origin_only & & ! dm - > activation & & ! dm - > track_pending_delete ) {
2014-07-09 19:24:34 +04:00
/* Setup callback for non-activation partial tree */
/* Activation gets own callback when needed */
2024-03-21 19:23:33 +03:00
if ( ( node = _cached_dm_tree_node ( dm - > mem , dtree , lv , lv_layer ( lv ) ) ) & &
2014-07-09 19:24:34 +04:00
! _pool_register_callback ( dm , node , lv ) )
2013-07-09 14:34:49 +04:00
return_0 ;
}
2013-02-21 13:39:47 +04:00
}
2006-04-28 21:01:07 +04:00
2024-03-21 18:12:06 +03:00
if ( lv_is_cache_vol ( lv ) )
/* Cachevol with cache LV spans some extra layers -cdata, -cmeta */
dm_list_iterate_items ( sl , & lv - > segs_using_this_lv )
if ( lv_is_cache ( sl - > seg - > lv ) & &
( ! _add_cvol_subdev_to_dtree ( dm , dtree , sl - > seg - > lv , 0 ) | |
! _add_cvol_subdev_to_dtree ( dm , dtree , sl - > seg - > lv , 1 ) | |
! _add_dev_to_dtree ( dm , dtree , lv , lv_layer ( lv ) ) ) )
return_0 ;
2013-02-21 13:39:47 +04:00
/* Add any snapshots of this LV */
2013-07-09 14:34:49 +04:00
if ( ! origin_only & & lv_is_origin ( lv ) )
2013-02-21 13:39:47 +04:00
dm_list_iterate ( snh , & lv - > snapshot_segs )
if ( ! _add_lv_to_dtree ( dm , dtree , dm_list_struct_base ( snh , struct lv_segment , origin_list ) - > cow , 0 ) )
2011-08-11 08:18:17 +04:00
return_0 ;
2016-05-18 16:22:26 +03:00
2013-11-30 00:18:34 +04:00
if ( dm - > activation & & ! origin_only & & lv_is_merging_origin ( lv ) & &
! _add_lv_to_dtree ( dm , dtree , find_snapshot ( lv ) - > lv , 1 ) )
return_0 ;
2011-08-11 08:18:17 +04:00
2011-06-11 04:03:06 +04:00
/* Add any LVs referencing a PVMOVE LV unless told not to. */
2017-11-30 15:26:44 +03:00
if ( ( dm - > track_pvmove_deps = = 1 ) & & lv_is_pvmove ( lv ) ) {
dm - > track_pvmove_deps = 2 ; /* Mark as already seen */
2017-11-10 23:18:55 +03:00
dm_list_iterate_items ( sl , & lv - > segs_using_this_lv ) {
/* If LV is snapshot COW - whole snapshot needs reload */
plv = lv_is_cow ( sl - > seg - > lv ) ? origin_from_cow ( sl - > seg - > lv ) : sl - > seg - > lv ;
if ( ! _add_lv_to_dtree ( dm , dtree , plv , 0 ) )
2011-06-11 04:03:06 +04:00
return_0 ;
2017-11-10 23:18:55 +03:00
}
2013-08-29 01:56:23 +04:00
dm - > track_pvmove_deps = 1 ;
}
2011-06-11 04:03:06 +04:00
2014-11-13 12:08:40 +03:00
if ( ! dm - > track_pending_delete )
dm_list_iterate_items ( sl , & lv - > segs_using_this_lv ) {
if ( lv_is_pending_delete ( sl - > seg - > lv ) ) {
/* LV is referenced by 'cache pending delete LV */
dm - > track_pending_delete = 1 ;
2024-03-21 23:02:24 +03:00
if ( ! _cached_dm_tree_node ( dm - > mem , dtree , sl - > seg - > lv , lv_layer ( sl - > seg - > lv ) ) & &
! _add_lv_to_dtree ( dm , dtree , sl - > seg - > lv , 0 ) )
2014-11-13 12:08:40 +03:00
return_0 ;
dm - > track_pending_delete = 0 ;
}
2014-11-10 12:56:43 +03:00
}
2013-02-21 13:39:47 +04:00
/* Add any LVs used by segments in this LV */
dm_list_iterate_items ( seg , & lv - > segments ) {
2024-02-27 19:23:01 +03:00
if ( seg - > external_lv & &
! _add_lv_to_dtree ( dm , dtree , seg - > external_lv ,
/* For origin LV check for complete device tree */
lv_is_origin ( seg - > external_lv ) ? 0 : 1 ) ) /* stack */
2013-02-21 13:25:44 +04:00
return_0 ;
2013-02-21 13:39:47 +04:00
if ( seg - > log_lv & &
2013-07-15 15:39:53 +04:00
! _add_lv_to_dtree ( dm , dtree , seg - > log_lv , 0 ) )
2011-11-03 18:52:09 +04:00
return_0 ;
2013-02-21 13:39:47 +04:00
if ( seg - > metadata_lv & &
2013-07-15 15:39:53 +04:00
! _add_lv_to_dtree ( dm , dtree , seg - > metadata_lv , 0 ) )
2012-03-03 01:49:43 +04:00
return_0 ;
2018-08-27 22:53:09 +03:00
if ( seg - > writecache & & seg_is_writecache ( seg ) ) {
if ( ! _add_lv_to_dtree ( dm , dtree , seg - > writecache , dm - > activation ? origin_only : 1 ) )
return_0 ;
}
2019-11-21 01:07:27 +03:00
if ( seg - > integrity_meta_dev & & seg_is_integrity ( seg ) ) {
if ( ! _add_lv_to_dtree ( dm , dtree , seg - > integrity_meta_dev , dm - > activation ? origin_only : 1 ) )
return_0 ;
}
2014-04-01 19:53:18 +04:00
if ( seg - > pool_lv & &
2015-11-25 01:24:11 +03:00
/* When activating and not origin_only detect linear 'overlay' over pool */
! _add_lv_to_dtree ( dm , dtree , seg - > pool_lv , dm - > activation ? origin_only : 1 ) )
2012-03-03 01:49:43 +04:00
return_0 ;
2013-02-21 13:39:47 +04:00
for ( s = 0 ; s < seg - > area_count ; s + + ) {
if ( seg_type ( seg , s ) = = AREA_LV & & seg_lv ( seg , s ) & &
2014-11-13 12:08:40 +03:00
/* origin only for cache without pending delete */
( ! dm - > track_pending_delete | | ! lv_is_cache ( lv ) ) & &
2019-09-14 02:13:33 +03:00
! _add_lv_to_dtree ( dm , dtree , seg_lv ( seg , s ) ,
2020-09-09 01:58:52 +03:00
lv_is_vdo_pool ( seg_lv ( seg , s ) ) ? 1 : 0 ) )
2013-02-21 13:39:47 +04:00
return_0 ;
2016-05-24 00:55:13 +03:00
if ( seg_is_raid_with_meta ( seg ) & & seg - > meta_areas & & seg_metalv ( seg , s ) & &
2020-09-09 01:58:52 +03:00
! _add_lv_to_dtree ( dm , dtree , seg_metalv ( seg , s ) , 0 ) )
2013-02-21 13:39:47 +04:00
return_0 ;
}
2013-11-30 00:18:34 +04:00
/* When activating, detect merging LV presence */
if ( dm - > activation & & seg - > merge_lv & &
! _add_lv_to_dtree ( dm , dtree , seg - > merge_lv , 1 ) )
return_0 ;
2012-01-25 13:10:13 +04:00
}
2011-10-03 22:24:47 +04:00
2003-04-25 02:09:13 +04:00
return 1 ;
}
2014-09-22 17:50:07 +04:00
static struct dm_tree * _create_partial_dtree ( struct dev_manager * dm , const struct logical_volume * lv , int origin_only )
2004-05-05 22:11:43 +04:00
{
2005-11-09 16:05:17 +03:00
struct dm_tree * dtree ;
2004-05-05 22:11:43 +04:00
2005-11-09 16:05:17 +03:00
if ( ! ( dtree = dm_tree_create ( ) ) ) {
2015-11-25 18:06:31 +03:00
log_debug_activation ( " Partial dtree creation failed for %s. " ,
display_lvname ( lv ) ) ;
2005-11-09 01:52:26 +03:00
return NULL ;
}
2004-05-05 22:11:43 +04:00
2024-05-03 13:43:14 +03:00
dm_tree_set_optional_uuid_suffixes ( dtree , ( const char * * ) _uuid_suffix_list ) ;
2014-07-31 00:55:11 +04:00
thin: move pool messaging from resume to suspend
Existing messaging intarface for thin-pool has a few 'weak' points:
* Message were posted with each 'resume' operation, thus not allowing
activation of thin-pool with the existing state.
* Acceleration skipped suspend step has not worked in cluster,
since clvmd resumes only nodes which are suspended (have proper lock
state).
* Resume may fail and code is not really designed to 'fail' in this
phase (generic rule here is resume DOES NOT fail unless something serious
is wrong and lvm2 tool usually doesn't handle recovery path in this case.)
* Full thin-pool suspend happened, when taken a thin-volume snapshot.
With this patch the new method relocates message passing into suspend
state.
This has a few drawbacks with current API, but overal it performs
better and gives are more posibilities to deal with errors.
Patch introduces a new logic for 'origin-only' suspend of thin-pool and
this also relates to thin-volume when taking snapshot.
When suspend_origin_only operation is invoked on a pool with
queued messages then only those messages are posted to thin-pool and
actual suspend of thin pool and data and metadata volume is skipped.
This makes taking a snapshot of thin-volume lighter operation and
avoids blocking of other unrelated active thin volumes.
Also fail now happens in 'suspend' state where the 'Fail' is more expected
and it is better handled through error paths.
Activation of thin-pool is now not sending any message and leaves upto a tool
to decided later how to finish unfinished double-commit transaction.
Problem which needs some API improvements relates to the lvm2 tree
construction. For the suspend tree we do not add target table line
into the tree, but only a device is inserted into a tree.
Current mechanism to attach messages for thin-pool requires the libdm
to know about thin-pool target, so lvm2 currently takes assumption, node
is really a thin-pool and fills in the table line for this node (which
should be ensured by the PRELOAD phase, but it's a misuse of internal API)
we would possibly need to be able to attach message to 'any' node.
Other thing to notice - current messaging interface in thin-pool
target requires to suspend thin volume origin first and then send
a create message, but this could not have any 'nice' solution on lvm2
side and IMHO we should introduce something like 'create_after_resume'
message.
Patch also changes the moment, where lvm2 transaction id is increased.
Now it happens only after successful finish of kernel transaction id
change. This change was needed to handle properly activation of pool,
which is in the middle of unfinished transaction, and also this corrects
usage of thin-pool by external apps like Docker.
2015-07-01 14:31:37 +03:00
if ( ! _add_lv_to_dtree ( dm , dtree , lv , ( lv_is_origin ( lv ) | | lv_is_thin_volume ( lv ) | | lv_is_thin_pool ( lv ) ) ? origin_only : 0 ) )
2008-01-30 16:19:47 +03:00
goto_bad ;
2004-05-05 22:11:43 +04:00
2005-11-09 01:52:26 +03:00
return dtree ;
2004-05-05 22:11:43 +04:00
2008-01-30 16:19:47 +03:00
bad :
2005-11-09 16:05:17 +03:00
dm_tree_free ( dtree ) ;
2005-11-09 01:52:26 +03:00
return NULL ;
2004-05-05 22:11:43 +04:00
}
2016-10-18 13:58:22 +03:00
static char * _add_error_or_zero_device ( struct dev_manager * dm , struct dm_tree * dtree ,
struct lv_segment * seg , int s , int use_zero )
2008-09-19 10:42:00 +04:00
{
2012-04-24 04:51:26 +04:00
char * dlid , * name ;
2008-09-19 10:42:00 +04:00
char errid [ 32 ] ;
struct dm_tree_node * node ;
struct lv_segment * seg_i ;
2012-04-24 04:51:26 +04:00
struct dm_info info ;
2011-03-30 00:30:05 +04:00
int segno = - 1 , i = 0 ;
2017-02-24 02:50:00 +03:00
uint64_t size = ( uint64_t ) _seg_len ( seg ) * seg - > lv - > vg - > extent_size ;
2008-09-19 10:42:00 +04:00
2008-11-04 01:14:30 +03:00
dm_list_iterate_items ( seg_i , & seg - > lv - > segments ) {
2016-10-18 13:58:22 +03:00
if ( seg = = seg_i ) {
2008-09-19 10:42:00 +04:00
segno = i ;
2016-10-18 13:58:22 +03:00
break ;
}
2008-09-19 10:42:00 +04:00
+ + i ;
}
if ( segno < 0 ) {
2016-10-18 13:58:22 +03:00
log_error ( INTERNAL_ERROR " _add_error_or_zero_device called with bad segment. " ) ;
2011-10-11 13:02:20 +04:00
return NULL ;
2008-09-19 10:42:00 +04:00
}
sprintf ( errid , " missing_%d_%d " , segno , s ) ;
2014-03-11 20:13:47 +04:00
if ( ! ( dlid = build_dm_uuid ( dm - > mem , seg - > lv , errid ) ) )
2008-09-19 10:42:00 +04:00
return_NULL ;
2011-08-30 18:55:15 +04:00
if ( ! ( name = dm_build_dm_name ( dm - > mem , seg - > lv - > vg - > name ,
2015-11-25 18:06:31 +03:00
seg - > lv - > name , errid ) ) )
2008-09-19 10:42:00 +04:00
return_NULL ;
2024-03-21 16:27:13 +03:00
if ( ! _info ( dm - > cmd , name , dlid , 0 , 0 , 0 , & info , NULL , NULL ) )
2016-12-18 16:57:16 +03:00
return_NULL ;
2012-04-24 04:51:26 +04:00
if ( ! info . exists ) {
/* Create new node */
if ( ! ( node = dm_tree_add_new_dev ( dtree , name , dlid , 0 , 0 , 0 , 0 , 0 ) ) )
return_NULL ;
2016-10-18 13:58:22 +03:00
if ( use_zero ) {
if ( ! dm_tree_node_add_zero_target ( node , size ) )
return_NULL ;
} else
if ( ! dm_tree_node_add_error_target ( node , size ) )
return_NULL ;
2012-04-24 04:51:26 +04:00
} else {
/* Already exists */
if ( ! dm_tree_add_dev ( dtree , info . major , info . minor ) ) {
2016-11-11 18:38:04 +03:00
log_error ( " Failed to add device (% " PRIu32 " :% " PRIu32 " ) to dtree. " ,
2012-04-24 04:51:26 +04:00
info . major , info . minor ) ;
2016-11-08 18:20:51 +03:00
return NULL ;
2012-04-24 04:51:26 +04:00
}
}
return dlid ;
2008-09-19 10:42:00 +04:00
}
static int _add_error_area ( struct dev_manager * dm , struct dm_tree_node * node ,
struct lv_segment * seg , int s )
{
char * dlid ;
uint64_t extent_size = seg - > lv - > vg - > extent_size ;
2016-10-18 13:58:22 +03:00
int use_zero = ! strcmp ( dm - > cmd - > stripe_filler , TARGET_NAME_ZERO ) ? 1 : 0 ;
2008-09-19 10:42:00 +04:00
2016-10-18 13:58:22 +03:00
if ( ! strcmp ( dm - > cmd - > stripe_filler , TARGET_NAME_ERROR ) | | use_zero ) {
2008-09-19 10:42:00 +04:00
/*
* FIXME , the tree pointer is first field of dm_tree_node , but
* we don ' t have the struct definition available .
*/
struct dm_tree * * tree = ( struct dm_tree * * ) node ;
2016-10-18 13:58:22 +03:00
if ( ! ( dlid = _add_error_or_zero_device ( dm , * tree , seg , s , use_zero ) ) )
2008-09-19 10:42:00 +04:00
return_0 ;
2011-07-06 03:10:14 +04:00
if ( ! dm_tree_node_add_target_area ( node , NULL , dlid , extent_size * seg_le ( seg , s ) ) )
return_0 ;
2008-09-19 11:18:03 +04:00
} else
2011-07-06 03:10:14 +04:00
if ( ! dm_tree_node_add_target_area ( node , dm - > cmd - > stripe_filler , NULL , UINT64_C ( 0 ) ) )
return_0 ;
2008-09-19 11:18:03 +04:00
2008-09-19 10:42:00 +04:00
return 1 ;
}
2023-04-24 23:47:45 +03:00
static int _bad_pv_area ( struct lv_segment * seg , uint32_t s )
{
struct stat info ;
const char * name ;
struct device * dev ;
if ( ! seg_pvseg ( seg , s ) )
return 1 ;
if ( ! seg_pv ( seg , s ) )
return 1 ;
if ( ! ( dev = seg_dev ( seg , s ) ) )
return 1 ;
if ( dm_list_empty ( & dev - > aliases ) )
return 1 ;
/* FIXME Avoid repeating identical stat in dm_tree_node_add_target_area */
name = dev_name ( dev ) ;
if ( stat ( name , & info ) < 0 )
return 1 ;
if ( ! S_ISBLK ( info . st_mode ) )
return 1 ;
return 0 ;
}
2005-11-09 01:52:26 +03:00
int add_areas_line ( struct dev_manager * dm , struct lv_segment * seg ,
2006-05-10 01:23:51 +04:00
struct dm_tree_node * node , uint32_t start_area ,
uint32_t areas )
2002-03-11 14:27:48 +03:00
{
2023-04-24 23:47:45 +03:00
struct cmd_context * cmd = seg - > lv - > vg - > cmd ;
2005-11-09 01:52:26 +03:00
uint64_t extent_size = seg - > lv - > vg - > extent_size ;
uint32_t s ;
2002-03-19 02:25:50 +03:00
char * dlid ;
2011-07-06 04:29:44 +04:00
const char * name ;
2012-02-01 17:47:27 +04:00
unsigned num_error_areas = 0 ;
unsigned num_existing_areas = 0 ;
2002-03-11 14:27:48 +03:00
2005-11-09 01:52:26 +03:00
for ( s = start_area ; s < areas ; s + + ) {
2023-04-24 23:47:45 +03:00
if ( ( ( seg_type ( seg , s ) = = AREA_PV ) & & _bad_pv_area ( seg , s ) ) | |
( ( seg_type ( seg , s ) = = AREA_LV ) & & ! seg_lv ( seg , s ) ) ) {
if ( ! cmd - > partial_activation ) {
2023-03-22 21:05:43 +03:00
if ( ! cmd - > degraded_activation | |
( ! lv_is_raid_type ( seg - > lv ) & &
! lv_is_integrity ( seg - > lv ) & &
! lv_is_integrity_metadata ( seg - > lv ) & &
! lv_is_integrity_origin ( seg - > lv ) ) ) {
2023-04-24 23:47:45 +03:00
log_error ( " Aborting. LV %s is incomplete and --activationmode partial was not specified. " ,
2015-11-25 18:06:31 +03:00
display_lvname ( seg - > lv ) ) ;
activation: Add "degraded" activation mode
Currently, we have two modes of activation, an unnamed nominal mode
(which I will refer to as "complete") and "partial" mode. The
"complete" mode requires that a volume group be 'complete' - that
is, no missing PVs. If there are any missing PVs, no affected LVs
are allowed to activate - even RAID LVs which might be able to
tolerate a failure. The "partial" mode allows anything to be
activated (or at least attempted). If a non-redundant LV is
missing a portion of its addressable space due to a device failure,
it will be replaced with an error target. RAID LVs will either
activate or fail to activate depending on how badly their
redundancy is compromised.
This patch adds a third option, "degraded" mode. This mode can
be selected via the '--activationmode {complete|degraded|partial}'
option to lvchange/vgchange. It can also be set in lvm.conf.
The "degraded" activation mode allows RAID LVs with a sufficient
level of redundancy to activate (e.g. a RAID5 LV with one device
failure, a RAID6 with two device failures, or RAID1 with n-1
failures). RAID LVs with too many device failures are not allowed
to activate - nor are any non-redundant LVs that may have been
affected. This patch also makes the "degraded" mode the default
activation mode.
The degraded activation mode does not yet work in a cluster. A
new cluster lock flag (LCK_DEGRADED_MODE) will need to be created
to make that work. Currently, there is limited space for this
extra flag and I am looking for possible solutions. One possible
solution is to usurp LCK_CONVERT, as it is not used. When the
locking_type is 3, the degraded mode flag simply gets dropped and
the old ("complete") behavior is exhibited.
2014-07-10 07:56:11 +04:00
return 0 ;
}
2011-07-06 04:29:44 +04:00
}
2008-09-19 10:42:00 +04:00
if ( ! _add_error_area ( dm , node , seg , s ) )
return_0 ;
2012-02-01 17:47:27 +04:00
num_error_areas + + ;
2011-07-06 03:10:14 +04:00
} else if ( seg_type ( seg , s ) = = AREA_PV ) {
2022-02-23 00:03:11 +03:00
struct device * dev = seg_dev ( seg , s ) ;
name = dm_list_empty ( & dev - > aliases ) ? NULL : dev_name ( dev ) ;
if ( ! dm_tree_node_add_target_area ( node , name , NULL ,
2011-07-06 03:10:14 +04:00
( seg_pv ( seg , s ) - > pe_start + ( extent_size * seg_pe ( seg , s ) ) ) ) )
return_0 ;
2012-02-01 17:47:27 +04:00
num_existing_areas + + ;
2011-08-18 23:38:26 +04:00
} else if ( seg_is_raid ( seg ) ) {
/*
* RAID can handle unassigned areas . It simple puts
* ' - - ' in for the metadata / data device pair . This
* is a valid way to indicate to the RAID target that
* the device is missing .
*
* If an image is marked as VISIBLE_LV and ! LVM_WRITE ,
* it means the device has temporarily been extracted
* from the array . It may come back at a future date ,
* so the bitmap must track differences . Again , ' - - '
* is used in the CTR table .
*/
if ( ( seg_type ( seg , s ) = = AREA_UNASSIGNED ) | |
2015-01-28 15:34:12 +03:00
( lv_is_visible ( seg_lv ( seg , s ) ) & &
2011-08-18 23:38:26 +04:00
! ( seg_lv ( seg , s ) - > status & LVM_WRITE ) ) ) {
/* One each for metadata area and data area */
if ( ! dm_tree_node_add_null_area ( node , 0 ) | |
! dm_tree_node_add_null_area ( node , 0 ) )
2011-08-11 23:38:00 +04:00
return_0 ;
2011-08-18 23:38:26 +04:00
continue ;
2011-08-03 02:07:20 +04:00
}
2016-05-23 18:46:38 +03:00
if ( seg - > meta_areas & & seg_metalv ( seg , s ) ) {
if ( ! ( dlid = build_dm_uuid ( dm - > mem , seg_metalv ( seg , s ) , NULL ) ) )
return_0 ;
if ( ! dm_tree_node_add_target_area ( node , NULL , dlid , extent_size * seg_metale ( seg , s ) ) )
return_0 ;
} else if ( ! dm_tree_node_add_null_area ( node , 0 ) )
2011-08-18 23:38:26 +04:00
return_0 ;
2014-03-11 20:13:47 +04:00
if ( ! ( dlid = build_dm_uuid ( dm - > mem , seg_lv ( seg , s ) , NULL ) ) )
2011-08-18 23:38:26 +04:00
return_0 ;
if ( ! dm_tree_node_add_target_area ( node , NULL , dlid , extent_size * seg_le ( seg , s ) ) )
return_0 ;
} else if ( seg_type ( seg , s ) = = AREA_LV ) {
2014-03-11 20:13:47 +04:00
if ( ! ( dlid = build_dm_uuid ( dm - > mem , seg_lv ( seg , s ) , NULL ) ) )
2011-07-06 03:10:14 +04:00
return_0 ;
if ( ! dm_tree_node_add_target_area ( node , NULL , dlid , extent_size * seg_le ( seg , s ) ) )
2005-11-09 01:52:26 +03:00
return_0 ;
} else {
2009-12-16 22:22:11 +03:00
log_error ( INTERNAL_ERROR " Unassigned area found in LV %s. " ,
2015-11-25 18:06:31 +03:00
display_lvname ( seg - > lv ) ) ;
2002-03-11 14:27:48 +03:00
return 0 ;
}
2005-11-09 01:52:26 +03:00
}
2002-03-11 14:27:48 +03:00
2012-02-01 17:47:27 +04:00
if ( num_error_areas ) {
/* Thins currently do not support partial activation */
if ( lv_is_thin_type ( seg - > lv ) ) {
2015-11-25 18:06:31 +03:00
log_error ( " Cannot activate %s: pool incomplete. " ,
display_lvname ( seg - > lv ) ) ;
2012-02-01 17:47:27 +04:00
return 0 ;
}
}
2005-11-09 01:52:26 +03:00
return 1 ;
}
2002-03-11 14:27:48 +03:00
2013-02-21 13:24:28 +04:00
static int _add_layer_target_to_dtree ( struct dev_manager * dm ,
struct dm_tree_node * dnode ,
2014-09-22 17:50:07 +04:00
const struct logical_volume * lv )
2013-02-21 13:24:28 +04:00
{
const char * layer_dlid ;
2014-03-11 20:13:47 +04:00
if ( ! ( layer_dlid = build_dm_uuid ( dm - > mem , lv , lv_layer ( lv ) ) ) )
2013-02-21 13:24:28 +04:00
return_0 ;
2019-09-14 02:13:33 +03:00
2013-02-21 13:24:28 +04:00
/* Add linear mapping over layered LV */
2019-09-14 02:13:33 +03:00
/* From VDO layer expose ONLY vdo pool header, we would need to use virtual size otherwise */
2023-03-06 16:52:59 +03:00
if ( ! add_linear_area_to_dtree ( dnode , lv_is_vdo_pool ( lv ) ? 8 : lv - > size ,
2019-09-14 02:13:33 +03:00
lv - > vg - > extent_size ,
2013-02-21 13:24:28 +04:00
lv - > vg - > cmd - > use_linear_target ,
lv - > vg - > name , lv - > name ) | |
! dm_tree_node_add_target_area ( dnode , NULL , layer_dlid , 0 ) )
return_0 ;
return 1 ;
}
2005-11-09 16:08:41 +03:00
static int _add_origin_target_to_dtree ( struct dev_manager * dm ,
2014-09-22 17:50:07 +04:00
struct dm_tree_node * dnode ,
const struct logical_volume * lv )
2005-11-09 01:52:26 +03:00
{
const char * real_dlid ;
2002-03-11 23:36:04 +03:00
2014-03-11 20:13:47 +04:00
if ( ! ( real_dlid = build_dm_uuid ( dm - > mem , lv , " real " ) ) )
2005-11-09 01:52:26 +03:00
return_0 ;
2003-04-25 02:09:13 +04:00
2005-11-09 16:05:17 +03:00
if ( ! dm_tree_node_add_snapshot_origin_target ( dnode , lv - > size , real_dlid ) )
2005-11-09 01:52:26 +03:00
return_0 ;
2002-03-11 14:27:48 +03:00
return 1 ;
}
2010-01-13 04:43:32 +03:00
static int _add_snapshot_merge_target_to_dtree ( struct dev_manager * dm ,
struct dm_tree_node * dnode ,
2014-09-22 17:50:07 +04:00
const struct logical_volume * lv )
2010-01-13 04:43:32 +03:00
{
const char * origin_dlid , * cow_dlid , * merge_dlid ;
2013-11-28 14:39:38 +04:00
struct lv_segment * merging_snap_seg = find_snapshot ( lv ) ;
2013-07-19 19:28:43 +04:00
2013-11-28 14:39:38 +04:00
if ( ! lv_is_merging_origin ( lv ) ) {
2015-11-25 18:06:31 +03:00
log_error ( INTERNAL_ERROR " LV %s is not merging snapshot. " ,
display_lvname ( lv ) ) ;
2013-07-19 19:28:43 +04:00
return 0 ;
}
2010-01-13 04:43:32 +03:00
2014-03-11 20:13:47 +04:00
if ( ! ( origin_dlid = build_dm_uuid ( dm - > mem , lv , " real " ) ) )
2010-01-13 04:43:32 +03:00
return_0 ;
2014-03-11 20:13:47 +04:00
if ( ! ( cow_dlid = build_dm_uuid ( dm - > mem , merging_snap_seg - > cow , " cow " ) ) )
2010-01-13 04:43:32 +03:00
return_0 ;
2014-03-11 20:13:47 +04:00
if ( ! ( merge_dlid = build_dm_uuid ( dm - > mem , merging_snap_seg - > cow , NULL ) ) )
2010-01-13 04:43:32 +03:00
return_0 ;
if ( ! dm_tree_node_add_snapshot_merge_target ( dnode , lv - > size , origin_dlid ,
cow_dlid , merge_dlid ,
2013-07-03 00:26:03 +04:00
merging_snap_seg - > chunk_size ) )
2010-01-13 04:43:32 +03:00
return_0 ;
return 1 ;
}
2005-11-09 16:08:41 +03:00
static int _add_snapshot_target_to_dtree ( struct dev_manager * dm ,
2011-06-17 18:14:19 +04:00
struct dm_tree_node * dnode ,
2014-09-22 17:50:07 +04:00
const struct logical_volume * lv ,
2011-06-17 18:14:19 +04:00
struct lv_activate_opts * laopts )
2002-03-11 14:27:48 +03:00
{
2005-11-09 01:52:26 +03:00
const char * origin_dlid ;
const char * cow_dlid ;
struct lv_segment * snap_seg ;
uint64_t size ;
2002-03-11 14:27:48 +03:00
2013-07-03 00:26:03 +04:00
if ( ! ( snap_seg = find_snapshot ( lv ) ) ) {
2015-11-25 18:06:31 +03:00
log_error ( " Couldn't find snapshot for '%s'. " ,
display_lvname ( lv ) ) ;
2005-10-26 23:50:00 +04:00
return 0 ;
2002-03-11 14:27:48 +03:00
}
2014-03-11 20:13:47 +04:00
if ( ! ( origin_dlid = build_dm_uuid ( dm - > mem , snap_seg - > origin , " real " ) ) )
2005-11-09 01:52:26 +03:00
return_0 ;
2003-04-25 02:09:13 +04:00
2014-03-11 20:13:47 +04:00
if ( ! ( cow_dlid = build_dm_uuid ( dm - > mem , snap_seg - > cow , " cow " ) ) )
2005-11-09 01:52:26 +03:00
return_0 ;
2002-03-27 21:17:43 +03:00
2005-11-09 01:52:26 +03:00
size = ( uint64_t ) snap_seg - > len * snap_seg - > origin - > vg - > extent_size ;
2004-05-05 22:11:43 +04:00
2011-06-17 18:22:48 +04:00
if ( ! laopts - > no_merging & & lv_is_merging_cow ( lv ) ) {
2010-02-18 01:59:46 +03:00
/* cow is to be merged so load the error target */
if ( ! dm_tree_node_add_error_target ( dnode , size ) )
return_0 ;
}
else if ( ! dm_tree_node_add_snapshot_target ( dnode , size , origin_dlid ,
cow_dlid , 1 , snap_seg - > chunk_size ) )
2005-11-09 01:52:26 +03:00
return_0 ;
2002-03-11 14:27:48 +03:00
return 1 ;
}
2005-11-09 16:08:41 +03:00
static int _add_target_to_dtree ( struct dev_manager * dm ,
2011-06-17 18:14:19 +04:00
struct dm_tree_node * dnode ,
struct lv_segment * seg ,
struct lv_activate_opts * laopts )
2002-03-11 14:27:48 +03:00
{
2005-11-09 01:52:26 +03:00
uint64_t extent_size = seg - > lv - > vg - > extent_size ;
if ( ! seg - > segtype - > ops - > add_target_line ) {
2009-12-16 22:22:11 +03:00
log_error ( INTERNAL_ERROR " _emit_target cannot handle "
2016-12-25 02:29:30 +03:00
" segment type %s. " , lvseg_name ( seg ) ) ;
2005-11-09 01:52:26 +03:00
return 0 ;
}
2006-05-16 20:48:31 +04:00
return seg - > segtype - > ops - > add_target_line ( dm , dm - > mem , dm - > cmd ,
2005-11-09 01:52:26 +03:00
& dm - > target_state , seg ,
2011-06-17 18:14:19 +04:00
laopts , dnode ,
2017-02-24 02:50:00 +03:00
extent_size * _seg_len ( seg ) ,
2014-08-26 13:39:51 +04:00
& dm - > pvmove_mirror_count ) ;
2002-03-16 01:59:12 +03:00
}
2002-03-11 14:27:48 +03:00
2013-07-09 14:34:49 +04:00
static int _add_new_external_lv_to_dtree ( struct dev_manager * dm ,
struct dm_tree * dtree ,
struct logical_volume * external_lv ,
struct lv_activate_opts * laopts )
2013-02-21 13:25:44 +04:00
{
struct seg_list * sl ;
2024-02-27 19:23:01 +03:00
struct dm_tree_node * dnode ;
2013-02-21 13:25:44 +04:00
2024-02-27 19:23:01 +03:00
/* We've already processed this node if it already has a context ptr */
2024-03-21 19:23:33 +03:00
if ( ( dnode = _cached_dm_tree_node ( dm - > mem , dtree , external_lv , lv_layer ( external_lv ) ) ) & &
2024-02-27 19:23:01 +03:00
dm_tree_node_get_context ( dnode ) ) {
/* Skip repeated invocation of external lv processing */
log_debug_activation ( " Skipping users for already added external origin LV %s. " ,
display_lvname ( external_lv ) ) ;
return 1 ;
}
2013-07-09 14:34:49 +04:00
2015-11-25 18:06:31 +03:00
log_debug_activation ( " Adding external origin LV %s and all active users. " ,
display_lvname ( external_lv ) ) ;
2024-02-27 19:23:01 +03:00
/* If there is active origin LV, add whole origin device, otherwise only -layer */
2013-07-09 14:34:49 +04:00
if ( ! _add_new_lv_to_dtree ( dm , dtree , external_lv , laopts ,
2024-02-27 19:23:01 +03:00
( lv_is_origin ( external_lv ) & &
_cached_dm_info ( dm - > mem , dtree , external_lv , NULL ) )
? NULL : lv_layer ( external_lv ) ) )
2013-07-09 14:34:49 +04:00
return_0 ;
/*
* Add all ACTIVE LVs using this external origin LV . This is
* needed because of conversion of thin which could have been
* also an old - snapshot to external origin .
*/
dm_list_iterate_items ( sl , & external_lv - > segs_using_this_lv )
2024-02-27 19:23:01 +03:00
/* Add only active layered devices (also avoids loop) */
if ( _cached_dm_info ( dm - > mem , dtree , sl - > seg - > lv ,
2014-11-02 22:59:57 +03:00
lv_layer ( sl - > seg - > lv ) ) & &
2013-02-21 13:25:44 +04:00
! _add_new_lv_to_dtree ( dm , dtree , sl - > seg - > lv ,
laopts , lv_layer ( sl - > seg - > lv ) ) )
return_0 ;
2013-07-09 14:34:49 +04:00
2015-11-25 18:06:31 +03:00
log_debug_activation ( " Finished adding external origin LV %s and all active users. " ,
display_lvname ( external_lv ) ) ;
2013-02-21 13:25:44 +04:00
return 1 ;
}
2024-03-21 23:03:44 +03:00
static int _add_new_cvol_subdev_to_dtree ( struct dev_manager * dm ,
struct dm_tree * dtree ,
const struct logical_volume * lv ,
struct lv_activate_opts * laopts ,
struct lv_layer * lvlayer ,
int meta_or_data )
{
const char * layer = meta_or_data ? " cmeta " : " cdata " ;
const struct lv_segment * lvseg = first_seg ( lv ) ;
uint64_t size = meta_or_data ? lvseg - > metadata_len : lvseg - > data_len ;
const struct logical_volume * pool_lv = lvseg - > pool_lv ;
struct dm_tree_node * dnode ;
char * dlid , * dlid_pool , * name ;
2024-04-04 02:29:30 +03:00
union lvid lvid = { { lv - > vg - > id , _get_id_for_meta_or_data ( lvseg , meta_or_data ) } } ;
2024-03-21 23:03:44 +03:00
if ( ! ( dlid = dm_build_dm_uuid ( dm - > mem , UUID_PREFIX , ( const char * ) & lvid . s , layer ) ) )
return_0 ;
if ( ! ( name = dm_build_dm_name ( dm - > mem , lv - > vg - > name , pool_lv - > name , layer ) ) )
return_0 ;
if ( ! ( dnode = dm_tree_add_new_dev_with_udev_flags ( dtree , name , dlid , - 1 , - 1 ,
read_only_lv ( lv , laopts , layer ) ,
( ( lv - > vg - > status & PRECOMMITTED ) | laopts - > revert ) ? 1 : 0 ,
lvlayer ,
_get_udev_flags ( dm , lv , layer , laopts - > noscan ,
laopts - > temporary , 0 ) ) ) )
return_0 ;
if ( dm - > track_pending_delete ) {
log_debug_activation ( " Using error for pending delete of %s-%s. " ,
display_lvname ( lv ) , layer ) ;
if ( ! dm_tree_node_add_error_target ( dnode , size ) )
return_0 ;
} else {
/* add load_segment to meta dnode: linear, size of meta area */
if ( ! add_linear_area_to_dtree ( dnode , size , lv - > vg - > extent_size ,
lv - > vg - > cmd - > use_linear_target ,
lv - > vg - > name , lv - > name ) )
return_0 ;
if ( ! ( dlid_pool = build_dm_uuid ( dm - > mem , pool_lv , NULL ) ) )
return_0 ;
/* add seg_area to prev load_seg: offset 0 maps to cachevol lv offset 0 */
if ( ! dm_tree_node_add_target_area ( dnode , NULL , dlid_pool ,
meta_or_data ? 0 : lvseg - > metadata_len ) )
return_0 ;
}
return 1 ;
}
2005-11-09 16:08:41 +03:00
static int _add_segment_to_dtree ( struct dev_manager * dm ,
2011-06-17 18:14:19 +04:00
struct dm_tree * dtree ,
struct dm_tree_node * dnode ,
struct lv_segment * seg ,
struct lv_activate_opts * laopts ,
const char * layer )
2003-11-12 22:16:48 +03:00
{
2005-11-09 01:52:26 +03:00
uint32_t s ;
2006-11-20 19:45:45 +03:00
struct lv_segment * seg_present ;
2014-11-13 12:08:40 +03:00
const struct segment_type * segtype ;
2010-10-14 01:26:37 +04:00
const char * target_name ;
2003-11-12 22:16:48 +03:00
2005-11-09 01:52:26 +03:00
/* Ensure required device-mapper targets are loaded */
2013-07-03 00:26:03 +04:00
seg_present = find_snapshot ( seg - > lv ) ? : seg ;
2014-11-13 12:08:40 +03:00
segtype = seg_present - > segtype ;
target_name = ( segtype - > ops - > target_name ?
segtype - > ops - > target_name ( seg_present , laopts ) :
segtype - > name ) ;
2006-11-20 19:45:45 +03:00
2013-01-08 02:30:29 +04:00
log_debug_activation ( " Checking kernel supports %s segment type for %s%s%s " ,
2015-11-25 18:06:31 +03:00
target_name , display_lvname ( seg - > lv ) ,
2013-01-08 02:30:29 +04:00
layer ? " - " : " " , layer ? : " " ) ;
2006-11-20 19:45:45 +03:00
2014-11-13 12:08:40 +03:00
if ( segtype - > ops - > target_present & &
! segtype - > ops - > target_present ( seg_present - > lv - > vg - > cmd ,
seg_present , NULL ) ) {
2010-10-14 01:26:37 +04:00
log_error ( " Can't process LV %s: %s target support missing "
2015-11-25 18:06:31 +03:00
" from kernel? " , display_lvname ( seg - > lv ) , target_name ) ;
2003-11-12 22:16:48 +03:00
return 0 ;
}
2013-02-21 13:25:44 +04:00
/* Add external origin layer */
2013-07-09 14:34:49 +04:00
if ( seg - > external_lv & &
! _add_new_external_lv_to_dtree ( dm , dtree , seg - > external_lv , laopts ) )
return_0 ;
2014-02-05 02:50:16 +04:00
2005-11-09 01:52:26 +03:00
/* Add mirror log */
if ( seg - > log_lv & &
2011-06-17 18:14:19 +04:00
! _add_new_lv_to_dtree ( dm , dtree , seg - > log_lv , laopts , NULL ) )
2005-11-09 01:52:26 +03:00
return_0 ;
2014-02-05 02:50:16 +04:00
/* Add pool metadata */
2013-02-21 13:39:47 +04:00
if ( seg - > metadata_lv & &
! _add_new_lv_to_dtree ( dm , dtree , seg - > metadata_lv , laopts , NULL ) )
return_0 ;
2014-02-05 02:50:16 +04:00
/* Add pool layer */
thin: move pool messaging from resume to suspend
Existing messaging intarface for thin-pool has a few 'weak' points:
* Message were posted with each 'resume' operation, thus not allowing
activation of thin-pool with the existing state.
* Acceleration skipped suspend step has not worked in cluster,
since clvmd resumes only nodes which are suspended (have proper lock
state).
* Resume may fail and code is not really designed to 'fail' in this
phase (generic rule here is resume DOES NOT fail unless something serious
is wrong and lvm2 tool usually doesn't handle recovery path in this case.)
* Full thin-pool suspend happened, when taken a thin-volume snapshot.
With this patch the new method relocates message passing into suspend
state.
This has a few drawbacks with current API, but overal it performs
better and gives are more posibilities to deal with errors.
Patch introduces a new logic for 'origin-only' suspend of thin-pool and
this also relates to thin-volume when taking snapshot.
When suspend_origin_only operation is invoked on a pool with
queued messages then only those messages are posted to thin-pool and
actual suspend of thin pool and data and metadata volume is skipped.
This makes taking a snapshot of thin-volume lighter operation and
avoids blocking of other unrelated active thin volumes.
Also fail now happens in 'suspend' state where the 'Fail' is more expected
and it is better handled through error paths.
Activation of thin-pool is now not sending any message and leaves upto a tool
to decided later how to finish unfinished double-commit transaction.
Problem which needs some API improvements relates to the lvm2 tree
construction. For the suspend tree we do not add target table line
into the tree, but only a device is inserted into a tree.
Current mechanism to attach messages for thin-pool requires the libdm
to know about thin-pool target, so lvm2 currently takes assumption, node
is really a thin-pool and fills in the table line for this node (which
should be ensured by the PRELOAD phase, but it's a misuse of internal API)
we would possibly need to be able to attach message to 'any' node.
Other thing to notice - current messaging interface in thin-pool
target requires to suspend thin volume origin first and then send
a create message, but this could not have any 'nice' solution on lvm2
side and IMHO we should introduce something like 'create_after_resume'
message.
Patch also changes the moment, where lvm2 transaction id is increased.
Now it happens only after successful finish of kernel transaction id
change. This change was needed to handle properly activation of pool,
which is in the middle of unfinished transaction, and also this corrects
usage of thin-pool by external apps like Docker.
2015-07-01 14:31:37 +03:00
if ( seg - > pool_lv & & ! laopts - > origin_only & &
2013-02-21 13:39:47 +04:00
! _add_new_lv_to_dtree ( dm , dtree , seg - > pool_lv , laopts ,
lv_layer ( seg - > pool_lv ) ) )
return_0 ;
2003-11-12 22:16:48 +03:00
2018-08-27 22:53:09 +03:00
if ( seg - > writecache & & ! laopts - > origin_only & &
! _add_new_lv_to_dtree ( dm , dtree , seg - > writecache , laopts ,
lv_layer ( seg - > writecache ) ) )
return_0 ;
2019-11-21 01:07:27 +03:00
if ( seg - > integrity_meta_dev & & ! laopts - > origin_only & &
! _add_new_lv_to_dtree ( dm , dtree , seg - > integrity_meta_dev , laopts ,
lv_layer ( seg - > integrity_meta_dev ) ) )
return_0 ;
2013-02-21 13:39:47 +04:00
/* Add any LVs used by this segment */
for ( s = 0 ; s < seg - > area_count ; + + s ) {
if ( ( seg_type ( seg , s ) = = AREA_LV ) & &
2016-12-14 23:47:38 +03:00
/* do not bring up tracked image */
! lv_is_raid_image_with_tracking ( seg_lv ( seg , s ) ) & &
2014-11-13 12:08:40 +03:00
/* origin only for cache without pending delete */
( ! dm - > track_pending_delete | | ! seg_is_cache ( seg ) ) & &
! _add_new_lv_to_dtree ( dm , dtree , seg_lv ( seg , s ) ,
2019-09-14 02:13:33 +03:00
laopts ,
lv_is_vdo_pool ( seg_lv ( seg , s ) ) ?
lv_layer ( seg_lv ( seg , s ) ) : NULL ) )
2011-09-29 12:56:38 +04:00
return_0 ;
2016-05-24 00:55:13 +03:00
if ( seg_is_raid_with_meta ( seg ) & & seg - > meta_areas & & seg_metalv ( seg , s ) & &
2016-12-14 23:47:38 +03:00
! lv_is_raid_image_with_tracking ( seg_lv ( seg , s ) ) & &
2013-02-21 13:39:47 +04:00
! _add_new_lv_to_dtree ( dm , dtree , seg_metalv ( seg , s ) ,
laopts , NULL ) )
2011-09-29 12:56:38 +04:00
return_0 ;
2005-11-09 01:52:26 +03:00
}
2014-11-13 12:08:40 +03:00
if ( dm - > track_pending_delete ) {
/* Replace target and all its used devs with error mapping */
log_debug_activation ( " Using error for pending delete %s. " ,
2015-11-25 18:06:31 +03:00
display_lvname ( seg - > lv ) ) ;
2017-02-24 02:50:00 +03:00
if ( ! dm_tree_node_add_error_target ( dnode , ( uint64_t ) seg - > lv - > vg - > extent_size * _seg_len ( seg ) ) )
2014-11-13 12:08:40 +03:00
return_0 ;
} else if ( ! _add_target_to_dtree ( dm , dnode , seg , laopts ) )
2005-11-09 01:52:26 +03:00
return_0 ;
return 1 ;
2003-07-05 02:34:56 +04:00
}
2005-10-17 22:00:02 +04:00
2005-11-09 16:08:41 +03:00
static int _add_new_lv_to_dtree ( struct dev_manager * dm , struct dm_tree * dtree ,
2014-09-22 17:50:07 +04:00
const struct logical_volume * lv , struct lv_activate_opts * laopts ,
2011-06-17 18:14:19 +04:00
const char * layer )
2005-10-17 22:21:05 +04:00
{
2005-11-09 01:52:26 +03:00
struct lv_segment * seg ;
struct lv_layer * lvlayer ;
2011-06-11 04:03:06 +04:00
struct seg_list * sl ;
2013-02-21 13:39:47 +04:00
struct dm_list * snh ;
2005-11-09 16:05:17 +03:00
struct dm_tree_node * dnode ;
2010-01-22 18:40:31 +03:00
const struct dm_info * dinfo ;
2010-02-18 01:59:46 +03:00
char * name , * dlid ;
2007-11-12 23:51:54 +03:00
uint32_t max_stripe_size = UINT32_C ( 0 ) ;
uint32_t read_ahead = lv - > read_ahead ;
2007-11-29 18:04:12 +03:00
uint32_t read_ahead_flags = UINT32_C ( 0 ) ;
2014-11-13 12:08:40 +03:00
int save_pending_delete = dm - > track_pending_delete ;
2016-04-18 23:54:27 +03:00
int merge_in_progress = 0 ;
2005-10-17 22:21:05 +04:00
2018-08-17 23:45:52 +03:00
if ( ! ( lvlayer = dm_pool_alloc ( dm - > mem , sizeof ( * lvlayer ) ) ) ) {
log_error ( " _add_new_lv_to_dtree: pool alloc failed for %s %s. " ,
display_lvname ( lv ) , layer ) ;
return 0 ;
}
lvlayer - > lv = lv ;
lvlayer - > visible_component = ( laopts - > component_lv = = lv ) ? 1 : 0 ;
2016-12-18 17:05:18 +03:00
log_debug_activation ( " Adding new LV %s%s%s to dtree " , display_lvname ( lv ) ,
layer ? " - " : " " , layer ? : " " ) ;
2014-11-11 13:00:51 +03:00
/* LV with pending delete is never put new into a table */
if ( lv_is_pending_delete ( lv ) & & ! _cached_dm_info ( dm - > mem , dtree , lv , NULL ) )
return 1 ; /* Replace with error only when already exists */
2014-11-02 21:34:50 +03:00
if ( lv_is_cache_pool ( lv ) & &
! dm_list_empty ( & lv - > segs_using_this_lv ) ) {
2014-04-01 19:53:18 +04:00
/* cache pool is 'meta' LV and does not have a real device node */
if ( ! _add_new_lv_to_dtree ( dm , dtree , seg_lv ( first_seg ( lv ) , 0 ) , laopts , NULL ) )
return_0 ;
if ( ! _add_new_lv_to_dtree ( dm , dtree , first_seg ( lv ) - > metadata_lv , laopts , NULL ) )
return_0 ;
return 1 ;
}
2010-01-14 17:39:57 +03:00
/* FIXME Seek a simpler way to lay out the snapshot-merge tree. */
2013-11-28 14:39:38 +04:00
if ( ! layer & & lv_is_merging_origin ( lv ) ) {
seg = find_snapshot ( lv ) ;
2010-01-13 04:54:34 +03:00
/*
2016-04-18 23:54:27 +03:00
* Prevent merge if merge isn ' t currently possible :
2010-01-13 04:54:34 +03:00
* either origin or merging snapshot are open
2016-04-18 12:04:16 +03:00
* - for old snaps use " snapshot-merge " if it is already in use
2010-01-22 18:40:31 +03:00
* - open_count is always retrieved ( as of dm - ioctl 4.7 .0 )
* so just use the tree ' s existing nodes ' info
2010-01-13 04:54:34 +03:00
*/
2016-04-18 23:54:27 +03:00
if ( ( dinfo = _cached_dm_info ( dm - > mem , dtree , lv , NULL ) ) ) {
2024-08-30 00:05:41 +03:00
/* Merging origin LV is present, check if merging is already running. */
2023-03-06 16:52:28 +03:00
if ( ( seg_is_thin_volume ( seg ) & & _lv_has_thin_device_id ( dm - > mem , lv , NULL , seg - > device_id ) ) | |
2016-04-18 23:54:27 +03:00
( ! seg_is_thin_volume ( seg ) & & lv_has_target_type ( dm - > mem , lv , NULL , TARGET_NAME_SNAPSHOT_MERGE ) ) ) {
log_debug_activation ( " Merging of snapshot volume %s to origin %s is in progress. " ,
display_lvname ( seg - > lv ) , display_lvname ( seg - > lv ) ) ;
merge_in_progress = 1 ; /* Merge is already running */
} /* Merge is not yet running, so check if it can be started */
else if ( laopts - > resuming ) {
log_debug_activation ( " Postponing pending snapshot merge for origin %s, "
" merge was not started before suspend. " ,
display_lvname ( lv ) ) ;
laopts - > no_merging = 1 ; /* Cannot be reloaded in suspend */
} /* Non-resuming merge requires origin to be unused */
else if ( dinfo - > open_count ) {
log_debug_activation ( " Postponing pending snapshot merge for origin %s, "
" origin volume is opened. " ,
display_lvname ( lv ) ) ;
laopts - > no_merging = 1 ;
}
}
/* If merge would be still undecided, look as snapshot */
if ( ! merge_in_progress & & ! laopts - > no_merging & &
( dinfo = _cached_dm_info ( dm - > mem , dtree ,
2016-04-18 12:04:16 +03:00
seg_is_thin_volume ( seg ) ?
seg - > lv : seg - > cow , NULL ) ) ) {
if ( seg_is_thin_volume ( seg ) ) {
/* Active thin snapshot prevents merge */
2016-04-18 23:54:27 +03:00
log_debug_activation ( " Postponing pending snapshot merge for origin volume %s, "
" merging thin snapshot volume %s is active. " ,
display_lvname ( lv ) , display_lvname ( seg - > lv ) ) ;
laopts - > no_merging = 1 ;
2016-04-18 12:04:16 +03:00
} else if ( dinfo - > open_count ) {
2016-04-18 23:54:27 +03:00
log_debug_activation ( " Postponing pending snapshot merge for origin volume %s, "
" merging snapshot volume %s is opened. " ,
display_lvname ( lv ) , display_lvname ( seg - > lv ) ) ;
laopts - > no_merging = 1 ;
2016-04-18 12:04:16 +03:00
}
}
2010-01-13 04:54:34 +03:00
}
2011-08-30 18:55:15 +04:00
if ( ! ( name = dm_build_dm_name ( dm - > mem , lv - > vg - > name , lv - > name , layer ) ) )
2005-11-09 01:52:26 +03:00
return_0 ;
2005-10-17 22:21:05 +04:00
2016-11-08 18:20:51 +03:00
/* Even unused thin-pool still needs to get layered UUID -suffix */
2014-11-04 12:33:35 +03:00
if ( ! layer & & lv_is_new_thin_pool ( lv ) )
layer = lv_layer ( lv ) ;
2020-02-19 22:30:02 +03:00
/* Adds -real to the dm uuid of wcorig LV. */
if ( ! layer & & lv_is_writecache_origin ( lv ) )
layer = lv_layer ( lv ) ; /* "real" */
2024-06-26 00:13:54 +03:00
/*
* FIXME : we would like to have - private suffixes used for device not processed by udev
* however ATM we also sometimes want to provide / dev / vg / lv symlinks to such devices
* and also be able to correctly report its status with lvs .
*
* Until problems are resolved this code path needs to be disabled .
*/
if ( 0 & & lvlayer - > visible_component ) {
2024-06-21 12:08:30 +03:00
/* Component LV will be public, do not add any layer suffixes */
if ( ! ( dlid = dm_build_dm_uuid ( dm - > mem , UUID_PREFIX , lv - > lvid . s , NULL ) ) )
return_0 ;
} else if ( ! ( dlid = build_dm_uuid ( dm - > mem , lv , layer ) ) )
2005-11-09 01:52:26 +03:00
return_0 ;
2005-10-17 22:21:05 +04:00
2005-11-09 01:52:26 +03:00
/* We've already processed this node if it already has a context ptr */
2005-11-09 16:05:17 +03:00
if ( ( dnode = dm_tree_find_node_by_uuid ( dtree , dlid ) ) & &
dm_tree_node_get_context ( dnode ) )
2005-11-09 01:52:26 +03:00
return 1 ;
2005-10-17 22:21:05 +04:00
2005-11-09 01:52:26 +03:00
/*
2005-11-09 16:08:41 +03:00
* Add LV to dtree .
2005-11-09 01:52:26 +03:00
* If we ' re working with precommitted metadata , clear any
* existing inactive table left behind .
* Major / minor settings only apply to the visible layer .
*/
2011-06-11 04:03:06 +04:00
/* FIXME Move the clear from here until later, so we can leave
* identical inactive tables untouched . ( For pvmove . )
*/
2009-10-22 17:00:07 +04:00
if ( ! ( dnode = dm_tree_add_new_dev_with_udev_flags ( dtree , name , dlid ,
2006-07-10 23:17:40 +04:00
layer ? UINT32_C ( 0 ) : ( uint32_t ) lv - > major ,
layer ? UINT32_C ( 0 ) : ( uint32_t ) lv - > minor ,
2018-03-08 12:27:04 +03:00
read_only_lv ( lv , laopts , layer ) ,
2011-09-28 02:43:40 +04:00
( ( lv - > vg - > status & PRECOMMITTED ) | laopts - > revert ) ? 1 : 0 ,
2009-10-22 17:00:07 +04:00
lvlayer ,
2018-02-28 19:22:09 +03:00
_get_udev_flags ( dm , lv , layer , laopts - > noscan , laopts - > temporary ,
lvlayer - > visible_component ) ) ) )
2005-11-09 01:52:26 +03:00
return_0 ;
/* Store existing name so we can do rename later */
2005-11-09 16:05:17 +03:00
lvlayer - > old_name = dm_tree_node_get_name ( dnode ) ;
2005-11-09 01:52:26 +03:00
/* Create table */
dm - > pvmove_mirror_count = 0u ;
2013-02-21 13:39:47 +04:00
2014-11-13 12:08:40 +03:00
if ( lv_is_pending_delete ( lv ) )
2014-11-10 12:56:43 +03:00
/* Handle LVs with pending delete */
2014-11-13 12:08:40 +03:00
/* Fow now used only by cache segtype, TODO snapshots */
dm - > track_pending_delete = 1 ;
2014-11-10 12:56:43 +03:00
2024-03-21 23:03:44 +03:00
if ( lv_is_cache_vol ( lv ) )
dm_list_iterate_items ( sl , & lv - > segs_using_this_lv )
if ( lv_is_cache ( sl - > seg - > lv ) & &
/* Cachevol is used by cache LV segment -> add cvol-cdata/cmeta extra layer */
( ! _add_new_cvol_subdev_to_dtree ( dm , dtree , sl - > seg - > lv , laopts , lvlayer , 0 ) | |
! _add_new_cvol_subdev_to_dtree ( dm , dtree , sl - > seg - > lv , laopts , lvlayer , 1 ) ) )
return_0 ;
2014-11-02 21:34:50 +03:00
/* This is unused cache-pool - make metadata accessible */
if ( lv_is_cache_pool ( lv ) )
lv = first_seg ( lv ) - > metadata_lv ;
2013-02-21 13:39:47 +04:00
/* If this is a snapshot origin, add real LV */
/* If this is a snapshot origin + merging snapshot, add cow + real LV */
2013-02-21 13:25:44 +04:00
/* Snapshot origin could be also external origin */
2013-02-21 13:39:47 +04:00
if ( lv_is_origin ( lv ) & & ! layer ) {
if ( ! _add_new_lv_to_dtree ( dm , dtree , lv , laopts , " real " ) )
2005-11-09 01:52:26 +03:00
return_0 ;
2013-02-21 13:39:47 +04:00
if ( ! laopts - > no_merging & & lv_is_merging_origin ( lv ) ) {
if ( ! _add_new_lv_to_dtree ( dm , dtree ,
2014-02-18 00:49:51 +04:00
find_snapshot ( lv ) - > cow , laopts , " cow " ) )
2013-02-21 13:39:47 +04:00
return_0 ;
/*
* Must also add " real " LV for use when
* snapshot - merge target is added
*/
if ( ! _add_snapshot_merge_target_to_dtree ( dm , dnode , lv ) )
return_0 ;
} else if ( ! _add_origin_target_to_dtree ( dm , dnode , lv ) )
return_0 ;
/* Add any snapshots of this LV */
dm_list_iterate ( snh , & lv - > snapshot_segs )
if ( ! _add_new_lv_to_dtree ( dm , dtree ,
dm_list_struct_base ( snh , struct lv_segment ,
origin_list ) - > cow ,
laopts , NULL ) )
return_0 ;
} else if ( lv_is_cow ( lv ) & & ! layer ) {
if ( ! _add_new_lv_to_dtree ( dm , dtree , lv , laopts , " cow " ) )
return_0 ;
if ( ! _add_snapshot_target_to_dtree ( dm , dnode , lv , laopts ) )
return_0 ;
2014-11-04 12:33:35 +03:00
} else if ( ! layer & & ( ( lv_is_thin_pool ( lv ) & & ! lv_is_new_thin_pool ( lv ) ) | |
2019-09-14 02:13:33 +03:00
lv_is_vdo_pool ( lv ) | |
2014-11-04 12:33:35 +03:00
lv_is_external_origin ( lv ) ) ) {
2019-09-14 02:13:33 +03:00
/* External origin or 'used' Thin pool or VDO pool is using layer */
2013-02-21 13:39:47 +04:00
if ( ! _add_new_lv_to_dtree ( dm , dtree , lv , laopts , lv_layer ( lv ) ) )
return_0 ;
if ( ! _add_layer_target_to_dtree ( dm , dnode , lv ) )
return_0 ;
} else {
/* Add 'real' segments for LVs */
dm_list_iterate_items ( seg , & lv - > segments ) {
if ( ! _add_segment_to_dtree ( dm , dtree , dnode , seg , laopts , layer ) )
return_0 ;
if ( max_stripe_size < seg - > stripe_size * seg - > area_count )
max_stripe_size = seg - > stripe_size * seg - > area_count ;
}
2019-09-14 02:13:33 +03:00
if ( ! layer & & lv_is_vdo_pool ( lv ) & &
! _add_layer_target_to_dtree ( dm , dnode , lv ) )
return_0 ;
2005-11-09 01:52:26 +03:00
}
2005-10-17 22:21:05 +04:00
2013-02-21 13:39:47 +04:00
/* Setup thin pool callback */
if ( lv_is_thin_pool ( lv ) & & layer & &
2014-07-09 19:24:34 +04:00
! _pool_register_callback ( dm , dnode , lv ) )
return_0 ;
2019-01-30 18:55:34 +03:00
if ( lv_is_cache ( lv ) & & ! lv_is_cache_vol ( first_seg ( lv ) - > pool_lv ) & &
2016-12-18 17:05:31 +03:00
/* Register callback only for layer activation or non-layered cache LV */
( layer | | ! lv_layer ( lv ) ) & &
/* Register callback when metadata LV is NOT already active */
! _cached_dm_info ( dm - > mem , dtree , first_seg ( first_seg ( lv ) - > pool_lv ) - > metadata_lv , NULL ) & &
2014-07-09 19:24:34 +04:00
! _pool_register_callback ( dm , dnode , lv ) )
2013-02-21 13:39:47 +04:00
return_0 ;
2020-12-10 02:36:09 +03:00
if ( lv_is_cache ( lv ) & & lv_is_cache_vol ( first_seg ( lv ) - > pool_lv ) & &
/* Register callback only for layer activation or non-layered cache LV */
( layer | | ! lv_layer ( lv ) ) & &
2023-04-10 19:41:28 +03:00
/* Register callback when cachevol LV is NOT already active */
! _cached_dm_info ( dm - > mem , dtree , first_seg ( lv ) - > pool_lv , NULL ) & &
2020-12-10 02:36:09 +03:00
! _pool_register_callback ( dm , dnode , lv ) )
return_0 ;
2019-08-20 13:30:25 +03:00
/*
* Update tables for ANY PVMOVE holders for active LV where the name starts with ' pvmove ' ,
* but it ' s not anymore PVMOVE LV and also it ' s not a PVMOVE _mimage LV .
* When resume happens , tables MUST be already preloaded with correct entries !
* ( since we can ' t preload different table while devices are suspended )
*/
if ( ! lv_is_pvmove ( lv ) & & ! strncmp ( lv - > name , " pvmove " , 6 ) & & ! strchr ( lv - > name , ' _ ' ) & &
( dinfo = _cached_dm_info ( dm - > mem , dtree , lv , NULL ) ) )
if ( ! _add_holders_to_dtree ( dm , dtree , lv , laopts , dinfo ) )
return_0 ;
2007-12-05 22:24:32 +03:00
if ( read_ahead = = DM_READ_AHEAD_AUTO ) {
2008-01-08 19:47:10 +03:00
/* we need RA at least twice a whole stripe - see the comment in md/raid0.c */
read_ahead = max_stripe_size * 2 ;
2013-02-21 13:39:47 +04:00
/* FIXME: layered device read-ahead */
2009-05-20 15:09:49 +04:00
if ( ! read_ahead )
2009-06-01 16:43:31 +04:00
lv_calculate_readahead ( lv , & read_ahead ) ;
2007-11-29 18:04:12 +03:00
read_ahead_flags = DM_READ_AHEAD_MINIMUM_FLAG ;
2007-12-05 22:24:32 +03:00
}
2007-11-12 23:51:54 +03:00
2007-11-29 18:04:12 +03:00
dm_tree_node_set_read_ahead ( dnode , read_ahead , read_ahead_flags ) ;
2007-11-12 23:51:54 +03:00
2011-06-11 04:03:06 +04:00
/* Add any LVs referencing a PVMOVE LV unless told not to */
2014-09-16 00:33:53 +04:00
if ( dm - > track_pvmove_deps & & lv_is_pvmove ( lv ) )
2011-06-11 04:03:06 +04:00
dm_list_iterate_items ( sl , & lv - > segs_using_this_lv )
2011-06-17 18:14:19 +04:00
if ( ! _add_new_lv_to_dtree ( dm , dtree , sl - > seg - > lv , laopts , NULL ) )
2011-06-11 04:03:06 +04:00
return_0 ;
2014-11-13 12:08:40 +03:00
dm - > track_pending_delete = save_pending_delete ; /* restore */
2005-10-18 16:39:20 +04:00
return 1 ;
2005-10-17 22:21:05 +04:00
}
2007-05-15 18:42:01 +04:00
/* FIXME: symlinks should be created/destroyed at the same time
* as the kernel devices but we can ' t do that from within libdevmapper
* at present so we must walk the tree twice instead . */
2005-11-09 01:52:26 +03:00
/*
* Create LV symlinks for children of supplied root node .
*/
2005-11-09 16:05:17 +03:00
static int _create_lv_symlinks ( struct dev_manager * dm , struct dm_tree_node * root )
2005-10-17 22:21:05 +04:00
{
2005-11-09 01:52:26 +03:00
void * handle = NULL ;
2005-11-09 16:05:17 +03:00
struct dm_tree_node * child ;
2005-11-09 01:52:26 +03:00
struct lv_layer * lvlayer ;
2008-12-19 17:22:48 +03:00
char * old_vgname , * old_lvname , * old_layer ;
char * new_vgname , * new_lvname , * new_layer ;
2005-11-09 01:52:26 +03:00
const char * name ;
int r = 1 ;
2005-10-17 22:21:05 +04:00
2011-06-17 18:50:53 +04:00
/* Nothing to do if udev fallback is disabled. */
2013-05-13 13:46:24 +04:00
if ( ! _check_udev_fallback ( dm - > cmd ) ) {
2011-10-14 17:23:47 +04:00
fs_set_create ( ) ;
2011-06-17 18:50:53 +04:00
return 1 ;
2011-10-14 17:23:47 +04:00
}
2011-06-17 18:50:53 +04:00
2005-11-09 16:05:17 +03:00
while ( ( child = dm_tree_next_child ( & handle , root , 0 ) ) ) {
2009-12-03 12:59:54 +03:00
if ( ! ( lvlayer = dm_tree_node_get_context ( child ) ) )
2005-11-09 01:52:26 +03:00
continue ;
2005-10-17 22:21:05 +04:00
2005-11-09 01:52:26 +03:00
/* Detect rename */
2005-11-09 16:05:17 +03:00
name = dm_tree_node_get_name ( child ) ;
2005-10-17 22:21:05 +04:00
2005-11-09 01:52:26 +03:00
if ( name & & lvlayer - > old_name & & * lvlayer - > old_name & & strcmp ( name , lvlayer - > old_name ) ) {
2008-12-19 17:22:48 +03:00
if ( ! dm_split_lvm_name ( dm - > mem , lvlayer - > old_name , & old_vgname , & old_lvname , & old_layer ) ) {
2016-12-01 16:53:35 +03:00
log_error ( " _create_lv_symlinks: Couldn't split up old device name %s. " , lvlayer - > old_name ) ;
2008-01-30 17:00:02 +03:00
return 0 ;
}
2008-12-19 17:22:48 +03:00
if ( ! dm_split_lvm_name ( dm - > mem , name , & new_vgname , & new_lvname , & new_layer ) ) {
2016-12-01 16:53:35 +03:00
log_error ( " _create_lv_symlinks: Couldn't split up new device name %s. " , name ) ;
2008-12-19 17:22:48 +03:00
return 0 ;
}
if ( ! fs_rename_lv ( lvlayer - > lv , name , old_vgname , old_lvname ) )
r = 0 ;
2009-05-28 05:11:29 +04:00
continue ;
}
2018-02-28 19:22:09 +03:00
if ( _lv_has_mknode ( lvlayer - > lv ) | | lvlayer - > visible_component ) {
2010-02-24 23:00:56 +03:00
if ( ! _dev_manager_lv_mknodes ( lvlayer - > lv ) )
2009-05-28 05:11:29 +04:00
r = 0 ;
continue ;
}
2010-02-24 23:00:56 +03:00
if ( ! _dev_manager_lv_rmnodes ( lvlayer - > lv ) )
2005-11-09 01:52:26 +03:00
r = 0 ;
2005-10-17 22:21:05 +04:00
}
2005-11-09 01:52:26 +03:00
return r ;
}
2005-10-18 16:39:20 +04:00
2007-05-15 18:42:01 +04:00
/*
* Remove LV symlinks for children of supplied root node .
*/
static int _remove_lv_symlinks ( struct dev_manager * dm , struct dm_tree_node * root )
{
void * handle = NULL ;
struct dm_tree_node * child ;
char * vgname , * lvname , * layer ;
int r = 1 ;
2011-06-17 18:50:53 +04:00
/* Nothing to do if udev fallback is disabled. */
2013-05-13 13:46:24 +04:00
if ( ! _check_udev_fallback ( dm - > cmd ) )
2011-06-17 18:50:53 +04:00
return 1 ;
2007-05-15 18:42:01 +04:00
while ( ( child = dm_tree_next_child ( & handle , root , 0 ) ) ) {
2008-01-30 17:00:02 +03:00
if ( ! dm_split_lvm_name ( dm - > mem , dm_tree_node_get_name ( child ) , & vgname , & lvname , & layer ) ) {
2007-05-15 18:42:01 +04:00
r = 0 ;
continue ;
}
2008-06-05 16:45:55 +04:00
if ( ! * vgname )
continue ;
2007-05-15 18:42:01 +04:00
/* only top level layer has symlinks */
if ( * layer )
continue ;
2010-01-07 22:54:21 +03:00
fs_del_lv_byname ( dm - > cmd - > dev_dir , vgname , lvname ,
dm - > cmd - > current_settings . udev_rules ) ;
2007-05-15 18:42:01 +04:00
}
return r ;
}
2014-11-13 12:08:40 +03:00
static int _clean_tree ( struct dev_manager * dm , struct dm_tree_node * root , const char * non_toplevel_tree_dlid )
2005-11-09 01:52:26 +03:00
{
void * handle = NULL ;
2005-11-09 16:05:17 +03:00
struct dm_tree_node * child ;
2005-11-09 01:52:26 +03:00
char * vgname , * lvname , * layer ;
const char * name , * uuid ;
2014-11-13 12:08:40 +03:00
struct dm_str_list * dl ;
2005-11-09 16:05:17 +03:00
while ( ( child = dm_tree_next_child ( & handle , root , 0 ) ) ) {
if ( ! ( name = dm_tree_node_get_name ( child ) ) )
2005-11-09 01:52:26 +03:00
continue ;
2005-11-09 16:05:17 +03:00
if ( ! ( uuid = dm_tree_node_get_uuid ( child ) ) )
2005-11-09 01:52:26 +03:00
continue ;
2008-01-30 17:00:02 +03:00
if ( ! dm_split_lvm_name ( dm - > mem , name , & vgname , & lvname , & layer ) ) {
log_error ( " _clean_tree: Couldn't split up device name %s. " , name ) ;
return 0 ;
}
2005-11-09 01:52:26 +03:00
/* Not meant to be top level? */
2014-11-13 12:08:40 +03:00
if ( ! * layer )
2014-11-11 02:50:13 +03:00
continue ;
2010-08-17 23:25:05 +04:00
/* If operation was performed on a partial tree, don't remove it */
if ( non_toplevel_tree_dlid & & ! strcmp ( non_toplevel_tree_dlid , uuid ) )
continue ;
2019-08-27 13:18:47 +03:00
if ( ! ( uuid = dm_pool_strdup ( dm - > cmd - > pending_delete_mem , uuid ) ) ) {
log_error ( " _clean_tree: Failed to duplicate uuid. " ) ;
return 0 ;
}
if ( ! str_list_add ( dm - > cmd - > pending_delete_mem , & dm - > cmd - > pending_delete , uuid ) )
2009-08-03 22:01:45 +04:00
return_0 ;
2005-11-09 01:52:26 +03:00
}
2019-08-17 00:49:59 +03:00
/* Deactivate any tracked pending delete nodes */
2019-08-27 13:18:47 +03:00
if ( ! dm_list_empty ( & dm - > cmd - > pending_delete ) & & ! dm_get_suspended_counter ( ) ) {
2019-08-17 00:49:59 +03:00
fs_unlock ( ) ;
dm_tree_set_cookie ( root , fs_get_cookie ( ) ) ;
2019-08-27 13:18:47 +03:00
dm_list_iterate_items ( dl , & dm - > cmd - > pending_delete ) {
2019-08-17 00:49:59 +03:00
log_debug_activation ( " Deleting tracked UUID %s. " , dl - > str ) ;
if ( ! dm_tree_deactivate_children ( root , dl - > str , strlen ( dl - > str ) ) )
return_0 ;
}
2019-08-27 13:18:47 +03:00
dm_list_init ( & dm - > cmd - > pending_delete ) ;
dm_pool_empty ( dm - > cmd - > pending_delete_mem ) ;
2019-08-17 00:49:59 +03:00
}
2005-11-09 01:52:26 +03:00
return 1 ;
2005-10-17 22:21:05 +04:00
}
2014-09-22 17:50:07 +04:00
static int _tree_action ( struct dev_manager * dm , const struct logical_volume * lv ,
2011-06-17 18:14:19 +04:00
struct lv_activate_opts * laopts , action_t action )
2005-10-17 22:21:05 +04:00
{
2014-11-13 15:15:58 +03:00
static const char _action_names [ ] [ 24 ] = {
2014-11-09 01:41:22 +03:00
" PRELOAD " , " ACTIVATE " , " DEACTIVATE " , " SUSPEND " , " SUSPEND_WITH_LOCKFS " , " CLEAN "
} ;
2011-10-11 14:02:28 +04:00
const size_t DLID_SIZE = ID_LEN + sizeof ( UUID_PREFIX ) - 1 ;
2005-11-09 16:05:17 +03:00
struct dm_tree * dtree ;
struct dm_tree_node * root ;
2005-10-18 16:39:20 +04:00
char * dlid ;
2005-10-17 22:21:05 +04:00
int r = 0 ;
2014-11-09 01:41:22 +03:00
if ( action < DM_ARRAY_SIZE ( _action_names ) )
2015-06-15 15:33:29 +03:00
log_debug_activation ( " Creating %s%s tree for %s. " ,
_action_names [ action ] ,
( laopts - > origin_only ) ? " origin-only " : " " ,
display_lvname ( lv ) ) ;
2014-11-09 01:41:22 +03:00
2017-02-04 17:44:25 +03:00
/* Some LV cannot be used for top level tree */
2014-04-07 22:33:50 +04:00
/* TODO: add more.... */
2014-11-02 21:34:50 +03:00
if ( lv_is_cache_pool ( lv ) & & ! dm_list_empty ( & lv - > segs_using_this_lv ) ) {
2015-11-25 18:06:31 +03:00
log_error ( INTERNAL_ERROR " Cannot create tree for %s. " ,
display_lvname ( lv ) ) ;
2014-04-07 22:33:50 +04:00
return 0 ;
}
2013-07-09 14:34:49 +04:00
/* Some targets may build bigger tree for activation */
dm - > activation = ( ( action = = PRELOAD ) | | ( action = = ACTIVATE ) ) ;
thin: move pool messaging from resume to suspend
Existing messaging intarface for thin-pool has a few 'weak' points:
* Message were posted with each 'resume' operation, thus not allowing
activation of thin-pool with the existing state.
* Acceleration skipped suspend step has not worked in cluster,
since clvmd resumes only nodes which are suspended (have proper lock
state).
* Resume may fail and code is not really designed to 'fail' in this
phase (generic rule here is resume DOES NOT fail unless something serious
is wrong and lvm2 tool usually doesn't handle recovery path in this case.)
* Full thin-pool suspend happened, when taken a thin-volume snapshot.
With this patch the new method relocates message passing into suspend
state.
This has a few drawbacks with current API, but overal it performs
better and gives are more posibilities to deal with errors.
Patch introduces a new logic for 'origin-only' suspend of thin-pool and
this also relates to thin-volume when taking snapshot.
When suspend_origin_only operation is invoked on a pool with
queued messages then only those messages are posted to thin-pool and
actual suspend of thin pool and data and metadata volume is skipped.
This makes taking a snapshot of thin-volume lighter operation and
avoids blocking of other unrelated active thin volumes.
Also fail now happens in 'suspend' state where the 'Fail' is more expected
and it is better handled through error paths.
Activation of thin-pool is now not sending any message and leaves upto a tool
to decided later how to finish unfinished double-commit transaction.
Problem which needs some API improvements relates to the lvm2 tree
construction. For the suspend tree we do not add target table line
into the tree, but only a device is inserted into a tree.
Current mechanism to attach messages for thin-pool requires the libdm
to know about thin-pool target, so lvm2 currently takes assumption, node
is really a thin-pool and fills in the table line for this node (which
should be ensured by the PRELOAD phase, but it's a misuse of internal API)
we would possibly need to be able to attach message to 'any' node.
Other thing to notice - current messaging interface in thin-pool
target requires to suspend thin volume origin first and then send
a create message, but this could not have any 'nice' solution on lvm2
side and IMHO we should introduce something like 'create_after_resume'
message.
Patch also changes the moment, where lvm2 transaction id is increased.
Now it happens only after successful finish of kernel transaction id
change. This change was needed to handle properly activation of pool,
which is in the middle of unfinished transaction, and also this corrects
usage of thin-pool by external apps like Docker.
2015-07-01 14:31:37 +03:00
dm - > suspend = ( action = = SUSPEND_WITH_LOCKFS ) | | ( action = = SUSPEND ) ;
2016-05-18 16:22:26 +03:00
2024-05-24 21:50:53 +03:00
/* Drop any cache before DM table manipulation within locked section
* TODO : check if it makes sense to manage cache within lock */
2024-07-08 23:32:41 +03:00
dm_devs_cache_destroy ( ) ;
2021-12-15 13:45:22 +03:00
2022-07-09 03:50:08 +03:00
dtree = _create_partial_dtree ( dm , lv , laopts - > origin_only ) ;
2021-12-15 13:45:22 +03:00
2022-07-09 03:50:08 +03:00
if ( ! dtree )
return_0 ;
2005-11-09 16:05:17 +03:00
if ( ! ( root = dm_tree_find_node ( dtree , 0 , 0 ) ) ) {
2016-12-01 16:53:35 +03:00
log_error ( " Lost dependency tree root node. " ) ;
2011-01-10 17:02:30 +03:00
goto out_no_root ;
2005-10-17 22:21:05 +04:00
}
2011-01-10 17:02:30 +03:00
/* Restore fs cookie */
dm_tree_set_cookie ( root , fs_get_cookie ( ) ) ;
2014-03-11 20:13:47 +04:00
if ( ! ( dlid = build_dm_uuid ( dm - > mem , lv , laopts - > origin_only ? lv_layer ( lv ) : NULL ) ) )
2005-11-09 01:52:26 +03:00
goto_out ;
2005-10-17 22:21:05 +04:00
2005-10-19 17:59:18 +04:00
/* Only process nodes with uuid of "LVM-" plus VG id. */
2005-10-25 23:08:21 +04:00
switch ( action ) {
2005-11-09 01:52:26 +03:00
case CLEAN :
2014-06-09 12:58:57 +04:00
if ( retry_deactivation ( ) )
dm_tree_retry_remove ( root ) ;
2005-11-09 01:52:26 +03:00
/* Deactivate any unused non-toplevel nodes */
2011-06-17 18:14:19 +04:00
if ( ! _clean_tree ( dm , root , laopts - > origin_only ? dlid : NULL ) )
2005-11-09 01:52:26 +03:00
goto_out ;
break ;
2005-10-25 23:08:21 +04:00
case DEACTIVATE :
2011-09-22 21:39:56 +04:00
if ( retry_deactivation ( ) )
dm_tree_retry_remove ( root ) ;
2011-10-11 12:54:01 +04:00
/* Deactivate LV and all devices it references that nothing else has open. */
2011-10-11 14:02:28 +04:00
if ( ! dm_tree_deactivate_children ( root , dlid , DLID_SIZE ) )
2009-08-03 22:01:45 +04:00
goto_out ;
2007-05-15 18:42:01 +04:00
if ( ! _remove_lv_symlinks ( dm , root ) )
2015-11-25 18:06:31 +03:00
log_warn ( " Failed to remove all device symlinks associated with %s. " ,
display_lvname ( lv ) ) ;
2005-10-25 23:08:21 +04:00
break ;
case SUSPEND :
2006-08-09 01:20:00 +04:00
dm_tree_skip_lockfs ( root ) ;
2016-04-06 11:29:05 +03:00
if ( ! dm - > flush_required )
2007-01-09 23:31:08 +03:00
dm_tree_use_no_flush_suspend ( root ) ;
2011-02-28 22:53:03 +03:00
/* Fall through */
2006-08-09 01:20:00 +04:00
case SUSPEND_WITH_LOCKFS :
2011-10-11 14:02:28 +04:00
if ( ! dm_tree_suspend_children ( root , dlid , DLID_SIZE ) )
2005-11-09 01:52:26 +03:00
goto_out ;
break ;
case PRELOAD :
case ACTIVATE :
/* Add all required new devices to tree */
thin: move pool messaging from resume to suspend
Existing messaging intarface for thin-pool has a few 'weak' points:
* Message were posted with each 'resume' operation, thus not allowing
activation of thin-pool with the existing state.
* Acceleration skipped suspend step has not worked in cluster,
since clvmd resumes only nodes which are suspended (have proper lock
state).
* Resume may fail and code is not really designed to 'fail' in this
phase (generic rule here is resume DOES NOT fail unless something serious
is wrong and lvm2 tool usually doesn't handle recovery path in this case.)
* Full thin-pool suspend happened, when taken a thin-volume snapshot.
With this patch the new method relocates message passing into suspend
state.
This has a few drawbacks with current API, but overal it performs
better and gives are more posibilities to deal with errors.
Patch introduces a new logic for 'origin-only' suspend of thin-pool and
this also relates to thin-volume when taking snapshot.
When suspend_origin_only operation is invoked on a pool with
queued messages then only those messages are posted to thin-pool and
actual suspend of thin pool and data and metadata volume is skipped.
This makes taking a snapshot of thin-volume lighter operation and
avoids blocking of other unrelated active thin volumes.
Also fail now happens in 'suspend' state where the 'Fail' is more expected
and it is better handled through error paths.
Activation of thin-pool is now not sending any message and leaves upto a tool
to decided later how to finish unfinished double-commit transaction.
Problem which needs some API improvements relates to the lvm2 tree
construction. For the suspend tree we do not add target table line
into the tree, but only a device is inserted into a tree.
Current mechanism to attach messages for thin-pool requires the libdm
to know about thin-pool target, so lvm2 currently takes assumption, node
is really a thin-pool and fills in the table line for this node (which
should be ensured by the PRELOAD phase, but it's a misuse of internal API)
we would possibly need to be able to attach message to 'any' node.
Other thing to notice - current messaging interface in thin-pool
target requires to suspend thin volume origin first and then send
a create message, but this could not have any 'nice' solution on lvm2
side and IMHO we should introduce something like 'create_after_resume'
message.
Patch also changes the moment, where lvm2 transaction id is increased.
Now it happens only after successful finish of kernel transaction id
change. This change was needed to handle properly activation of pool,
which is in the middle of unfinished transaction, and also this corrects
usage of thin-pool by external apps like Docker.
2015-07-01 14:31:37 +03:00
if ( ! _add_new_lv_to_dtree ( dm , dtree , lv , laopts ,
( lv_is_origin ( lv ) & & laopts - > origin_only ) ? " real " :
2019-09-14 02:13:33 +03:00
( laopts - > origin_only & &
( lv_is_thin_pool ( lv ) | |
lv_is_vdo_pool ( lv ) ) ) ?
lv_layer ( lv ) : NULL ) )
2005-11-09 01:52:26 +03:00
goto_out ;
/* Preload any devices required before any suspensions */
2013-12-17 18:17:44 +04:00
if ( ! dm_tree_preload_children ( root , dlid , DLID_SIZE ) )
2009-08-03 22:01:45 +04:00
goto_out ;
2005-11-09 01:52:26 +03:00
2015-10-25 22:41:19 +03:00
if ( ( dm_tree_node_size_changed ( root ) < 0 ) )
2009-05-20 13:52:37 +04:00
dm - > flush_required = 1 ;
2015-10-26 23:46:54 +03:00
/* Currently keep the code require flush for any
2016-04-05 22:34:42 +03:00
* non ' thin pool / volume ' and size increase */
else if ( ! lv_is_thin_volume ( lv ) & &
! lv_is_thin_pool ( lv ) & &
2022-08-19 15:48:01 +03:00
! lv_is_vdo ( lv ) & &
! lv_is_vdo_pool ( lv ) & &
2016-04-05 22:34:42 +03:00
dm_tree_node_size_changed ( root ) )
2015-10-26 23:46:54 +03:00
dm - > flush_required = 1 ;
2009-07-31 22:30:31 +04:00
if ( action = = ACTIVATE ) {
2011-10-11 14:02:28 +04:00
if ( ! dm_tree_activate_children ( root , dlid , DLID_SIZE ) )
2013-12-17 18:17:44 +04:00
goto_out ;
2011-10-11 12:57:13 +04:00
if ( ! _create_lv_symlinks ( dm , root ) )
2015-11-25 18:06:31 +03:00
log_warn ( " Failed to create symlinks for %s. " ,
display_lvname ( lv ) ) ;
2009-07-31 22:30:31 +04:00
}
2005-11-09 01:52:26 +03:00
2005-10-25 23:08:21 +04:00
break ;
default :
2014-11-09 01:41:22 +03:00
log_error ( INTERNAL_ERROR " _tree_action: Action %u not supported. " , action ) ;
2005-10-17 22:21:05 +04:00
goto out ;
2009-12-03 12:58:30 +03:00
}
2005-10-17 22:21:05 +04:00
r = 1 ;
out :
2011-01-10 17:02:30 +03:00
/* Save fs cookie for udev settle, do not wait here */
fs_set_cookie ( dm_tree_get_cookie ( root ) ) ;
out_no_root :
2005-11-09 16:05:17 +03:00
dm_tree_free ( dtree ) ;
2005-10-17 22:21:05 +04:00
return r ;
}
2010-08-17 20:25:32 +04:00
/* origin_only may only be set if we are resuming (not activating) an origin LV */
2014-09-22 17:50:07 +04:00
int dev_manager_activate ( struct dev_manager * dm , const struct logical_volume * lv ,
2011-06-17 18:14:19 +04:00
struct lv_activate_opts * laopts )
2005-11-09 01:52:26 +03:00
{
2011-06-17 18:14:19 +04:00
if ( ! _tree_action ( dm , lv , laopts , ACTIVATE ) )
2005-11-09 01:52:26 +03:00
return_0 ;
2011-10-11 12:59:42 +04:00
if ( ! _tree_action ( dm , lv , laopts , CLEAN ) )
return_0 ;
return 1 ;
2005-11-09 01:52:26 +03:00
}
2010-08-17 20:25:32 +04:00
/* origin_only may only be set if we are resuming (not activating) an origin LV */
2014-09-22 17:50:07 +04:00
int dev_manager_preload ( struct dev_manager * dm , const struct logical_volume * lv ,
2011-06-17 18:14:19 +04:00
struct lv_activate_opts * laopts , int * flush_required )
2005-11-09 01:52:26 +03:00
{
2016-04-06 11:29:57 +03:00
dm - > flush_required = * flush_required ;
2011-06-17 18:14:19 +04:00
if ( ! _tree_action ( dm , lv , laopts , PRELOAD ) )
2011-10-11 12:59:42 +04:00
return_0 ;
2009-05-20 13:52:37 +04:00
* flush_required = dm - > flush_required ;
return 1 ;
2005-11-09 01:52:26 +03:00
}
2005-10-26 18:13:52 +04:00
2014-09-22 17:50:07 +04:00
int dev_manager_deactivate ( struct dev_manager * dm , const struct logical_volume * lv )
2005-10-25 23:08:21 +04:00
{
2011-06-17 18:14:19 +04:00
struct lv_activate_opts laopts = { 0 } ;
2005-10-25 23:08:21 +04:00
2011-10-11 12:59:42 +04:00
if ( ! _tree_action ( dm , lv , & laopts , DEACTIVATE ) )
return_0 ;
2005-10-25 23:08:21 +04:00
2011-10-11 12:59:42 +04:00
return 1 ;
2005-10-25 23:08:21 +04:00
}
2014-09-22 17:50:07 +04:00
int dev_manager_suspend ( struct dev_manager * dm , const struct logical_volume * lv ,
2011-06-17 18:14:19 +04:00
struct lv_activate_opts * laopts , int lockfs , int flush_required )
2005-10-26 18:13:52 +04:00
{
2009-05-20 13:52:37 +04:00
dm - > flush_required = flush_required ;
2011-10-11 12:59:42 +04:00
if ( ! _tree_action ( dm , lv , laopts , lockfs ? SUSPEND_WITH_LOCKFS : SUSPEND ) )
return_0 ;
return 1 ;
2005-10-26 18:13:52 +04:00
}
2005-10-25 23:08:21 +04:00
/*
* Does device use VG somewhere in its construction ?
* Returns 1 if uncertain .
*/
2006-05-11 21:58:58 +04:00
int dev_manager_device_uses_vg ( struct device * dev ,
2005-10-25 23:08:21 +04:00
struct volume_group * vg )
{
2005-11-09 16:05:17 +03:00
struct dm_tree * dtree ;
struct dm_tree_node * root ;
2010-07-09 19:34:40 +04:00
char dlid [ sizeof ( UUID_PREFIX ) + sizeof ( struct id ) - 1 ] __attribute__ ( ( aligned ( 8 ) ) ) ;
2005-10-25 23:08:21 +04:00
int r = 1 ;
2005-11-09 16:05:17 +03:00
if ( ! ( dtree = dm_tree_create ( ) ) ) {
2016-12-01 16:53:35 +03:00
log_error ( " Failed to create partial dtree. " ) ;
2005-10-25 23:08:21 +04:00
return r ;
}
2024-05-03 13:43:14 +03:00
dm_tree_set_optional_uuid_suffixes ( dtree , ( const char * * ) _uuid_suffix_list ) ;
2014-07-31 00:55:11 +04:00
2024-05-25 23:36:18 +03:00
if ( ! dm_tree_add_dev ( dtree , MAJOR ( dev - > dev ) , MINOR ( dev - > dev ) ) ) {
log_error ( " Failed to add device %s (%u:%u) to dtree. " ,
dev_name ( dev ) , MAJOR ( dev - > dev ) , MINOR ( dev - > dev ) ) ;
2005-10-25 23:08:21 +04:00
goto out ;
}
memcpy ( dlid , UUID_PREFIX , sizeof ( UUID_PREFIX ) - 1 ) ;
memcpy ( dlid + sizeof ( UUID_PREFIX ) - 1 , & vg - > id . uuid [ 0 ] , sizeof ( vg - > id ) ) ;
2005-11-09 16:05:17 +03:00
if ( ! ( root = dm_tree_find_node ( dtree , 0 , 0 ) ) ) {
2016-12-01 16:53:35 +03:00
log_error ( " Lost dependency tree root node. " ) ;
2005-10-25 23:08:21 +04:00
goto out ;
}
2005-11-09 16:05:17 +03:00
if ( dm_tree_children_use_uuid ( root , dlid , sizeof ( UUID_PREFIX ) + sizeof ( vg - > id ) - 1 ) )
2005-11-09 01:52:26 +03:00
goto_out ;
2005-10-25 23:08:21 +04:00
r = 0 ;
out :
2005-11-09 16:05:17 +03:00
dm_tree_free ( dtree ) ;
2016-12-01 16:53:35 +03:00
2005-10-25 23:08:21 +04:00
return r ;
}
lvresize: add new options and defaults for fs handling
The new option "--fs String" for lvresize/lvreduce/lvextend
controls the handling of file systems before/after resizing
the LV. --resizefs is the same as --fs resize.
The new option "--fsmode String" can be used to control
mounting and unmounting of the fs during resizing.
Possible --fs values:
checksize
Only applies to reducing size; does nothing for extend.
Check the fs size and reduce the LV if the fs is not using
the affected space, i.e. the fs does not need to be shrunk.
Fail the command without reducing the fs or LV if the fs is
using the affected space.
resize
Resize the fs using the fs-specific resize command.
This may include mounting, unmounting, or running fsck.
See --fsmode to control mounting behavior, and --nofsck to
disable fsck.
resize_fsadm
Use the old method of calling fsadm to handle the fs
(deprecated.) Warning: this option does not prevent lvreduce
from destroying file systems that are unmounted (or mounted
if prompts are skipped.)
ignore
Resize the LV without checking for or handling a file system.
Warning: using ignore when reducing the LV size may destroy the
file system.
Possible --fsmode values:
manage
Mount or unmount the fs as needed to resize the fs,
and attempt to restore the original mount state at the end.
nochange
Do not mount or unmount the fs. If mounting or unmounting
is required to resize the fs, then do not resize the fs or
the LV and fail the command.
offline
Unmount the fs if it is mounted, and resize the fs while it
is unmounted. If mounting is required to resize the fs,
then do not resize the fs or the LV and fail the command.
Notes on lvreduce:
When no --fs or --resizefs option is specified:
. lvextend default behavior is fs ignore.
. lvreduce default behavior is fs checksize
(includes activating the LV.)
With the exception of --fs resize_fsadm|ignore, lvreduce requires
the recent libblkid fields FSLASTBLOCK and FSBLOCKSIZE.
FSLASTBLOCK*FSBLOCKSIZE is the last byte used by the fs on the LV,
which determines if reducing the fs is necessary.
2022-06-14 23:20:21 +03:00
/*
* crypt offset is usually the LUKS header size but can be larger .
* The LUKS header is usually 2 MB for LUKS1 and 16 MB for LUKS2 .
* The offset needs to be subtracted from the LV size to get the
* size used to resize the crypt device .
*/
int get_crypt_table_offset ( dev_t crypt_devt , uint32_t * offset_bytes )
{
2023-11-06 16:39:30 +03:00
struct dm_task * dmt ;
lvresize: add new options and defaults for fs handling
The new option "--fs String" for lvresize/lvreduce/lvextend
controls the handling of file systems before/after resizing
the LV. --resizefs is the same as --fs resize.
The new option "--fsmode String" can be used to control
mounting and unmounting of the fs during resizing.
Possible --fs values:
checksize
Only applies to reducing size; does nothing for extend.
Check the fs size and reduce the LV if the fs is not using
the affected space, i.e. the fs does not need to be shrunk.
Fail the command without reducing the fs or LV if the fs is
using the affected space.
resize
Resize the fs using the fs-specific resize command.
This may include mounting, unmounting, or running fsck.
See --fsmode to control mounting behavior, and --nofsck to
disable fsck.
resize_fsadm
Use the old method of calling fsadm to handle the fs
(deprecated.) Warning: this option does not prevent lvreduce
from destroying file systems that are unmounted (or mounted
if prompts are skipped.)
ignore
Resize the LV without checking for or handling a file system.
Warning: using ignore when reducing the LV size may destroy the
file system.
Possible --fsmode values:
manage
Mount or unmount the fs as needed to resize the fs,
and attempt to restore the original mount state at the end.
nochange
Do not mount or unmount the fs. If mounting or unmounting
is required to resize the fs, then do not resize the fs or
the LV and fail the command.
offline
Unmount the fs if it is mounted, and resize the fs while it
is unmounted. If mounting is required to resize the fs,
then do not resize the fs or the LV and fail the command.
Notes on lvreduce:
When no --fs or --resizefs option is specified:
. lvextend default behavior is fs ignore.
. lvreduce default behavior is fs checksize
(includes activating the LV.)
With the exception of --fs resize_fsadm|ignore, lvreduce requires
the recent libblkid fields FSLASTBLOCK and FSBLOCKSIZE.
FSLASTBLOCK*FSBLOCKSIZE is the last byte used by the fs on the LV,
which determines if reducing the fs is necessary.
2022-06-14 23:20:21 +03:00
uint64_t start , length ;
char * target_type = NULL ;
void * next = NULL ;
char * params = NULL ;
char offset_str [ 32 ] = { 0 } ;
int copy_offset = 0 ;
int spaces = 0 ;
2023-07-13 13:32:44 +03:00
unsigned i , i_off = 0 ;
lvresize: add new options and defaults for fs handling
The new option "--fs String" for lvresize/lvreduce/lvextend
controls the handling of file systems before/after resizing
the LV. --resizefs is the same as --fs resize.
The new option "--fsmode String" can be used to control
mounting and unmounting of the fs during resizing.
Possible --fs values:
checksize
Only applies to reducing size; does nothing for extend.
Check the fs size and reduce the LV if the fs is not using
the affected space, i.e. the fs does not need to be shrunk.
Fail the command without reducing the fs or LV if the fs is
using the affected space.
resize
Resize the fs using the fs-specific resize command.
This may include mounting, unmounting, or running fsck.
See --fsmode to control mounting behavior, and --nofsck to
disable fsck.
resize_fsadm
Use the old method of calling fsadm to handle the fs
(deprecated.) Warning: this option does not prevent lvreduce
from destroying file systems that are unmounted (or mounted
if prompts are skipped.)
ignore
Resize the LV without checking for or handling a file system.
Warning: using ignore when reducing the LV size may destroy the
file system.
Possible --fsmode values:
manage
Mount or unmount the fs as needed to resize the fs,
and attempt to restore the original mount state at the end.
nochange
Do not mount or unmount the fs. If mounting or unmounting
is required to resize the fs, then do not resize the fs or
the LV and fail the command.
offline
Unmount the fs if it is mounted, and resize the fs while it
is unmounted. If mounting is required to resize the fs,
then do not resize the fs or the LV and fail the command.
Notes on lvreduce:
When no --fs or --resizefs option is specified:
. lvextend default behavior is fs ignore.
. lvreduce default behavior is fs checksize
(includes activating the LV.)
With the exception of --fs resize_fsadm|ignore, lvreduce requires
the recent libblkid fields FSLASTBLOCK and FSBLOCKSIZE.
FSLASTBLOCK*FSBLOCKSIZE is the last byte used by the fs on the LV,
which determines if reducing the fs is necessary.
2022-06-14 23:20:21 +03:00
2023-11-06 16:39:30 +03:00
if ( ! ( dmt = _setup_task_run ( DM_DEVICE_TABLE , NULL , NULL , NULL , NULL ,
MAJOR ( crypt_devt ) , MINOR ( crypt_devt ) , 0 , 0 , 0 ) ) )
lvresize: add new options and defaults for fs handling
The new option "--fs String" for lvresize/lvreduce/lvextend
controls the handling of file systems before/after resizing
the LV. --resizefs is the same as --fs resize.
The new option "--fsmode String" can be used to control
mounting and unmounting of the fs during resizing.
Possible --fs values:
checksize
Only applies to reducing size; does nothing for extend.
Check the fs size and reduce the LV if the fs is not using
the affected space, i.e. the fs does not need to be shrunk.
Fail the command without reducing the fs or LV if the fs is
using the affected space.
resize
Resize the fs using the fs-specific resize command.
This may include mounting, unmounting, or running fsck.
See --fsmode to control mounting behavior, and --nofsck to
disable fsck.
resize_fsadm
Use the old method of calling fsadm to handle the fs
(deprecated.) Warning: this option does not prevent lvreduce
from destroying file systems that are unmounted (or mounted
if prompts are skipped.)
ignore
Resize the LV without checking for or handling a file system.
Warning: using ignore when reducing the LV size may destroy the
file system.
Possible --fsmode values:
manage
Mount or unmount the fs as needed to resize the fs,
and attempt to restore the original mount state at the end.
nochange
Do not mount or unmount the fs. If mounting or unmounting
is required to resize the fs, then do not resize the fs or
the LV and fail the command.
offline
Unmount the fs if it is mounted, and resize the fs while it
is unmounted. If mounting is required to resize the fs,
then do not resize the fs or the LV and fail the command.
Notes on lvreduce:
When no --fs or --resizefs option is specified:
. lvextend default behavior is fs ignore.
. lvreduce default behavior is fs checksize
(includes activating the LV.)
With the exception of --fs resize_fsadm|ignore, lvreduce requires
the recent libblkid fields FSLASTBLOCK and FSBLOCKSIZE.
FSLASTBLOCK*FSBLOCKSIZE is the last byte used by the fs on the LV,
which determines if reducing the fs is necessary.
2022-06-14 23:20:21 +03:00
return_0 ;
next = dm_get_next_target ( dmt , next , & start , & length , & target_type , & params ) ;
if ( ! target_type | | ! params | | strcmp ( target_type , " crypt " ) ) {
dm_task_destroy ( dmt ) ;
return_0 ;
}
/*
* get offset from params string :
* < cipher > < key > < iv_offset > < device > < offset > [ < # opt_params > < opt_params > ]
* < offset > is reported in 512 byte sectors .
*/
2024-05-24 01:34:00 +03:00
for ( i = 0 ; params [ i ] ; i + + ) {
lvresize: add new options and defaults for fs handling
The new option "--fs String" for lvresize/lvreduce/lvextend
controls the handling of file systems before/after resizing
the LV. --resizefs is the same as --fs resize.
The new option "--fsmode String" can be used to control
mounting and unmounting of the fs during resizing.
Possible --fs values:
checksize
Only applies to reducing size; does nothing for extend.
Check the fs size and reduce the LV if the fs is not using
the affected space, i.e. the fs does not need to be shrunk.
Fail the command without reducing the fs or LV if the fs is
using the affected space.
resize
Resize the fs using the fs-specific resize command.
This may include mounting, unmounting, or running fsck.
See --fsmode to control mounting behavior, and --nofsck to
disable fsck.
resize_fsadm
Use the old method of calling fsadm to handle the fs
(deprecated.) Warning: this option does not prevent lvreduce
from destroying file systems that are unmounted (or mounted
if prompts are skipped.)
ignore
Resize the LV without checking for or handling a file system.
Warning: using ignore when reducing the LV size may destroy the
file system.
Possible --fsmode values:
manage
Mount or unmount the fs as needed to resize the fs,
and attempt to restore the original mount state at the end.
nochange
Do not mount or unmount the fs. If mounting or unmounting
is required to resize the fs, then do not resize the fs or
the LV and fail the command.
offline
Unmount the fs if it is mounted, and resize the fs while it
is unmounted. If mounting is required to resize the fs,
then do not resize the fs or the LV and fail the command.
Notes on lvreduce:
When no --fs or --resizefs option is specified:
. lvextend default behavior is fs ignore.
. lvreduce default behavior is fs checksize
(includes activating the LV.)
With the exception of --fs resize_fsadm|ignore, lvreduce requires
the recent libblkid fields FSLASTBLOCK and FSBLOCKSIZE.
FSLASTBLOCK*FSBLOCKSIZE is the last byte used by the fs on the LV,
which determines if reducing the fs is necessary.
2022-06-14 23:20:21 +03:00
if ( params [ i ] = = ' ' ) {
spaces + + ;
if ( spaces = = 4 )
copy_offset = 1 ;
if ( spaces = = 5 )
break ;
continue ;
}
if ( ! copy_offset )
continue ;
offset_str [ i_off + + ] = params [ i ] ;
if ( i_off = = sizeof ( offset_str ) ) {
offset_str [ 0 ] = ' \0 ' ;
break ;
}
}
dm_task_destroy ( dmt ) ;
if ( ! offset_str [ 0 ] )
return_0 ;
* offset_bytes = ( ( uint32_t ) strtoul ( offset_str , NULL , 0 ) * 512 ) ;
return 1 ;
}