2001-12-11 14:42:30 +03:00
/*
2004-03-30 23:35:44 +04:00
* Copyright ( C ) 2002 - 2004 Sistina Software , Inc . All rights reserved .
2018-04-20 18:43:50 +03:00
* Copyright ( C ) 2004 - 2007 Red Hat , Inc . All rights reserved .
2001-12-11 14:42:30 +03:00
*
2004-03-30 23:35:44 +04:00
* This file is part of LVM2 .
*
* This copyrighted material is made available to anyone wishing to use ,
* modify , copy , or redistribute it subject to the terms and conditions
2007-08-21 00:55:30 +04:00
* of the GNU Lesser General Public License v .2 .1 .
2004-03-30 23:35:44 +04:00
*
2007-08-21 00:55:30 +04:00
* You should have received a copy of the GNU Lesser General Public License
2004-03-30 23:35:44 +04:00
* along with this program ; if not , write to the Free Software Foundation ,
2016-01-21 13:49:46 +03:00
* Inc . , 51 Franklin Street , Fifth Floor , Boston , MA 02110 - 1301 USA
2001-12-11 14:42:30 +03:00
*/
2002-11-18 17:04:08 +03:00
# include "lib.h"
2001-12-11 14:42:30 +03:00
# include "label.h"
2002-11-18 17:04:08 +03:00
# include "crc.h"
# include "xlate.h"
2003-07-05 02:34:56 +04:00
# include "lvmcache.h"
2018-02-07 00:18:11 +03:00
# include "bcache.h"
# include "toolcontext.h"
2002-11-18 17:04:08 +03:00
# include <sys/stat.h>
# include <fcntl.h>
# include <unistd.h>
2018-02-07 00:18:11 +03:00
# include <sys/time.h>
2002-11-18 17:04:08 +03:00
/* FIXME Allow for larger labels? Restricted to single sector currently */
2001-12-11 14:42:30 +03:00
2002-01-10 18:01:58 +03:00
/*
* Internal labeller struct .
*/
struct labeller_i {
2008-11-04 01:14:30 +03:00
struct dm_list list ;
2001-12-11 14:42:30 +03:00
2002-01-10 18:01:58 +03:00
struct labeller * l ;
char name [ 0 ] ;
} ;
2001-12-13 11:40:47 +03:00
2008-11-04 01:14:30 +03:00
static struct dm_list _labellers ;
2001-12-11 14:42:30 +03:00
2002-01-10 18:01:58 +03:00
static struct labeller_i * _alloc_li ( const char * name , struct labeller * l )
2001-12-11 14:42:30 +03:00
{
2002-01-10 18:01:58 +03:00
struct labeller_i * li ;
size_t len ;
2001-12-11 14:42:30 +03:00
2002-01-10 18:01:58 +03:00
len = sizeof ( * li ) + strlen ( name ) + 1 ;
2001-12-11 14:42:30 +03:00
2005-10-17 03:03:59 +04:00
if ( ! ( li = dm_malloc ( len ) ) ) {
2002-01-16 03:01:36 +03:00
log_error ( " Couldn't allocate memory for labeller list object. " ) ;
2002-01-10 18:01:58 +03:00
return NULL ;
}
li - > l = l ;
strcpy ( li - > name , name ) ;
return li ;
2001-12-12 12:05:44 +03:00
}
2001-12-11 14:42:30 +03:00
2002-01-10 18:01:58 +03:00
int label_init ( void )
{
2008-11-04 01:14:30 +03:00
dm_list_init ( & _labellers ) ;
2002-01-10 19:48:28 +03:00
return 1 ;
2001-12-11 14:42:30 +03:00
}
2002-01-10 18:01:58 +03:00
void label_exit ( void )
2001-12-14 16:15:15 +03:00
{
2012-02-08 15:12:18 +04:00
struct labeller_i * li , * tli ;
2002-01-10 18:01:58 +03:00
2012-02-08 15:12:18 +04:00
dm_list_iterate_items_safe ( li , tli , & _labellers ) {
dm_list_del ( & li - > list ) ;
2002-11-18 17:04:08 +03:00
li - > l - > ops - > destroy ( li - > l ) ;
2012-02-08 15:12:18 +04:00
dm_free ( li ) ;
2002-01-10 18:01:58 +03:00
}
2004-05-05 01:25:57 +04:00
2008-11-04 01:14:30 +03:00
dm_list_init ( & _labellers ) ;
2001-12-14 16:15:15 +03:00
}
2013-07-29 17:58:18 +04:00
int label_register_handler ( struct labeller * handler )
2001-12-11 14:42:30 +03:00
{
2002-01-10 18:01:58 +03:00
struct labeller_i * li ;
2013-07-29 17:58:18 +04:00
if ( ! ( li = _alloc_li ( handler - > fmt - > name , handler ) ) )
2008-01-30 16:19:47 +03:00
return_0 ;
2001-12-11 14:42:30 +03:00
2008-11-04 01:14:30 +03:00
dm_list_add ( & _labellers , & li - > list ) ;
2002-01-10 18:01:58 +03:00
return 1 ;
2001-12-11 14:42:30 +03:00
}
2002-01-10 18:01:58 +03:00
struct labeller * label_get_handler ( const char * name )
2001-12-11 14:42:30 +03:00
{
2002-01-10 18:01:58 +03:00
struct labeller_i * li ;
2001-12-11 14:42:30 +03:00
2008-11-04 01:14:30 +03:00
dm_list_iterate_items ( li , & _labellers )
2002-01-10 18:01:58 +03:00
if ( ! strcmp ( li - > name , name ) )
return li - > l ;
2001-12-11 14:42:30 +03:00
2002-01-10 18:01:58 +03:00
return NULL ;
}
2001-12-11 14:42:30 +03:00
2002-11-18 17:04:08 +03:00
/* FIXME Also wipe associated metadata area headers? */
2002-01-11 13:43:32 +03:00
int label_remove ( struct device * dev )
2001-12-11 14:42:30 +03:00
{
2018-04-20 18:43:50 +03:00
char buf [ LABEL_SIZE ] __attribute__ ( ( aligned ( 8 ) ) ) ;
char readbuf [ LABEL_SCAN_SIZE ] __attribute__ ( ( aligned ( 8 ) ) ) ;
2002-11-18 17:04:08 +03:00
int r = 1 ;
uint64_t sector ;
int wipe ;
struct labeller_i * li ;
struct label_header * lh ;
2017-08-11 22:32:04 +03:00
struct lvmcache_info * info ;
2002-11-18 17:04:08 +03:00
2018-04-20 18:43:50 +03:00
memset ( buf , 0 , LABEL_SIZE ) ;
2002-01-10 18:01:58 +03:00
2002-11-18 17:04:08 +03:00
log_very_verbose ( " Scanning for labels to wipe from %s " , dev_name ( dev ) ) ;
2008-01-30 16:19:47 +03:00
if ( ! dev_open ( dev ) )
return_0 ;
2001-12-11 14:42:30 +03:00
2003-07-05 02:34:56 +04:00
/*
* We flush the device just in case someone is stupid
* enough to be trying to import an open pv into lvm .
*/
dev_flush ( dev ) ;
2018-02-07 00:18:11 +03:00
label_scan_invalidate ( dev ) ;
2018-04-20 18:43:50 +03:00
if ( ! dev_read ( dev , UINT64_C ( 0 ) , LABEL_SCAN_SIZE , DEV_IO_LABEL , readbuf ) ) {
2013-01-08 02:30:29 +04:00
log_debug_devs ( " %s: Failed to read label area " , dev_name ( dev ) ) ;
2002-11-18 17:04:08 +03:00
goto out ;
}
/* Scan first few sectors for anything looking like a label */
for ( sector = 0 ; sector < LABEL_SCAN_SECTORS ;
sector + = LABEL_SIZE > > SECTOR_SHIFT ) {
2018-04-20 18:43:50 +03:00
lh = ( struct label_header * ) ( readbuf +
( sector < < SECTOR_SHIFT ) ) ;
2002-11-18 17:04:08 +03:00
wipe = 0 ;
2006-05-10 01:23:51 +04:00
if ( ! strncmp ( ( char * ) lh - > id , LABEL_ID , sizeof ( lh - > id ) ) ) {
2002-11-18 17:04:08 +03:00
if ( xlate64 ( lh - > sector_xl ) = = sector )
wipe = 1 ;
} else {
2008-11-04 01:14:30 +03:00
dm_list_iterate_items ( li , & _labellers ) {
2002-11-18 17:04:08 +03:00
if ( li - > l - > ops - > can_handle ( li - > l , ( char * ) lh ,
sector ) ) {
wipe = 1 ;
break ;
}
}
}
if ( wipe ) {
2016-11-01 13:02:01 +03:00
log_very_verbose ( " %s: Wiping label at sector % " PRIu64 ,
dev_name ( dev ) , sector ) ;
2018-04-20 18:43:50 +03:00
if ( dev_write ( dev , sector < < SECTOR_SHIFT , LABEL_SIZE , DEV_IO_LABEL ,
buf ) ) {
2017-08-11 22:32:04 +03:00
/* Also remove the PV record from cache. */
info = lvmcache_info_from_pvid ( dev - > pvid , dev , 0 ) ;
if ( info )
lvmcache_del ( info ) ;
} else {
2002-11-18 17:04:08 +03:00
log_error ( " Failed to remove label from %s at "
" sector % " PRIu64 , dev_name ( dev ) ,
sector ) ;
r = 0 ;
}
}
}
out :
if ( ! dev_close ( dev ) )
stack ;
return r ;
2001-12-11 14:42:30 +03:00
}
2002-11-18 17:04:08 +03:00
/* Caller may need to use label_get_handler to create label struct! */
int label_write ( struct device * dev , struct label * label )
{
2010-07-09 19:34:40 +04:00
char buf [ LABEL_SIZE ] __attribute__ ( ( aligned ( 8 ) ) ) ;
2002-11-18 17:04:08 +03:00
struct label_header * lh = ( struct label_header * ) buf ;
int r = 1 ;
2003-08-27 01:12:06 +04:00
if ( ! label - > labeller - > ops - > write ) {
2009-07-16 00:02:46 +04:00
log_error ( " Label handler does not support label writes " ) ;
2003-08-27 01:12:06 +04:00
return 0 ;
}
2002-11-18 17:04:08 +03:00
if ( ( LABEL_SIZE + ( label - > sector < < SECTOR_SHIFT ) ) > LABEL_SCAN_SIZE ) {
log_error ( " Label sector % " PRIu64 " beyond range (%ld) " ,
label - > sector , LABEL_SCAN_SECTORS ) ;
return 0 ;
}
2018-02-07 00:18:11 +03:00
label_scan_invalidate ( dev ) ;
2002-11-18 17:04:08 +03:00
memset ( buf , 0 , LABEL_SIZE ) ;
2006-05-10 01:23:51 +04:00
strncpy ( ( char * ) lh - > id , LABEL_ID , sizeof ( lh - > id ) ) ;
2002-11-18 17:04:08 +03:00
lh - > sector_xl = xlate64 ( label - > sector ) ;
lh - > offset_xl = xlate32 ( sizeof ( * lh ) ) ;
2008-01-30 16:19:47 +03:00
if ( ! ( label - > labeller - > ops - > write ) ( label , buf ) )
return_0 ;
2002-11-18 17:04:08 +03:00
2010-09-27 23:09:34 +04:00
lh - > crc_xl = xlate32 ( calc_crc ( INITIAL_CRC , ( uint8_t * ) & lh - > offset_xl , LABEL_SIZE -
( ( uint8_t * ) & lh - > offset_xl - ( uint8_t * ) lh ) ) ) ;
2002-11-18 17:04:08 +03:00
2008-01-30 16:19:47 +03:00
if ( ! dev_open ( dev ) )
return_0 ;
2002-11-18 17:04:08 +03:00
2016-11-01 13:02:01 +03:00
log_very_verbose ( " %s: Writing label to sector % " PRIu64 " with stored offset % "
PRIu32 " . " , dev_name ( dev ) , label - > sector ,
xlate32 ( lh - > offset_xl ) ) ;
2017-12-05 02:18:56 +03:00
if ( ! dev_write ( dev , label - > sector < < SECTOR_SHIFT , LABEL_SIZE , DEV_IO_LABEL , buf ) ) {
2013-01-08 02:30:29 +04:00
log_debug_devs ( " Failed to write label to %s " , dev_name ( dev ) ) ;
2002-11-18 17:04:08 +03:00
r = 0 ;
}
2003-07-05 02:34:56 +04:00
if ( ! dev_close ( dev ) )
2002-11-18 17:04:08 +03:00
stack ;
return r ;
2001-12-11 14:42:30 +03:00
}
2002-11-18 17:04:08 +03:00
void label_destroy ( struct label * label )
2001-12-11 19:49:40 +03:00
{
2002-11-18 17:04:08 +03:00
label - > labeller - > ops - > destroy_label ( label - > labeller , label ) ;
2005-10-17 03:03:59 +04:00
dm_free ( label ) ;
2002-11-18 17:04:08 +03:00
}
struct label * label_create ( struct labeller * labeller )
{
struct label * label ;
2010-10-01 01:06:50 +04:00
if ( ! ( label = dm_zalloc ( sizeof ( * label ) ) ) ) {
2002-11-18 17:04:08 +03:00
log_error ( " label allocaction failed " ) ;
return NULL ;
}
label - > labeller = labeller ;
labeller - > ops - > initialise_label ( labeller , label ) ;
return label ;
2001-12-11 19:49:40 +03:00
}
2018-02-07 00:18:11 +03:00
/* global variable for accessing the bcache populated by label scan */
struct bcache * scan_bcache ;
# define BCACHE_BLOCK_SIZE_IN_SECTORS 2048 /* 1MB */
static bool _in_bcache ( struct device * dev )
{
return ( dev - > flags & DEV_IN_BCACHE ) ? true : false ;
}
static struct labeller * _find_lvm_header ( struct device * dev ,
char * scan_buf ,
char * label_buf ,
uint64_t * label_sector ,
uint64_t scan_sector )
{
struct labeller_i * li ;
struct labeller * labeller_ret = NULL ;
struct label_header * lh ;
uint64_t sector ;
int found = 0 ;
/*
* Find which sector in scan_buf starts with a valid label ,
* and copy it into label_buf .
*/
for ( sector = 0 ; sector < LABEL_SCAN_SECTORS ;
sector + = LABEL_SIZE > > SECTOR_SHIFT ) {
lh = ( struct label_header * ) ( scan_buf + ( sector < < SECTOR_SHIFT ) ) ;
if ( ! strncmp ( ( char * ) lh - > id , LABEL_ID , sizeof ( lh - > id ) ) ) {
if ( found ) {
log_error ( " Ignoring additional label on %s at sector %llu " ,
dev_name ( dev ) , ( unsigned long long ) ( sector + scan_sector ) ) ;
}
if ( xlate64 ( lh - > sector_xl ) ! = sector + scan_sector ) {
log_very_verbose ( " %s: Label for sector %llu found at sector %llu - ignoring. " ,
dev_name ( dev ) ,
( unsigned long long ) xlate64 ( lh - > sector_xl ) ,
( unsigned long long ) ( sector + scan_sector ) ) ;
continue ;
}
if ( calc_crc ( INITIAL_CRC , ( uint8_t * ) & lh - > offset_xl , LABEL_SIZE -
( ( uint8_t * ) & lh - > offset_xl - ( uint8_t * ) lh ) ) ! =
xlate32 ( lh - > crc_xl ) ) {
log_very_verbose ( " Label checksum incorrect on %s - ignoring " , dev_name ( dev ) ) ;
continue ;
}
if ( found )
continue ;
}
dm_list_iterate_items ( li , & _labellers ) {
if ( li - > l - > ops - > can_handle ( li - > l , ( char * ) lh , sector + scan_sector ) ) {
log_very_verbose ( " %s: %s label detected at sector %llu " ,
dev_name ( dev ) , li - > name ,
( unsigned long long ) ( sector + scan_sector ) ) ;
if ( found ) {
log_error ( " Ignoring additional label on %s at sector %llu " ,
dev_name ( dev ) ,
( unsigned long long ) ( sector + scan_sector ) ) ;
continue ;
}
labeller_ret = li - > l ;
found = 1 ;
memcpy ( label_buf , lh , LABEL_SIZE ) ;
if ( label_sector )
* label_sector = sector + scan_sector ;
break ;
}
}
}
return labeller_ret ;
}
/*
* Process / parse the headers from the data read from a device .
* Populates lvmcache with device / mda locations / vgname
* so that vg_read ( vgname ) will know which devices / locations
* to read metadata from .
*
* If during processing , headers / metadata are found to be needed
* beyond the range of the scanned block , then additional reads
* are performed in the processing functions to get that data .
*/
static int _process_block ( struct device * dev , struct block * bb , int * is_lvm_device )
{
char label_buf [ LABEL_SIZE ] __attribute__ ( ( aligned ( 8 ) ) ) ;
struct label * label = NULL ;
struct labeller * labeller ;
struct lvmcache_info * info ;
uint64_t sector ;
int ret = 0 ;
/*
* Finds the data sector containing the label and copies into label_buf .
* label_buf : struct label_header + struct pv_header + struct pv_header_extension
*
* FIXME : we don ' t need to copy one sector from bb - > data into label_buf ,
* we can just point label_buf at one sector in ld - > buf .
*/
if ( ! ( labeller = _find_lvm_header ( dev , bb - > data , label_buf , & sector , 0 ) ) ) {
/*
* Non - PVs exit here
*
* FIXME : check for PVs with errors that also exit here !
* i . e . this code cannot distinguish between a non - lvm
* device an an lvm device with errors .
*/
log_very_verbose ( " %s: No lvm label detected " , dev_name ( dev ) ) ;
if ( ( info = lvmcache_info_from_pvid ( dev - > pvid , dev , 0 ) ) ) {
/* FIXME: if this case is actually happening, fix it. */
log_warn ( " Device %s has no label, removing PV info from lvmcache. " , dev_name ( dev ) ) ;
lvmcache_del ( info ) ;
}
* is_lvm_device = 0 ;
goto_out ;
}
* is_lvm_device = 1 ;
/*
* This is the point where the scanning code dives into the rest of
* lvm . ops - > read ( ) is usually _text_read ( ) which reads the pv_header ,
* mda locations , mda contents . As these bits of data are read , they
* are saved into lvmcache as info / vginfo structs .
*/
if ( ( ret = ( labeller - > ops - > read ) ( labeller , dev , label_buf , & label ) ) & & label ) {
label - > dev = dev ;
label - > sector = sector ;
} else {
/* FIXME: handle errors */
}
out :
return ret ;
}
/*
* Read or reread label / metadata from selected devs .
*
* Reads and looks at label_header , pv_header , pv_header_extension ,
* mda_header , raw_locns , vg metadata from each device .
*
* Effect is populating lvmcache with latest info / vginfo ( PV / VG ) data
* from the devs . If a scanned device does not have a label_header ,
* its info is removed from lvmcache .
*/
static int _scan_list ( struct dm_list * devs )
{
struct dm_list wait_devs ;
struct dm_list done_devs ;
struct device_list * devl , * devl2 ;
struct block * bb ;
int scan_failed_count = 0 ;
int scan_lvm_count = 0 ;
int rem_prefetches ;
int scan_failed ;
int is_lvm_device ;
dm_list_init ( & wait_devs ) ;
dm_list_init ( & done_devs ) ;
log_debug_devs ( " Scanning %d devices. " , dm_list_size ( devs ) ) ;
scan_more :
rem_prefetches = bcache_max_prefetches ( scan_bcache ) ;
dm_list_iterate_items_safe ( devl , devl2 , devs ) {
/*
* If we prefetch more devs than blocks in the cache , then the
* cache will wait for earlier reads to complete , toss the
* results , and reuse those blocks before we ' ve had a chance to
* use them . So , prefetch as many as are available , wait for
* and process them , then repeat .
*/
if ( ! rem_prefetches )
break ;
/*
* The in - bcache flag corresponds with this dev_open .
* Clearing the in - bcache flag should be paired with
* a dev_close . ( This dev may already be in bcache . )
*/
if ( ! _in_bcache ( devl - > dev ) ) {
if ( ! dev_open_readonly ( devl - > dev ) ) {
log_debug_devs ( " %s: Failed to open device. " , dev_name ( devl - > dev ) ) ;
continue ;
}
}
bcache_prefetch ( scan_bcache , devl - > dev - > fd , 0 ) ;
rem_prefetches - - ;
dm_list_del ( & devl - > list ) ;
dm_list_add ( & wait_devs , & devl - > list ) ;
}
dm_list_iterate_items_safe ( devl , devl2 , & wait_devs ) {
bb = NULL ;
if ( ! bcache_get ( scan_bcache , devl - > dev - > fd , 0 , 0 , & bb ) ) {
log_debug_devs ( " %s: Failed to scan device. " , dev_name ( devl - > dev ) ) ;
scan_failed_count + + ;
scan_failed = 1 ;
} else {
log_debug_devs ( " Processing data from device %s fd %d block %p " , dev_name ( devl - > dev ) , devl - > dev - > fd , bb ) ;
_process_block ( devl - > dev , bb , & is_lvm_device ) ;
scan_lvm_count + + ;
scan_failed = 0 ;
}
if ( bb )
bcache_put ( bb ) ;
/*
* Keep the bcache block of lvm devices we have processed so
* that the vg_read phase can reuse it . If bcache failed to
* read the block , or the device does not belong to lvm , then
* drop it from bcache .
*/
if ( scan_failed | | ! is_lvm_device ) {
devl - > dev - > flags & = ~ DEV_IN_BCACHE ;
bcache_invalidate_fd ( scan_bcache , devl - > dev - > fd ) ;
dev_close ( devl - > dev ) ;
} else {
/* The device must be kept open while it's in bcache. */
devl - > dev - > flags | = DEV_IN_BCACHE ;
}
dm_list_del ( & devl - > list ) ;
dm_list_add ( & done_devs , & devl - > list ) ;
}
if ( ! dm_list_empty ( devs ) )
goto scan_more ;
/* FIXME: let the caller know if some lvm devices failed to be scanned. */
log_debug_devs ( " Scanned %d devices: %d for lvm, %d failed. " ,
dm_list_size ( & done_devs ) , scan_lvm_count , scan_failed_count ) ;
return 0 ;
}
/*
* Scan and cache lvm data from all devices on the system .
* The cache should be empty / reset before calling this .
*/
int label_scan ( struct cmd_context * cmd )
{
struct dm_list all_devs ;
struct dev_iter * iter ;
struct device_list * devl ;
struct device * dev ;
struct io_engine * ioe ;
log_debug_devs ( " Finding devices to scan " ) ;
dm_list_init ( & all_devs ) ;
/*
* Iterate through all the devices in dev - cache ( block devs that appear
* under / dev that could possibly hold a PV and are not excluded by
* filters ) . Read each to see if it ' s an lvm device , and if so
* populate lvmcache with some basic info about the device and the VG
* on it . This info will be used by the vg_read ( ) phase of the
* command .
*/
dev_cache_full_scan ( cmd - > full_filter ) ;
if ( ! ( iter = dev_iter_create ( cmd - > full_filter , 0 ) ) ) {
log_error ( " Scanning failed to get devices. " ) ;
return 0 ;
}
while ( ( dev = dev_iter_get ( iter ) ) ) {
if ( ! ( devl = dm_pool_zalloc ( cmd - > mem , sizeof ( * devl ) ) ) )
return 0 ;
devl - > dev = dev ;
dm_list_add ( & all_devs , & devl - > list ) ;
/*
* label_scan should not generally be called a second time ,
* so this will usually not be true .
*/
if ( _in_bcache ( dev ) )
bcache_invalidate_fd ( scan_bcache , dev - > fd ) ;
} ;
dev_iter_destroy ( iter ) ;
if ( ! scan_bcache ) {
/*
* 100 is arbitrary , it ' s the max number of concurrent aio ' s
* possible , i . e , the number of devices that can be read at
* once . Should this be configurable ?
*/
if ( ! ( ioe = create_async_io_engine ( 100 ) ) )
return 0 ;
/*
* Configure one cache block for each device on the system .
* We won ' t generally need to cache that many because some
* of the devs will not be lvm devices , and we don ' t need
* an entry for those . We might want to change this .
*/
if ( ! ( scan_bcache = bcache_create ( BCACHE_BLOCK_SIZE_IN_SECTORS , dm_list_size ( & all_devs ) , ioe ) ) )
return 0 ;
}
return _scan_list ( & all_devs ) ;
}
/*
* Scan and cache lvm data from the listed devices . If a device is already
* scanned and cached , this replaces the previously cached lvm data for the
* device . This is called when vg_read ( ) wants to guarantee that it is using
* the latest data from the devices in the VG ( since the scan populated bcache
* without a lock . )
*/
int label_scan_devs ( struct cmd_context * cmd , struct dm_list * devs )
{
struct device_list * devl ;
dm_list_iterate_items ( devl , devs ) {
if ( _in_bcache ( devl - > dev ) )
bcache_invalidate_fd ( scan_bcache , devl - > dev - > fd ) ;
}
return _scan_list ( devs ) ;
}
void label_scan_invalidate ( struct device * dev )
{
if ( _in_bcache ( dev ) ) {
dev - > flags & = ~ DEV_IN_BCACHE ;
bcache_invalidate_fd ( scan_bcache , dev - > fd ) ;
dev_close ( dev ) ;
}
}
/*
* Undo label_scan ( )
*
* Close devices that are open because bcache is holding blocks for them .
* Destroy the bcache .
*/
void label_scan_destroy ( struct cmd_context * cmd )
{
struct dev_iter * iter ;
struct device * dev ;
if ( ! scan_bcache )
return ;
if ( ! ( iter = dev_iter_create ( cmd - > full_filter , 0 ) ) ) {
return ;
}
while ( ( dev = dev_iter_get ( iter ) ) )
label_scan_invalidate ( dev ) ;
dev_iter_destroy ( iter ) ;
bcache_destroy ( scan_bcache ) ;
scan_bcache = NULL ;
}
/*
* Read ( or re - read ) and process ( or re - process ) the data for a device . This
* will reset ( clear and repopulate ) the bcache and lvmcache info for this
* device . There are only a couple odd places that want to reread a specific
* device , this is not a commonly used function .
*/
/* FIXME: remove unused_sector arg */
int label_read ( struct device * dev , struct label * * labelp , uint64_t unused_sector )
{
struct dm_list one_dev ;
struct device_list * devl ;
int ret ;
/* scanning is done by list, so make a single item list for this dev */
if ( ! ( devl = dm_zalloc ( sizeof ( * devl ) ) ) )
return 0 ;
devl - > dev = dev ;
dm_list_init ( & one_dev ) ;
dm_list_add ( & one_dev , & devl - > list ) ;
if ( _in_bcache ( dev ) )
bcache_invalidate_fd ( scan_bcache , dev - > fd ) ;
ret = _scan_list ( & one_dev ) ;
/*
* FIXME : this ugliness of returning a pointer to the label is
* temporary until the callers can be updated to not use this .
*/
if ( labelp ) {
struct lvmcache_info * info ;
info = lvmcache_info_from_pvid ( dev - > pvid , dev , 1 ) ;
if ( info )
* labelp = lvmcache_get_label ( info ) ;
}
return ret ;
}
/*
* Read a label from a specfic , non - zero sector . This is used in only
* one place : pvck - > pv_analyze .
*/
int label_read_sector ( struct device * dev , struct label * * labelp , uint64_t scan_sector )
{
if ( scan_sector ) {
/* TODO: not yet implemented */
/* When is this done? When does it make sense? Is it actually possible? */
return 0 ;
}
return label_read ( dev , labelp , 0 ) ;
}