2018-05-14 14:16:43 +03:00
/*
* Copyright ( C ) 2001 - 2004 Sistina Software , Inc . All rights reserved .
* Copyright ( C ) 2004 - 2011 Red Hat , Inc . All rights reserved .
*
* This file is part of the device - mapper userspace tools .
*
* This copyrighted material is made available to anyone wishing to use ,
* modify , copy , or redistribute it subject to the terms and conditions
* of the GNU Lesser General Public License v .2 .1 .
*
* You should have received a copy of the GNU Lesser General Public License
* along with this program ; if not , write to the Free Software Foundation ,
* Inc . , 51 Franklin Street , Fifth Floor , Boston , MA 02110 - 1301 USA
*/
# ifdef VALGRIND_POOL
# include "memcheck.h"
# endif
2018-06-08 15:40:53 +03:00
# include "base/memory/zalloc.h"
2018-06-04 15:22:14 +03:00
# include "device_mapper/misc/dmlib.h"
2018-05-14 14:16:43 +03:00
# include <stddef.h> /* For musl libc */
# include <malloc.h>
struct chunk {
char * begin , * end ;
struct chunk * prev ;
} __attribute__ ( ( aligned ( 8 ) ) ) ;
struct dm_pool {
struct dm_list list ;
struct chunk * chunk , * spare_chunk ; /* spare_chunk is a one entry free
list to stop ' bobbling ' */
const char * name ;
size_t chunk_size ;
size_t object_len ;
unsigned object_alignment ;
int locked ;
long crc ;
} ;
static void _align_chunk ( struct chunk * c , unsigned alignment ) ;
static struct chunk * _new_chunk ( struct dm_pool * p , size_t s ) ;
static void _free_chunk ( struct chunk * c ) ;
/* by default things come out aligned for doubles */
# define DEFAULT_ALIGNMENT __alignof__ (double)
struct dm_pool * dm_pool_create ( const char * name , size_t chunk_hint )
{
size_t new_size = 1024 ;
2018-06-08 15:40:53 +03:00
struct dm_pool * p = zalloc ( sizeof ( * p ) ) ;
2018-05-14 14:16:43 +03:00
if ( ! p ) {
log_error ( " Couldn't create memory pool %s (size % "
PRIsize_t " ) " , name , sizeof ( * p ) ) ;
return 0 ;
}
p - > name = name ;
/* round chunk_hint up to the next power of 2 */
p - > chunk_size = chunk_hint + sizeof ( struct chunk ) ;
while ( new_size < p - > chunk_size )
new_size < < = 1 ;
p - > chunk_size = new_size ;
pthread_mutex_lock ( & _dm_pools_mutex ) ;
dm_list_add ( & _dm_pools , & p - > list ) ;
pthread_mutex_unlock ( & _dm_pools_mutex ) ;
return p ;
}
void dm_pool_destroy ( struct dm_pool * p )
{
struct chunk * c , * pr ;
_free_chunk ( p - > spare_chunk ) ;
c = p - > chunk ;
while ( c ) {
pr = c - > prev ;
_free_chunk ( c ) ;
c = pr ;
}
pthread_mutex_lock ( & _dm_pools_mutex ) ;
dm_list_del ( & p - > list ) ;
pthread_mutex_unlock ( & _dm_pools_mutex ) ;
2018-06-08 15:40:53 +03:00
free ( p ) ;
2018-05-14 14:16:43 +03:00
}
void * dm_pool_alloc ( struct dm_pool * p , size_t s )
{
return dm_pool_alloc_aligned ( p , s , DEFAULT_ALIGNMENT ) ;
}
void * dm_pool_alloc_aligned ( struct dm_pool * p , size_t s , unsigned alignment )
{
struct chunk * c = p - > chunk ;
void * r ;
/* realign begin */
if ( c )
_align_chunk ( c , alignment ) ;
/* have we got room ? */
if ( ! c | | ( c - > begin > c - > end ) | | ( ( c - > end - c - > begin ) < ( int ) s ) ) {
/* allocate new chunk */
size_t needed = s + alignment + sizeof ( struct chunk ) ;
c = _new_chunk ( p , ( needed > p - > chunk_size ) ?
needed : p - > chunk_size ) ;
if ( ! c )
return_NULL ;
_align_chunk ( c , alignment ) ;
}
r = c - > begin ;
c - > begin + = s ;
# ifdef VALGRIND_POOL
VALGRIND_MAKE_MEM_UNDEFINED ( r , s ) ;
# endif
return r ;
}
void dm_pool_empty ( struct dm_pool * p )
{
struct chunk * c ;
for ( c = p - > chunk ; c & & c - > prev ; c = c - > prev )
;
if ( c )
dm_pool_free ( p , ( char * ) ( c + 1 ) ) ;
}
void dm_pool_free ( struct dm_pool * p , void * ptr )
{
struct chunk * c = p - > chunk ;
while ( c ) {
if ( ( ( char * ) c < ( char * ) ptr ) & &
( ( char * ) c - > end > ( char * ) ptr ) ) {
c - > begin = ptr ;
# ifdef VALGRIND_POOL
VALGRIND_MAKE_MEM_NOACCESS ( c - > begin , c - > end - c - > begin ) ;
# endif
break ;
}
if ( p - > spare_chunk )
_free_chunk ( p - > spare_chunk ) ;
c - > begin = ( char * ) ( c + 1 ) ;
# ifdef VALGRIND_POOL
VALGRIND_MAKE_MEM_NOACCESS ( c - > begin , c - > end - c - > begin ) ;
# endif
p - > spare_chunk = c ;
c = c - > prev ;
}
if ( ! c )
log_error ( INTERNAL_ERROR " pool_free asked to free pointer "
" not in pool " ) ;
else
p - > chunk = c ;
}
int dm_pool_begin_object ( struct dm_pool * p , size_t hint )
{
struct chunk * c = p - > chunk ;
const size_t align = DEFAULT_ALIGNMENT ;
p - > object_len = 0 ;
p - > object_alignment = align ;
if ( c )
_align_chunk ( c , align ) ;
if ( ! c | | ( c - > begin > c - > end ) | | ( ( c - > end - c - > begin ) < ( int ) hint ) ) {
/* allocate a new chunk */
c = _new_chunk ( p ,
hint > ( p - > chunk_size - sizeof ( struct chunk ) ) ?
hint + sizeof ( struct chunk ) + align :
p - > chunk_size ) ;
if ( ! c )
return 0 ;
_align_chunk ( c , align ) ;
}
return 1 ;
}
int dm_pool_grow_object ( struct dm_pool * p , const void * extra , size_t delta )
{
struct chunk * c = p - > chunk , * nc ;
if ( ! delta )
delta = strlen ( extra ) ;
if ( ( c - > end - ( c - > begin + p - > object_len ) ) < ( int ) delta ) {
/* move into a new chunk */
if ( p - > object_len + delta > ( p - > chunk_size / 2 ) )
nc = _new_chunk ( p , ( p - > object_len + delta ) * 2 ) ;
else
nc = _new_chunk ( p , p - > chunk_size ) ;
if ( ! nc )
return 0 ;
_align_chunk ( p - > chunk , p - > object_alignment ) ;
# ifdef VALGRIND_POOL
VALGRIND_MAKE_MEM_UNDEFINED ( p - > chunk - > begin , p - > object_len ) ;
# endif
memcpy ( p - > chunk - > begin , c - > begin , p - > object_len ) ;
# ifdef VALGRIND_POOL
VALGRIND_MAKE_MEM_NOACCESS ( c - > begin , p - > object_len ) ;
# endif
c = p - > chunk ;
}
# ifdef VALGRIND_POOL
VALGRIND_MAKE_MEM_UNDEFINED ( p - > chunk - > begin + p - > object_len , delta ) ;
# endif
memcpy ( c - > begin + p - > object_len , extra , delta ) ;
p - > object_len + = delta ;
return 1 ;
}
void * dm_pool_end_object ( struct dm_pool * p )
{
struct chunk * c = p - > chunk ;
void * r = c - > begin ;
c - > begin + = p - > object_len ;
p - > object_len = 0u ;
p - > object_alignment = DEFAULT_ALIGNMENT ;
return r ;
}
void dm_pool_abandon_object ( struct dm_pool * p )
{
# ifdef VALGRIND_POOL
VALGRIND_MAKE_MEM_NOACCESS ( p - > chunk , p - > object_len ) ;
# endif
p - > object_len = 0 ;
p - > object_alignment = DEFAULT_ALIGNMENT ;
}
static void _align_chunk ( struct chunk * c , unsigned alignment )
{
c - > begin + = alignment - ( ( unsigned long ) c - > begin & ( alignment - 1 ) ) ;
}
static struct chunk * _new_chunk ( struct dm_pool * p , size_t s )
{
struct chunk * c ;
if ( p - > spare_chunk & &
( ( p - > spare_chunk - > end - p - > spare_chunk - > begin ) > = ( ptrdiff_t ) s ) ) {
/* reuse old chunk */
c = p - > spare_chunk ;
p - > spare_chunk = 0 ;
} else {
# ifdef DEBUG_ENFORCE_POOL_LOCKING
if ( ! _pagesize ) {
_pagesize = getpagesize ( ) ; /* lvm_pagesize(); */
_pagesize_mask = _pagesize - 1 ;
}
/*
* Allocate page aligned size so malloc could work .
* Otherwise page fault would happen from pool unrelated
* memory writes of internal malloc pointers .
*/
# define aligned_malloc(s) (posix_memalign((void**)&c, _pagesize, \
ALIGN_ON_PAGE ( s ) ) = = 0 )
# else
2018-06-08 15:40:53 +03:00
# define aligned_malloc(s) (c = malloc(s))
2018-05-14 14:16:43 +03:00
# endif /* DEBUG_ENFORCE_POOL_LOCKING */
if ( ! aligned_malloc ( s ) ) {
# undef aligned_malloc
log_error ( " Out of memory. Requested % " PRIsize_t
" bytes. " , s ) ;
return NULL ;
}
c - > begin = ( char * ) ( c + 1 ) ;
c - > end = ( char * ) c + s ;
# ifdef VALGRIND_POOL
VALGRIND_MAKE_MEM_NOACCESS ( c - > begin , c - > end - c - > begin ) ;
# endif
}
c - > prev = p - > chunk ;
p - > chunk = c ;
return c ;
}
static void _free_chunk ( struct chunk * c )
{
# ifdef VALGRIND_POOL
# ifdef DEBUG_MEM
if ( c )
VALGRIND_MAKE_MEM_UNDEFINED ( c + 1 , c - > end - ( char * ) ( c + 1 ) ) ;
# endif
# endif
# ifdef DEBUG_ENFORCE_POOL_LOCKING
/* since DEBUG_MEM is using own memory list */
free ( c ) ; /* for posix_memalign() */
# else
2018-06-08 15:40:53 +03:00
free ( c ) ;
2018-05-14 14:16:43 +03:00
# endif
}
/**
* Calc crc / hash from pool ' s memory chunks with internal pointers
*/
static long _pool_crc ( const struct dm_pool * p )
{
long crc_hash = 0 ;
# ifndef DEBUG_ENFORCE_POOL_LOCKING
const struct chunk * c ;
const long * ptr , * end ;
for ( c = p - > chunk ; c ; c = c - > prev ) {
end = ( const long * ) ( c - > begin < c - > end ? ( long ) c - > begin & ~ 7 : ( long ) c - > end ) ;
ptr = ( const long * ) c ;
# ifdef VALGRIND_POOL
VALGRIND_MAKE_MEM_DEFINED ( ptr , ( end - ptr ) * sizeof ( * end ) ) ;
# endif
while ( ptr < end ) {
crc_hash + = * ptr + + ;
crc_hash + = ( crc_hash < < 10 ) ;
crc_hash ^ = ( crc_hash > > 6 ) ;
}
}
# endif /* DEBUG_ENFORCE_POOL_LOCKING */
return crc_hash ;
}
static int _pool_protect ( struct dm_pool * p , int prot )
{
# ifdef DEBUG_ENFORCE_POOL_LOCKING
struct chunk * c ;
for ( c = p - > chunk ; c ; c = c - > prev ) {
if ( mprotect ( c , ( size_t ) ( ( c - > end - ( char * ) c ) - 1 ) , prot ) ! = 0 ) {
log_sys_error ( " mprotect " , " " ) ;
return 0 ;
}
}
# endif
return 1 ;
}