2007-06-12 17:07:21 +04:00
/*
* Copyright ( C ) 2007 Oracle . All rights reserved .
*
* This program is free software ; you can redistribute it and / or
* modify it under the terms of the GNU General Public
* License v2 as published by the Free Software Foundation .
*
* This program is distributed in the hope that it will be useful ,
* but WITHOUT ANY WARRANTY ; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE . See the GNU
* General Public License for more details .
*
* You should have received a copy of the GNU General Public
* License along with this program ; if not , write to the
* Free Software Foundation , Inc . , 59 Temple Place - Suite 330 ,
* Boston , MA 021110 - 1307 , USA .
*/
2007-03-22 19:13:20 +03:00
# include <linux/fs.h>
2007-03-28 21:57:48 +04:00
# include <linux/blkdev.h>
2007-03-29 03:44:27 +04:00
# include <linux/scatterlist.h>
2007-03-30 16:47:31 +04:00
# include <linux/swap.h>
2007-04-09 18:42:37 +04:00
# include <linux/radix-tree.h>
2007-05-02 23:53:43 +04:00
# include <linux/writeback.h>
2009-01-06 05:25:51 +03:00
# include <linux/buffer_head.h>
2008-04-10 00:28:12 +04:00
# include <linux/workqueue.h>
2008-06-26 00:01:31 +04:00
# include <linux/kthread.h>
2008-11-20 18:22:27 +03:00
# include <linux/freezer.h>
2009-04-19 16:02:41 +04:00
# include <linux/crc32c.h>
include cleanup: Update gfp.h and slab.h includes to prepare for breaking implicit slab.h inclusion from percpu.h
percpu.h is included by sched.h and module.h and thus ends up being
included when building most .c files. percpu.h includes slab.h which
in turn includes gfp.h making everything defined by the two files
universally available and complicating inclusion dependencies.
percpu.h -> slab.h dependency is about to be removed. Prepare for
this change by updating users of gfp and slab facilities include those
headers directly instead of assuming availability. As this conversion
needs to touch large number of source files, the following script is
used as the basis of conversion.
http://userweb.kernel.org/~tj/misc/slabh-sweep.py
The script does the followings.
* Scan files for gfp and slab usages and update includes such that
only the necessary includes are there. ie. if only gfp is used,
gfp.h, if slab is used, slab.h.
* When the script inserts a new include, it looks at the include
blocks and try to put the new include such that its order conforms
to its surrounding. It's put in the include block which contains
core kernel includes, in the same order that the rest are ordered -
alphabetical, Christmas tree, rev-Xmas-tree or at the end if there
doesn't seem to be any matching order.
* If the script can't find a place to put a new include (mostly
because the file doesn't have fitting include block), it prints out
an error message indicating which .h file needs to be added to the
file.
The conversion was done in the following steps.
1. The initial automatic conversion of all .c files updated slightly
over 4000 files, deleting around 700 includes and adding ~480 gfp.h
and ~3000 slab.h inclusions. The script emitted errors for ~400
files.
2. Each error was manually checked. Some didn't need the inclusion,
some needed manual addition while adding it to implementation .h or
embedding .c file was more appropriate for others. This step added
inclusions to around 150 files.
3. The script was run again and the output was compared to the edits
from #2 to make sure no file was left behind.
4. Several build tests were done and a couple of problems were fixed.
e.g. lib/decompress_*.c used malloc/free() wrappers around slab
APIs requiring slab.h to be added manually.
5. The script was run on all .h files but without automatically
editing them as sprinkling gfp.h and slab.h inclusions around .h
files could easily lead to inclusion dependency hell. Most gfp.h
inclusion directives were ignored as stuff from gfp.h was usually
wildly available and often used in preprocessor macros. Each
slab.h inclusion directive was examined and added manually as
necessary.
6. percpu.h was updated not to include slab.h.
7. Build test were done on the following configurations and failures
were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my
distributed build env didn't work with gcov compiles) and a few
more options had to be turned off depending on archs to make things
build (like ipr on powerpc/64 which failed due to missing writeq).
* x86 and x86_64 UP and SMP allmodconfig and a custom test config.
* powerpc and powerpc64 SMP allmodconfig
* sparc and sparc64 SMP allmodconfig
* ia64 SMP allmodconfig
* s390 SMP allmodconfig
* alpha SMP allmodconfig
* um on x86_64 SMP allmodconfig
8. percpu.h modifications were reverted so that it could be applied as
a separate patch and serve as bisection point.
Given the fact that I had only a couple of failures from tests on step
6, I'm fairly confident about the coverage of this conversion patch.
If there is a breakage, it's likely to be something in one of the arch
headers which should be easily discoverable easily on most builds of
the specific arch.
Signed-off-by: Tejun Heo <tj@kernel.org>
Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
2010-03-24 11:04:11 +03:00
# include <linux/slab.h>
2008-11-20 18:22:27 +03:00
# include "compat.h"
2007-02-02 17:18:22 +03:00
# include "ctree.h"
# include "disk-io.h"
2007-03-16 23:20:31 +03:00
# include "transaction.h"
2007-04-09 18:42:37 +04:00
# include "btrfs_inode.h"
2008-03-24 22:01:56 +03:00
# include "volumes.h"
2007-10-16 00:15:53 +04:00
# include "print-tree.h"
2008-06-12 00:50:36 +04:00
# include "async-thread.h"
2008-06-26 00:01:30 +04:00
# include "locking.h"
2008-09-06 00:13:11 +04:00
# include "tree-log.h"
2009-04-03 17:47:43 +04:00
# include "free-space-cache.h"
2007-02-02 17:18:22 +03:00
2008-01-25 00:13:08 +03:00
static struct extent_io_ops btree_extent_io_ops ;
2008-06-12 00:50:36 +04:00
static void end_workqueue_fn ( struct btrfs_work * work ) ;
2009-09-21 23:56:00 +04:00
static void free_fs_root ( struct btrfs_root * root ) ;
2008-04-10 00:28:12 +04:00
2008-09-29 23:18:18 +04:00
/*
* end_io_wq structs are used to do processing in task context when an IO is
* complete . This is used during reads to verify checksums , and it is used
* by writes to insert metadata for new file extents after IO is complete .
*/
2008-04-10 00:28:12 +04:00
struct end_io_wq {
struct bio * bio ;
bio_end_io_t * end_io ;
void * private ;
struct btrfs_fs_info * info ;
int error ;
2008-04-10 00:28:12 +04:00
int metadata ;
2008-04-10 00:28:12 +04:00
struct list_head list ;
2008-06-12 00:50:36 +04:00
struct btrfs_work work ;
2008-04-10 00:28:12 +04:00
} ;
2007-11-08 05:08:01 +03:00
2008-09-29 23:18:18 +04:00
/*
* async submit bios are used to offload expensive checksumming
* onto the worker threads . They checksum file and metadata bios
* just before they are sent down the IO stack .
*/
2008-04-16 19:14:51 +04:00
struct async_submit_bio {
struct inode * inode ;
struct bio * bio ;
struct list_head list ;
Btrfs: Add ordered async work queues
Btrfs uses kernel threads to create async work queues for cpu intensive
operations such as checksumming and decompression. These work well,
but they make it difficult to keep IO order intact.
A single writepages call from pdflush or fsync will turn into a number
of bios, and each bio is checksummed in parallel. Once the checksum is
computed, the bio is sent down to the disk, and since we don't control
the order in which the parallel operations happen, they might go down to
the disk in almost any order.
The code deals with this somewhat by having deep work queues for a single
kernel thread, making it very likely that a single thread will process all
the bios for a single inode.
This patch introduces an explicitly ordered work queue. As work structs
are placed into the queue they are put onto the tail of a list. They have
three callbacks:
->func (cpu intensive processing here)
->ordered_func (order sensitive processing here)
->ordered_free (free the work struct, all processing is done)
The work struct has three callbacks. The func callback does the cpu intensive
work, and when it completes the work struct is marked as done.
Every time a work struct completes, the list is checked to see if the head
is marked as done. If so the ordered_func callback is used to do the
order sensitive processing and the ordered_free callback is used to do
any cleanup. Then we loop back and check the head of the list again.
This patch also changes the checksumming code to use the ordered workqueues.
One a 4 drive array, it increases streaming writes from 280MB/s to 350MB/s.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-11-07 06:03:00 +03:00
extent_submit_bio_hook_t * submit_bio_start ;
extent_submit_bio_hook_t * submit_bio_done ;
2008-04-16 19:14:51 +04:00
int rw ;
int mirror_num ;
Btrfs: Add zlib compression support
This is a large change for adding compression on reading and writing,
both for inline and regular extents. It does some fairly large
surgery to the writeback paths.
Compression is off by default and enabled by mount -o compress. Even
when the -o compress mount option is not used, it is possible to read
compressed extents off the disk.
If compression for a given set of pages fails to make them smaller, the
file is flagged to avoid future compression attempts later.
* While finding delalloc extents, the pages are locked before being sent down
to the delalloc handler. This allows the delalloc handler to do complex things
such as cleaning the pages, marking them writeback and starting IO on their
behalf.
* Inline extents are inserted at delalloc time now. This allows us to compress
the data before inserting the inline extent, and it allows us to insert
an inline extent that spans multiple pages.
* All of the in-memory extent representations (extent_map.c, ordered-data.c etc)
are changed to record both an in-memory size and an on disk size, as well
as a flag for compression.
From a disk format point of view, the extent pointers in the file are changed
to record the on disk size of a given extent and some encoding flags.
Space in the disk format is allocated for compression encoding, as well
as encryption and a generic 'other' field. Neither the encryption or the
'other' field are currently used.
In order to limit the amount of data read for a single random read in the
file, the size of a compressed extent is limited to 128k. This is a
software only limit, the disk format supports u64 sized compressed extents.
In order to limit the ram consumed while processing extents, the uncompressed
size of a compressed extent is limited to 256k. This is a software only limit
and will be subject to tuning later.
Checksumming is still done on compressed extents, and it is done on the
uncompressed version of the data. This way additional encodings can be
layered on without having to figure out which encoding to checksum.
Compression happens at delalloc time, which is basically singled threaded because
it is usually done by a single pdflush thread. This makes it tricky to
spread the compression load across all the cpus on the box. We'll have to
look at parallel pdflush walks of dirty inodes at a later time.
Decompression is hooked into readpages and it does spread across CPUs nicely.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-10-29 21:49:59 +03:00
unsigned long bio_flags ;
2010-05-25 17:48:28 +04:00
/*
* bio_offset is optional , can be used if the pages in the bio
* can ' t tell us where in the file the bio should go
*/
u64 bio_offset ;
2008-06-12 00:50:36 +04:00
struct btrfs_work work ;
2008-04-16 19:14:51 +04:00
} ;
2009-02-12 22:09:45 +03:00
/* These are used to set the lockdep class on the extent buffer locks.
* The class is set by the readpage_end_io_hook after the buffer has
* passed csum validation but before the pages are unlocked .
*
* The lockdep class is also set by btrfs_init_new_buffer on freshly
* allocated blocks .
*
* The class is based on the level in the tree block , which allows lockdep
* to know that lower nodes nest inside the locks of higher nodes .
*
* We also add a check to make sure the highest level of the tree is
* the same as our lockdep setup here . If BTRFS_MAX_LEVEL changes , this
* code needs update as well .
*/
# ifdef CONFIG_DEBUG_LOCK_ALLOC
# if BTRFS_MAX_LEVEL != 8
# error
# endif
static struct lock_class_key btrfs_eb_class [ BTRFS_MAX_LEVEL + 1 ] ;
static const char * btrfs_eb_name [ BTRFS_MAX_LEVEL + 1 ] = {
/* leaf */
" btrfs-extent-00 " ,
" btrfs-extent-01 " ,
" btrfs-extent-02 " ,
" btrfs-extent-03 " ,
" btrfs-extent-04 " ,
" btrfs-extent-05 " ,
" btrfs-extent-06 " ,
" btrfs-extent-07 " ,
/* highest possible level */
" btrfs-extent-08 " ,
} ;
# endif
2008-09-29 23:18:18 +04:00
/*
* extents on the btree inode are pretty simple , there ' s one extent
* that covers the entire device
*/
2008-12-02 17:54:17 +03:00
static struct extent_map * btree_get_extent ( struct inode * inode ,
struct page * page , size_t page_offset , u64 start , u64 len ,
int create )
2007-04-11 23:53:25 +04:00
{
2007-10-16 00:14:19 +04:00
struct extent_map_tree * em_tree = & BTRFS_I ( inode ) - > extent_tree ;
struct extent_map * em ;
int ret ;
2009-09-03 00:24:52 +04:00
read_lock ( & em_tree - > lock ) ;
2008-01-25 00:13:08 +03:00
em = lookup_extent_mapping ( em_tree , start , len ) ;
2008-05-07 19:43:44 +04:00
if ( em ) {
em - > bdev =
BTRFS_I ( inode ) - > root - > fs_info - > fs_devices - > latest_bdev ;
2009-09-03 00:24:52 +04:00
read_unlock ( & em_tree - > lock ) ;
2007-10-16 00:14:19 +04:00
goto out ;
2008-05-07 19:43:44 +04:00
}
2009-09-03 00:24:52 +04:00
read_unlock ( & em_tree - > lock ) ;
2008-04-18 18:29:50 +04:00
2007-10-16 00:14:19 +04:00
em = alloc_extent_map ( GFP_NOFS ) ;
if ( ! em ) {
em = ERR_PTR ( - ENOMEM ) ;
goto out ;
}
em - > start = 0 ;
2008-04-18 22:17:20 +04:00
em - > len = ( u64 ) - 1 ;
Btrfs: Add zlib compression support
This is a large change for adding compression on reading and writing,
both for inline and regular extents. It does some fairly large
surgery to the writeback paths.
Compression is off by default and enabled by mount -o compress. Even
when the -o compress mount option is not used, it is possible to read
compressed extents off the disk.
If compression for a given set of pages fails to make them smaller, the
file is flagged to avoid future compression attempts later.
* While finding delalloc extents, the pages are locked before being sent down
to the delalloc handler. This allows the delalloc handler to do complex things
such as cleaning the pages, marking them writeback and starting IO on their
behalf.
* Inline extents are inserted at delalloc time now. This allows us to compress
the data before inserting the inline extent, and it allows us to insert
an inline extent that spans multiple pages.
* All of the in-memory extent representations (extent_map.c, ordered-data.c etc)
are changed to record both an in-memory size and an on disk size, as well
as a flag for compression.
From a disk format point of view, the extent pointers in the file are changed
to record the on disk size of a given extent and some encoding flags.
Space in the disk format is allocated for compression encoding, as well
as encryption and a generic 'other' field. Neither the encryption or the
'other' field are currently used.
In order to limit the amount of data read for a single random read in the
file, the size of a compressed extent is limited to 128k. This is a
software only limit, the disk format supports u64 sized compressed extents.
In order to limit the ram consumed while processing extents, the uncompressed
size of a compressed extent is limited to 256k. This is a software only limit
and will be subject to tuning later.
Checksumming is still done on compressed extents, and it is done on the
uncompressed version of the data. This way additional encodings can be
layered on without having to figure out which encoding to checksum.
Compression happens at delalloc time, which is basically singled threaded because
it is usually done by a single pdflush thread. This makes it tricky to
spread the compression load across all the cpus on the box. We'll have to
look at parallel pdflush walks of dirty inodes at a later time.
Decompression is hooked into readpages and it does spread across CPUs nicely.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-10-29 21:49:59 +03:00
em - > block_len = ( u64 ) - 1 ;
2007-10-16 00:14:19 +04:00
em - > block_start = 0 ;
2008-05-07 19:43:44 +04:00
em - > bdev = BTRFS_I ( inode ) - > root - > fs_info - > fs_devices - > latest_bdev ;
2008-01-25 00:13:08 +03:00
2009-09-03 00:24:52 +04:00
write_lock ( & em_tree - > lock ) ;
2007-10-16 00:14:19 +04:00
ret = add_extent_mapping ( em_tree , em ) ;
if ( ret = = - EEXIST ) {
2008-04-18 22:17:20 +04:00
u64 failed_start = em - > start ;
u64 failed_len = em - > len ;
2007-10-16 00:14:19 +04:00
free_extent_map ( em ) ;
2008-04-18 18:29:50 +04:00
em = lookup_extent_mapping ( em_tree , start , len ) ;
2008-04-18 22:17:20 +04:00
if ( em ) {
2008-04-18 18:29:50 +04:00
ret = 0 ;
2008-04-18 22:17:20 +04:00
} else {
em = lookup_extent_mapping ( em_tree , failed_start ,
failed_len ) ;
2008-04-18 18:29:50 +04:00
ret = - EIO ;
2008-04-18 22:17:20 +04:00
}
2007-10-16 00:14:19 +04:00
} else if ( ret ) {
2008-04-18 18:29:50 +04:00
free_extent_map ( em ) ;
em = NULL ;
2007-10-16 00:14:19 +04:00
}
2009-09-03 00:24:52 +04:00
write_unlock ( & em_tree - > lock ) ;
2008-04-18 18:29:50 +04:00
if ( ret )
em = ERR_PTR ( ret ) ;
2007-10-16 00:14:19 +04:00
out :
return em ;
2007-04-11 23:53:25 +04:00
}
2007-10-16 00:19:22 +04:00
u32 btrfs_csum_data ( struct btrfs_root * root , char * data , u32 seed , size_t len )
{
2009-04-19 16:02:41 +04:00
return crc32c ( seed , data , len ) ;
2007-10-16 00:19:22 +04:00
}
void btrfs_csum_final ( u32 crc , char * result )
{
* ( __le32 * ) result = ~ cpu_to_le32 ( crc ) ;
}
2008-09-29 23:18:18 +04:00
/*
* compute the csum for a btree block , and either verify it or write it
* into the csum field of the block .
*/
2007-10-16 00:19:22 +04:00
static int csum_tree_block ( struct btrfs_root * root , struct extent_buffer * buf ,
int verify )
{
2008-12-02 15:17:45 +03:00
u16 csum_size =
btrfs_super_csum_size ( & root - > fs_info - > super_copy ) ;
char * result = NULL ;
2007-10-16 00:19:22 +04:00
unsigned long len ;
unsigned long cur_len ;
unsigned long offset = BTRFS_CSUM_SIZE ;
char * map_token = NULL ;
char * kaddr ;
unsigned long map_start ;
unsigned long map_len ;
int err ;
u32 crc = ~ ( u32 ) 0 ;
2008-12-02 15:17:45 +03:00
unsigned long inline_result ;
2007-10-16 00:19:22 +04:00
len = buf - > len - offset ;
2009-01-06 05:25:51 +03:00
while ( len > 0 ) {
2007-10-16 00:19:22 +04:00
err = map_private_extent_buffer ( buf , offset , 32 ,
& map_token , & kaddr ,
& map_start , & map_len , KM_USER0 ) ;
2009-01-06 05:25:51 +03:00
if ( err )
2007-10-16 00:19:22 +04:00
return 1 ;
cur_len = min ( len , map_len - ( offset - map_start ) ) ;
crc = btrfs_csum_data ( root , kaddr + offset - map_start ,
crc , cur_len ) ;
len - = cur_len ;
offset + = cur_len ;
unmap_extent_buffer ( buf , map_token , KM_USER0 ) ;
}
2008-12-02 15:17:45 +03:00
if ( csum_size > sizeof ( inline_result ) ) {
result = kzalloc ( csum_size * sizeof ( char ) , GFP_NOFS ) ;
if ( ! result )
return 1 ;
} else {
result = ( char * ) & inline_result ;
}
2007-10-16 00:19:22 +04:00
btrfs_csum_final ( crc , result ) ;
if ( verify ) {
2008-12-02 15:17:45 +03:00
if ( memcmp_extent_buffer ( buf , result , 0 , csum_size ) ) {
2008-01-08 23:46:27 +03:00
u32 val ;
u32 found = 0 ;
2008-12-02 15:17:45 +03:00
memcpy ( & found , result , csum_size ) ;
2008-01-08 23:46:27 +03:00
2008-12-02 15:17:45 +03:00
read_extent_buffer ( buf , & val , 0 , csum_size ) ;
2009-04-27 15:29:05 +04:00
if ( printk_ratelimit ( ) ) {
printk ( KERN_INFO " btrfs: %s checksum verify "
" failed on %llu wanted %X found %X "
" level %d \n " ,
root - > fs_info - > sb - > s_id ,
( unsigned long long ) buf - > start , val , found ,
btrfs_header_level ( buf ) ) ;
}
2008-12-02 15:17:45 +03:00
if ( result ! = ( char * ) & inline_result )
kfree ( result ) ;
2007-10-16 00:19:22 +04:00
return 1 ;
}
} else {
2008-12-02 15:17:45 +03:00
write_extent_buffer ( buf , result , 0 , csum_size ) ;
2007-10-16 00:19:22 +04:00
}
2008-12-02 15:17:45 +03:00
if ( result ! = ( char * ) & inline_result )
kfree ( result ) ;
2007-10-16 00:19:22 +04:00
return 0 ;
}
2008-09-29 23:18:18 +04:00
/*
* we can ' t consider a given block up to date unless the transid of the
* block matches the transid in the parent node ' s pointer . This is how we
* detect blocks that either didn ' t get written at all or got written
* in the wrong place .
*/
2008-05-12 21:39:03 +04:00
static int verify_parent_transid ( struct extent_io_tree * io_tree ,
struct extent_buffer * eb , u64 parent_transid )
{
2010-02-03 22:33:23 +03:00
struct extent_state * cached_state = NULL ;
2008-05-12 21:39:03 +04:00
int ret ;
if ( ! parent_transid | | btrfs_header_generation ( eb ) = = parent_transid )
return 0 ;
2010-02-03 22:33:23 +03:00
lock_extent_bits ( io_tree , eb - > start , eb - > start + eb - > len - 1 ,
0 , & cached_state , GFP_NOFS ) ;
if ( extent_buffer_uptodate ( io_tree , eb , cached_state ) & &
2008-05-12 21:39:03 +04:00
btrfs_header_generation ( eb ) = = parent_transid ) {
ret = 0 ;
goto out ;
}
2009-04-27 15:29:05 +04:00
if ( printk_ratelimit ( ) ) {
printk ( " parent transid verify failed on %llu wanted %llu "
" found %llu \n " ,
( unsigned long long ) eb - > start ,
( unsigned long long ) parent_transid ,
( unsigned long long ) btrfs_header_generation ( eb ) ) ;
}
2008-05-12 21:39:03 +04:00
ret = 1 ;
2010-02-03 22:33:23 +03:00
clear_extent_buffer_uptodate ( io_tree , eb , & cached_state ) ;
2008-07-30 18:29:12 +04:00
out :
2010-02-03 22:33:23 +03:00
unlock_extent_cached ( io_tree , eb - > start , eb - > start + eb - > len - 1 ,
& cached_state , GFP_NOFS ) ;
2008-05-12 21:39:03 +04:00
return ret ;
}
2008-09-29 23:18:18 +04:00
/*
* helper to read a given tree block , doing retries as required when
* the checksums don ' t match and we have alternate mirrors to try .
*/
2008-04-10 00:28:12 +04:00
static int btree_read_extent_buffer_pages ( struct btrfs_root * root ,
struct extent_buffer * eb ,
2008-05-12 20:59:19 +04:00
u64 start , u64 parent_transid )
2008-04-10 00:28:12 +04:00
{
struct extent_io_tree * io_tree ;
int ret ;
int num_copies = 0 ;
int mirror_num = 0 ;
io_tree = & BTRFS_I ( root - > fs_info - > btree_inode ) - > io_tree ;
while ( 1 ) {
ret = read_extent_buffer_pages ( io_tree , eb , start , 1 ,
btree_get_extent , mirror_num ) ;
2008-05-12 21:39:03 +04:00
if ( ! ret & &
! verify_parent_transid ( io_tree , eb , parent_transid ) )
2008-04-10 00:28:12 +04:00
return ret ;
2009-01-06 05:25:51 +03:00
2008-04-10 00:28:12 +04:00
num_copies = btrfs_num_copies ( & root - > fs_info - > mapping_tree ,
eb - > start , eb - > len ) ;
2008-04-29 00:40:52 +04:00
if ( num_copies = = 1 )
2008-04-10 00:28:12 +04:00
return ret ;
2008-04-29 00:40:52 +04:00
2008-04-10 00:28:12 +04:00
mirror_num + + ;
2008-04-29 00:40:52 +04:00
if ( mirror_num > num_copies )
2008-04-10 00:28:12 +04:00
return ret ;
}
return - EIO ;
}
2007-10-16 00:19:22 +04:00
2008-09-29 23:18:18 +04:00
/*
2009-01-06 05:25:51 +03:00
* checksum a dirty tree block before IO . This has extra checks to make sure
* we only fill in the checksum field in the first page of a multi - page block
2008-09-29 23:18:18 +04:00
*/
2009-01-06 05:25:51 +03:00
2008-12-02 17:54:17 +03:00
static int csum_dirty_buffer ( struct btrfs_root * root , struct page * page )
2007-10-16 00:19:22 +04:00
{
2008-01-25 00:13:08 +03:00
struct extent_io_tree * tree ;
2007-10-30 23:56:53 +03:00
u64 start = ( u64 ) page - > index < < PAGE_CACHE_SHIFT ;
2007-10-16 00:19:22 +04:00
u64 found_start ;
int found_level ;
unsigned long len ;
struct extent_buffer * eb ;
2008-04-10 00:28:12 +04:00
int ret ;
2008-01-25 00:13:08 +03:00
tree = & BTRFS_I ( page - > mapping - > host ) - > io_tree ;
2007-10-16 00:19:22 +04:00
if ( page - > private = = EXTENT_PAGE_PRIVATE )
goto out ;
if ( ! page - > private )
goto out ;
len = page - > private > > 2 ;
2009-01-06 05:25:51 +03:00
WARN_ON ( len = = 0 ) ;
2007-10-16 00:19:22 +04:00
eb = alloc_extent_buffer ( tree , start , len , page , GFP_NOFS ) ;
2008-05-12 20:59:19 +04:00
ret = btree_read_extent_buffer_pages ( root , eb , start + PAGE_CACHE_SIZE ,
btrfs_header_generation ( eb ) ) ;
2008-04-10 00:28:12 +04:00
BUG_ON ( ret ) ;
2007-10-16 00:19:22 +04:00
found_start = btrfs_header_bytenr ( eb ) ;
if ( found_start ! = start ) {
2008-01-09 23:55:33 +03:00
WARN_ON ( 1 ) ;
goto err ;
}
if ( eb - > first_page ! = page ) {
WARN_ON ( 1 ) ;
goto err ;
}
if ( ! PageUptodate ( page ) ) {
WARN_ON ( 1 ) ;
goto err ;
2007-10-16 00:19:22 +04:00
}
found_level = btrfs_header_level ( eb ) ;
2008-09-08 19:18:08 +04:00
2007-10-16 00:19:22 +04:00
csum_tree_block ( root , eb , 0 ) ;
2008-01-09 23:55:33 +03:00
err :
2007-10-16 00:19:22 +04:00
free_extent_buffer ( eb ) ;
out :
return 0 ;
}
2008-11-18 05:11:30 +03:00
static int check_tree_block_fsid ( struct btrfs_root * root ,
struct extent_buffer * eb )
{
struct btrfs_fs_devices * fs_devices = root - > fs_info - > fs_devices ;
u8 fsid [ BTRFS_UUID_SIZE ] ;
int ret = 1 ;
read_extent_buffer ( eb , fsid , ( unsigned long ) btrfs_header_fsid ( eb ) ,
BTRFS_FSID_SIZE ) ;
while ( fs_devices ) {
if ( ! memcmp ( fsid , fs_devices - > fsid , BTRFS_FSID_SIZE ) ) {
ret = 0 ;
break ;
}
fs_devices = fs_devices - > seed ;
}
return ret ;
}
2009-02-12 22:09:45 +03:00
# ifdef CONFIG_DEBUG_LOCK_ALLOC
void btrfs_set_buffer_lockdep_class ( struct extent_buffer * eb , int level )
{
lockdep_set_class_and_name ( & eb - > lock ,
& btrfs_eb_class [ level ] ,
btrfs_eb_name [ level ] ) ;
}
# endif
2008-12-02 17:54:17 +03:00
static int btree_readpage_end_io_hook ( struct page * page , u64 start , u64 end ,
2008-04-10 00:28:12 +04:00
struct extent_state * state )
{
struct extent_io_tree * tree ;
u64 found_start ;
int found_level ;
unsigned long len ;
struct extent_buffer * eb ;
struct btrfs_root * root = BTRFS_I ( page - > mapping - > host ) - > root ;
2008-04-10 00:28:12 +04:00
int ret = 0 ;
2008-04-10 00:28:12 +04:00
tree = & BTRFS_I ( page - > mapping - > host ) - > io_tree ;
if ( page - > private = = EXTENT_PAGE_PRIVATE )
goto out ;
if ( ! page - > private )
goto out ;
2009-01-06 05:25:51 +03:00
2008-04-10 00:28:12 +04:00
len = page - > private > > 2 ;
2009-01-06 05:25:51 +03:00
WARN_ON ( len = = 0 ) ;
2008-04-10 00:28:12 +04:00
eb = alloc_extent_buffer ( tree , start , len , page , GFP_NOFS ) ;
2008-04-10 00:28:12 +04:00
2008-04-10 00:28:12 +04:00
found_start = btrfs_header_bytenr ( eb ) ;
2008-09-12 16:57:47 +04:00
if ( found_start ! = start ) {
2009-04-27 15:29:05 +04:00
if ( printk_ratelimit ( ) ) {
printk ( KERN_INFO " btrfs bad tree block start "
" %llu %llu \n " ,
( unsigned long long ) found_start ,
( unsigned long long ) eb - > start ) ;
}
2008-04-10 00:28:12 +04:00
ret = - EIO ;
2008-04-10 00:28:12 +04:00
goto err ;
}
if ( eb - > first_page ! = page ) {
2009-01-06 05:25:51 +03:00
printk ( KERN_INFO " btrfs bad first page %lu %lu \n " ,
eb - > first_page - > index , page - > index ) ;
2008-04-10 00:28:12 +04:00
WARN_ON ( 1 ) ;
2008-04-10 00:28:12 +04:00
ret = - EIO ;
2008-04-10 00:28:12 +04:00
goto err ;
}
2008-11-18 05:11:30 +03:00
if ( check_tree_block_fsid ( root , eb ) ) {
2009-04-27 15:29:05 +04:00
if ( printk_ratelimit ( ) ) {
printk ( KERN_INFO " btrfs bad fsid on block %llu \n " ,
( unsigned long long ) eb - > start ) ;
}
2008-05-12 21:39:03 +04:00
ret = - EIO ;
goto err ;
}
2008-04-10 00:28:12 +04:00
found_level = btrfs_header_level ( eb ) ;
2009-02-12 22:09:45 +03:00
btrfs_set_buffer_lockdep_class ( eb , found_level ) ;
2008-04-10 00:28:12 +04:00
ret = csum_tree_block ( root , eb , 1 ) ;
2008-04-10 00:28:12 +04:00
if ( ret )
ret = - EIO ;
2008-04-10 00:28:12 +04:00
end = min_t ( u64 , eb - > len , PAGE_CACHE_SIZE ) ;
end = eb - > start + end - 1 ;
err :
free_extent_buffer ( eb ) ;
out :
2008-04-10 00:28:12 +04:00
return ret ;
2008-04-10 00:28:12 +04:00
}
static void end_workqueue_bio ( struct bio * bio , int err )
{
struct end_io_wq * end_io_wq = bio - > bi_private ;
struct btrfs_fs_info * fs_info ;
fs_info = end_io_wq - > info ;
end_io_wq - > error = err ;
2008-06-12 00:50:36 +04:00
end_io_wq - > work . func = end_workqueue_fn ;
end_io_wq - > work . flags = 0 ;
Btrfs: move data checksumming into a dedicated tree
Btrfs stores checksums for each data block. Until now, they have
been stored in the subvolume trees, indexed by the inode that is
referencing the data block. This means that when we read the inode,
we've probably read in at least some checksums as well.
But, this has a few problems:
* The checksums are indexed by logical offset in the file. When
compression is on, this means we have to do the expensive checksumming
on the uncompressed data. It would be faster if we could checksum
the compressed data instead.
* If we implement encryption, we'll be checksumming the plain text and
storing that on disk. This is significantly less secure.
* For either compression or encryption, we have to get the plain text
back before we can verify the checksum as correct. This makes the raid
layer balancing and extent moving much more expensive.
* It makes the front end caching code more complex, as we have touch
the subvolume and inodes as we cache extents.
* There is potentitally one copy of the checksum in each subvolume
referencing an extent.
The solution used here is to store the extent checksums in a dedicated
tree. This allows us to index the checksums by phyiscal extent
start and length. It means:
* The checksum is against the data stored on disk, after any compression
or encryption is done.
* The checksum is stored in a central location, and can be verified without
following back references, or reading inodes.
This makes compression significantly faster by reducing the amount of
data that needs to be checksummed. It will also allow much faster
raid management code in general.
The checksums are indexed by a key with a fixed objectid (a magic value
in ctree.h) and offset set to the starting byte of the extent. This
allows us to copy the checksum items into the fsync log tree directly (or
any other tree), without having to invent a second format for them.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-12-09 00:58:54 +03:00
if ( bio - > bi_rw & ( 1 < < BIO_RW ) ) {
2008-12-17 22:51:42 +03:00
if ( end_io_wq - > metadata )
btrfs_queue_worker ( & fs_info - > endio_meta_write_workers ,
& end_io_wq - > work ) ;
else
btrfs_queue_worker ( & fs_info - > endio_write_workers ,
& end_io_wq - > work ) ;
Btrfs: move data checksumming into a dedicated tree
Btrfs stores checksums for each data block. Until now, they have
been stored in the subvolume trees, indexed by the inode that is
referencing the data block. This means that when we read the inode,
we've probably read in at least some checksums as well.
But, this has a few problems:
* The checksums are indexed by logical offset in the file. When
compression is on, this means we have to do the expensive checksumming
on the uncompressed data. It would be faster if we could checksum
the compressed data instead.
* If we implement encryption, we'll be checksumming the plain text and
storing that on disk. This is significantly less secure.
* For either compression or encryption, we have to get the plain text
back before we can verify the checksum as correct. This makes the raid
layer balancing and extent moving much more expensive.
* It makes the front end caching code more complex, as we have touch
the subvolume and inodes as we cache extents.
* There is potentitally one copy of the checksum in each subvolume
referencing an extent.
The solution used here is to store the extent checksums in a dedicated
tree. This allows us to index the checksums by phyiscal extent
start and length. It means:
* The checksum is against the data stored on disk, after any compression
or encryption is done.
* The checksum is stored in a central location, and can be verified without
following back references, or reading inodes.
This makes compression significantly faster by reducing the amount of
data that needs to be checksummed. It will also allow much faster
raid management code in general.
The checksums are indexed by a key with a fixed objectid (a magic value
in ctree.h) and offset set to the starting byte of the extent. This
allows us to copy the checksum items into the fsync log tree directly (or
any other tree), without having to invent a second format for them.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-12-09 00:58:54 +03:00
} else {
if ( end_io_wq - > metadata )
btrfs_queue_worker ( & fs_info - > endio_meta_workers ,
& end_io_wq - > work ) ;
else
btrfs_queue_worker ( & fs_info - > endio_workers ,
& end_io_wq - > work ) ;
}
2008-04-10 00:28:12 +04:00
}
2008-04-10 00:28:12 +04:00
int btrfs_bio_wq_end_io ( struct btrfs_fs_info * info , struct bio * bio ,
int metadata )
2008-03-24 22:01:56 +03:00
{
2008-04-10 00:28:12 +04:00
struct end_io_wq * end_io_wq ;
end_io_wq = kmalloc ( sizeof ( * end_io_wq ) , GFP_NOFS ) ;
if ( ! end_io_wq )
return - ENOMEM ;
end_io_wq - > private = bio - > bi_private ;
end_io_wq - > end_io = bio - > bi_end_io ;
2008-04-10 00:28:12 +04:00
end_io_wq - > info = info ;
2008-04-10 00:28:12 +04:00
end_io_wq - > error = 0 ;
end_io_wq - > bio = bio ;
2008-04-10 00:28:12 +04:00
end_io_wq - > metadata = metadata ;
2008-04-10 00:28:12 +04:00
bio - > bi_private = end_io_wq ;
bio - > bi_end_io = end_workqueue_bio ;
2008-04-10 00:28:12 +04:00
return 0 ;
}
2008-08-20 21:39:41 +04:00
unsigned long btrfs_async_submit_limit ( struct btrfs_fs_info * info )
2008-08-15 23:34:15 +04:00
{
2008-08-15 23:34:17 +04:00
unsigned long limit = min_t ( unsigned long ,
info - > workers . max_workers ,
info - > fs_devices - > open_devices ) ;
return 256 * limit ;
}
2008-08-15 23:34:15 +04:00
2008-08-15 23:34:17 +04:00
int btrfs_congested_async ( struct btrfs_fs_info * info , int iodone )
{
2008-08-20 21:39:41 +04:00
return atomic_read ( & info - > nr_async_bios ) >
btrfs_async_submit_limit ( info ) ;
2008-08-15 23:34:15 +04:00
}
Btrfs: Add ordered async work queues
Btrfs uses kernel threads to create async work queues for cpu intensive
operations such as checksumming and decompression. These work well,
but they make it difficult to keep IO order intact.
A single writepages call from pdflush or fsync will turn into a number
of bios, and each bio is checksummed in parallel. Once the checksum is
computed, the bio is sent down to the disk, and since we don't control
the order in which the parallel operations happen, they might go down to
the disk in almost any order.
The code deals with this somewhat by having deep work queues for a single
kernel thread, making it very likely that a single thread will process all
the bios for a single inode.
This patch introduces an explicitly ordered work queue. As work structs
are placed into the queue they are put onto the tail of a list. They have
three callbacks:
->func (cpu intensive processing here)
->ordered_func (order sensitive processing here)
->ordered_free (free the work struct, all processing is done)
The work struct has three callbacks. The func callback does the cpu intensive
work, and when it completes the work struct is marked as done.
Every time a work struct completes, the list is checked to see if the head
is marked as done. If so the ordered_func callback is used to do the
order sensitive processing and the ordered_free callback is used to do
any cleanup. Then we loop back and check the head of the list again.
This patch also changes the checksumming code to use the ordered workqueues.
One a 4 drive array, it increases streaming writes from 280MB/s to 350MB/s.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-11-07 06:03:00 +03:00
static void run_one_async_start ( struct btrfs_work * work )
{
struct btrfs_fs_info * fs_info ;
struct async_submit_bio * async ;
async = container_of ( work , struct async_submit_bio , work ) ;
fs_info = BTRFS_I ( async - > inode ) - > root - > fs_info ;
async - > submit_bio_start ( async - > inode , async - > rw , async - > bio ,
2010-05-25 17:48:28 +04:00
async - > mirror_num , async - > bio_flags ,
async - > bio_offset ) ;
Btrfs: Add ordered async work queues
Btrfs uses kernel threads to create async work queues for cpu intensive
operations such as checksumming and decompression. These work well,
but they make it difficult to keep IO order intact.
A single writepages call from pdflush or fsync will turn into a number
of bios, and each bio is checksummed in parallel. Once the checksum is
computed, the bio is sent down to the disk, and since we don't control
the order in which the parallel operations happen, they might go down to
the disk in almost any order.
The code deals with this somewhat by having deep work queues for a single
kernel thread, making it very likely that a single thread will process all
the bios for a single inode.
This patch introduces an explicitly ordered work queue. As work structs
are placed into the queue they are put onto the tail of a list. They have
three callbacks:
->func (cpu intensive processing here)
->ordered_func (order sensitive processing here)
->ordered_free (free the work struct, all processing is done)
The work struct has three callbacks. The func callback does the cpu intensive
work, and when it completes the work struct is marked as done.
Every time a work struct completes, the list is checked to see if the head
is marked as done. If so the ordered_func callback is used to do the
order sensitive processing and the ordered_free callback is used to do
any cleanup. Then we loop back and check the head of the list again.
This patch also changes the checksumming code to use the ordered workqueues.
One a 4 drive array, it increases streaming writes from 280MB/s to 350MB/s.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-11-07 06:03:00 +03:00
}
static void run_one_async_done ( struct btrfs_work * work )
2008-06-12 00:50:36 +04:00
{
struct btrfs_fs_info * fs_info ;
struct async_submit_bio * async ;
2008-08-15 23:34:17 +04:00
int limit ;
2008-06-12 00:50:36 +04:00
async = container_of ( work , struct async_submit_bio , work ) ;
fs_info = BTRFS_I ( async - > inode ) - > root - > fs_info ;
2008-08-15 23:34:17 +04:00
2008-08-20 21:39:41 +04:00
limit = btrfs_async_submit_limit ( fs_info ) ;
2008-08-15 23:34:17 +04:00
limit = limit * 2 / 3 ;
2008-06-12 00:50:36 +04:00
atomic_dec ( & fs_info - > nr_async_submits ) ;
2008-08-15 23:34:15 +04:00
2008-08-20 21:39:41 +04:00
if ( atomic_read ( & fs_info - > nr_async_submits ) < limit & &
waitqueue_active ( & fs_info - > async_submit_wait ) )
2008-08-15 23:34:17 +04:00
wake_up ( & fs_info - > async_submit_wait ) ;
Btrfs: Add ordered async work queues
Btrfs uses kernel threads to create async work queues for cpu intensive
operations such as checksumming and decompression. These work well,
but they make it difficult to keep IO order intact.
A single writepages call from pdflush or fsync will turn into a number
of bios, and each bio is checksummed in parallel. Once the checksum is
computed, the bio is sent down to the disk, and since we don't control
the order in which the parallel operations happen, they might go down to
the disk in almost any order.
The code deals with this somewhat by having deep work queues for a single
kernel thread, making it very likely that a single thread will process all
the bios for a single inode.
This patch introduces an explicitly ordered work queue. As work structs
are placed into the queue they are put onto the tail of a list. They have
three callbacks:
->func (cpu intensive processing here)
->ordered_func (order sensitive processing here)
->ordered_free (free the work struct, all processing is done)
The work struct has three callbacks. The func callback does the cpu intensive
work, and when it completes the work struct is marked as done.
Every time a work struct completes, the list is checked to see if the head
is marked as done. If so the ordered_func callback is used to do the
order sensitive processing and the ordered_free callback is used to do
any cleanup. Then we loop back and check the head of the list again.
This patch also changes the checksumming code to use the ordered workqueues.
One a 4 drive array, it increases streaming writes from 280MB/s to 350MB/s.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-11-07 06:03:00 +03:00
async - > submit_bio_done ( async - > inode , async - > rw , async - > bio ,
2010-05-25 17:48:28 +04:00
async - > mirror_num , async - > bio_flags ,
async - > bio_offset ) ;
Btrfs: Add ordered async work queues
Btrfs uses kernel threads to create async work queues for cpu intensive
operations such as checksumming and decompression. These work well,
but they make it difficult to keep IO order intact.
A single writepages call from pdflush or fsync will turn into a number
of bios, and each bio is checksummed in parallel. Once the checksum is
computed, the bio is sent down to the disk, and since we don't control
the order in which the parallel operations happen, they might go down to
the disk in almost any order.
The code deals with this somewhat by having deep work queues for a single
kernel thread, making it very likely that a single thread will process all
the bios for a single inode.
This patch introduces an explicitly ordered work queue. As work structs
are placed into the queue they are put onto the tail of a list. They have
three callbacks:
->func (cpu intensive processing here)
->ordered_func (order sensitive processing here)
->ordered_free (free the work struct, all processing is done)
The work struct has three callbacks. The func callback does the cpu intensive
work, and when it completes the work struct is marked as done.
Every time a work struct completes, the list is checked to see if the head
is marked as done. If so the ordered_func callback is used to do the
order sensitive processing and the ordered_free callback is used to do
any cleanup. Then we loop back and check the head of the list again.
This patch also changes the checksumming code to use the ordered workqueues.
One a 4 drive array, it increases streaming writes from 280MB/s to 350MB/s.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-11-07 06:03:00 +03:00
}
static void run_one_async_free ( struct btrfs_work * work )
{
struct async_submit_bio * async ;
async = container_of ( work , struct async_submit_bio , work ) ;
2008-06-12 00:50:36 +04:00
kfree ( async ) ;
}
2008-04-16 19:14:51 +04:00
int btrfs_wq_submit_bio ( struct btrfs_fs_info * fs_info , struct inode * inode ,
int rw , struct bio * bio , int mirror_num ,
Btrfs: Add zlib compression support
This is a large change for adding compression on reading and writing,
both for inline and regular extents. It does some fairly large
surgery to the writeback paths.
Compression is off by default and enabled by mount -o compress. Even
when the -o compress mount option is not used, it is possible to read
compressed extents off the disk.
If compression for a given set of pages fails to make them smaller, the
file is flagged to avoid future compression attempts later.
* While finding delalloc extents, the pages are locked before being sent down
to the delalloc handler. This allows the delalloc handler to do complex things
such as cleaning the pages, marking them writeback and starting IO on their
behalf.
* Inline extents are inserted at delalloc time now. This allows us to compress
the data before inserting the inline extent, and it allows us to insert
an inline extent that spans multiple pages.
* All of the in-memory extent representations (extent_map.c, ordered-data.c etc)
are changed to record both an in-memory size and an on disk size, as well
as a flag for compression.
From a disk format point of view, the extent pointers in the file are changed
to record the on disk size of a given extent and some encoding flags.
Space in the disk format is allocated for compression encoding, as well
as encryption and a generic 'other' field. Neither the encryption or the
'other' field are currently used.
In order to limit the amount of data read for a single random read in the
file, the size of a compressed extent is limited to 128k. This is a
software only limit, the disk format supports u64 sized compressed extents.
In order to limit the ram consumed while processing extents, the uncompressed
size of a compressed extent is limited to 256k. This is a software only limit
and will be subject to tuning later.
Checksumming is still done on compressed extents, and it is done on the
uncompressed version of the data. This way additional encodings can be
layered on without having to figure out which encoding to checksum.
Compression happens at delalloc time, which is basically singled threaded because
it is usually done by a single pdflush thread. This makes it tricky to
spread the compression load across all the cpus on the box. We'll have to
look at parallel pdflush walks of dirty inodes at a later time.
Decompression is hooked into readpages and it does spread across CPUs nicely.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-10-29 21:49:59 +03:00
unsigned long bio_flags ,
2010-05-25 17:48:28 +04:00
u64 bio_offset ,
Btrfs: Add ordered async work queues
Btrfs uses kernel threads to create async work queues for cpu intensive
operations such as checksumming and decompression. These work well,
but they make it difficult to keep IO order intact.
A single writepages call from pdflush or fsync will turn into a number
of bios, and each bio is checksummed in parallel. Once the checksum is
computed, the bio is sent down to the disk, and since we don't control
the order in which the parallel operations happen, they might go down to
the disk in almost any order.
The code deals with this somewhat by having deep work queues for a single
kernel thread, making it very likely that a single thread will process all
the bios for a single inode.
This patch introduces an explicitly ordered work queue. As work structs
are placed into the queue they are put onto the tail of a list. They have
three callbacks:
->func (cpu intensive processing here)
->ordered_func (order sensitive processing here)
->ordered_free (free the work struct, all processing is done)
The work struct has three callbacks. The func callback does the cpu intensive
work, and when it completes the work struct is marked as done.
Every time a work struct completes, the list is checked to see if the head
is marked as done. If so the ordered_func callback is used to do the
order sensitive processing and the ordered_free callback is used to do
any cleanup. Then we loop back and check the head of the list again.
This patch also changes the checksumming code to use the ordered workqueues.
One a 4 drive array, it increases streaming writes from 280MB/s to 350MB/s.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-11-07 06:03:00 +03:00
extent_submit_bio_hook_t * submit_bio_start ,
extent_submit_bio_hook_t * submit_bio_done )
2008-04-16 19:14:51 +04:00
{
struct async_submit_bio * async ;
async = kmalloc ( sizeof ( * async ) , GFP_NOFS ) ;
if ( ! async )
return - ENOMEM ;
async - > inode = inode ;
async - > rw = rw ;
async - > bio = bio ;
async - > mirror_num = mirror_num ;
Btrfs: Add ordered async work queues
Btrfs uses kernel threads to create async work queues for cpu intensive
operations such as checksumming and decompression. These work well,
but they make it difficult to keep IO order intact.
A single writepages call from pdflush or fsync will turn into a number
of bios, and each bio is checksummed in parallel. Once the checksum is
computed, the bio is sent down to the disk, and since we don't control
the order in which the parallel operations happen, they might go down to
the disk in almost any order.
The code deals with this somewhat by having deep work queues for a single
kernel thread, making it very likely that a single thread will process all
the bios for a single inode.
This patch introduces an explicitly ordered work queue. As work structs
are placed into the queue they are put onto the tail of a list. They have
three callbacks:
->func (cpu intensive processing here)
->ordered_func (order sensitive processing here)
->ordered_free (free the work struct, all processing is done)
The work struct has three callbacks. The func callback does the cpu intensive
work, and when it completes the work struct is marked as done.
Every time a work struct completes, the list is checked to see if the head
is marked as done. If so the ordered_func callback is used to do the
order sensitive processing and the ordered_free callback is used to do
any cleanup. Then we loop back and check the head of the list again.
This patch also changes the checksumming code to use the ordered workqueues.
One a 4 drive array, it increases streaming writes from 280MB/s to 350MB/s.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-11-07 06:03:00 +03:00
async - > submit_bio_start = submit_bio_start ;
async - > submit_bio_done = submit_bio_done ;
async - > work . func = run_one_async_start ;
async - > work . ordered_func = run_one_async_done ;
async - > work . ordered_free = run_one_async_free ;
2008-06-12 00:50:36 +04:00
async - > work . flags = 0 ;
Btrfs: Add zlib compression support
This is a large change for adding compression on reading and writing,
both for inline and regular extents. It does some fairly large
surgery to the writeback paths.
Compression is off by default and enabled by mount -o compress. Even
when the -o compress mount option is not used, it is possible to read
compressed extents off the disk.
If compression for a given set of pages fails to make them smaller, the
file is flagged to avoid future compression attempts later.
* While finding delalloc extents, the pages are locked before being sent down
to the delalloc handler. This allows the delalloc handler to do complex things
such as cleaning the pages, marking them writeback and starting IO on their
behalf.
* Inline extents are inserted at delalloc time now. This allows us to compress
the data before inserting the inline extent, and it allows us to insert
an inline extent that spans multiple pages.
* All of the in-memory extent representations (extent_map.c, ordered-data.c etc)
are changed to record both an in-memory size and an on disk size, as well
as a flag for compression.
From a disk format point of view, the extent pointers in the file are changed
to record the on disk size of a given extent and some encoding flags.
Space in the disk format is allocated for compression encoding, as well
as encryption and a generic 'other' field. Neither the encryption or the
'other' field are currently used.
In order to limit the amount of data read for a single random read in the
file, the size of a compressed extent is limited to 128k. This is a
software only limit, the disk format supports u64 sized compressed extents.
In order to limit the ram consumed while processing extents, the uncompressed
size of a compressed extent is limited to 256k. This is a software only limit
and will be subject to tuning later.
Checksumming is still done on compressed extents, and it is done on the
uncompressed version of the data. This way additional encodings can be
layered on without having to figure out which encoding to checksum.
Compression happens at delalloc time, which is basically singled threaded because
it is usually done by a single pdflush thread. This makes it tricky to
spread the compression load across all the cpus on the box. We'll have to
look at parallel pdflush walks of dirty inodes at a later time.
Decompression is hooked into readpages and it does spread across CPUs nicely.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-10-29 21:49:59 +03:00
async - > bio_flags = bio_flags ;
2010-05-25 17:48:28 +04:00
async - > bio_offset = bio_offset ;
2008-09-29 19:19:10 +04:00
2008-05-16 00:15:45 +04:00
atomic_inc ( & fs_info - > nr_async_submits ) ;
2009-04-20 23:50:09 +04:00
if ( rw & ( 1 < < BIO_RW_SYNCIO ) )
btrfs_set_work_high_prio ( & async - > work ) ;
2008-06-12 00:50:36 +04:00
btrfs_queue_worker ( & fs_info - > workers , & async - > work ) ;
2008-08-28 14:15:24 +04:00
2009-01-06 05:25:51 +03:00
while ( atomic_read ( & fs_info - > async_submit_draining ) & &
2008-11-07 06:02:51 +03:00
atomic_read ( & fs_info - > nr_async_submits ) ) {
wait_event ( fs_info - > async_submit_wait ,
( atomic_read ( & fs_info - > nr_async_submits ) = = 0 ) ) ;
}
2008-04-16 19:14:51 +04:00
return 0 ;
}
2008-09-23 21:14:12 +04:00
static int btree_csum_one_bio ( struct bio * bio )
{
struct bio_vec * bvec = bio - > bi_io_vec ;
int bio_index = 0 ;
struct btrfs_root * root ;
WARN_ON ( bio - > bi_vcnt < = 0 ) ;
2009-01-06 05:25:51 +03:00
while ( bio_index < bio - > bi_vcnt ) {
2008-09-23 21:14:12 +04:00
root = BTRFS_I ( bvec - > bv_page - > mapping - > host ) - > root ;
csum_dirty_buffer ( root , bvec - > bv_page ) ;
bio_index + + ;
bvec + + ;
}
return 0 ;
}
Btrfs: Add ordered async work queues
Btrfs uses kernel threads to create async work queues for cpu intensive
operations such as checksumming and decompression. These work well,
but they make it difficult to keep IO order intact.
A single writepages call from pdflush or fsync will turn into a number
of bios, and each bio is checksummed in parallel. Once the checksum is
computed, the bio is sent down to the disk, and since we don't control
the order in which the parallel operations happen, they might go down to
the disk in almost any order.
The code deals with this somewhat by having deep work queues for a single
kernel thread, making it very likely that a single thread will process all
the bios for a single inode.
This patch introduces an explicitly ordered work queue. As work structs
are placed into the queue they are put onto the tail of a list. They have
three callbacks:
->func (cpu intensive processing here)
->ordered_func (order sensitive processing here)
->ordered_free (free the work struct, all processing is done)
The work struct has three callbacks. The func callback does the cpu intensive
work, and when it completes the work struct is marked as done.
Every time a work struct completes, the list is checked to see if the head
is marked as done. If so the ordered_func callback is used to do the
order sensitive processing and the ordered_free callback is used to do
any cleanup. Then we loop back and check the head of the list again.
This patch also changes the checksumming code to use the ordered workqueues.
One a 4 drive array, it increases streaming writes from 280MB/s to 350MB/s.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-11-07 06:03:00 +03:00
static int __btree_submit_bio_start ( struct inode * inode , int rw ,
struct bio * bio , int mirror_num ,
2010-05-25 17:48:28 +04:00
unsigned long bio_flags ,
u64 bio_offset )
2008-04-10 00:28:12 +04:00
{
2008-06-12 00:50:36 +04:00
/*
* when we ' re called for a write , we ' re already in the async
2008-08-15 23:34:16 +04:00
* submission context . Just jump into btrfs_map_bio
2008-06-12 00:50:36 +04:00
*/
Btrfs: Add ordered async work queues
Btrfs uses kernel threads to create async work queues for cpu intensive
operations such as checksumming and decompression. These work well,
but they make it difficult to keep IO order intact.
A single writepages call from pdflush or fsync will turn into a number
of bios, and each bio is checksummed in parallel. Once the checksum is
computed, the bio is sent down to the disk, and since we don't control
the order in which the parallel operations happen, they might go down to
the disk in almost any order.
The code deals with this somewhat by having deep work queues for a single
kernel thread, making it very likely that a single thread will process all
the bios for a single inode.
This patch introduces an explicitly ordered work queue. As work structs
are placed into the queue they are put onto the tail of a list. They have
three callbacks:
->func (cpu intensive processing here)
->ordered_func (order sensitive processing here)
->ordered_free (free the work struct, all processing is done)
The work struct has three callbacks. The func callback does the cpu intensive
work, and when it completes the work struct is marked as done.
Every time a work struct completes, the list is checked to see if the head
is marked as done. If so the ordered_func callback is used to do the
order sensitive processing and the ordered_free callback is used to do
any cleanup. Then we loop back and check the head of the list again.
This patch also changes the checksumming code to use the ordered workqueues.
One a 4 drive array, it increases streaming writes from 280MB/s to 350MB/s.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-11-07 06:03:00 +03:00
btree_csum_one_bio ( bio ) ;
return 0 ;
}
2008-04-10 00:28:12 +04:00
Btrfs: Add ordered async work queues
Btrfs uses kernel threads to create async work queues for cpu intensive
operations such as checksumming and decompression. These work well,
but they make it difficult to keep IO order intact.
A single writepages call from pdflush or fsync will turn into a number
of bios, and each bio is checksummed in parallel. Once the checksum is
computed, the bio is sent down to the disk, and since we don't control
the order in which the parallel operations happen, they might go down to
the disk in almost any order.
The code deals with this somewhat by having deep work queues for a single
kernel thread, making it very likely that a single thread will process all
the bios for a single inode.
This patch introduces an explicitly ordered work queue. As work structs
are placed into the queue they are put onto the tail of a list. They have
three callbacks:
->func (cpu intensive processing here)
->ordered_func (order sensitive processing here)
->ordered_free (free the work struct, all processing is done)
The work struct has three callbacks. The func callback does the cpu intensive
work, and when it completes the work struct is marked as done.
Every time a work struct completes, the list is checked to see if the head
is marked as done. If so the ordered_func callback is used to do the
order sensitive processing and the ordered_free callback is used to do
any cleanup. Then we loop back and check the head of the list again.
This patch also changes the checksumming code to use the ordered workqueues.
One a 4 drive array, it increases streaming writes from 280MB/s to 350MB/s.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-11-07 06:03:00 +03:00
static int __btree_submit_bio_done ( struct inode * inode , int rw , struct bio * bio ,
2010-05-25 17:48:28 +04:00
int mirror_num , unsigned long bio_flags ,
u64 bio_offset )
Btrfs: Add ordered async work queues
Btrfs uses kernel threads to create async work queues for cpu intensive
operations such as checksumming and decompression. These work well,
but they make it difficult to keep IO order intact.
A single writepages call from pdflush or fsync will turn into a number
of bios, and each bio is checksummed in parallel. Once the checksum is
computed, the bio is sent down to the disk, and since we don't control
the order in which the parallel operations happen, they might go down to
the disk in almost any order.
The code deals with this somewhat by having deep work queues for a single
kernel thread, making it very likely that a single thread will process all
the bios for a single inode.
This patch introduces an explicitly ordered work queue. As work structs
are placed into the queue they are put onto the tail of a list. They have
three callbacks:
->func (cpu intensive processing here)
->ordered_func (order sensitive processing here)
->ordered_free (free the work struct, all processing is done)
The work struct has three callbacks. The func callback does the cpu intensive
work, and when it completes the work struct is marked as done.
Every time a work struct completes, the list is checked to see if the head
is marked as done. If so the ordered_func callback is used to do the
order sensitive processing and the ordered_free callback is used to do
any cleanup. Then we loop back and check the head of the list again.
This patch also changes the checksumming code to use the ordered workqueues.
One a 4 drive array, it increases streaming writes from 280MB/s to 350MB/s.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-11-07 06:03:00 +03:00
{
2008-06-12 00:50:36 +04:00
/*
Btrfs: Add ordered async work queues
Btrfs uses kernel threads to create async work queues for cpu intensive
operations such as checksumming and decompression. These work well,
but they make it difficult to keep IO order intact.
A single writepages call from pdflush or fsync will turn into a number
of bios, and each bio is checksummed in parallel. Once the checksum is
computed, the bio is sent down to the disk, and since we don't control
the order in which the parallel operations happen, they might go down to
the disk in almost any order.
The code deals with this somewhat by having deep work queues for a single
kernel thread, making it very likely that a single thread will process all
the bios for a single inode.
This patch introduces an explicitly ordered work queue. As work structs
are placed into the queue they are put onto the tail of a list. They have
three callbacks:
->func (cpu intensive processing here)
->ordered_func (order sensitive processing here)
->ordered_free (free the work struct, all processing is done)
The work struct has three callbacks. The func callback does the cpu intensive
work, and when it completes the work struct is marked as done.
Every time a work struct completes, the list is checked to see if the head
is marked as done. If so the ordered_func callback is used to do the
order sensitive processing and the ordered_free callback is used to do
any cleanup. Then we loop back and check the head of the list again.
This patch also changes the checksumming code to use the ordered workqueues.
One a 4 drive array, it increases streaming writes from 280MB/s to 350MB/s.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-11-07 06:03:00 +03:00
* when we ' re called for a write , we ' re already in the async
* submission context . Just jump into btrfs_map_bio
2008-06-12 00:50:36 +04:00
*/
return btrfs_map_bio ( BTRFS_I ( inode ) - > root , rw , bio , mirror_num , 1 ) ;
2008-03-24 22:01:56 +03:00
}
2008-04-16 19:14:51 +04:00
static int btree_submit_bio_hook ( struct inode * inode , int rw , struct bio * bio ,
2010-05-25 17:48:28 +04:00
int mirror_num , unsigned long bio_flags ,
u64 bio_offset )
2008-04-16 19:14:51 +04:00
{
2008-12-17 22:51:42 +03:00
int ret ;
ret = btrfs_bio_wq_end_io ( BTRFS_I ( inode ) - > root - > fs_info ,
bio , 1 ) ;
BUG_ON ( ret ) ;
2008-04-16 19:14:51 +04:00
if ( ! ( rw & ( 1 < < BIO_RW ) ) ) {
Btrfs: Add ordered async work queues
Btrfs uses kernel threads to create async work queues for cpu intensive
operations such as checksumming and decompression. These work well,
but they make it difficult to keep IO order intact.
A single writepages call from pdflush or fsync will turn into a number
of bios, and each bio is checksummed in parallel. Once the checksum is
computed, the bio is sent down to the disk, and since we don't control
the order in which the parallel operations happen, they might go down to
the disk in almost any order.
The code deals with this somewhat by having deep work queues for a single
kernel thread, making it very likely that a single thread will process all
the bios for a single inode.
This patch introduces an explicitly ordered work queue. As work structs
are placed into the queue they are put onto the tail of a list. They have
three callbacks:
->func (cpu intensive processing here)
->ordered_func (order sensitive processing here)
->ordered_free (free the work struct, all processing is done)
The work struct has three callbacks. The func callback does the cpu intensive
work, and when it completes the work struct is marked as done.
Every time a work struct completes, the list is checked to see if the head
is marked as done. If so the ordered_func callback is used to do the
order sensitive processing and the ordered_free callback is used to do
any cleanup. Then we loop back and check the head of the list again.
This patch also changes the checksumming code to use the ordered workqueues.
One a 4 drive array, it increases streaming writes from 280MB/s to 350MB/s.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-11-07 06:03:00 +03:00
/*
* called for a read , do the setup so that checksum validation
* can happen in the async kernel threads
*/
return btrfs_map_bio ( BTRFS_I ( inode ) - > root , rw , bio ,
2008-11-13 17:59:36 +03:00
mirror_num , 0 ) ;
2008-04-16 19:14:51 +04:00
}
2009-04-20 23:50:09 +04:00
2008-12-17 22:51:42 +03:00
/*
* kthread helpers are used to submit writes so that checksumming
* can happen in parallel across all CPUs
*/
2008-04-16 19:14:51 +04:00
return btrfs_wq_submit_bio ( BTRFS_I ( inode ) - > root - > fs_info ,
Btrfs: Add zlib compression support
This is a large change for adding compression on reading and writing,
both for inline and regular extents. It does some fairly large
surgery to the writeback paths.
Compression is off by default and enabled by mount -o compress. Even
when the -o compress mount option is not used, it is possible to read
compressed extents off the disk.
If compression for a given set of pages fails to make them smaller, the
file is flagged to avoid future compression attempts later.
* While finding delalloc extents, the pages are locked before being sent down
to the delalloc handler. This allows the delalloc handler to do complex things
such as cleaning the pages, marking them writeback and starting IO on their
behalf.
* Inline extents are inserted at delalloc time now. This allows us to compress
the data before inserting the inline extent, and it allows us to insert
an inline extent that spans multiple pages.
* All of the in-memory extent representations (extent_map.c, ordered-data.c etc)
are changed to record both an in-memory size and an on disk size, as well
as a flag for compression.
From a disk format point of view, the extent pointers in the file are changed
to record the on disk size of a given extent and some encoding flags.
Space in the disk format is allocated for compression encoding, as well
as encryption and a generic 'other' field. Neither the encryption or the
'other' field are currently used.
In order to limit the amount of data read for a single random read in the
file, the size of a compressed extent is limited to 128k. This is a
software only limit, the disk format supports u64 sized compressed extents.
In order to limit the ram consumed while processing extents, the uncompressed
size of a compressed extent is limited to 256k. This is a software only limit
and will be subject to tuning later.
Checksumming is still done on compressed extents, and it is done on the
uncompressed version of the data. This way additional encodings can be
layered on without having to figure out which encoding to checksum.
Compression happens at delalloc time, which is basically singled threaded because
it is usually done by a single pdflush thread. This makes it tricky to
spread the compression load across all the cpus on the box. We'll have to
look at parallel pdflush walks of dirty inodes at a later time.
Decompression is hooked into readpages and it does spread across CPUs nicely.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-10-29 21:49:59 +03:00
inode , rw , bio , mirror_num , 0 ,
2010-05-25 17:48:28 +04:00
bio_offset ,
Btrfs: Add ordered async work queues
Btrfs uses kernel threads to create async work queues for cpu intensive
operations such as checksumming and decompression. These work well,
but they make it difficult to keep IO order intact.
A single writepages call from pdflush or fsync will turn into a number
of bios, and each bio is checksummed in parallel. Once the checksum is
computed, the bio is sent down to the disk, and since we don't control
the order in which the parallel operations happen, they might go down to
the disk in almost any order.
The code deals with this somewhat by having deep work queues for a single
kernel thread, making it very likely that a single thread will process all
the bios for a single inode.
This patch introduces an explicitly ordered work queue. As work structs
are placed into the queue they are put onto the tail of a list. They have
three callbacks:
->func (cpu intensive processing here)
->ordered_func (order sensitive processing here)
->ordered_free (free the work struct, all processing is done)
The work struct has three callbacks. The func callback does the cpu intensive
work, and when it completes the work struct is marked as done.
Every time a work struct completes, the list is checked to see if the head
is marked as done. If so the ordered_func callback is used to do the
order sensitive processing and the ordered_free callback is used to do
any cleanup. Then we loop back and check the head of the list again.
This patch also changes the checksumming code to use the ordered workqueues.
One a 4 drive array, it increases streaming writes from 280MB/s to 350MB/s.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-11-07 06:03:00 +03:00
__btree_submit_bio_start ,
__btree_submit_bio_done ) ;
2008-04-16 19:14:51 +04:00
}
2007-11-08 05:08:01 +03:00
static int btree_writepage ( struct page * page , struct writeback_control * wbc )
{
2008-01-25 00:13:08 +03:00
struct extent_io_tree * tree ;
2009-03-13 18:00:37 +03:00
struct btrfs_root * root = BTRFS_I ( page - > mapping - > host ) - > root ;
struct extent_buffer * eb ;
int was_dirty ;
2008-01-25 00:13:08 +03:00
tree = & BTRFS_I ( page - > mapping - > host ) - > io_tree ;
2009-03-13 18:00:37 +03:00
if ( ! ( current - > flags & PF_MEMALLOC ) ) {
return extent_write_full_page ( tree , page ,
btree_get_extent , wbc ) ;
}
2008-08-15 23:34:16 +04:00
2009-03-13 18:00:37 +03:00
redirty_page_for_writepage ( wbc , page ) ;
eb = btrfs_find_tree_block ( root , page_offset ( page ) ,
PAGE_CACHE_SIZE ) ;
WARN_ON ( ! eb ) ;
was_dirty = test_and_set_bit ( EXTENT_BUFFER_DIRTY , & eb - > bflags ) ;
if ( ! was_dirty ) {
spin_lock ( & root - > fs_info - > delalloc_lock ) ;
root - > fs_info - > dirty_metadata_bytes + = PAGE_CACHE_SIZE ;
spin_unlock ( & root - > fs_info - > delalloc_lock ) ;
2008-08-15 23:34:16 +04:00
}
2009-03-13 18:00:37 +03:00
free_extent_buffer ( eb ) ;
unlock_page ( page ) ;
return 0 ;
2007-10-16 00:14:19 +04:00
}
2007-11-08 05:08:01 +03:00
static int btree_writepages ( struct address_space * mapping ,
struct writeback_control * wbc )
{
2008-01-25 00:13:08 +03:00
struct extent_io_tree * tree ;
tree = & BTRFS_I ( mapping - > host ) - > io_tree ;
2007-12-11 20:42:00 +03:00
if ( wbc - > sync_mode = = WB_SYNC_NONE ) {
2009-03-13 18:00:37 +03:00
struct btrfs_root * root = BTRFS_I ( mapping - > host ) - > root ;
2007-11-27 03:34:41 +03:00
u64 num_dirty ;
2008-09-24 22:51:30 +04:00
unsigned long thresh = 32 * 1024 * 1024 ;
2007-11-27 18:52:01 +03:00
if ( wbc - > for_kupdate )
return 0 ;
2009-03-13 18:00:37 +03:00
/* this is a bit racy, but that's ok */
num_dirty = root - > fs_info - > dirty_metadata_bytes ;
2009-01-06 05:25:51 +03:00
if ( num_dirty < thresh )
2007-11-27 03:34:41 +03:00
return 0 ;
}
2007-11-08 05:08:01 +03:00
return extent_writepages ( tree , mapping , btree_get_extent , wbc ) ;
}
2008-12-02 17:54:17 +03:00
static int btree_readpage ( struct file * file , struct page * page )
2007-10-16 00:14:19 +04:00
{
2008-01-25 00:13:08 +03:00
struct extent_io_tree * tree ;
tree = & BTRFS_I ( page - > mapping - > host ) - > io_tree ;
2007-10-16 00:14:19 +04:00
return extent_read_full_page ( tree , page , btree_get_extent ) ;
}
2007-03-30 16:47:31 +04:00
2008-01-29 17:59:12 +03:00
static int btree_releasepage ( struct page * page , gfp_t gfp_flags )
2007-10-16 00:14:19 +04:00
{
2008-01-25 00:13:08 +03:00
struct extent_io_tree * tree ;
struct extent_map_tree * map ;
2007-10-16 00:14:19 +04:00
int ret ;
2007-03-28 21:57:48 +04:00
2008-09-11 23:51:43 +04:00
if ( PageWriteback ( page ) | | PageDirty ( page ) )
2009-01-06 05:25:51 +03:00
return 0 ;
2008-09-11 23:51:43 +04:00
2008-01-25 00:13:08 +03:00
tree = & BTRFS_I ( page - > mapping - > host ) - > io_tree ;
map = & BTRFS_I ( page - > mapping - > host ) - > extent_tree ;
2008-07-22 19:18:07 +04:00
2008-04-18 18:29:50 +04:00
ret = try_release_extent_state ( map , tree , page , gfp_flags ) ;
2009-01-06 05:25:51 +03:00
if ( ! ret )
2008-07-22 19:18:07 +04:00
return 0 ;
ret = try_release_extent_buffer ( tree , page ) ;
2007-10-16 00:14:19 +04:00
if ( ret = = 1 ) {
ClearPagePrivate ( page ) ;
set_page_private ( page , 0 ) ;
page_cache_release ( page ) ;
}
2008-07-22 19:18:07 +04:00
2007-03-28 21:57:48 +04:00
return ret ;
}
2007-10-16 00:14:19 +04:00
static void btree_invalidatepage ( struct page * page , unsigned long offset )
2007-03-28 21:57:48 +04:00
{
2008-01-25 00:13:08 +03:00
struct extent_io_tree * tree ;
tree = & BTRFS_I ( page - > mapping - > host ) - > io_tree ;
2007-10-16 00:14:19 +04:00
extent_invalidatepage ( tree , page , offset ) ;
btree_releasepage ( page , GFP_NOFS ) ;
2008-04-19 00:11:30 +04:00
if ( PagePrivate ( page ) ) {
2009-01-06 05:25:51 +03:00
printk ( KERN_WARNING " btrfs warning page private not zero "
" on page %llu \n " , ( unsigned long long ) page_offset ( page ) ) ;
2008-04-19 00:11:30 +04:00
ClearPagePrivate ( page ) ;
set_page_private ( page , 0 ) ;
page_cache_release ( page ) ;
}
2007-03-28 21:57:48 +04:00
}
2009-09-22 04:01:10 +04:00
static const struct address_space_operations btree_aops = {
2007-03-28 21:57:48 +04:00
. readpage = btree_readpage ,
. writepage = btree_writepage ,
2007-11-08 05:08:01 +03:00
. writepages = btree_writepages ,
2007-10-16 00:14:19 +04:00
. releasepage = btree_releasepage ,
. invalidatepage = btree_invalidatepage ,
2007-03-28 21:57:48 +04:00
. sync_page = block_sync_page ,
} ;
2008-05-12 20:59:19 +04:00
int readahead_tree_block ( struct btrfs_root * root , u64 bytenr , u32 blocksize ,
u64 parent_transid )
2007-05-01 16:53:32 +04:00
{
2007-10-16 00:14:19 +04:00
struct extent_buffer * buf = NULL ;
struct inode * btree_inode = root - > fs_info - > btree_inode ;
2007-05-18 21:28:27 +04:00
int ret = 0 ;
2007-05-01 16:53:32 +04:00
2007-10-16 00:15:53 +04:00
buf = btrfs_find_create_tree_block ( root , bytenr , blocksize ) ;
2007-10-16 00:14:19 +04:00
if ( ! buf )
2007-05-01 16:53:32 +04:00
return 0 ;
2008-01-25 00:13:08 +03:00
read_extent_buffer_pages ( & BTRFS_I ( btree_inode ) - > io_tree ,
2008-04-10 00:28:12 +04:00
buf , 0 , 0 , btree_get_extent , 0 ) ;
2007-10-16 00:14:19 +04:00
free_extent_buffer ( buf ) ;
2007-05-18 21:28:27 +04:00
return ret ;
2007-05-01 16:53:32 +04:00
}
2008-04-01 21:48:14 +04:00
struct extent_buffer * btrfs_find_tree_block ( struct btrfs_root * root ,
u64 bytenr , u32 blocksize )
{
struct inode * btree_inode = root - > fs_info - > btree_inode ;
struct extent_buffer * eb ;
eb = find_extent_buffer ( & BTRFS_I ( btree_inode ) - > io_tree ,
bytenr , blocksize , GFP_NOFS ) ;
return eb ;
}
struct extent_buffer * btrfs_find_create_tree_block ( struct btrfs_root * root ,
u64 bytenr , u32 blocksize )
{
struct inode * btree_inode = root - > fs_info - > btree_inode ;
struct extent_buffer * eb ;
eb = alloc_extent_buffer ( & BTRFS_I ( btree_inode ) - > io_tree ,
bytenr , blocksize , NULL , GFP_NOFS ) ;
return eb ;
}
2008-09-06 00:13:11 +04:00
int btrfs_write_tree_block ( struct extent_buffer * buf )
{
2009-10-01 20:58:30 +04:00
return filemap_fdatawrite_range ( buf - > first_page - > mapping , buf - > start ,
buf - > start + buf - > len - 1 ) ;
2008-09-06 00:13:11 +04:00
}
int btrfs_wait_tree_block_writeback ( struct extent_buffer * buf )
{
2009-10-01 20:58:30 +04:00
return filemap_fdatawait_range ( buf - > first_page - > mapping ,
buf - > start , buf - > start + buf - > len - 1 ) ;
2008-09-06 00:13:11 +04:00
}
2008-04-01 21:48:14 +04:00
struct extent_buffer * read_tree_block ( struct btrfs_root * root , u64 bytenr ,
2008-05-12 20:59:19 +04:00
u32 blocksize , u64 parent_transid )
2008-04-01 21:48:14 +04:00
{
struct extent_buffer * buf = NULL ;
struct inode * btree_inode = root - > fs_info - > btree_inode ;
struct extent_io_tree * io_tree ;
int ret ;
io_tree = & BTRFS_I ( btree_inode ) - > io_tree ;
buf = btrfs_find_create_tree_block ( root , bytenr , blocksize ) ;
if ( ! buf )
return NULL ;
2008-05-12 20:59:19 +04:00
ret = btree_read_extent_buffer_pages ( root , buf , 0 , parent_transid ) ;
2008-04-10 00:28:12 +04:00
2009-01-06 05:25:51 +03:00
if ( ret = = 0 )
Btrfs: Change btree locking to use explicit blocking points
Most of the btrfs metadata operations can be protected by a spinlock,
but some operations still need to schedule.
So far, btrfs has been using a mutex along with a trylock loop,
most of the time it is able to avoid going for the full mutex, so
the trylock loop is a big performance gain.
This commit is step one for getting rid of the blocking locks entirely.
btrfs_tree_lock takes a spinlock, and the code explicitly switches
to a blocking lock when it starts an operation that can schedule.
We'll be able get rid of the blocking locks in smaller pieces over time.
Tracing allows us to find the most common cause of blocking, so we
can start with the hot spots first.
The basic idea is:
btrfs_tree_lock() returns with the spin lock held
btrfs_set_lock_blocking() sets the EXTENT_BUFFER_BLOCKING bit in
the extent buffer flags, and then drops the spin lock. The buffer is
still considered locked by all of the btrfs code.
If btrfs_tree_lock gets the spinlock but finds the blocking bit set, it drops
the spin lock and waits on a wait queue for the blocking bit to go away.
Much of the code that needs to set the blocking bit finishes without actually
blocking a good percentage of the time. So, an adaptive spin is still
used against the blocking bit to avoid very high context switch rates.
btrfs_clear_lock_blocking() clears the blocking bit and returns
with the spinlock held again.
btrfs_tree_unlock() can be called on either blocking or spinning locks,
it does the right thing based on the blocking bit.
ctree.c has a helper function to set/clear all the locked buffers in a
path as blocking.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-02-04 17:25:08 +03:00
set_bit ( EXTENT_BUFFER_UPTODATE , & buf - > bflags ) ;
2007-10-16 00:14:19 +04:00
return buf ;
2008-04-10 00:28:12 +04:00
2007-02-02 17:18:22 +03:00
}
2007-03-16 23:20:31 +03:00
int clean_tree_block ( struct btrfs_trans_handle * trans , struct btrfs_root * root ,
2007-10-16 00:14:19 +04:00
struct extent_buffer * buf )
2007-03-02 02:59:40 +03:00
{
2007-10-16 00:14:19 +04:00
struct inode * btree_inode = root - > fs_info - > btree_inode ;
2008-01-09 23:55:33 +03:00
if ( btrfs_header_generation ( buf ) = =
2008-06-26 00:01:30 +04:00
root - > fs_info - > running_transaction - > transid ) {
2009-03-09 18:45:38 +03:00
btrfs_assert_tree_locked ( buf ) ;
Btrfs: Change btree locking to use explicit blocking points
Most of the btrfs metadata operations can be protected by a spinlock,
but some operations still need to schedule.
So far, btrfs has been using a mutex along with a trylock loop,
most of the time it is able to avoid going for the full mutex, so
the trylock loop is a big performance gain.
This commit is step one for getting rid of the blocking locks entirely.
btrfs_tree_lock takes a spinlock, and the code explicitly switches
to a blocking lock when it starts an operation that can schedule.
We'll be able get rid of the blocking locks in smaller pieces over time.
Tracing allows us to find the most common cause of blocking, so we
can start with the hot spots first.
The basic idea is:
btrfs_tree_lock() returns with the spin lock held
btrfs_set_lock_blocking() sets the EXTENT_BUFFER_BLOCKING bit in
the extent buffer flags, and then drops the spin lock. The buffer is
still considered locked by all of the btrfs code.
If btrfs_tree_lock gets the spinlock but finds the blocking bit set, it drops
the spin lock and waits on a wait queue for the blocking bit to go away.
Much of the code that needs to set the blocking bit finishes without actually
blocking a good percentage of the time. So, an adaptive spin is still
used against the blocking bit to avoid very high context switch rates.
btrfs_clear_lock_blocking() clears the blocking bit and returns
with the spinlock held again.
btrfs_tree_unlock() can be called on either blocking or spinning locks,
it does the right thing based on the blocking bit.
ctree.c has a helper function to set/clear all the locked buffers in a
path as blocking.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-02-04 17:25:08 +03:00
2009-03-13 18:00:37 +03:00
if ( test_and_clear_bit ( EXTENT_BUFFER_DIRTY , & buf - > bflags ) ) {
spin_lock ( & root - > fs_info - > delalloc_lock ) ;
if ( root - > fs_info - > dirty_metadata_bytes > = buf - > len )
root - > fs_info - > dirty_metadata_bytes - = buf - > len ;
else
WARN_ON ( 1 ) ;
spin_unlock ( & root - > fs_info - > delalloc_lock ) ;
}
Btrfs: Change btree locking to use explicit blocking points
Most of the btrfs metadata operations can be protected by a spinlock,
but some operations still need to schedule.
So far, btrfs has been using a mutex along with a trylock loop,
most of the time it is able to avoid going for the full mutex, so
the trylock loop is a big performance gain.
This commit is step one for getting rid of the blocking locks entirely.
btrfs_tree_lock takes a spinlock, and the code explicitly switches
to a blocking lock when it starts an operation that can schedule.
We'll be able get rid of the blocking locks in smaller pieces over time.
Tracing allows us to find the most common cause of blocking, so we
can start with the hot spots first.
The basic idea is:
btrfs_tree_lock() returns with the spin lock held
btrfs_set_lock_blocking() sets the EXTENT_BUFFER_BLOCKING bit in
the extent buffer flags, and then drops the spin lock. The buffer is
still considered locked by all of the btrfs code.
If btrfs_tree_lock gets the spinlock but finds the blocking bit set, it drops
the spin lock and waits on a wait queue for the blocking bit to go away.
Much of the code that needs to set the blocking bit finishes without actually
blocking a good percentage of the time. So, an adaptive spin is still
used against the blocking bit to avoid very high context switch rates.
btrfs_clear_lock_blocking() clears the blocking bit and returns
with the spinlock held again.
btrfs_tree_unlock() can be called on either blocking or spinning locks,
it does the right thing based on the blocking bit.
ctree.c has a helper function to set/clear all the locked buffers in a
path as blocking.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-02-04 17:25:08 +03:00
2009-03-13 18:00:37 +03:00
/* ugh, clear_extent_buffer_dirty needs to lock the page */
btrfs_set_lock_blocking ( buf ) ;
2008-01-25 00:13:08 +03:00
clear_extent_buffer_dirty ( & BTRFS_I ( btree_inode ) - > io_tree ,
2008-01-09 23:55:33 +03:00
buf ) ;
2008-06-26 00:01:30 +04:00
}
2007-10-16 00:14:19 +04:00
return 0 ;
}
2007-10-16 00:15:53 +04:00
static int __setup_root ( u32 nodesize , u32 leafsize , u32 sectorsize ,
2007-11-30 19:30:34 +03:00
u32 stripesize , struct btrfs_root * root ,
2007-03-20 21:38:32 +03:00
struct btrfs_fs_info * fs_info ,
2007-03-22 19:13:20 +03:00
u64 objectid )
2007-02-21 00:40:44 +03:00
{
2007-02-22 01:04:57 +03:00
root - > node = NULL ;
2007-03-07 04:08:01 +03:00
root - > commit_root = NULL ;
2007-10-16 00:15:53 +04:00
root - > sectorsize = sectorsize ;
root - > nodesize = nodesize ;
root - > leafsize = leafsize ;
2007-11-30 19:30:34 +03:00
root - > stripesize = stripesize ;
2007-03-14 21:14:43 +03:00
root - > ref_cows = 0 ;
2008-03-24 22:01:56 +03:00
root - > track_dirty = 0 ;
2009-11-12 12:34:40 +03:00
root - > in_radix = 0 ;
2010-05-16 18:49:58 +04:00
root - > orphan_item_inserted = 0 ;
root - > orphan_cleanup_state = 0 ;
2008-03-24 22:01:56 +03:00
2007-03-20 21:38:32 +03:00
root - > fs_info = fs_info ;
2007-04-09 18:42:37 +04:00
root - > objectid = objectid ;
root - > last_trans = 0 ;
2009-09-21 23:56:00 +04:00
root - > highest_objectid = 0 ;
2007-08-29 23:47:34 +04:00
root - > name = NULL ;
2008-01-03 17:08:48 +03:00
root - > in_sysfs = 0 ;
2010-02-23 22:43:04 +03:00
root - > inode_tree = RB_ROOT ;
2010-05-16 18:46:25 +04:00
root - > block_rsv = NULL ;
2010-05-16 18:49:58 +04:00
root - > orphan_block_rsv = NULL ;
2008-03-24 22:01:56 +03:00
INIT_LIST_HEAD ( & root - > dirty_list ) ;
2008-07-24 20:17:14 +04:00
INIT_LIST_HEAD ( & root - > orphan_list ) ;
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
INIT_LIST_HEAD ( & root - > root_list ) ;
2008-06-26 00:01:30 +04:00
spin_lock_init ( & root - > node_lock ) ;
2010-05-16 18:49:58 +04:00
spin_lock_init ( & root - > orphan_lock ) ;
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
spin_lock_init ( & root - > inode_lock ) ;
2010-05-16 18:46:25 +04:00
spin_lock_init ( & root - > accounting_lock ) ;
2008-06-26 00:01:30 +04:00
mutex_init ( & root - > objectid_mutex ) ;
2008-09-06 00:13:11 +04:00
mutex_init ( & root - > log_mutex ) ;
2009-01-21 20:54:03 +03:00
init_waitqueue_head ( & root - > log_writer_wait ) ;
init_waitqueue_head ( & root - > log_commit_wait [ 0 ] ) ;
init_waitqueue_head ( & root - > log_commit_wait [ 1 ] ) ;
atomic_set ( & root - > log_commit [ 0 ] , 0 ) ;
atomic_set ( & root - > log_commit [ 1 ] , 0 ) ;
atomic_set ( & root - > log_writers , 0 ) ;
root - > log_batch = 0 ;
root - > log_transid = 0 ;
2009-10-13 21:21:08 +04:00
root - > last_log_commit = 0 ;
2008-09-12 00:17:57 +04:00
extent_io_tree_init ( & root - > dirty_log_pages ,
fs_info - > btree_inode - > i_mapping , GFP_NOFS ) ;
2008-07-28 23:32:51 +04:00
2007-03-13 23:47:54 +03:00
memset ( & root - > root_key , 0 , sizeof ( root - > root_key ) ) ;
memset ( & root - > root_item , 0 , sizeof ( root - > root_item ) ) ;
2007-08-08 00:15:09 +04:00
memset ( & root - > defrag_progress , 0 , sizeof ( root - > defrag_progress ) ) ;
2007-08-29 23:47:34 +04:00
memset ( & root - > root_kobj , 0 , sizeof ( root - > root_kobj ) ) ;
2008-06-26 00:01:31 +04:00
root - > defrag_trans_start = fs_info - > generation ;
2007-08-29 23:47:34 +04:00
init_completion ( & root - > kobj_unregister ) ;
2007-08-08 00:15:09 +04:00
root - > defrag_running = 0 ;
2007-04-21 04:23:12 +04:00
root - > root_key . objectid = objectid ;
2008-11-18 04:42:26 +03:00
root - > anon_super . s_root = NULL ;
root - > anon_super . s_dev = 0 ;
INIT_LIST_HEAD ( & root - > anon_super . s_list ) ;
INIT_LIST_HEAD ( & root - > anon_super . s_instances ) ;
init_rwsem ( & root - > anon_super . s_umount ) ;
2007-03-13 23:47:54 +03:00
return 0 ;
}
2007-10-16 00:15:53 +04:00
static int find_and_setup_root ( struct btrfs_root * tree_root ,
2007-03-20 21:38:32 +03:00
struct btrfs_fs_info * fs_info ,
u64 objectid ,
2007-03-22 19:13:20 +03:00
struct btrfs_root * root )
2007-03-13 23:47:54 +03:00
{
int ret ;
2007-10-16 00:15:53 +04:00
u32 blocksize ;
2008-10-29 21:49:05 +03:00
u64 generation ;
2007-03-13 23:47:54 +03:00
2007-10-16 00:15:53 +04:00
__setup_root ( tree_root - > nodesize , tree_root - > leafsize ,
2007-11-30 19:30:34 +03:00
tree_root - > sectorsize , tree_root - > stripesize ,
root , fs_info , objectid ) ;
2007-03-13 23:47:54 +03:00
ret = btrfs_find_last_root ( tree_root , objectid ,
& root - > root_item , & root - > root_key ) ;
2009-09-21 23:56:00 +04:00
if ( ret > 0 )
return - ENOENT ;
2007-03-13 23:47:54 +03:00
BUG_ON ( ret ) ;
2008-10-29 21:49:05 +03:00
generation = btrfs_root_generation ( & root - > root_item ) ;
2007-10-16 00:15:53 +04:00
blocksize = btrfs_level_size ( root , btrfs_root_level ( & root - > root_item ) ) ;
root - > node = read_tree_block ( root , btrfs_root_bytenr ( & root - > root_item ) ,
2008-10-29 21:49:05 +03:00
blocksize , generation ) ;
2007-03-13 23:47:54 +03:00
BUG_ON ( ! root - > node ) ;
2009-09-21 23:56:00 +04:00
root - > commit_root = btrfs_root_node ( root ) ;
2007-02-21 00:40:44 +03:00
return 0 ;
}
2009-01-21 20:54:03 +03:00
static struct btrfs_root * alloc_log_tree ( struct btrfs_trans_handle * trans ,
struct btrfs_fs_info * fs_info )
2007-04-09 18:42:37 +04:00
{
struct btrfs_root * root ;
struct btrfs_root * tree_root = fs_info - > tree_root ;
2009-01-21 20:54:03 +03:00
struct extent_buffer * leaf ;
2008-09-06 00:13:11 +04:00
root = kzalloc ( sizeof ( * root ) , GFP_NOFS ) ;
if ( ! root )
2009-01-21 20:54:03 +03:00
return ERR_PTR ( - ENOMEM ) ;
2008-09-06 00:13:11 +04:00
__setup_root ( tree_root - > nodesize , tree_root - > leafsize ,
tree_root - > sectorsize , tree_root - > stripesize ,
root , fs_info , BTRFS_TREE_LOG_OBJECTID ) ;
root - > root_key . objectid = BTRFS_TREE_LOG_OBJECTID ;
root - > root_key . type = BTRFS_ROOT_ITEM_KEY ;
root - > root_key . offset = BTRFS_TREE_LOG_OBJECTID ;
2009-01-21 20:54:03 +03:00
/*
* log trees do not get reference counted because they go away
* before a real commit is actually done . They do store pointers
* to file data extents , and those reference counts still get
* updated ( along with back refs to the log tree ) .
*/
2008-09-06 00:13:11 +04:00
root - > ref_cows = 0 ;
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
leaf = btrfs_alloc_free_block ( trans , root , root - > leafsize , 0 ,
BTRFS_TREE_LOG_OBJECTID , NULL , 0 , 0 , 0 ) ;
2009-01-21 20:54:03 +03:00
if ( IS_ERR ( leaf ) ) {
kfree ( root ) ;
return ERR_CAST ( leaf ) ;
}
2008-09-06 00:13:11 +04:00
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
memset_extent_buffer ( leaf , 0 , 0 , sizeof ( struct btrfs_header ) ) ;
btrfs_set_header_bytenr ( leaf , leaf - > start ) ;
btrfs_set_header_generation ( leaf , trans - > transid ) ;
btrfs_set_header_backref_rev ( leaf , BTRFS_MIXED_BACKREF_REV ) ;
btrfs_set_header_owner ( leaf , BTRFS_TREE_LOG_OBJECTID ) ;
2009-01-21 20:54:03 +03:00
root - > node = leaf ;
2008-09-06 00:13:11 +04:00
write_extent_buffer ( root - > node , root - > fs_info - > fsid ,
( unsigned long ) btrfs_header_fsid ( root - > node ) ,
BTRFS_FSID_SIZE ) ;
btrfs_mark_buffer_dirty ( root - > node ) ;
btrfs_tree_unlock ( root - > node ) ;
2009-01-21 20:54:03 +03:00
return root ;
}
int btrfs_init_log_root_tree ( struct btrfs_trans_handle * trans ,
struct btrfs_fs_info * fs_info )
{
struct btrfs_root * log_root ;
log_root = alloc_log_tree ( trans , fs_info ) ;
if ( IS_ERR ( log_root ) )
return PTR_ERR ( log_root ) ;
WARN_ON ( fs_info - > log_root_tree ) ;
fs_info - > log_root_tree = log_root ;
return 0 ;
}
int btrfs_add_log_tree ( struct btrfs_trans_handle * trans ,
struct btrfs_root * root )
{
struct btrfs_root * log_root ;
struct btrfs_inode_item * inode_item ;
log_root = alloc_log_tree ( trans , root - > fs_info ) ;
if ( IS_ERR ( log_root ) )
return PTR_ERR ( log_root ) ;
log_root - > last_trans = trans - > transid ;
log_root - > root_key . offset = root - > root_key . objectid ;
inode_item = & log_root - > root_item . inode ;
inode_item - > generation = cpu_to_le64 ( 1 ) ;
inode_item - > size = cpu_to_le64 ( 3 ) ;
inode_item - > nlink = cpu_to_le32 ( 1 ) ;
inode_item - > nbytes = cpu_to_le64 ( root - > leafsize ) ;
inode_item - > mode = cpu_to_le32 ( S_IFDIR | 0755 ) ;
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
btrfs_set_root_node ( & log_root - > root_item , log_root - > node ) ;
2009-01-21 20:54:03 +03:00
WARN_ON ( root - > log_root ) ;
root - > log_root = log_root ;
root - > log_transid = 0 ;
2009-10-13 21:21:08 +04:00
root - > last_log_commit = 0 ;
2008-09-06 00:13:11 +04:00
return 0 ;
}
struct btrfs_root * btrfs_read_fs_root_no_radix ( struct btrfs_root * tree_root ,
struct btrfs_key * location )
{
struct btrfs_root * root ;
struct btrfs_fs_info * fs_info = tree_root - > fs_info ;
2007-04-09 18:42:37 +04:00
struct btrfs_path * path ;
2007-10-16 00:14:19 +04:00
struct extent_buffer * l ;
2008-10-29 21:49:05 +03:00
u64 generation ;
2007-10-16 00:15:53 +04:00
u32 blocksize ;
2007-04-09 18:42:37 +04:00
int ret = 0 ;
2007-06-22 22:16:25 +04:00
root = kzalloc ( sizeof ( * root ) , GFP_NOFS ) ;
2007-06-09 17:22:25 +04:00
if ( ! root )
2007-04-09 18:42:37 +04:00
return ERR_PTR ( - ENOMEM ) ;
if ( location - > offset = = ( u64 ) - 1 ) {
2007-10-16 00:15:53 +04:00
ret = find_and_setup_root ( tree_root , fs_info ,
2007-04-09 18:42:37 +04:00
location - > objectid , root ) ;
if ( ret ) {
kfree ( root ) ;
return ERR_PTR ( ret ) ;
}
2009-09-21 23:56:00 +04:00
goto out ;
2007-04-09 18:42:37 +04:00
}
2007-10-16 00:15:53 +04:00
__setup_root ( tree_root - > nodesize , tree_root - > leafsize ,
2007-11-30 19:30:34 +03:00
tree_root - > sectorsize , tree_root - > stripesize ,
root , fs_info , location - > objectid ) ;
2007-04-09 18:42:37 +04:00
path = btrfs_alloc_path ( ) ;
BUG_ON ( ! path ) ;
ret = btrfs_search_slot ( NULL , tree_root , location , path , 0 , 0 ) ;
2009-09-21 23:56:00 +04:00
if ( ret = = 0 ) {
l = path - > nodes [ 0 ] ;
read_extent_buffer ( l , & root - > root_item ,
btrfs_item_ptr_offset ( l , path - > slots [ 0 ] ) ,
sizeof ( root - > root_item ) ) ;
memcpy ( & root - > root_key , location , sizeof ( * location ) ) ;
2007-04-09 18:42:37 +04:00
}
btrfs_free_path ( path ) ;
if ( ret ) {
2009-09-21 23:56:00 +04:00
if ( ret > 0 )
ret = - ENOENT ;
2007-04-09 18:42:37 +04:00
return ERR_PTR ( ret ) ;
}
2009-09-21 23:56:00 +04:00
2008-10-29 21:49:05 +03:00
generation = btrfs_root_generation ( & root - > root_item ) ;
2007-10-16 00:15:53 +04:00
blocksize = btrfs_level_size ( root , btrfs_root_level ( & root - > root_item ) ) ;
root - > node = read_tree_block ( root , btrfs_root_bytenr ( & root - > root_item ) ,
2008-10-29 21:49:05 +03:00
blocksize , generation ) ;
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
root - > commit_root = btrfs_root_node ( root ) ;
2007-04-09 18:42:37 +04:00
BUG_ON ( ! root - > node ) ;
2009-09-21 23:56:00 +04:00
out :
if ( location - > objectid ! = BTRFS_TREE_LOG_OBJECTID )
2008-09-06 00:13:11 +04:00
root - > ref_cows = 1 ;
2009-09-21 23:56:00 +04:00
2007-06-22 22:16:25 +04:00
return root ;
}
2008-01-08 23:46:30 +03:00
struct btrfs_root * btrfs_lookup_fs_root ( struct btrfs_fs_info * fs_info ,
u64 root_objectid )
{
struct btrfs_root * root ;
if ( root_objectid = = BTRFS_ROOT_TREE_OBJECTID )
return fs_info - > tree_root ;
if ( root_objectid = = BTRFS_EXTENT_TREE_OBJECTID )
return fs_info - > extent_root ;
root = radix_tree_lookup ( & fs_info - > fs_roots_radix ,
( unsigned long ) root_objectid ) ;
return root ;
}
2007-12-22 00:27:24 +03:00
struct btrfs_root * btrfs_read_fs_root_no_name ( struct btrfs_fs_info * fs_info ,
struct btrfs_key * location )
2007-06-22 22:16:25 +04:00
{
struct btrfs_root * root ;
int ret ;
2007-12-22 00:27:24 +03:00
if ( location - > objectid = = BTRFS_ROOT_TREE_OBJECTID )
return fs_info - > tree_root ;
if ( location - > objectid = = BTRFS_EXTENT_TREE_OBJECTID )
return fs_info - > extent_root ;
2008-04-26 00:53:30 +04:00
if ( location - > objectid = = BTRFS_CHUNK_TREE_OBJECTID )
return fs_info - > chunk_root ;
if ( location - > objectid = = BTRFS_DEV_TREE_OBJECTID )
return fs_info - > dev_root ;
2008-12-11 04:32:51 +03:00
if ( location - > objectid = = BTRFS_CSUM_TREE_OBJECTID )
return fs_info - > csum_root ;
2009-09-21 23:56:00 +04:00
again :
spin_lock ( & fs_info - > fs_roots_radix_lock ) ;
2007-06-22 22:16:25 +04:00
root = radix_tree_lookup ( & fs_info - > fs_roots_radix ,
( unsigned long ) location - > objectid ) ;
2009-09-21 23:56:00 +04:00
spin_unlock ( & fs_info - > fs_roots_radix_lock ) ;
2007-06-22 22:16:25 +04:00
if ( root )
return root ;
2008-09-06 00:13:11 +04:00
root = btrfs_read_fs_root_no_radix ( fs_info - > tree_root , location ) ;
2007-06-22 22:16:25 +04:00
if ( IS_ERR ( root ) )
return root ;
2008-11-18 04:42:26 +03:00
set_anon_super ( & root - > anon_super , NULL ) ;
2010-05-16 18:49:58 +04:00
if ( btrfs_root_refs ( & root - > root_item ) = = 0 ) {
ret = - ENOENT ;
goto fail ;
}
ret = btrfs_find_orphan_item ( fs_info - > tree_root , location - > objectid ) ;
if ( ret < 0 )
goto fail ;
if ( ret = = 0 )
root - > orphan_item_inserted = 1 ;
2009-09-21 23:56:00 +04:00
ret = radix_tree_preload ( GFP_NOFS & ~ __GFP_HIGHMEM ) ;
if ( ret )
goto fail ;
spin_lock ( & fs_info - > fs_roots_radix_lock ) ;
2007-04-11 00:58:11 +04:00
ret = radix_tree_insert ( & fs_info - > fs_roots_radix ,
( unsigned long ) root - > root_key . objectid ,
2007-04-09 18:42:37 +04:00
root ) ;
2010-05-16 18:49:58 +04:00
if ( ret = = 0 )
2009-09-21 23:56:00 +04:00
root - > in_radix = 1 ;
2010-05-16 18:49:58 +04:00
2009-09-21 23:56:00 +04:00
spin_unlock ( & fs_info - > fs_roots_radix_lock ) ;
radix_tree_preload_end ( ) ;
2007-04-09 18:42:37 +04:00
if ( ret ) {
2009-09-21 23:56:00 +04:00
if ( ret = = - EEXIST ) {
free_fs_root ( root ) ;
goto again ;
}
goto fail ;
2007-04-09 18:42:37 +04:00
}
2009-09-21 23:56:00 +04:00
ret = btrfs_find_dead_roots ( fs_info - > tree_root ,
root - > root_key . objectid ) ;
WARN_ON ( ret ) ;
2007-12-22 00:27:24 +03:00
return root ;
2009-09-21 23:56:00 +04:00
fail :
free_fs_root ( root ) ;
return ERR_PTR ( ret ) ;
2007-12-22 00:27:24 +03:00
}
struct btrfs_root * btrfs_read_fs_root ( struct btrfs_fs_info * fs_info ,
struct btrfs_key * location ,
const char * name , int namelen )
{
2009-09-21 23:56:00 +04:00
return btrfs_read_fs_root_no_name ( fs_info , location ) ;
#if 0
2007-12-22 00:27:24 +03:00
struct btrfs_root * root ;
int ret ;
root = btrfs_read_fs_root_no_name ( fs_info , location ) ;
if ( ! root )
return NULL ;
2007-08-29 23:47:34 +04:00
2008-01-03 17:08:48 +03:00
if ( root - > in_sysfs )
return root ;
2007-08-29 23:47:34 +04:00
ret = btrfs_set_root_name ( root , name , namelen ) ;
if ( ret ) {
2007-10-16 00:14:19 +04:00
free_extent_buffer ( root - > node ) ;
2007-08-29 23:47:34 +04:00
kfree ( root ) ;
return ERR_PTR ( ret ) ;
}
2009-09-21 23:56:00 +04:00
2007-08-29 23:47:34 +04:00
ret = btrfs_sysfs_add_root ( root ) ;
if ( ret ) {
2007-10-16 00:14:19 +04:00
free_extent_buffer ( root - > node ) ;
2007-08-29 23:47:34 +04:00
kfree ( root - > name ) ;
kfree ( root ) ;
return ERR_PTR ( ret ) ;
}
2008-01-03 17:08:48 +03:00
root - > in_sysfs = 1 ;
2007-04-09 18:42:37 +04:00
return root ;
2009-09-21 23:56:00 +04:00
# endif
2007-04-09 18:42:37 +04:00
}
2008-03-26 17:28:07 +03:00
static int btrfs_congested_fn ( void * congested_data , int bdi_bits )
{
struct btrfs_fs_info * info = ( struct btrfs_fs_info * ) congested_data ;
int ret = 0 ;
struct btrfs_device * device ;
struct backing_dev_info * bdi ;
2009-04-27 15:29:04 +04:00
2009-01-21 18:59:08 +03:00
list_for_each_entry ( device , & info - > fs_devices - > devices , dev_list ) {
2008-05-13 21:46:40 +04:00
if ( ! device - > bdev )
continue ;
2008-03-26 17:28:07 +03:00
bdi = blk_get_backing_dev_info ( device - > bdev ) ;
if ( bdi & & bdi_congested ( bdi , bdi_bits ) ) {
ret = 1 ;
break ;
}
}
return ret ;
}
2008-04-22 17:22:11 +04:00
/*
* this unplugs every device on the box , and it is only used when page
* is null
*/
static void __unplug_io_fn ( struct backing_dev_info * bdi , struct page * page )
{
struct btrfs_device * device ;
struct btrfs_fs_info * info ;
info = ( struct btrfs_fs_info * ) bdi - > unplug_io_data ;
2009-01-21 18:59:08 +03:00
list_for_each_entry ( device , & info - > fs_devices - > devices , dev_list ) {
Btrfs: move data checksumming into a dedicated tree
Btrfs stores checksums for each data block. Until now, they have
been stored in the subvolume trees, indexed by the inode that is
referencing the data block. This means that when we read the inode,
we've probably read in at least some checksums as well.
But, this has a few problems:
* The checksums are indexed by logical offset in the file. When
compression is on, this means we have to do the expensive checksumming
on the uncompressed data. It would be faster if we could checksum
the compressed data instead.
* If we implement encryption, we'll be checksumming the plain text and
storing that on disk. This is significantly less secure.
* For either compression or encryption, we have to get the plain text
back before we can verify the checksum as correct. This makes the raid
layer balancing and extent moving much more expensive.
* It makes the front end caching code more complex, as we have touch
the subvolume and inodes as we cache extents.
* There is potentitally one copy of the checksum in each subvolume
referencing an extent.
The solution used here is to store the extent checksums in a dedicated
tree. This allows us to index the checksums by phyiscal extent
start and length. It means:
* The checksum is against the data stored on disk, after any compression
or encryption is done.
* The checksum is stored in a central location, and can be verified without
following back references, or reading inodes.
This makes compression significantly faster by reducing the amount of
data that needs to be checksummed. It will also allow much faster
raid management code in general.
The checksums are indexed by a key with a fixed objectid (a magic value
in ctree.h) and offset set to the starting byte of the extent. This
allows us to copy the checksum items into the fsync log tree directly (or
any other tree), without having to invent a second format for them.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-12-09 00:58:54 +03:00
if ( ! device - > bdev )
continue ;
2008-04-22 17:22:11 +04:00
bdi = blk_get_backing_dev_info ( device - > bdev ) ;
2009-01-06 05:25:51 +03:00
if ( bdi - > unplug_io_fn )
2008-04-22 17:22:11 +04:00
bdi - > unplug_io_fn ( bdi , page ) ;
}
}
2008-12-02 17:54:17 +03:00
static void btrfs_unplug_io_fn ( struct backing_dev_info * bdi , struct page * page )
2008-03-26 17:28:07 +03:00
{
2008-04-22 17:22:11 +04:00
struct inode * inode ;
2008-04-21 18:03:05 +04:00
struct extent_map_tree * em_tree ;
struct extent_map * em ;
2008-04-22 21:26:47 +04:00
struct address_space * mapping ;
2008-04-22 17:22:11 +04:00
u64 offset ;
2008-04-22 21:26:47 +04:00
/* the generic O_DIRECT read code does this */
2008-11-18 18:31:22 +03:00
if ( 1 | | ! page ) {
2008-04-22 17:22:11 +04:00
__unplug_io_fn ( bdi , page ) ;
return ;
}
2008-04-22 21:26:47 +04:00
/*
* page - > mapping may change at any time . Get a consistent copy
* and use that for everything below
*/
smp_mb ( ) ;
mapping = page - > mapping ;
if ( ! mapping )
return ;
inode = mapping - > host ;
2008-11-10 21:08:31 +03:00
/*
* don ' t do the expensive searching for a small number of
* devices
*/
if ( BTRFS_I ( inode ) - > root - > fs_info - > fs_devices - > open_devices < = 2 ) {
__unplug_io_fn ( bdi , page ) ;
return ;
}
2008-04-22 17:22:11 +04:00
offset = page_offset ( page ) ;
2008-03-26 17:28:07 +03:00
2008-04-21 18:03:05 +04:00
em_tree = & BTRFS_I ( inode ) - > extent_tree ;
2009-09-03 00:24:52 +04:00
read_lock ( & em_tree - > lock ) ;
2008-04-21 18:03:05 +04:00
em = lookup_extent_mapping ( em_tree , offset , PAGE_CACHE_SIZE ) ;
2009-09-03 00:24:52 +04:00
read_unlock ( & em_tree - > lock ) ;
2008-07-24 17:41:53 +04:00
if ( ! em ) {
__unplug_io_fn ( bdi , page ) ;
2008-04-21 18:03:05 +04:00
return ;
2008-07-24 17:41:53 +04:00
}
2008-04-21 18:03:05 +04:00
2008-07-24 17:41:53 +04:00
if ( em - > block_start > = EXTENT_MAP_LAST_BYTE ) {
free_extent_map ( em ) ;
__unplug_io_fn ( bdi , page ) ;
return ;
}
2008-04-21 18:03:05 +04:00
offset = offset - em - > start ;
btrfs_unplug_page ( & BTRFS_I ( inode ) - > root - > fs_info - > mapping_tree ,
em - > block_start + offset , page ) ;
free_extent_map ( em ) ;
2008-03-26 17:28:07 +03:00
}
2009-06-12 16:43:40 +04:00
/*
* If this fails , caller must call bdi_destroy ( ) to get rid of the
* bdi again .
*/
2008-03-26 17:28:07 +03:00
static int setup_bdi ( struct btrfs_fs_info * info , struct backing_dev_info * bdi )
{
2009-06-12 16:43:40 +04:00
int err ;
bdi - > capabilities = BDI_CAP_MAP_COPY ;
2010-04-26 12:27:54 +04:00
err = bdi_setup_and_register ( bdi , " btrfs " , BDI_CAP_MAP_COPY ) ;
2009-06-12 16:43:40 +04:00
if ( err )
return err ;
2008-04-19 00:13:31 +04:00
bdi - > ra_pages = default_backing_dev_info . ra_pages ;
2008-03-26 17:28:07 +03:00
bdi - > unplug_io_fn = btrfs_unplug_io_fn ;
bdi - > unplug_io_data = info ;
bdi - > congested_fn = btrfs_congested_fn ;
bdi - > congested_data = info ;
return 0 ;
}
2008-04-10 00:28:12 +04:00
static int bio_ready_for_csum ( struct bio * bio )
{
u64 length = 0 ;
u64 buf_len = 0 ;
u64 start = 0 ;
struct page * page ;
struct extent_io_tree * io_tree = NULL ;
struct btrfs_fs_info * info = NULL ;
struct bio_vec * bvec ;
int i ;
int ret ;
bio_for_each_segment ( bvec , bio , i ) {
page = bvec - > bv_page ;
if ( page - > private = = EXTENT_PAGE_PRIVATE ) {
length + = bvec - > bv_len ;
continue ;
}
if ( ! page - > private ) {
length + = bvec - > bv_len ;
continue ;
}
length = bvec - > bv_len ;
buf_len = page - > private > > 2 ;
start = page_offset ( page ) + bvec - > bv_offset ;
io_tree = & BTRFS_I ( page - > mapping - > host ) - > io_tree ;
info = BTRFS_I ( page - > mapping - > host ) - > root - > fs_info ;
}
/* are we fully contained in this bio? */
if ( buf_len < = length )
return 1 ;
ret = extent_range_uptodate ( io_tree , start + length ,
start + buf_len - 1 ) ;
return ret ;
}
2008-06-12 00:50:36 +04:00
/*
* called by the kthread helper functions to finally call the bio end_io
* functions . This is where read checksum verification actually happens
*/
static void end_workqueue_fn ( struct btrfs_work * work )
2008-04-10 00:28:12 +04:00
{
struct bio * bio ;
2008-06-12 00:50:36 +04:00
struct end_io_wq * end_io_wq ;
struct btrfs_fs_info * fs_info ;
2008-04-10 00:28:12 +04:00
int error ;
2008-06-12 00:50:36 +04:00
end_io_wq = container_of ( work , struct end_io_wq , work ) ;
bio = end_io_wq - > bio ;
fs_info = end_io_wq - > info ;
2008-04-10 00:28:12 +04:00
2008-12-17 22:51:42 +03:00
/* metadata bio reads are special because the whole tree block must
2008-06-12 00:50:36 +04:00
* be checksummed at once . This makes sure the entire block is in
* ram and up to date before trying to verify things . For
* blocksize < = pagesize , it is basically a noop
*/
2008-12-17 22:51:42 +03:00
if ( ! ( bio - > bi_rw & ( 1 < < BIO_RW ) ) & & end_io_wq - > metadata & &
! bio_ready_for_csum ( bio ) ) {
Btrfs: move data checksumming into a dedicated tree
Btrfs stores checksums for each data block. Until now, they have
been stored in the subvolume trees, indexed by the inode that is
referencing the data block. This means that when we read the inode,
we've probably read in at least some checksums as well.
But, this has a few problems:
* The checksums are indexed by logical offset in the file. When
compression is on, this means we have to do the expensive checksumming
on the uncompressed data. It would be faster if we could checksum
the compressed data instead.
* If we implement encryption, we'll be checksumming the plain text and
storing that on disk. This is significantly less secure.
* For either compression or encryption, we have to get the plain text
back before we can verify the checksum as correct. This makes the raid
layer balancing and extent moving much more expensive.
* It makes the front end caching code more complex, as we have touch
the subvolume and inodes as we cache extents.
* There is potentitally one copy of the checksum in each subvolume
referencing an extent.
The solution used here is to store the extent checksums in a dedicated
tree. This allows us to index the checksums by phyiscal extent
start and length. It means:
* The checksum is against the data stored on disk, after any compression
or encryption is done.
* The checksum is stored in a central location, and can be verified without
following back references, or reading inodes.
This makes compression significantly faster by reducing the amount of
data that needs to be checksummed. It will also allow much faster
raid management code in general.
The checksums are indexed by a key with a fixed objectid (a magic value
in ctree.h) and offset set to the starting byte of the extent. This
allows us to copy the checksum items into the fsync log tree directly (or
any other tree), without having to invent a second format for them.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-12-09 00:58:54 +03:00
btrfs_queue_worker ( & fs_info - > endio_meta_workers ,
2008-06-12 00:50:36 +04:00
& end_io_wq - > work ) ;
return ;
}
error = end_io_wq - > error ;
bio - > bi_private = end_io_wq - > private ;
bio - > bi_end_io = end_io_wq - > end_io ;
kfree ( end_io_wq ) ;
bio_endio ( bio , error ) ;
2008-04-16 19:14:51 +04:00
}
2008-06-26 00:01:31 +04:00
static int cleaner_kthread ( void * arg )
{
struct btrfs_root * root = arg ;
do {
vfs_check_frozen ( root - > fs_info - > sb , SB_FREEZE_WRITE ) ;
2009-09-22 00:00:26 +04:00
if ( ! ( root - > fs_info - > sb - > s_flags & MS_RDONLY ) & &
mutex_trylock ( & root - > fs_info - > cleaner_mutex ) ) {
2009-11-12 12:36:34 +03:00
btrfs_run_delayed_iputs ( root ) ;
2009-09-22 00:00:26 +04:00
btrfs_clean_old_snapshots ( root ) ;
mutex_unlock ( & root - > fs_info - > cleaner_mutex ) ;
}
2008-06-26 00:01:31 +04:00
if ( freezing ( current ) ) {
refrigerator ( ) ;
} else {
set_current_state ( TASK_INTERRUPTIBLE ) ;
2010-05-16 18:49:58 +04:00
if ( ! kthread_should_stop ( ) )
schedule ( ) ;
2008-06-26 00:01:31 +04:00
__set_current_state ( TASK_RUNNING ) ;
}
} while ( ! kthread_should_stop ( ) ) ;
return 0 ;
}
static int transaction_kthread ( void * arg )
{
struct btrfs_root * root = arg ;
struct btrfs_trans_handle * trans ;
struct btrfs_transaction * cur ;
2010-05-16 18:49:58 +04:00
u64 transid ;
2008-06-26 00:01:31 +04:00
unsigned long now ;
unsigned long delay ;
int ret ;
do {
delay = HZ * 30 ;
vfs_check_frozen ( root - > fs_info - > sb , SB_FREEZE_WRITE ) ;
mutex_lock ( & root - > fs_info - > transaction_kthread_mutex ) ;
2010-05-16 18:49:58 +04:00
spin_lock ( & root - > fs_info - > new_trans_lock ) ;
2008-06-26 00:01:31 +04:00
cur = root - > fs_info - > running_transaction ;
if ( ! cur ) {
2010-05-16 18:49:58 +04:00
spin_unlock ( & root - > fs_info - > new_trans_lock ) ;
2008-06-26 00:01:31 +04:00
goto sleep ;
}
2008-07-28 23:32:19 +04:00
2008-06-26 00:01:31 +04:00
now = get_seconds ( ) ;
2010-05-16 18:49:58 +04:00
if ( ! cur - > blocked & &
( now < cur - > start_time | | now - cur - > start_time < 30 ) ) {
spin_unlock ( & root - > fs_info - > new_trans_lock ) ;
2008-06-26 00:01:31 +04:00
delay = HZ * 5 ;
goto sleep ;
}
2010-05-16 18:49:58 +04:00
transid = cur - > transid ;
spin_unlock ( & root - > fs_info - > new_trans_lock ) ;
2009-03-13 17:10:06 +03:00
2010-05-16 18:49:58 +04:00
trans = btrfs_join_transaction ( root , 1 ) ;
if ( transid = = trans - > transid ) {
ret = btrfs_commit_transaction ( trans , root ) ;
BUG_ON ( ret ) ;
} else {
btrfs_end_transaction ( trans , root ) ;
}
2008-06-26 00:01:31 +04:00
sleep :
wake_up_process ( root - > fs_info - > cleaner_kthread ) ;
mutex_unlock ( & root - > fs_info - > transaction_kthread_mutex ) ;
if ( freezing ( current ) ) {
refrigerator ( ) ;
} else {
set_current_state ( TASK_INTERRUPTIBLE ) ;
2010-05-16 18:49:58 +04:00
if ( ! kthread_should_stop ( ) & &
! btrfs_transaction_blocked ( root - > fs_info ) )
schedule_timeout ( delay ) ;
2008-06-26 00:01:31 +04:00
__set_current_state ( TASK_RUNNING ) ;
}
} while ( ! kthread_should_stop ( ) ) ;
return 0 ;
}
2008-03-24 22:02:07 +03:00
struct btrfs_root * open_ctree ( struct super_block * sb ,
2008-05-13 21:46:40 +04:00
struct btrfs_fs_devices * fs_devices ,
char * options )
2007-03-21 18:12:56 +03:00
{
2007-10-16 00:15:53 +04:00
u32 sectorsize ;
u32 nodesize ;
u32 leafsize ;
u32 blocksize ;
2007-11-30 19:30:34 +03:00
u32 stripesize ;
2008-10-29 21:49:05 +03:00
u64 generation ;
2008-12-02 14:36:08 +03:00
u64 features ;
2008-11-18 05:02:50 +03:00
struct btrfs_key location ;
2008-05-07 19:43:44 +04:00
struct buffer_head * bh ;
2008-09-06 00:13:11 +04:00
struct btrfs_root * extent_root = kzalloc ( sizeof ( struct btrfs_root ) ,
2007-03-22 19:13:20 +03:00
GFP_NOFS ) ;
Btrfs: move data checksumming into a dedicated tree
Btrfs stores checksums for each data block. Until now, they have
been stored in the subvolume trees, indexed by the inode that is
referencing the data block. This means that when we read the inode,
we've probably read in at least some checksums as well.
But, this has a few problems:
* The checksums are indexed by logical offset in the file. When
compression is on, this means we have to do the expensive checksumming
on the uncompressed data. It would be faster if we could checksum
the compressed data instead.
* If we implement encryption, we'll be checksumming the plain text and
storing that on disk. This is significantly less secure.
* For either compression or encryption, we have to get the plain text
back before we can verify the checksum as correct. This makes the raid
layer balancing and extent moving much more expensive.
* It makes the front end caching code more complex, as we have touch
the subvolume and inodes as we cache extents.
* There is potentitally one copy of the checksum in each subvolume
referencing an extent.
The solution used here is to store the extent checksums in a dedicated
tree. This allows us to index the checksums by phyiscal extent
start and length. It means:
* The checksum is against the data stored on disk, after any compression
or encryption is done.
* The checksum is stored in a central location, and can be verified without
following back references, or reading inodes.
This makes compression significantly faster by reducing the amount of
data that needs to be checksummed. It will also allow much faster
raid management code in general.
The checksums are indexed by a key with a fixed objectid (a magic value
in ctree.h) and offset set to the starting byte of the extent. This
allows us to copy the checksum items into the fsync log tree directly (or
any other tree), without having to invent a second format for them.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-12-09 00:58:54 +03:00
struct btrfs_root * csum_root = kzalloc ( sizeof ( struct btrfs_root ) ,
GFP_NOFS ) ;
2008-09-06 00:13:11 +04:00
struct btrfs_root * tree_root = kzalloc ( sizeof ( struct btrfs_root ) ,
2007-03-22 19:13:20 +03:00
GFP_NOFS ) ;
2008-04-04 00:29:03 +04:00
struct btrfs_fs_info * fs_info = kzalloc ( sizeof ( * fs_info ) ,
2007-03-22 19:13:20 +03:00
GFP_NOFS ) ;
2008-09-06 00:13:11 +04:00
struct btrfs_root * chunk_root = kzalloc ( sizeof ( struct btrfs_root ) ,
2008-03-24 22:01:56 +03:00
GFP_NOFS ) ;
2008-09-06 00:13:11 +04:00
struct btrfs_root * dev_root = kzalloc ( sizeof ( struct btrfs_root ) ,
2008-03-24 22:01:56 +03:00
GFP_NOFS ) ;
2008-09-06 00:13:11 +04:00
struct btrfs_root * log_tree_root ;
2007-02-02 17:18:22 +03:00
int ret ;
2008-04-01 19:21:34 +04:00
int err = - EINVAL ;
2008-06-12 05:47:56 +04:00
2007-04-02 18:50:19 +04:00
struct btrfs_super_block * disk_super ;
2008-04-04 00:29:03 +04:00
2008-10-02 03:09:04 +04:00
if ( ! extent_root | | ! tree_root | | ! fs_info | |
Btrfs: move data checksumming into a dedicated tree
Btrfs stores checksums for each data block. Until now, they have
been stored in the subvolume trees, indexed by the inode that is
referencing the data block. This means that when we read the inode,
we've probably read in at least some checksums as well.
But, this has a few problems:
* The checksums are indexed by logical offset in the file. When
compression is on, this means we have to do the expensive checksumming
on the uncompressed data. It would be faster if we could checksum
the compressed data instead.
* If we implement encryption, we'll be checksumming the plain text and
storing that on disk. This is significantly less secure.
* For either compression or encryption, we have to get the plain text
back before we can verify the checksum as correct. This makes the raid
layer balancing and extent moving much more expensive.
* It makes the front end caching code more complex, as we have touch
the subvolume and inodes as we cache extents.
* There is potentitally one copy of the checksum in each subvolume
referencing an extent.
The solution used here is to store the extent checksums in a dedicated
tree. This allows us to index the checksums by phyiscal extent
start and length. It means:
* The checksum is against the data stored on disk, after any compression
or encryption is done.
* The checksum is stored in a central location, and can be verified without
following back references, or reading inodes.
This makes compression significantly faster by reducing the amount of
data that needs to be checksummed. It will also allow much faster
raid management code in general.
The checksums are indexed by a key with a fixed objectid (a magic value
in ctree.h) and offset set to the starting byte of the extent. This
allows us to copy the checksum items into the fsync log tree directly (or
any other tree), without having to invent a second format for them.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-12-09 00:58:54 +03:00
! chunk_root | | ! dev_root | | ! csum_root ) {
2007-06-12 14:35:45 +04:00
err = - ENOMEM ;
goto fail ;
}
2009-09-22 00:00:26 +04:00
ret = init_srcu_struct ( & fs_info - > subvol_srcu ) ;
if ( ret ) {
err = ret ;
goto fail ;
}
ret = setup_bdi ( fs_info , & fs_info - > bdi ) ;
if ( ret ) {
err = ret ;
goto fail_srcu ;
}
fs_info - > btree_inode = new_inode ( sb ) ;
if ( ! fs_info - > btree_inode ) {
err = - ENOMEM ;
goto fail_bdi ;
}
INIT_RADIX_TREE ( & fs_info - > fs_roots_radix , GFP_ATOMIC ) ;
2007-04-20 05:01:03 +04:00
INIT_LIST_HEAD ( & fs_info - > trans_list ) ;
2007-06-09 02:11:48 +04:00
INIT_LIST_HEAD ( & fs_info - > dead_roots ) ;
2009-11-12 12:36:34 +03:00
INIT_LIST_HEAD ( & fs_info - > delayed_iputs ) ;
2007-10-16 00:19:22 +04:00
INIT_LIST_HEAD ( & fs_info - > hashers ) ;
2008-08-05 07:17:27 +04:00
INIT_LIST_HEAD ( & fs_info - > delalloc_inodes ) ;
2009-03-31 21:27:11 +04:00
INIT_LIST_HEAD ( & fs_info - > ordered_operations ) ;
2009-09-12 00:11:19 +04:00
INIT_LIST_HEAD ( & fs_info - > caching_block_groups ) ;
2007-12-22 00:27:21 +03:00
spin_lock_init ( & fs_info - > delalloc_lock ) ;
2008-01-15 16:40:48 +03:00
spin_lock_init ( & fs_info - > new_trans_lock ) ;
2008-07-28 23:32:19 +04:00
spin_lock_init ( & fs_info - > ref_cache_lock ) ;
2009-09-22 00:00:26 +04:00
spin_lock_init ( & fs_info - > fs_roots_radix_lock ) ;
2009-11-12 12:36:34 +03:00
spin_lock_init ( & fs_info - > delayed_iput_lock ) ;
2007-10-16 00:19:22 +04:00
2007-08-29 23:47:34 +04:00
init_completion ( & fs_info - > kobj_unregister ) ;
2007-03-20 21:38:32 +03:00
fs_info - > tree_root = tree_root ;
fs_info - > extent_root = extent_root ;
Btrfs: move data checksumming into a dedicated tree
Btrfs stores checksums for each data block. Until now, they have
been stored in the subvolume trees, indexed by the inode that is
referencing the data block. This means that when we read the inode,
we've probably read in at least some checksums as well.
But, this has a few problems:
* The checksums are indexed by logical offset in the file. When
compression is on, this means we have to do the expensive checksumming
on the uncompressed data. It would be faster if we could checksum
the compressed data instead.
* If we implement encryption, we'll be checksumming the plain text and
storing that on disk. This is significantly less secure.
* For either compression or encryption, we have to get the plain text
back before we can verify the checksum as correct. This makes the raid
layer balancing and extent moving much more expensive.
* It makes the front end caching code more complex, as we have touch
the subvolume and inodes as we cache extents.
* There is potentitally one copy of the checksum in each subvolume
referencing an extent.
The solution used here is to store the extent checksums in a dedicated
tree. This allows us to index the checksums by phyiscal extent
start and length. It means:
* The checksum is against the data stored on disk, after any compression
or encryption is done.
* The checksum is stored in a central location, and can be verified without
following back references, or reading inodes.
This makes compression significantly faster by reducing the amount of
data that needs to be checksummed. It will also allow much faster
raid management code in general.
The checksums are indexed by a key with a fixed objectid (a magic value
in ctree.h) and offset set to the starting byte of the extent. This
allows us to copy the checksum items into the fsync log tree directly (or
any other tree), without having to invent a second format for them.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-12-09 00:58:54 +03:00
fs_info - > csum_root = csum_root ;
2008-03-24 22:01:56 +03:00
fs_info - > chunk_root = chunk_root ;
fs_info - > dev_root = dev_root ;
2008-03-24 22:02:07 +03:00
fs_info - > fs_devices = fs_devices ;
2008-03-24 22:01:56 +03:00
INIT_LIST_HEAD ( & fs_info - > dirty_cowonly_roots ) ;
2008-03-24 22:01:59 +03:00
INIT_LIST_HEAD ( & fs_info - > space_info ) ;
2008-03-24 22:01:56 +03:00
btrfs_mapping_init ( & fs_info - > mapping_tree ) ;
2010-05-16 18:46:25 +04:00
btrfs_init_block_rsv ( & fs_info - > global_block_rsv ) ;
btrfs_init_block_rsv ( & fs_info - > delalloc_block_rsv ) ;
btrfs_init_block_rsv ( & fs_info - > trans_block_rsv ) ;
btrfs_init_block_rsv ( & fs_info - > chunk_block_rsv ) ;
btrfs_init_block_rsv ( & fs_info - > empty_block_rsv ) ;
INIT_LIST_HEAD ( & fs_info - > durable_block_rsv_list ) ;
mutex_init ( & fs_info - > durable_block_rsv_mutex ) ;
2008-05-16 00:15:45 +04:00
atomic_set ( & fs_info - > nr_async_submits , 0 ) ;
2008-11-07 06:02:51 +03:00
atomic_set ( & fs_info - > async_delalloc_pages , 0 ) ;
2008-09-29 19:19:10 +04:00
atomic_set ( & fs_info - > async_submit_draining , 0 ) ;
2008-08-15 23:34:15 +04:00
atomic_set ( & fs_info - > nr_async_bios , 0 ) ;
2007-03-22 19:13:20 +03:00
fs_info - > sb = sb ;
2008-01-30 00:03:38 +03:00
fs_info - > max_inline = 8192 * 1024 ;
2009-09-12 00:12:44 +04:00
fs_info - > metadata_ratio = 0 ;
Btrfs: Add zlib compression support
This is a large change for adding compression on reading and writing,
both for inline and regular extents. It does some fairly large
surgery to the writeback paths.
Compression is off by default and enabled by mount -o compress. Even
when the -o compress mount option is not used, it is possible to read
compressed extents off the disk.
If compression for a given set of pages fails to make them smaller, the
file is flagged to avoid future compression attempts later.
* While finding delalloc extents, the pages are locked before being sent down
to the delalloc handler. This allows the delalloc handler to do complex things
such as cleaning the pages, marking them writeback and starting IO on their
behalf.
* Inline extents are inserted at delalloc time now. This allows us to compress
the data before inserting the inline extent, and it allows us to insert
an inline extent that spans multiple pages.
* All of the in-memory extent representations (extent_map.c, ordered-data.c etc)
are changed to record both an in-memory size and an on disk size, as well
as a flag for compression.
From a disk format point of view, the extent pointers in the file are changed
to record the on disk size of a given extent and some encoding flags.
Space in the disk format is allocated for compression encoding, as well
as encryption and a generic 'other' field. Neither the encryption or the
'other' field are currently used.
In order to limit the amount of data read for a single random read in the
file, the size of a compressed extent is limited to 128k. This is a
software only limit, the disk format supports u64 sized compressed extents.
In order to limit the ram consumed while processing extents, the uncompressed
size of a compressed extent is limited to 256k. This is a software only limit
and will be subject to tuning later.
Checksumming is still done on compressed extents, and it is done on the
uncompressed version of the data. This way additional encodings can be
layered on without having to figure out which encoding to checksum.
Compression happens at delalloc time, which is basically singled threaded because
it is usually done by a single pdflush thread. This makes it tricky to
spread the compression load across all the cpus on the box. We'll have to
look at parallel pdflush walks of dirty inodes at a later time.
Decompression is hooked into readpages and it does spread across CPUs nicely.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-10-29 21:49:59 +03:00
2008-12-19 23:43:22 +03:00
fs_info - > thread_pool_size = min_t ( unsigned long ,
num_online_cpus ( ) + 2 , 8 ) ;
2008-04-18 22:17:20 +04:00
2008-07-24 19:57:52 +04:00
INIT_LIST_HEAD ( & fs_info - > ordered_extents ) ;
spin_lock_init ( & fs_info - > ordered_extent_lock ) ;
2008-05-07 19:43:44 +04:00
sb - > s_blocksize = 4096 ;
sb - > s_blocksize_bits = blksize_bits ( 4096 ) ;
2009-09-16 17:02:33 +04:00
sb - > s_bdi = & fs_info - > bdi ;
2008-05-07 19:43:44 +04:00
2009-09-22 00:00:26 +04:00
fs_info - > btree_inode - > i_ino = BTRFS_BTREE_INODE_OBJECTID ;
fs_info - > btree_inode - > i_nlink = 1 ;
2008-04-18 22:17:20 +04:00
/*
* we set the i_size on the btree inode to the max possible int .
* the real end of the address space is determined by all of
* the devices in the system
*/
fs_info - > btree_inode - > i_size = OFFSET_MAX ;
2007-03-28 21:57:48 +04:00
fs_info - > btree_inode - > i_mapping - > a_ops = & btree_aops ;
2008-03-26 17:28:07 +03:00
fs_info - > btree_inode - > i_mapping - > backing_dev_info = & fs_info - > bdi ;
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
RB_CLEAR_NODE ( & BTRFS_I ( fs_info - > btree_inode ) - > rb_node ) ;
2008-01-25 00:13:08 +03:00
extent_io_tree_init ( & BTRFS_I ( fs_info - > btree_inode ) - > io_tree ,
2007-10-16 00:14:19 +04:00
fs_info - > btree_inode - > i_mapping ,
GFP_NOFS ) ;
2008-01-25 00:13:08 +03:00
extent_map_tree_init ( & BTRFS_I ( fs_info - > btree_inode ) - > extent_tree ,
GFP_NOFS ) ;
BTRFS_I ( fs_info - > btree_inode ) - > io_tree . ops = & btree_extent_io_ops ;
2007-11-08 05:08:01 +03:00
2009-09-22 00:00:26 +04:00
BTRFS_I ( fs_info - > btree_inode ) - > root = tree_root ;
memset ( & BTRFS_I ( fs_info - > btree_inode ) - > location , 0 ,
sizeof ( struct btrfs_key ) ) ;
BTRFS_I ( fs_info - > btree_inode ) - > dummy_inode = 1 ;
2009-09-24 17:24:43 +04:00
insert_inode_hash ( fs_info - > btree_inode ) ;
2009-09-22 00:00:26 +04:00
Btrfs: free space accounting redo
1) replace the per fs_info extent_io_tree that tracked free space with two
rb-trees per block group to track free space areas via offset and size. The
reason to do this is because most allocations come with a hint byte where to
start, so we can usually find a chunk of free space at that hint byte to satisfy
the allocation and get good space packing. If we cannot find free space at or
after the given offset we fall back on looking for a chunk of the given size as
close to that given offset as possible. When we fall back on the size search we
also try to find a slot as close to the size we want as possible, to avoid
breaking small chunks off of huge areas if possible.
2) remove the extent_io_tree that tracked the block group cache from fs_info and
replaced it with an rb-tree thats tracks block group cache via offset. also
added a per space_info list that tracks the block group cache for the particular
space so we can lookup related block groups easily.
3) cleaned up the allocation code to make it a little easier to read and a
little less complicated. Basically there are 3 steps, first look from our
provided hint. If we couldn't find from that given hint, start back at our
original search start and look for space from there. If that fails try to
allocate space if we can and start looking again. If not we're screwed and need
to start over again.
4) small fixes. there were some issues in volumes.c where we wouldn't allocate
the rest of the disk. fixed cow_file_range to actually pass the alloc_hint,
which has helped a good bit in making the fs_mark test I run have semi-normal
results as we run out of space. Generally with data allocations we don't track
where we last allocated from, so everytime we did a data allocation we'd search
through every block group that we have looking for free space. Now searching a
block group with no free space isn't terribly time consuming, it was causing a
slight degradation as we got more data block groups. The alloc_hint has fixed
this slight degredation and made things semi-normal.
There is still one nagging problem I'm working on where we will get ENOSPC when
there is definitely plenty of space. This only happens with metadata
allocations, and only when we are almost full. So you generally hit the 85%
mark first, but sometimes you'll hit the BUG before you hit the 85% wall. I'm
still tracking it down, but until then this seems to be pretty stable and make a
significant performance gain.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-09-23 21:14:11 +04:00
spin_lock_init ( & fs_info - > block_group_cache_lock ) ;
2010-02-23 22:43:04 +03:00
fs_info - > block_group_cache_tree = RB_ROOT ;
Btrfs: free space accounting redo
1) replace the per fs_info extent_io_tree that tracked free space with two
rb-trees per block group to track free space areas via offset and size. The
reason to do this is because most allocations come with a hint byte where to
start, so we can usually find a chunk of free space at that hint byte to satisfy
the allocation and get good space packing. If we cannot find free space at or
after the given offset we fall back on looking for a chunk of the given size as
close to that given offset as possible. When we fall back on the size search we
also try to find a slot as close to the size we want as possible, to avoid
breaking small chunks off of huge areas if possible.
2) remove the extent_io_tree that tracked the block group cache from fs_info and
replaced it with an rb-tree thats tracks block group cache via offset. also
added a per space_info list that tracks the block group cache for the particular
space so we can lookup related block groups easily.
3) cleaned up the allocation code to make it a little easier to read and a
little less complicated. Basically there are 3 steps, first look from our
provided hint. If we couldn't find from that given hint, start back at our
original search start and look for space from there. If that fails try to
allocate space if we can and start looking again. If not we're screwed and need
to start over again.
4) small fixes. there were some issues in volumes.c where we wouldn't allocate
the rest of the disk. fixed cow_file_range to actually pass the alloc_hint,
which has helped a good bit in making the fs_mark test I run have semi-normal
results as we run out of space. Generally with data allocations we don't track
where we last allocated from, so everytime we did a data allocation we'd search
through every block group that we have looking for free space. Now searching a
block group with no free space isn't terribly time consuming, it was causing a
slight degradation as we got more data block groups. The alloc_hint has fixed
this slight degredation and made things semi-normal.
There is still one nagging problem I'm working on where we will get ENOSPC when
there is definitely plenty of space. This only happens with metadata
allocations, and only when we are almost full. So you generally hit the 85%
mark first, but sometimes you'll hit the BUG before you hit the 85% wall. I'm
still tracking it down, but until then this seems to be pretty stable and make a
significant performance gain.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-09-23 21:14:11 +04:00
2009-09-12 00:11:19 +04:00
extent_io_tree_init ( & fs_info - > freed_extents [ 0 ] ,
2007-10-16 00:15:26 +04:00
fs_info - > btree_inode - > i_mapping , GFP_NOFS ) ;
2009-09-12 00:11:19 +04:00
extent_io_tree_init ( & fs_info - > freed_extents [ 1 ] ,
fs_info - > btree_inode - > i_mapping , GFP_NOFS ) ;
fs_info - > pinned_extents = & fs_info - > freed_extents [ 0 ] ;
2007-04-20 21:16:02 +04:00
fs_info - > do_barriers = 1 ;
2008-01-18 18:54:22 +03:00
2007-06-12 14:35:45 +04:00
2007-03-22 22:59:16 +03:00
mutex_init ( & fs_info - > trans_mutex ) ;
2009-03-31 21:27:11 +04:00
mutex_init ( & fs_info - > ordered_operations_mutex ) ;
2008-09-06 00:13:11 +04:00
mutex_init ( & fs_info - > tree_log_mutex ) ;
2008-06-26 00:01:30 +04:00
mutex_init ( & fs_info - > chunk_mutex ) ;
2008-06-26 00:01:31 +04:00
mutex_init ( & fs_info - > transaction_kthread_mutex ) ;
mutex_init ( & fs_info - > cleaner_mutex ) ;
2008-07-08 22:19:17 +04:00
mutex_init ( & fs_info - > volume_mutex ) ;
2009-07-30 17:40:40 +04:00
init_rwsem ( & fs_info - > extent_commit_sem ) ;
2009-11-12 12:34:40 +03:00
init_rwsem ( & fs_info - > cleanup_work_sem ) ;
2009-09-22 00:00:26 +04:00
init_rwsem ( & fs_info - > subvol_sem ) ;
2009-04-03 17:47:43 +04:00
btrfs_init_free_cluster ( & fs_info - > meta_alloc_cluster ) ;
btrfs_init_free_cluster ( & fs_info - > data_alloc_cluster ) ;
2008-07-17 20:53:50 +04:00
init_waitqueue_head ( & fs_info - > transaction_throttle ) ;
2008-07-17 20:54:14 +04:00
init_waitqueue_head ( & fs_info - > transaction_wait ) ;
2008-08-15 23:34:17 +04:00
init_waitqueue_head ( & fs_info - > async_submit_wait ) ;
2007-03-13 23:47:54 +03:00
2008-03-24 22:01:56 +03:00
__setup_root ( 4096 , 4096 , 4096 , 4096 , tree_root ,
2007-04-02 18:50:19 +04:00
fs_info , BTRFS_ROOT_TREE_OBJECTID ) ;
2007-04-11 23:53:25 +04:00
2007-03-28 21:57:48 +04:00
2008-12-09 00:46:26 +03:00
bh = btrfs_read_dev_super ( fs_devices - > latest_bdev ) ;
2008-05-07 19:43:44 +04:00
if ( ! bh )
2007-06-12 14:35:45 +04:00
goto fail_iput ;
2008-05-07 19:43:44 +04:00
memcpy ( & fs_info - > super_copy , bh - > b_data , sizeof ( fs_info - > super_copy ) ) ;
2009-01-05 23:43:42 +03:00
memcpy ( & fs_info - > super_for_commit , & fs_info - > super_copy ,
sizeof ( fs_info - > super_for_commit ) ) ;
2008-05-07 19:43:44 +04:00
brelse ( bh ) ;
2007-10-16 00:14:19 +04:00
2008-05-07 19:43:44 +04:00
memcpy ( fs_info - > fsid , fs_info - > super_copy . fsid , BTRFS_FSID_SIZE ) ;
2008-03-24 22:01:56 +03:00
2007-10-16 00:14:19 +04:00
disk_super = & fs_info - > super_copy ;
2007-04-09 18:42:37 +04:00
if ( ! btrfs_super_root ( disk_super ) )
2008-12-02 14:36:10 +03:00
goto fail_iput ;
2007-04-09 18:42:37 +04:00
2008-11-18 05:11:30 +03:00
ret = btrfs_parse_options ( tree_root , options ) ;
if ( ret ) {
err = ret ;
2008-12-02 14:36:10 +03:00
goto fail_iput ;
2008-11-18 05:11:30 +03:00
}
2008-05-13 21:46:40 +04:00
2008-12-02 14:36:08 +03:00
features = btrfs_super_incompat_flags ( disk_super ) &
~ BTRFS_FEATURE_INCOMPAT_SUPP ;
if ( features ) {
printk ( KERN_ERR " BTRFS: couldn't mount because of "
" unsupported optional features (%Lx). \n " ,
2009-04-21 23:38:29 +04:00
( unsigned long long ) features ) ;
2008-12-02 14:36:08 +03:00
err = - EINVAL ;
2008-12-02 14:36:10 +03:00
goto fail_iput ;
2008-12-02 14:36:08 +03:00
}
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
features = btrfs_super_incompat_flags ( disk_super ) ;
if ( ! ( features & BTRFS_FEATURE_INCOMPAT_MIXED_BACKREF ) ) {
features | = BTRFS_FEATURE_INCOMPAT_MIXED_BACKREF ;
btrfs_set_super_incompat_flags ( disk_super , features ) ;
}
2008-12-02 14:36:08 +03:00
features = btrfs_super_compat_ro_flags ( disk_super ) &
~ BTRFS_FEATURE_COMPAT_RO_SUPP ;
if ( ! ( sb - > s_flags & MS_RDONLY ) & & features ) {
printk ( KERN_ERR " BTRFS: couldn't mount RDWR because of "
" unsupported option features (%Lx). \n " ,
2009-04-21 23:38:29 +04:00
( unsigned long long ) features ) ;
2008-12-02 14:36:08 +03:00
err = - EINVAL ;
2008-12-02 14:36:10 +03:00
goto fail_iput ;
2008-12-02 14:36:08 +03:00
}
2009-10-03 03:11:56 +04:00
btrfs_init_workers ( & fs_info - > generic_worker ,
" genwork " , 1 , NULL ) ;
2008-08-15 23:34:16 +04:00
btrfs_init_workers ( & fs_info - > workers , " worker " ,
2009-10-03 03:11:56 +04:00
fs_info - > thread_pool_size ,
& fs_info - > generic_worker ) ;
Btrfs: Add zlib compression support
This is a large change for adding compression on reading and writing,
both for inline and regular extents. It does some fairly large
surgery to the writeback paths.
Compression is off by default and enabled by mount -o compress. Even
when the -o compress mount option is not used, it is possible to read
compressed extents off the disk.
If compression for a given set of pages fails to make them smaller, the
file is flagged to avoid future compression attempts later.
* While finding delalloc extents, the pages are locked before being sent down
to the delalloc handler. This allows the delalloc handler to do complex things
such as cleaning the pages, marking them writeback and starting IO on their
behalf.
* Inline extents are inserted at delalloc time now. This allows us to compress
the data before inserting the inline extent, and it allows us to insert
an inline extent that spans multiple pages.
* All of the in-memory extent representations (extent_map.c, ordered-data.c etc)
are changed to record both an in-memory size and an on disk size, as well
as a flag for compression.
From a disk format point of view, the extent pointers in the file are changed
to record the on disk size of a given extent and some encoding flags.
Space in the disk format is allocated for compression encoding, as well
as encryption and a generic 'other' field. Neither the encryption or the
'other' field are currently used.
In order to limit the amount of data read for a single random read in the
file, the size of a compressed extent is limited to 128k. This is a
software only limit, the disk format supports u64 sized compressed extents.
In order to limit the ram consumed while processing extents, the uncompressed
size of a compressed extent is limited to 256k. This is a software only limit
and will be subject to tuning later.
Checksumming is still done on compressed extents, and it is done on the
uncompressed version of the data. This way additional encodings can be
layered on without having to figure out which encoding to checksum.
Compression happens at delalloc time, which is basically singled threaded because
it is usually done by a single pdflush thread. This makes it tricky to
spread the compression load across all the cpus on the box. We'll have to
look at parallel pdflush walks of dirty inodes at a later time.
Decompression is hooked into readpages and it does spread across CPUs nicely.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-10-29 21:49:59 +03:00
2008-11-07 06:02:51 +03:00
btrfs_init_workers ( & fs_info - > delalloc_workers , " delalloc " ,
2009-10-03 03:11:56 +04:00
fs_info - > thread_pool_size ,
& fs_info - > generic_worker ) ;
2008-11-07 06:02:51 +03:00
2008-08-15 23:34:16 +04:00
btrfs_init_workers ( & fs_info - > submit_workers , " submit " ,
2008-08-15 23:34:14 +04:00
min_t ( u64 , fs_devices - > num_devices ,
2009-10-03 03:11:56 +04:00
fs_info - > thread_pool_size ) ,
& fs_info - > generic_worker ) ;
2008-07-31 23:42:53 +04:00
/* a higher idle thresh on the submit workers makes it much more
* likely that bios will be send down in a sane order to the
* devices
*/
fs_info - > submit_workers . idle_thresh = 64 ;
2008-08-15 23:34:18 +04:00
2008-11-07 06:02:51 +03:00
fs_info - > workers . idle_thresh = 16 ;
Btrfs: Add ordered async work queues
Btrfs uses kernel threads to create async work queues for cpu intensive
operations such as checksumming and decompression. These work well,
but they make it difficult to keep IO order intact.
A single writepages call from pdflush or fsync will turn into a number
of bios, and each bio is checksummed in parallel. Once the checksum is
computed, the bio is sent down to the disk, and since we don't control
the order in which the parallel operations happen, they might go down to
the disk in almost any order.
The code deals with this somewhat by having deep work queues for a single
kernel thread, making it very likely that a single thread will process all
the bios for a single inode.
This patch introduces an explicitly ordered work queue. As work structs
are placed into the queue they are put onto the tail of a list. They have
three callbacks:
->func (cpu intensive processing here)
->ordered_func (order sensitive processing here)
->ordered_free (free the work struct, all processing is done)
The work struct has three callbacks. The func callback does the cpu intensive
work, and when it completes the work struct is marked as done.
Every time a work struct completes, the list is checked to see if the head
is marked as done. If so the ordered_func callback is used to do the
order sensitive processing and the ordered_free callback is used to do
any cleanup. Then we loop back and check the head of the list again.
This patch also changes the checksumming code to use the ordered workqueues.
One a 4 drive array, it increases streaming writes from 280MB/s to 350MB/s.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-11-07 06:03:00 +03:00
fs_info - > workers . ordered = 1 ;
2008-07-31 23:42:53 +04:00
2008-11-07 06:02:51 +03:00
fs_info - > delalloc_workers . idle_thresh = 2 ;
fs_info - > delalloc_workers . ordered = 1 ;
2009-10-03 03:11:56 +04:00
btrfs_init_workers ( & fs_info - > fixup_workers , " fixup " , 1 ,
& fs_info - > generic_worker ) ;
2008-08-15 23:34:16 +04:00
btrfs_init_workers ( & fs_info - > endio_workers , " endio " ,
2009-10-03 03:11:56 +04:00
fs_info - > thread_pool_size ,
& fs_info - > generic_worker ) ;
Btrfs: move data checksumming into a dedicated tree
Btrfs stores checksums for each data block. Until now, they have
been stored in the subvolume trees, indexed by the inode that is
referencing the data block. This means that when we read the inode,
we've probably read in at least some checksums as well.
But, this has a few problems:
* The checksums are indexed by logical offset in the file. When
compression is on, this means we have to do the expensive checksumming
on the uncompressed data. It would be faster if we could checksum
the compressed data instead.
* If we implement encryption, we'll be checksumming the plain text and
storing that on disk. This is significantly less secure.
* For either compression or encryption, we have to get the plain text
back before we can verify the checksum as correct. This makes the raid
layer balancing and extent moving much more expensive.
* It makes the front end caching code more complex, as we have touch
the subvolume and inodes as we cache extents.
* There is potentitally one copy of the checksum in each subvolume
referencing an extent.
The solution used here is to store the extent checksums in a dedicated
tree. This allows us to index the checksums by phyiscal extent
start and length. It means:
* The checksum is against the data stored on disk, after any compression
or encryption is done.
* The checksum is stored in a central location, and can be verified without
following back references, or reading inodes.
This makes compression significantly faster by reducing the amount of
data that needs to be checksummed. It will also allow much faster
raid management code in general.
The checksums are indexed by a key with a fixed objectid (a magic value
in ctree.h) and offset set to the starting byte of the extent. This
allows us to copy the checksum items into the fsync log tree directly (or
any other tree), without having to invent a second format for them.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-12-09 00:58:54 +03:00
btrfs_init_workers ( & fs_info - > endio_meta_workers , " endio-meta " ,
2009-10-03 03:11:56 +04:00
fs_info - > thread_pool_size ,
& fs_info - > generic_worker ) ;
2008-12-17 22:51:42 +03:00
btrfs_init_workers ( & fs_info - > endio_meta_write_workers ,
2009-10-03 03:11:56 +04:00
" endio-meta-write " , fs_info - > thread_pool_size ,
& fs_info - > generic_worker ) ;
2008-08-15 23:34:16 +04:00
btrfs_init_workers ( & fs_info - > endio_write_workers , " endio-write " ,
2009-10-03 03:11:56 +04:00
fs_info - > thread_pool_size ,
& fs_info - > generic_worker ) ;
2008-07-31 23:42:53 +04:00
/*
* endios are largely parallel and should have a very
* low idle thresh
*/
fs_info - > endio_workers . idle_thresh = 4 ;
2009-02-04 17:23:24 +03:00
fs_info - > endio_meta_workers . idle_thresh = 4 ;
2009-08-05 00:56:34 +04:00
fs_info - > endio_write_workers . idle_thresh = 2 ;
fs_info - > endio_meta_write_workers . idle_thresh = 2 ;
2008-06-12 05:47:56 +04:00
btrfs_start_workers ( & fs_info - > workers , 1 ) ;
2009-10-03 03:11:56 +04:00
btrfs_start_workers ( & fs_info - > generic_worker , 1 ) ;
2008-06-12 22:46:17 +04:00
btrfs_start_workers ( & fs_info - > submit_workers , 1 ) ;
2008-11-07 06:02:51 +03:00
btrfs_start_workers ( & fs_info - > delalloc_workers , 1 ) ;
2008-07-17 20:53:51 +04:00
btrfs_start_workers ( & fs_info - > fixup_workers , 1 ) ;
2009-08-05 00:56:34 +04:00
btrfs_start_workers ( & fs_info - > endio_workers , 1 ) ;
btrfs_start_workers ( & fs_info - > endio_meta_workers , 1 ) ;
btrfs_start_workers ( & fs_info - > endio_meta_write_workers , 1 ) ;
btrfs_start_workers ( & fs_info - > endio_write_workers , 1 ) ;
2008-06-12 05:47:56 +04:00
2008-04-19 00:13:31 +04:00
fs_info - > bdi . ra_pages * = btrfs_super_num_devices ( disk_super ) ;
Btrfs: Add zlib compression support
This is a large change for adding compression on reading and writing,
both for inline and regular extents. It does some fairly large
surgery to the writeback paths.
Compression is off by default and enabled by mount -o compress. Even
when the -o compress mount option is not used, it is possible to read
compressed extents off the disk.
If compression for a given set of pages fails to make them smaller, the
file is flagged to avoid future compression attempts later.
* While finding delalloc extents, the pages are locked before being sent down
to the delalloc handler. This allows the delalloc handler to do complex things
such as cleaning the pages, marking them writeback and starting IO on their
behalf.
* Inline extents are inserted at delalloc time now. This allows us to compress
the data before inserting the inline extent, and it allows us to insert
an inline extent that spans multiple pages.
* All of the in-memory extent representations (extent_map.c, ordered-data.c etc)
are changed to record both an in-memory size and an on disk size, as well
as a flag for compression.
From a disk format point of view, the extent pointers in the file are changed
to record the on disk size of a given extent and some encoding flags.
Space in the disk format is allocated for compression encoding, as well
as encryption and a generic 'other' field. Neither the encryption or the
'other' field are currently used.
In order to limit the amount of data read for a single random read in the
file, the size of a compressed extent is limited to 128k. This is a
software only limit, the disk format supports u64 sized compressed extents.
In order to limit the ram consumed while processing extents, the uncompressed
size of a compressed extent is limited to 256k. This is a software only limit
and will be subject to tuning later.
Checksumming is still done on compressed extents, and it is done on the
uncompressed version of the data. This way additional encodings can be
layered on without having to figure out which encoding to checksum.
Compression happens at delalloc time, which is basically singled threaded because
it is usually done by a single pdflush thread. This makes it tricky to
spread the compression load across all the cpus on the box. We'll have to
look at parallel pdflush walks of dirty inodes at a later time.
Decompression is hooked into readpages and it does spread across CPUs nicely.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-10-29 21:49:59 +03:00
fs_info - > bdi . ra_pages = max ( fs_info - > bdi . ra_pages ,
4 * 1024 * 1024 / PAGE_CACHE_SIZE ) ;
2008-04-19 00:13:31 +04:00
2007-10-16 00:15:53 +04:00
nodesize = btrfs_super_nodesize ( disk_super ) ;
leafsize = btrfs_super_leafsize ( disk_super ) ;
sectorsize = btrfs_super_sectorsize ( disk_super ) ;
2007-11-30 19:30:34 +03:00
stripesize = btrfs_super_stripesize ( disk_super ) ;
2007-10-16 00:15:53 +04:00
tree_root - > nodesize = nodesize ;
tree_root - > leafsize = leafsize ;
tree_root - > sectorsize = sectorsize ;
2007-11-30 19:30:34 +03:00
tree_root - > stripesize = stripesize ;
2008-05-07 19:43:44 +04:00
sb - > s_blocksize = sectorsize ;
sb - > s_blocksize_bits = blksize_bits ( sectorsize ) ;
2007-10-16 00:15:53 +04:00
2007-06-12 14:35:45 +04:00
if ( strncmp ( ( char * ) ( & disk_super - > magic ) , BTRFS_MAGIC ,
sizeof ( disk_super - > magic ) ) ) {
2009-01-06 05:25:51 +03:00
printk ( KERN_INFO " btrfs: valid FS not found on %s \n " , sb - > s_id ) ;
2007-06-12 14:35:45 +04:00
goto fail_sb_buffer ;
}
2007-10-16 00:19:22 +04:00
2008-06-26 00:01:30 +04:00
mutex_lock ( & fs_info - > chunk_mutex ) ;
2008-12-12 18:03:26 +03:00
ret = btrfs_read_sys_array ( tree_root ) ;
2008-06-26 00:01:30 +04:00
mutex_unlock ( & fs_info - > chunk_mutex ) ;
2008-04-25 17:04:37 +04:00
if ( ret ) {
2009-01-06 05:25:51 +03:00
printk ( KERN_WARNING " btrfs: failed to read the system "
" array on %s \n " , sb - > s_id ) ;
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
goto fail_sb_buffer ;
2008-04-25 17:04:37 +04:00
}
2008-03-24 22:01:56 +03:00
blocksize = btrfs_level_size ( tree_root ,
btrfs_super_chunk_root_level ( disk_super ) ) ;
2008-10-29 21:49:05 +03:00
generation = btrfs_super_chunk_root_generation ( disk_super ) ;
2008-03-24 22:01:56 +03:00
__setup_root ( nodesize , leafsize , sectorsize , stripesize ,
chunk_root , fs_info , BTRFS_CHUNK_TREE_OBJECTID ) ;
chunk_root - > node = read_tree_block ( chunk_root ,
btrfs_super_chunk_root ( disk_super ) ,
2008-10-29 21:49:05 +03:00
blocksize , generation ) ;
2008-03-24 22:01:56 +03:00
BUG_ON ( ! chunk_root - > node ) ;
2009-07-23 00:52:13 +04:00
if ( ! test_bit ( EXTENT_BUFFER_UPTODATE , & chunk_root - > node - > bflags ) ) {
printk ( KERN_WARNING " btrfs: failed to read chunk root on %s \n " ,
sb - > s_id ) ;
goto fail_chunk_root ;
}
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
btrfs_set_root_node ( & chunk_root - > root_item , chunk_root - > node ) ;
chunk_root - > commit_root = btrfs_root_node ( chunk_root ) ;
2008-03-24 22:01:56 +03:00
2008-04-15 23:41:47 +04:00
read_extent_buffer ( chunk_root - > node , fs_info - > chunk_tree_uuid ,
2009-01-06 05:25:51 +03:00
( unsigned long ) btrfs_header_chunk_tree_uuid ( chunk_root - > node ) ,
BTRFS_UUID_SIZE ) ;
2008-04-15 23:41:47 +04:00
2008-06-26 00:01:30 +04:00
mutex_lock ( & fs_info - > chunk_mutex ) ;
2008-03-24 22:01:56 +03:00
ret = btrfs_read_chunk_tree ( chunk_root ) ;
2008-06-26 00:01:30 +04:00
mutex_unlock ( & fs_info - > chunk_mutex ) ;
2008-11-18 05:11:30 +03:00
if ( ret ) {
2009-01-06 05:25:51 +03:00
printk ( KERN_WARNING " btrfs: failed to read chunk tree on %s \n " ,
sb - > s_id ) ;
2008-11-18 05:11:30 +03:00
goto fail_chunk_root ;
}
2008-03-24 22:01:56 +03:00
2008-05-13 21:46:40 +04:00
btrfs_close_extra_devices ( fs_devices ) ;
2007-10-16 00:15:53 +04:00
blocksize = btrfs_level_size ( tree_root ,
btrfs_super_root_level ( disk_super ) ) ;
2008-10-29 21:49:05 +03:00
generation = btrfs_super_generation ( disk_super ) ;
2008-03-24 22:01:56 +03:00
2007-03-22 19:13:20 +03:00
tree_root - > node = read_tree_block ( tree_root ,
2007-10-16 00:15:53 +04:00
btrfs_super_root ( disk_super ) ,
2008-10-29 21:49:05 +03:00
blocksize , generation ) ;
2007-06-12 14:35:45 +04:00
if ( ! tree_root - > node )
2008-11-18 05:11:30 +03:00
goto fail_chunk_root ;
2009-07-23 00:52:13 +04:00
if ( ! test_bit ( EXTENT_BUFFER_UPTODATE , & tree_root - > node - > bflags ) ) {
printk ( KERN_WARNING " btrfs: failed to read tree root on %s \n " ,
sb - > s_id ) ;
goto fail_tree_root ;
}
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
btrfs_set_root_node ( & tree_root - > root_item , tree_root - > node ) ;
tree_root - > commit_root = btrfs_root_node ( tree_root ) ;
2007-10-16 00:15:53 +04:00
ret = find_and_setup_root ( tree_root , fs_info ,
2007-03-22 19:13:20 +03:00
BTRFS_EXTENT_TREE_OBJECTID , extent_root ) ;
2008-03-24 22:01:56 +03:00
if ( ret )
2007-06-12 14:35:45 +04:00
goto fail_tree_root ;
2008-03-24 22:01:56 +03:00
extent_root - > track_dirty = 1 ;
ret = find_and_setup_root ( tree_root , fs_info ,
BTRFS_DEV_TREE_OBJECTID , dev_root ) ;
if ( ret )
goto fail_extent_root ;
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
dev_root - > track_dirty = 1 ;
2007-03-13 23:47:54 +03:00
Btrfs: move data checksumming into a dedicated tree
Btrfs stores checksums for each data block. Until now, they have
been stored in the subvolume trees, indexed by the inode that is
referencing the data block. This means that when we read the inode,
we've probably read in at least some checksums as well.
But, this has a few problems:
* The checksums are indexed by logical offset in the file. When
compression is on, this means we have to do the expensive checksumming
on the uncompressed data. It would be faster if we could checksum
the compressed data instead.
* If we implement encryption, we'll be checksumming the plain text and
storing that on disk. This is significantly less secure.
* For either compression or encryption, we have to get the plain text
back before we can verify the checksum as correct. This makes the raid
layer balancing and extent moving much more expensive.
* It makes the front end caching code more complex, as we have touch
the subvolume and inodes as we cache extents.
* There is potentitally one copy of the checksum in each subvolume
referencing an extent.
The solution used here is to store the extent checksums in a dedicated
tree. This allows us to index the checksums by phyiscal extent
start and length. It means:
* The checksum is against the data stored on disk, after any compression
or encryption is done.
* The checksum is stored in a central location, and can be verified without
following back references, or reading inodes.
This makes compression significantly faster by reducing the amount of
data that needs to be checksummed. It will also allow much faster
raid management code in general.
The checksums are indexed by a key with a fixed objectid (a magic value
in ctree.h) and offset set to the starting byte of the extent. This
allows us to copy the checksum items into the fsync log tree directly (or
any other tree), without having to invent a second format for them.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-12-09 00:58:54 +03:00
ret = find_and_setup_root ( tree_root , fs_info ,
BTRFS_CSUM_TREE_OBJECTID , csum_root ) ;
if ( ret )
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
goto fail_dev_root ;
Btrfs: move data checksumming into a dedicated tree
Btrfs stores checksums for each data block. Until now, they have
been stored in the subvolume trees, indexed by the inode that is
referencing the data block. This means that when we read the inode,
we've probably read in at least some checksums as well.
But, this has a few problems:
* The checksums are indexed by logical offset in the file. When
compression is on, this means we have to do the expensive checksumming
on the uncompressed data. It would be faster if we could checksum
the compressed data instead.
* If we implement encryption, we'll be checksumming the plain text and
storing that on disk. This is significantly less secure.
* For either compression or encryption, we have to get the plain text
back before we can verify the checksum as correct. This makes the raid
layer balancing and extent moving much more expensive.
* It makes the front end caching code more complex, as we have touch
the subvolume and inodes as we cache extents.
* There is potentitally one copy of the checksum in each subvolume
referencing an extent.
The solution used here is to store the extent checksums in a dedicated
tree. This allows us to index the checksums by phyiscal extent
start and length. It means:
* The checksum is against the data stored on disk, after any compression
or encryption is done.
* The checksum is stored in a central location, and can be verified without
following back references, or reading inodes.
This makes compression significantly faster by reducing the amount of
data that needs to be checksummed. It will also allow much faster
raid management code in general.
The checksums are indexed by a key with a fixed objectid (a magic value
in ctree.h) and offset set to the starting byte of the extent. This
allows us to copy the checksum items into the fsync log tree directly (or
any other tree), without having to invent a second format for them.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-12-09 00:58:54 +03:00
csum_root - > track_dirty = 1 ;
2010-05-16 18:49:58 +04:00
fs_info - > generation = generation ;
fs_info - > last_trans_committed = generation ;
fs_info - > data_alloc_profile = ( u64 ) - 1 ;
fs_info - > metadata_alloc_profile = ( u64 ) - 1 ;
fs_info - > system_alloc_profile = fs_info - > metadata_alloc_profile ;
2010-03-19 23:49:55 +03:00
ret = btrfs_read_block_groups ( extent_root ) ;
if ( ret ) {
printk ( KERN_ERR " Failed to read block groups: %d \n " , ret ) ;
goto fail_block_groups ;
}
2007-04-27 00:46:15 +04:00
2008-06-26 00:01:31 +04:00
fs_info - > cleaner_kthread = kthread_run ( cleaner_kthread , tree_root ,
" btrfs-cleaner " ) ;
2009-01-21 18:49:16 +03:00
if ( IS_ERR ( fs_info - > cleaner_kthread ) )
2010-03-19 23:49:55 +03:00
goto fail_block_groups ;
2008-06-26 00:01:31 +04:00
fs_info - > transaction_kthread = kthread_run ( transaction_kthread ,
tree_root ,
" btrfs-transaction " ) ;
2009-01-21 18:49:16 +03:00
if ( IS_ERR ( fs_info - > transaction_kthread ) )
2008-06-26 00:01:31 +04:00
goto fail_cleaner ;
2008-06-26 00:01:31 +04:00
2009-06-10 17:51:32 +04:00
if ( ! btrfs_test_opt ( tree_root , SSD ) & &
! btrfs_test_opt ( tree_root , NOSSD ) & &
! fs_info - > fs_devices - > rotating ) {
printk ( KERN_INFO " Btrfs detected SSD devices, enabling SSD "
" mode \n " ) ;
btrfs_set_opt ( fs_info - > mount_opt , SSD ) ;
}
2008-09-06 00:13:11 +04:00
if ( btrfs_super_log_root ( disk_super ) ! = 0 ) {
u64 bytenr = btrfs_super_log_root ( disk_super ) ;
2008-11-19 23:13:35 +03:00
if ( fs_devices - > rw_devices = = 0 ) {
2009-01-06 05:25:51 +03:00
printk ( KERN_WARNING " Btrfs log replay required "
" on RO media \n " ) ;
2008-11-19 23:13:35 +03:00
err = - EIO ;
goto fail_trans_kthread ;
}
2008-09-06 00:13:11 +04:00
blocksize =
btrfs_level_size ( tree_root ,
btrfs_super_log_root_level ( disk_super ) ) ;
2008-04-04 23:40:00 +04:00
2010-05-29 13:43:07 +04:00
log_tree_root = kzalloc ( sizeof ( struct btrfs_root ) , GFP_NOFS ) ;
if ( ! log_tree_root ) {
err = - ENOMEM ;
goto fail_trans_kthread ;
}
2008-09-06 00:13:11 +04:00
__setup_root ( nodesize , leafsize , sectorsize , stripesize ,
log_tree_root , fs_info , BTRFS_TREE_LOG_OBJECTID ) ;
log_tree_root - > node = read_tree_block ( tree_root , bytenr ,
2008-10-29 21:49:05 +03:00
blocksize ,
generation + 1 ) ;
2008-09-06 00:13:11 +04:00
ret = btrfs_recover_log_trees ( log_tree_root ) ;
BUG_ON ( ret ) ;
2008-11-20 18:25:19 +03:00
if ( sb - > s_flags & MS_RDONLY ) {
ret = btrfs_commit_super ( tree_root ) ;
BUG_ON ( ret ) ;
}
2008-09-06 00:13:11 +04:00
}
Btrfs: update space balancing code
This patch updates the space balancing code to utilize the new
backref format. Before, btrfs-vol -b would break any COW links
on data blocks or metadata. This was slow and caused the amount
of space used to explode if a large number of snapshots were present.
The new code can keeps the sharing of all data extents and
most of the tree blocks.
To maintain the sharing of data extents, the space balance code uses
a seperate inode hold data extent pointers, then updates the references
to point to the new location.
To maintain the sharing of tree blocks, the space balance code uses
reloc trees to relocate tree blocks in reference counted roots.
There is one reloc tree for each subvol, and all reloc trees share
same root key objectid. Reloc trees are snapshots of the latest
committed roots of subvols (root->commit_root).
To relocate a tree block referenced by a subvol, there are two steps.
COW the block through subvol's reloc tree, then update block pointer in
the subvol to point to the new block. Since all reloc trees share
same root key objectid, doing special handing for tree blocks
owned by them is easy. Once a tree block has been COWed in one
reloc tree, we can use the resulting new block directly when the
same block is required to COW again through other reloc trees.
In this way, relocated tree blocks are shared between reloc trees,
so they are also shared between subvols.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-09-26 18:09:34 +04:00
2009-09-22 00:00:26 +04:00
ret = btrfs_find_orphan_roots ( tree_root ) ;
BUG_ON ( ret ) ;
2008-11-19 23:13:35 +03:00
if ( ! ( sb - > s_flags & MS_RDONLY ) ) {
2010-05-16 18:49:58 +04:00
ret = btrfs_cleanup_fs_roots ( fs_info ) ;
BUG_ON ( ret ) ;
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
ret = btrfs_recover_relocation ( tree_root ) ;
2010-02-02 11:46:44 +03:00
if ( ret < 0 ) {
printk ( KERN_WARNING
" btrfs: failed to recover relocation \n " ) ;
err = - EINVAL ;
goto fail_trans_kthread ;
}
2008-11-19 23:13:35 +03:00
}
Btrfs: update space balancing code
This patch updates the space balancing code to utilize the new
backref format. Before, btrfs-vol -b would break any COW links
on data blocks or metadata. This was slow and caused the amount
of space used to explode if a large number of snapshots were present.
The new code can keeps the sharing of all data extents and
most of the tree blocks.
To maintain the sharing of data extents, the space balance code uses
a seperate inode hold data extent pointers, then updates the references
to point to the new location.
To maintain the sharing of tree blocks, the space balance code uses
reloc trees to relocate tree blocks in reference counted roots.
There is one reloc tree for each subvol, and all reloc trees share
same root key objectid. Reloc trees are snapshots of the latest
committed roots of subvols (root->commit_root).
To relocate a tree block referenced by a subvol, there are two steps.
COW the block through subvol's reloc tree, then update block pointer in
the subvol to point to the new block. Since all reloc trees share
same root key objectid, doing special handing for tree blocks
owned by them is easy. Once a tree block has been COWed in one
reloc tree, we can use the resulting new block directly when the
same block is required to COW again through other reloc trees.
In this way, relocated tree blocks are shared between reloc trees,
so they are also shared between subvols.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-09-26 18:09:34 +04:00
2008-11-18 05:02:50 +03:00
location . objectid = BTRFS_FS_TREE_OBJECTID ;
location . type = BTRFS_ROOT_ITEM_KEY ;
location . offset = ( u64 ) - 1 ;
fs_info - > fs_root = btrfs_read_fs_root_no_name ( fs_info , & location ) ;
if ( ! fs_info - > fs_root )
2008-11-19 23:13:35 +03:00
goto fail_trans_kthread ;
2010-05-29 13:44:10 +04:00
if ( IS_ERR ( fs_info - > fs_root ) ) {
err = PTR_ERR ( fs_info - > fs_root ) ;
goto fail_trans_kthread ;
}
2009-06-10 17:51:32 +04:00
2010-01-26 17:30:53 +03:00
if ( ! ( sb - > s_flags & MS_RDONLY ) ) {
down_read ( & fs_info - > cleanup_work_sem ) ;
btrfs_orphan_cleanup ( fs_info - > fs_root ) ;
up_read ( & fs_info - > cleanup_work_sem ) ;
}
2007-04-09 18:42:37 +04:00
return tree_root ;
2007-06-12 14:35:45 +04:00
2008-11-19 23:13:35 +03:00
fail_trans_kthread :
kthread_stop ( fs_info - > transaction_kthread ) ;
2008-06-26 00:01:31 +04:00
fail_cleaner :
2008-06-26 00:01:31 +04:00
kthread_stop ( fs_info - > cleaner_kthread ) ;
2008-11-19 23:13:35 +03:00
/*
* make sure we ' re done with the btree inode before we stop our
* kthreads
*/
filemap_write_and_wait ( fs_info - > btree_inode - > i_mapping ) ;
invalidate_inode_pages2 ( fs_info - > btree_inode - > i_mapping ) ;
2010-03-19 23:49:55 +03:00
fail_block_groups :
btrfs_free_block_groups ( fs_info ) ;
Btrfs: move data checksumming into a dedicated tree
Btrfs stores checksums for each data block. Until now, they have
been stored in the subvolume trees, indexed by the inode that is
referencing the data block. This means that when we read the inode,
we've probably read in at least some checksums as well.
But, this has a few problems:
* The checksums are indexed by logical offset in the file. When
compression is on, this means we have to do the expensive checksumming
on the uncompressed data. It would be faster if we could checksum
the compressed data instead.
* If we implement encryption, we'll be checksumming the plain text and
storing that on disk. This is significantly less secure.
* For either compression or encryption, we have to get the plain text
back before we can verify the checksum as correct. This makes the raid
layer balancing and extent moving much more expensive.
* It makes the front end caching code more complex, as we have touch
the subvolume and inodes as we cache extents.
* There is potentitally one copy of the checksum in each subvolume
referencing an extent.
The solution used here is to store the extent checksums in a dedicated
tree. This allows us to index the checksums by phyiscal extent
start and length. It means:
* The checksum is against the data stored on disk, after any compression
or encryption is done.
* The checksum is stored in a central location, and can be verified without
following back references, or reading inodes.
This makes compression significantly faster by reducing the amount of
data that needs to be checksummed. It will also allow much faster
raid management code in general.
The checksums are indexed by a key with a fixed objectid (a magic value
in ctree.h) and offset set to the starting byte of the extent. This
allows us to copy the checksum items into the fsync log tree directly (or
any other tree), without having to invent a second format for them.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-12-09 00:58:54 +03:00
free_extent_buffer ( csum_root - > node ) ;
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
free_extent_buffer ( csum_root - > commit_root ) ;
fail_dev_root :
free_extent_buffer ( dev_root - > node ) ;
free_extent_buffer ( dev_root - > commit_root ) ;
2008-03-24 22:01:56 +03:00
fail_extent_root :
free_extent_buffer ( extent_root - > node ) ;
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
free_extent_buffer ( extent_root - > commit_root ) ;
2007-06-12 14:35:45 +04:00
fail_tree_root :
2007-10-16 00:14:19 +04:00
free_extent_buffer ( tree_root - > node ) ;
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
free_extent_buffer ( tree_root - > commit_root ) ;
2008-11-18 05:11:30 +03:00
fail_chunk_root :
free_extent_buffer ( chunk_root - > node ) ;
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
free_extent_buffer ( chunk_root - > commit_root ) ;
2007-06-12 14:35:45 +04:00
fail_sb_buffer :
2009-10-03 03:11:56 +04:00
btrfs_stop_workers ( & fs_info - > generic_worker ) ;
2008-07-17 20:53:51 +04:00
btrfs_stop_workers ( & fs_info - > fixup_workers ) ;
2008-11-07 06:02:51 +03:00
btrfs_stop_workers ( & fs_info - > delalloc_workers ) ;
2008-06-12 00:50:36 +04:00
btrfs_stop_workers ( & fs_info - > workers ) ;
btrfs_stop_workers ( & fs_info - > endio_workers ) ;
Btrfs: move data checksumming into a dedicated tree
Btrfs stores checksums for each data block. Until now, they have
been stored in the subvolume trees, indexed by the inode that is
referencing the data block. This means that when we read the inode,
we've probably read in at least some checksums as well.
But, this has a few problems:
* The checksums are indexed by logical offset in the file. When
compression is on, this means we have to do the expensive checksumming
on the uncompressed data. It would be faster if we could checksum
the compressed data instead.
* If we implement encryption, we'll be checksumming the plain text and
storing that on disk. This is significantly less secure.
* For either compression or encryption, we have to get the plain text
back before we can verify the checksum as correct. This makes the raid
layer balancing and extent moving much more expensive.
* It makes the front end caching code more complex, as we have touch
the subvolume and inodes as we cache extents.
* There is potentitally one copy of the checksum in each subvolume
referencing an extent.
The solution used here is to store the extent checksums in a dedicated
tree. This allows us to index the checksums by phyiscal extent
start and length. It means:
* The checksum is against the data stored on disk, after any compression
or encryption is done.
* The checksum is stored in a central location, and can be verified without
following back references, or reading inodes.
This makes compression significantly faster by reducing the amount of
data that needs to be checksummed. It will also allow much faster
raid management code in general.
The checksums are indexed by a key with a fixed objectid (a magic value
in ctree.h) and offset set to the starting byte of the extent. This
allows us to copy the checksum items into the fsync log tree directly (or
any other tree), without having to invent a second format for them.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-12-09 00:58:54 +03:00
btrfs_stop_workers ( & fs_info - > endio_meta_workers ) ;
2008-12-17 22:51:42 +03:00
btrfs_stop_workers ( & fs_info - > endio_meta_write_workers ) ;
2008-07-17 20:53:50 +04:00
btrfs_stop_workers ( & fs_info - > endio_write_workers ) ;
2008-06-12 22:46:17 +04:00
btrfs_stop_workers ( & fs_info - > submit_workers ) ;
2008-06-12 05:47:56 +04:00
fail_iput :
2008-11-19 23:13:35 +03:00
invalidate_inode_pages2 ( fs_info - > btree_inode - > i_mapping ) ;
2008-06-12 05:47:56 +04:00
iput ( fs_info - > btree_inode ) ;
2009-01-21 18:49:16 +03:00
2008-05-13 21:46:40 +04:00
btrfs_close_devices ( fs_info - > fs_devices ) ;
2008-04-25 17:04:37 +04:00
btrfs_mapping_tree_free ( & fs_info - > mapping_tree ) ;
2009-06-12 16:43:40 +04:00
fail_bdi :
2009-01-21 18:49:16 +03:00
bdi_destroy ( & fs_info - > bdi ) ;
2009-09-22 00:00:26 +04:00
fail_srcu :
cleanup_srcu_struct ( & fs_info - > subvol_srcu ) ;
2009-01-21 18:49:16 +03:00
fail :
2007-06-12 14:35:45 +04:00
kfree ( extent_root ) ;
kfree ( tree_root ) ;
kfree ( fs_info ) ;
2008-10-02 03:09:51 +04:00
kfree ( chunk_root ) ;
kfree ( dev_root ) ;
Btrfs: move data checksumming into a dedicated tree
Btrfs stores checksums for each data block. Until now, they have
been stored in the subvolume trees, indexed by the inode that is
referencing the data block. This means that when we read the inode,
we've probably read in at least some checksums as well.
But, this has a few problems:
* The checksums are indexed by logical offset in the file. When
compression is on, this means we have to do the expensive checksumming
on the uncompressed data. It would be faster if we could checksum
the compressed data instead.
* If we implement encryption, we'll be checksumming the plain text and
storing that on disk. This is significantly less secure.
* For either compression or encryption, we have to get the plain text
back before we can verify the checksum as correct. This makes the raid
layer balancing and extent moving much more expensive.
* It makes the front end caching code more complex, as we have touch
the subvolume and inodes as we cache extents.
* There is potentitally one copy of the checksum in each subvolume
referencing an extent.
The solution used here is to store the extent checksums in a dedicated
tree. This allows us to index the checksums by phyiscal extent
start and length. It means:
* The checksum is against the data stored on disk, after any compression
or encryption is done.
* The checksum is stored in a central location, and can be verified without
following back references, or reading inodes.
This makes compression significantly faster by reducing the amount of
data that needs to be checksummed. It will also allow much faster
raid management code in general.
The checksums are indexed by a key with a fixed objectid (a magic value
in ctree.h) and offset set to the starting byte of the extent. This
allows us to copy the checksum items into the fsync log tree directly (or
any other tree), without having to invent a second format for them.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-12-09 00:58:54 +03:00
kfree ( csum_root ) ;
2007-06-12 14:35:45 +04:00
return ERR_PTR ( err ) ;
2007-02-02 17:18:22 +03:00
}
2008-04-11 00:19:33 +04:00
static void btrfs_end_buffer_write_sync ( struct buffer_head * bh , int uptodate )
{
char b [ BDEVNAME_SIZE ] ;
if ( uptodate ) {
set_buffer_uptodate ( bh ) ;
} else {
if ( ! buffer_eopnotsupp ( bh ) & & printk_ratelimit ( ) ) {
printk ( KERN_WARNING " lost page write due to "
" I/O error on %s \n " ,
bdevname ( bh - > b_bdev , b ) ) ;
}
2008-05-12 21:39:03 +04:00
/* note, we dont' set_buffer_write_io_error because we have
* our own ways of dealing with the IO errors
*/
2008-04-11 00:19:33 +04:00
clear_buffer_uptodate ( bh ) ;
}
unlock_buffer ( bh ) ;
put_bh ( bh ) ;
}
2008-12-09 00:46:26 +03:00
struct buffer_head * btrfs_read_dev_super ( struct block_device * bdev )
{
struct buffer_head * bh ;
struct buffer_head * latest = NULL ;
struct btrfs_super_block * super ;
int i ;
u64 transid = 0 ;
u64 bytenr ;
/* we would like to check all the supers, but that would make
* a btrfs mount succeed after a mkfs from a different FS .
* So , we need to add a special mount option to scan for
* later supers , using BTRFS_SUPER_MIRROR_MAX instead
*/
for ( i = 0 ; i < 1 ; i + + ) {
bytenr = btrfs_sb_offset ( i ) ;
if ( bytenr + 4096 > = i_size_read ( bdev - > bd_inode ) )
break ;
bh = __bread ( bdev , bytenr / 4096 , 4096 ) ;
if ( ! bh )
continue ;
super = ( struct btrfs_super_block * ) bh - > b_data ;
if ( btrfs_super_bytenr ( super ) ! = bytenr | |
strncmp ( ( char * ) ( & super - > magic ) , BTRFS_MAGIC ,
sizeof ( super - > magic ) ) ) {
brelse ( bh ) ;
continue ;
}
if ( ! latest | | btrfs_super_generation ( super ) > transid ) {
brelse ( latest ) ;
latest = bh ;
transid = btrfs_super_generation ( super ) ;
} else {
brelse ( bh ) ;
}
}
return latest ;
}
2009-06-10 23:28:55 +04:00
/*
* this should be called twice , once with wait = = 0 and
* once with wait = = 1. When wait = = 0 is done , all the buffer heads
* we write are pinned .
*
* They are released when wait = = 1 is done .
* max_mirrors must be the same for both runs , and it indicates how
* many supers on this one device should be written .
*
* max_mirrors = = 0 means to write them all .
*/
2008-12-09 00:46:26 +03:00
static int write_dev_supers ( struct btrfs_device * device ,
struct btrfs_super_block * sb ,
int do_barriers , int wait , int max_mirrors )
{
struct buffer_head * bh ;
int i ;
int ret ;
int errors = 0 ;
u32 crc ;
u64 bytenr ;
int last_barrier = 0 ;
if ( max_mirrors = = 0 )
max_mirrors = BTRFS_SUPER_MIRROR_MAX ;
/* make sure only the last submit_bh does a barrier */
if ( do_barriers ) {
for ( i = 0 ; i < max_mirrors ; i + + ) {
bytenr = btrfs_sb_offset ( i ) ;
if ( bytenr + BTRFS_SUPER_INFO_SIZE > =
device - > total_bytes )
break ;
last_barrier = i ;
}
}
for ( i = 0 ; i < max_mirrors ; i + + ) {
bytenr = btrfs_sb_offset ( i ) ;
if ( bytenr + BTRFS_SUPER_INFO_SIZE > = device - > total_bytes )
break ;
if ( wait ) {
bh = __find_get_block ( device - > bdev , bytenr / 4096 ,
BTRFS_SUPER_INFO_SIZE ) ;
BUG_ON ( ! bh ) ;
wait_on_buffer ( bh ) ;
2009-06-10 23:28:55 +04:00
if ( ! buffer_uptodate ( bh ) )
errors + + ;
/* drop our reference */
brelse ( bh ) ;
/* drop the reference from the wait == 0 run */
brelse ( bh ) ;
continue ;
2008-12-09 00:46:26 +03:00
} else {
btrfs_set_super_bytenr ( sb , bytenr ) ;
crc = ~ ( u32 ) 0 ;
crc = btrfs_csum_data ( NULL , ( char * ) sb +
BTRFS_CSUM_SIZE , crc ,
BTRFS_SUPER_INFO_SIZE -
BTRFS_CSUM_SIZE ) ;
btrfs_csum_final ( crc , sb - > csum ) ;
2009-06-10 23:28:55 +04:00
/*
* one reference for us , and we leave it for the
* caller
*/
2008-12-09 00:46:26 +03:00
bh = __getblk ( device - > bdev , bytenr / 4096 ,
BTRFS_SUPER_INFO_SIZE ) ;
memcpy ( bh - > b_data , sb , BTRFS_SUPER_INFO_SIZE ) ;
2009-06-10 23:28:55 +04:00
/* one reference for submit_bh */
2008-12-09 00:46:26 +03:00
get_bh ( bh ) ;
2009-06-10 23:28:55 +04:00
set_buffer_uptodate ( bh ) ;
2008-12-09 00:46:26 +03:00
lock_buffer ( bh ) ;
bh - > b_end_io = btrfs_end_buffer_write_sync ;
}
if ( i = = last_barrier & & do_barriers & & device - > barriers ) {
ret = submit_bh ( WRITE_BARRIER , bh ) ;
if ( ret = = - EOPNOTSUPP ) {
printk ( " btrfs: disabling barriers on dev %s \n " ,
device - > name ) ;
set_buffer_uptodate ( bh ) ;
device - > barriers = 0 ;
2009-06-10 23:28:55 +04:00
/* one reference for submit_bh */
2008-12-09 00:46:26 +03:00
get_bh ( bh ) ;
lock_buffer ( bh ) ;
2009-04-20 23:50:09 +04:00
ret = submit_bh ( WRITE_SYNC , bh ) ;
2008-12-09 00:46:26 +03:00
}
} else {
2009-04-20 23:50:09 +04:00
ret = submit_bh ( WRITE_SYNC , bh ) ;
2008-12-09 00:46:26 +03:00
}
2009-06-10 23:28:55 +04:00
if ( ret )
2008-12-09 00:46:26 +03:00
errors + + ;
}
return errors < i ? 0 : - 1 ;
}
int write_all_supers ( struct btrfs_root * root , int max_mirrors )
2008-04-11 00:19:33 +04:00
{
2009-06-10 23:17:02 +04:00
struct list_head * head ;
2008-04-11 00:19:33 +04:00
struct btrfs_device * dev ;
2008-05-07 19:43:44 +04:00
struct btrfs_super_block * sb ;
2008-04-11 00:19:33 +04:00
struct btrfs_dev_item * dev_item ;
int ret ;
int do_barriers ;
2008-04-29 17:38:00 +04:00
int max_errors ;
int total_errors = 0 ;
2008-05-07 19:43:44 +04:00
u64 flags ;
2008-04-11 00:19:33 +04:00
2008-04-29 17:38:00 +04:00
max_errors = btrfs_super_num_devices ( & root - > fs_info - > super_copy ) - 1 ;
2008-04-11 00:19:33 +04:00
do_barriers = ! btrfs_test_opt ( root , NOBARRIER ) ;
2008-05-07 19:43:44 +04:00
sb = & root - > fs_info - > super_for_commit ;
dev_item = & sb - > dev_item ;
2009-06-10 23:17:02 +04:00
mutex_lock ( & root - > fs_info - > fs_devices - > device_list_mutex ) ;
head = & root - > fs_info - > fs_devices - > devices ;
2009-01-21 18:59:08 +03:00
list_for_each_entry ( dev , head , dev_list ) {
2008-05-13 21:46:40 +04:00
if ( ! dev - > bdev ) {
total_errors + + ;
continue ;
}
2008-11-18 05:11:30 +03:00
if ( ! dev - > in_fs_metadata | | ! dev - > writeable )
2008-05-13 21:46:40 +04:00
continue ;
2008-11-18 05:11:30 +03:00
btrfs_set_stack_device_generation ( dev_item , 0 ) ;
2008-05-07 19:43:44 +04:00
btrfs_set_stack_device_type ( dev_item , dev - > type ) ;
btrfs_set_stack_device_id ( dev_item , dev - > devid ) ;
btrfs_set_stack_device_total_bytes ( dev_item , dev - > total_bytes ) ;
btrfs_set_stack_device_bytes_used ( dev_item , dev - > bytes_used ) ;
btrfs_set_stack_device_io_align ( dev_item , dev - > io_align ) ;
btrfs_set_stack_device_io_width ( dev_item , dev - > io_width ) ;
btrfs_set_stack_device_sector_size ( dev_item , dev - > sector_size ) ;
memcpy ( dev_item - > uuid , dev - > uuid , BTRFS_UUID_SIZE ) ;
2008-11-18 05:11:30 +03:00
memcpy ( dev_item - > fsid , dev - > fs_devices - > fsid , BTRFS_UUID_SIZE ) ;
2008-12-09 00:46:26 +03:00
2008-05-07 19:43:44 +04:00
flags = btrfs_super_flags ( sb ) ;
btrfs_set_super_flags ( sb , flags | BTRFS_HEADER_FLAG_WRITTEN ) ;
2008-12-09 00:46:26 +03:00
ret = write_dev_supers ( dev , sb , do_barriers , 0 , max_mirrors ) ;
2008-04-29 17:38:00 +04:00
if ( ret )
total_errors + + ;
2008-04-11 00:19:33 +04:00
}
2008-04-29 17:38:00 +04:00
if ( total_errors > max_errors ) {
2009-01-06 05:25:51 +03:00
printk ( KERN_ERR " btrfs: %d errors while writing supers \n " ,
total_errors ) ;
2008-04-29 17:38:00 +04:00
BUG ( ) ;
}
2008-04-11 00:19:33 +04:00
2008-12-09 00:46:26 +03:00
total_errors = 0 ;
2009-01-21 18:59:08 +03:00
list_for_each_entry ( dev , head , dev_list ) {
2008-05-13 21:46:40 +04:00
if ( ! dev - > bdev )
continue ;
2008-11-18 05:11:30 +03:00
if ( ! dev - > in_fs_metadata | | ! dev - > writeable )
2008-05-13 21:46:40 +04:00
continue ;
2008-12-09 00:46:26 +03:00
ret = write_dev_supers ( dev , sb , do_barriers , 1 , max_mirrors ) ;
if ( ret )
total_errors + + ;
2008-04-11 00:19:33 +04:00
}
2009-06-10 23:17:02 +04:00
mutex_unlock ( & root - > fs_info - > fs_devices - > device_list_mutex ) ;
2008-04-29 17:38:00 +04:00
if ( total_errors > max_errors ) {
2009-01-06 05:25:51 +03:00
printk ( KERN_ERR " btrfs: %d errors while writing supers \n " ,
total_errors ) ;
2008-04-29 17:38:00 +04:00
BUG ( ) ;
}
2008-04-11 00:19:33 +04:00
return 0 ;
}
2008-12-09 00:46:26 +03:00
int write_ctree_super ( struct btrfs_trans_handle * trans ,
struct btrfs_root * root , int max_mirrors )
2007-02-02 17:18:22 +03:00
{
2007-04-20 21:16:02 +04:00
int ret ;
2007-10-16 00:14:19 +04:00
2008-12-09 00:46:26 +03:00
ret = write_all_supers ( root , max_mirrors ) ;
2007-10-16 00:14:19 +04:00
return ret ;
2007-02-22 01:04:57 +03:00
}
2007-06-22 22:16:25 +04:00
int btrfs_free_fs_root ( struct btrfs_fs_info * fs_info , struct btrfs_root * root )
2007-04-11 00:58:11 +04:00
{
2009-09-21 23:56:00 +04:00
spin_lock ( & fs_info - > fs_roots_radix_lock ) ;
2007-04-11 00:58:11 +04:00
radix_tree_delete ( & fs_info - > fs_roots_radix ,
( unsigned long ) root - > root_key . objectid ) ;
2009-09-21 23:56:00 +04:00
spin_unlock ( & fs_info - > fs_roots_radix_lock ) ;
2009-09-22 00:00:26 +04:00
if ( btrfs_root_refs ( & root - > root_item ) = = 0 )
synchronize_srcu ( & fs_info - > subvol_srcu ) ;
2009-09-21 23:56:00 +04:00
free_fs_root ( root ) ;
return 0 ;
}
static void free_fs_root ( struct btrfs_root * root )
{
WARN_ON ( ! RB_EMPTY_ROOT ( & root - > inode_tree ) ) ;
2008-11-18 04:42:26 +03:00
if ( root - > anon_super . s_dev ) {
down_write ( & root - > anon_super . s_umount ) ;
kill_anon_super ( & root - > anon_super ) ;
}
2009-09-21 23:56:00 +04:00
free_extent_buffer ( root - > node ) ;
free_extent_buffer ( root - > commit_root ) ;
2009-01-06 05:25:51 +03:00
kfree ( root - > name ) ;
2007-04-11 00:58:11 +04:00
kfree ( root ) ;
}
2007-05-02 23:53:43 +04:00
static int del_fs_roots ( struct btrfs_fs_info * fs_info )
2007-04-09 18:42:37 +04:00
{
int ret ;
struct btrfs_root * gang [ 8 ] ;
int i ;
2009-09-22 00:00:26 +04:00
while ( ! list_empty ( & fs_info - > dead_roots ) ) {
gang [ 0 ] = list_entry ( fs_info - > dead_roots . next ,
struct btrfs_root , root_list ) ;
list_del ( & gang [ 0 ] - > root_list ) ;
if ( gang [ 0 ] - > in_radix ) {
btrfs_free_fs_root ( fs_info , gang [ 0 ] ) ;
} else {
free_extent_buffer ( gang [ 0 ] - > node ) ;
free_extent_buffer ( gang [ 0 ] - > commit_root ) ;
kfree ( gang [ 0 ] ) ;
}
}
2009-01-06 05:25:51 +03:00
while ( 1 ) {
2007-04-09 18:42:37 +04:00
ret = radix_tree_gang_lookup ( & fs_info - > fs_roots_radix ,
( void * * ) gang , 0 ,
ARRAY_SIZE ( gang ) ) ;
if ( ! ret )
break ;
2007-04-11 00:58:11 +04:00
for ( i = 0 ; i < ret ; i + + )
2007-06-22 22:16:25 +04:00
btrfs_free_fs_root ( fs_info , gang [ i ] ) ;
2007-04-09 18:42:37 +04:00
}
return 0 ;
}
2007-04-12 20:14:00 +04:00
2008-11-12 22:34:12 +03:00
int btrfs_cleanup_fs_roots ( struct btrfs_fs_info * fs_info )
2007-02-22 01:04:57 +03:00
{
2008-11-12 22:34:12 +03:00
u64 root_objectid = 0 ;
struct btrfs_root * gang [ 8 ] ;
int i ;
2007-03-13 23:47:54 +03:00
int ret ;
2007-03-16 23:20:31 +03:00
2008-11-12 22:34:12 +03:00
while ( 1 ) {
ret = radix_tree_gang_lookup ( & fs_info - > fs_roots_radix ,
( void * * ) gang , root_objectid ,
ARRAY_SIZE ( gang ) ) ;
if ( ! ret )
break ;
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
root_objectid = gang [ ret - 1 ] - > root_key . objectid + 1 ;
2008-11-12 22:34:12 +03:00
for ( i = 0 ; i < ret ; i + + ) {
root_objectid = gang [ i ] - > root_key . objectid ;
btrfs_orphan_cleanup ( gang [ i ] ) ;
}
root_objectid + + ;
}
return 0 ;
}
2008-06-26 00:01:30 +04:00
2008-11-12 22:34:12 +03:00
int btrfs_commit_super ( struct btrfs_root * root )
{
struct btrfs_trans_handle * trans ;
int ret ;
2008-06-26 00:01:31 +04:00
2008-11-12 22:34:12 +03:00
mutex_lock ( & root - > fs_info - > cleaner_mutex ) ;
2009-11-12 12:36:34 +03:00
btrfs_run_delayed_iputs ( root ) ;
2008-06-26 00:01:31 +04:00
btrfs_clean_old_snapshots ( root ) ;
2008-11-12 22:34:12 +03:00
mutex_unlock ( & root - > fs_info - > cleaner_mutex ) ;
2009-11-12 12:34:40 +03:00
/* wait until ongoing cleanup work done */
down_write ( & root - > fs_info - > cleanup_work_sem ) ;
up_write ( & root - > fs_info - > cleanup_work_sem ) ;
2010-05-16 18:48:46 +04:00
trans = btrfs_join_transaction ( root , 1 ) ;
2007-06-22 22:16:25 +04:00
ret = btrfs_commit_transaction ( trans , root ) ;
2008-11-12 22:34:12 +03:00
BUG_ON ( ret ) ;
/* run commit again to drop the original snapshot */
2010-05-16 18:48:46 +04:00
trans = btrfs_join_transaction ( root , 1 ) ;
2007-03-22 22:59:16 +03:00
btrfs_commit_transaction ( trans , root ) ;
ret = btrfs_write_and_wait_transaction ( NULL , root ) ;
2007-03-13 23:47:54 +03:00
BUG_ON ( ret ) ;
2008-04-30 21:59:35 +04:00
2008-12-09 00:46:26 +03:00
ret = write_ctree_super ( NULL , root , 0 ) ;
2008-11-12 22:34:12 +03:00
return ret ;
}
int close_ctree ( struct btrfs_root * root )
{
struct btrfs_fs_info * fs_info = root - > fs_info ;
int ret ;
fs_info - > closing = 1 ;
smp_mb ( ) ;
if ( ! ( fs_info - > sb - > s_flags & MS_RDONLY ) ) {
ret = btrfs_commit_super ( root ) ;
2009-01-06 05:25:51 +03:00
if ( ret )
printk ( KERN_ERR " btrfs: commit super ret %d \n " , ret ) ;
2008-11-12 22:34:12 +03:00
}
2007-04-09 18:42:37 +04:00
2010-05-16 18:49:58 +04:00
kthread_stop ( root - > fs_info - > transaction_kthread ) ;
kthread_stop ( root - > fs_info - > cleaner_kthread ) ;
2009-07-28 16:41:57 +04:00
fs_info - > closing = 2 ;
smp_mb ( ) ;
2008-01-31 19:05:37 +03:00
if ( fs_info - > delalloc_bytes ) {
2009-01-06 05:25:51 +03:00
printk ( KERN_INFO " btrfs: at unmount delalloc count %llu \n " ,
2009-04-21 23:38:29 +04:00
( unsigned long long ) fs_info - > delalloc_bytes ) ;
2008-01-31 19:05:37 +03:00
}
2008-07-28 23:32:19 +04:00
if ( fs_info - > total_ref_cache_size ) {
2009-01-06 05:25:51 +03:00
printk ( KERN_INFO " btrfs: at umount reference cache size %llu \n " ,
( unsigned long long ) fs_info - > total_ref_cache_size ) ;
2008-07-28 23:32:19 +04:00
}
2008-07-31 00:29:20 +04:00
Btrfs: Mixed back reference (FORWARD ROLLING FORMAT CHANGE)
This commit introduces a new kind of back reference for btrfs metadata.
Once a filesystem has been mounted with this commit, IT WILL NO LONGER
BE MOUNTABLE BY OLDER KERNELS.
When a tree block in subvolume tree is cow'd, the reference counts of all
extents it points to are increased by one. At transaction commit time,
the old root of the subvolume is recorded in a "dead root" data structure,
and the btree it points to is later walked, dropping reference counts
and freeing any blocks where the reference count goes to 0.
The increments done during cow and decrements done after commit cancel out,
and the walk is a very expensive way to go about freeing the blocks that
are no longer referenced by the new btree root. This commit reduces the
transaction overhead by avoiding the need for dead root records.
When a non-shared tree block is cow'd, we free the old block at once, and the
new block inherits old block's references. When a tree block with reference
count > 1 is cow'd, we increase the reference counts of all extents
the new block points to by one, and decrease the old block's reference count by
one.
This dead tree avoidance code removes the need to modify the reference
counts of lower level extents when a non-shared tree block is cow'd.
But we still need to update back ref for all pointers in the block.
This is because the location of the block is recorded in the back ref
item.
We can solve this by introducing a new type of back ref. The new
back ref provides information about pointer's key, level and in which
tree the pointer lives. This information allow us to find the pointer
by searching the tree. The shortcoming of the new back ref is that it
only works for pointers in tree blocks referenced by their owner trees.
This is mostly a problem for snapshots, where resolving one of these
fuzzy back references would be O(number_of_snapshots) and quite slow.
The solution used here is to use the fuzzy back references in the common
case where a given tree block is only referenced by one root,
and use the full back references when multiple roots have a reference
on a given block.
This commit adds per subvolume red-black tree to keep trace of cached
inodes. The red-black tree helps the balancing code to find cached
inodes whose inode numbers within a given range.
This commit improves the balancing code by introducing several data
structures to keep the state of balancing. The most important one
is the back ref cache. It caches how the upper level tree blocks are
referenced. This greatly reduce the overhead of checking back ref.
The improved balancing code scales significantly better with a large
number of snapshots.
This is a very large commit and was written in a number of
pieces. But, they depend heavily on the disk format change and were
squashed together to make sure git bisect didn't end up in a
bad state wrt space balancing or the format change.
Signed-off-by: Yan Zheng <zheng.yan@oracle.com>
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-06-10 18:45:14 +04:00
free_extent_buffer ( fs_info - > extent_root - > node ) ;
free_extent_buffer ( fs_info - > extent_root - > commit_root ) ;
free_extent_buffer ( fs_info - > tree_root - > node ) ;
free_extent_buffer ( fs_info - > tree_root - > commit_root ) ;
free_extent_buffer ( root - > fs_info - > chunk_root - > node ) ;
free_extent_buffer ( root - > fs_info - > chunk_root - > commit_root ) ;
free_extent_buffer ( root - > fs_info - > dev_root - > node ) ;
free_extent_buffer ( root - > fs_info - > dev_root - > commit_root ) ;
free_extent_buffer ( root - > fs_info - > csum_root - > node ) ;
free_extent_buffer ( root - > fs_info - > csum_root - > commit_root ) ;
Btrfs: move data checksumming into a dedicated tree
Btrfs stores checksums for each data block. Until now, they have
been stored in the subvolume trees, indexed by the inode that is
referencing the data block. This means that when we read the inode,
we've probably read in at least some checksums as well.
But, this has a few problems:
* The checksums are indexed by logical offset in the file. When
compression is on, this means we have to do the expensive checksumming
on the uncompressed data. It would be faster if we could checksum
the compressed data instead.
* If we implement encryption, we'll be checksumming the plain text and
storing that on disk. This is significantly less secure.
* For either compression or encryption, we have to get the plain text
back before we can verify the checksum as correct. This makes the raid
layer balancing and extent moving much more expensive.
* It makes the front end caching code more complex, as we have touch
the subvolume and inodes as we cache extents.
* There is potentitally one copy of the checksum in each subvolume
referencing an extent.
The solution used here is to store the extent checksums in a dedicated
tree. This allows us to index the checksums by phyiscal extent
start and length. It means:
* The checksum is against the data stored on disk, after any compression
or encryption is done.
* The checksum is stored in a central location, and can be verified without
following back references, or reading inodes.
This makes compression significantly faster by reducing the amount of
data that needs to be checksummed. It will also allow much faster
raid management code in general.
The checksums are indexed by a key with a fixed objectid (a magic value
in ctree.h) and offset set to the starting byte of the extent. This
allows us to copy the checksum items into the fsync log tree directly (or
any other tree), without having to invent a second format for them.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-12-09 00:58:54 +03:00
2007-04-27 00:46:15 +04:00
btrfs_free_block_groups ( root - > fs_info ) ;
2007-12-18 04:14:04 +03:00
2008-11-12 22:34:12 +03:00
del_fs_roots ( fs_info ) ;
2007-12-18 04:14:04 +03:00
2008-11-12 22:34:12 +03:00
iput ( fs_info - > btree_inode ) ;
2008-04-19 00:11:30 +04:00
2009-10-03 03:11:56 +04:00
btrfs_stop_workers ( & fs_info - > generic_worker ) ;
2008-07-17 20:53:51 +04:00
btrfs_stop_workers ( & fs_info - > fixup_workers ) ;
2008-11-07 06:02:51 +03:00
btrfs_stop_workers ( & fs_info - > delalloc_workers ) ;
2008-06-12 00:50:36 +04:00
btrfs_stop_workers ( & fs_info - > workers ) ;
btrfs_stop_workers ( & fs_info - > endio_workers ) ;
Btrfs: move data checksumming into a dedicated tree
Btrfs stores checksums for each data block. Until now, they have
been stored in the subvolume trees, indexed by the inode that is
referencing the data block. This means that when we read the inode,
we've probably read in at least some checksums as well.
But, this has a few problems:
* The checksums are indexed by logical offset in the file. When
compression is on, this means we have to do the expensive checksumming
on the uncompressed data. It would be faster if we could checksum
the compressed data instead.
* If we implement encryption, we'll be checksumming the plain text and
storing that on disk. This is significantly less secure.
* For either compression or encryption, we have to get the plain text
back before we can verify the checksum as correct. This makes the raid
layer balancing and extent moving much more expensive.
* It makes the front end caching code more complex, as we have touch
the subvolume and inodes as we cache extents.
* There is potentitally one copy of the checksum in each subvolume
referencing an extent.
The solution used here is to store the extent checksums in a dedicated
tree. This allows us to index the checksums by phyiscal extent
start and length. It means:
* The checksum is against the data stored on disk, after any compression
or encryption is done.
* The checksum is stored in a central location, and can be verified without
following back references, or reading inodes.
This makes compression significantly faster by reducing the amount of
data that needs to be checksummed. It will also allow much faster
raid management code in general.
The checksums are indexed by a key with a fixed objectid (a magic value
in ctree.h) and offset set to the starting byte of the extent. This
allows us to copy the checksum items into the fsync log tree directly (or
any other tree), without having to invent a second format for them.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-12-09 00:58:54 +03:00
btrfs_stop_workers ( & fs_info - > endio_meta_workers ) ;
2008-12-17 22:51:42 +03:00
btrfs_stop_workers ( & fs_info - > endio_meta_write_workers ) ;
2008-07-17 20:53:50 +04:00
btrfs_stop_workers ( & fs_info - > endio_write_workers ) ;
2008-06-12 22:46:17 +04:00
btrfs_stop_workers ( & fs_info - > submit_workers ) ;
2008-04-30 21:59:35 +04:00
2008-05-13 21:46:40 +04:00
btrfs_close_devices ( fs_info - > fs_devices ) ;
2008-03-24 22:01:56 +03:00
btrfs_mapping_tree_free ( & fs_info - > mapping_tree ) ;
2008-04-14 17:48:18 +04:00
2008-03-26 17:28:07 +03:00
bdi_destroy ( & fs_info - > bdi ) ;
2009-09-22 00:00:26 +04:00
cleanup_srcu_struct ( & fs_info - > subvol_srcu ) ;
2008-03-24 22:01:56 +03:00
2007-04-09 18:42:37 +04:00
kfree ( fs_info - > extent_root ) ;
kfree ( fs_info - > tree_root ) ;
2008-03-24 22:01:56 +03:00
kfree ( fs_info - > chunk_root ) ;
kfree ( fs_info - > dev_root ) ;
Btrfs: move data checksumming into a dedicated tree
Btrfs stores checksums for each data block. Until now, they have
been stored in the subvolume trees, indexed by the inode that is
referencing the data block. This means that when we read the inode,
we've probably read in at least some checksums as well.
But, this has a few problems:
* The checksums are indexed by logical offset in the file. When
compression is on, this means we have to do the expensive checksumming
on the uncompressed data. It would be faster if we could checksum
the compressed data instead.
* If we implement encryption, we'll be checksumming the plain text and
storing that on disk. This is significantly less secure.
* For either compression or encryption, we have to get the plain text
back before we can verify the checksum as correct. This makes the raid
layer balancing and extent moving much more expensive.
* It makes the front end caching code more complex, as we have touch
the subvolume and inodes as we cache extents.
* There is potentitally one copy of the checksum in each subvolume
referencing an extent.
The solution used here is to store the extent checksums in a dedicated
tree. This allows us to index the checksums by phyiscal extent
start and length. It means:
* The checksum is against the data stored on disk, after any compression
or encryption is done.
* The checksum is stored in a central location, and can be verified without
following back references, or reading inodes.
This makes compression significantly faster by reducing the amount of
data that needs to be checksummed. It will also allow much faster
raid management code in general.
The checksums are indexed by a key with a fixed objectid (a magic value
in ctree.h) and offset set to the starting byte of the extent. This
allows us to copy the checksum items into the fsync log tree directly (or
any other tree), without having to invent a second format for them.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2008-12-09 00:58:54 +03:00
kfree ( fs_info - > csum_root ) ;
2007-02-02 17:18:22 +03:00
return 0 ;
}
2008-05-12 21:39:03 +04:00
int btrfs_buffer_uptodate ( struct extent_buffer * buf , u64 parent_transid )
2007-10-16 00:14:19 +04:00
{
2008-05-12 21:39:03 +04:00
int ret ;
2007-10-16 00:18:55 +04:00
struct inode * btree_inode = buf - > first_page - > mapping - > host ;
2008-05-12 21:39:03 +04:00
2010-02-03 22:33:23 +03:00
ret = extent_buffer_uptodate ( & BTRFS_I ( btree_inode ) - > io_tree , buf ,
NULL ) ;
2008-05-12 21:39:03 +04:00
if ( ! ret )
return ret ;
ret = verify_parent_transid ( & BTRFS_I ( btree_inode ) - > io_tree , buf ,
parent_transid ) ;
return ! ret ;
2007-10-16 00:14:19 +04:00
}
int btrfs_set_buffer_uptodate ( struct extent_buffer * buf )
2007-06-28 23:57:36 +04:00
{
2007-10-16 00:18:55 +04:00
struct inode * btree_inode = buf - > first_page - > mapping - > host ;
2008-01-25 00:13:08 +03:00
return set_extent_buffer_uptodate ( & BTRFS_I ( btree_inode ) - > io_tree ,
2007-10-16 00:14:19 +04:00
buf ) ;
}
2007-08-08 00:15:09 +04:00
2007-10-16 00:14:19 +04:00
void btrfs_mark_buffer_dirty ( struct extent_buffer * buf )
{
2007-10-16 00:18:55 +04:00
struct btrfs_root * root = BTRFS_I ( buf - > first_page - > mapping - > host ) - > root ;
2007-10-16 00:14:19 +04:00
u64 transid = btrfs_header_generation ( buf ) ;
struct inode * btree_inode = root - > fs_info - > btree_inode ;
2009-03-13 18:00:37 +03:00
int was_dirty ;
Btrfs: Change btree locking to use explicit blocking points
Most of the btrfs metadata operations can be protected by a spinlock,
but some operations still need to schedule.
So far, btrfs has been using a mutex along with a trylock loop,
most of the time it is able to avoid going for the full mutex, so
the trylock loop is a big performance gain.
This commit is step one for getting rid of the blocking locks entirely.
btrfs_tree_lock takes a spinlock, and the code explicitly switches
to a blocking lock when it starts an operation that can schedule.
We'll be able get rid of the blocking locks in smaller pieces over time.
Tracing allows us to find the most common cause of blocking, so we
can start with the hot spots first.
The basic idea is:
btrfs_tree_lock() returns with the spin lock held
btrfs_set_lock_blocking() sets the EXTENT_BUFFER_BLOCKING bit in
the extent buffer flags, and then drops the spin lock. The buffer is
still considered locked by all of the btrfs code.
If btrfs_tree_lock gets the spinlock but finds the blocking bit set, it drops
the spin lock and waits on a wait queue for the blocking bit to go away.
Much of the code that needs to set the blocking bit finishes without actually
blocking a good percentage of the time. So, an adaptive spin is still
used against the blocking bit to avoid very high context switch rates.
btrfs_clear_lock_blocking() clears the blocking bit and returns
with the spinlock held again.
btrfs_tree_unlock() can be called on either blocking or spinning locks,
it does the right thing based on the blocking bit.
ctree.c has a helper function to set/clear all the locked buffers in a
path as blocking.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-02-04 17:25:08 +03:00
2009-03-09 18:45:38 +03:00
btrfs_assert_tree_locked ( buf ) ;
2007-06-28 23:57:36 +04:00
if ( transid ! = root - > fs_info - > generation ) {
2009-01-06 05:25:51 +03:00
printk ( KERN_CRIT " btrfs transid mismatch buffer %llu, "
" found %llu running %llu \n " ,
2007-10-16 00:15:53 +04:00
( unsigned long long ) buf - > start ,
2009-01-06 05:25:51 +03:00
( unsigned long long ) transid ,
( unsigned long long ) root - > fs_info - > generation ) ;
2007-06-28 23:57:36 +04:00
WARN_ON ( 1 ) ;
}
2009-03-13 18:00:37 +03:00
was_dirty = set_extent_buffer_dirty ( & BTRFS_I ( btree_inode ) - > io_tree ,
buf ) ;
if ( ! was_dirty ) {
spin_lock ( & root - > fs_info - > delalloc_lock ) ;
root - > fs_info - > dirty_metadata_bytes + = buf - > len ;
spin_unlock ( & root - > fs_info - > delalloc_lock ) ;
}
2007-02-02 17:18:22 +03:00
}
2007-09-17 18:58:06 +04:00
void btrfs_btree_balance_dirty ( struct btrfs_root * root , unsigned long nr )
2007-05-02 23:53:43 +04:00
{
2008-05-09 19:52:25 +04:00
/*
* looks as though older kernels can get into trouble with
* this code , they end up stuck in balance_dirty_pages forever
*/
2008-04-30 21:59:35 +04:00
u64 num_dirty ;
2008-11-07 06:02:51 +03:00
unsigned long thresh = 32 * 1024 * 1024 ;
2008-04-30 21:59:35 +04:00
2009-03-17 11:36:37 +03:00
if ( current - > flags & PF_MEMALLOC )
2008-04-30 21:59:35 +04:00
return ;
2009-05-18 18:41:58 +04:00
num_dirty = root - > fs_info - > dirty_metadata_bytes ;
2008-04-30 21:59:35 +04:00
if ( num_dirty > thresh ) {
balance_dirty_pages_ratelimited_nr (
2008-02-18 20:12:38 +03:00
root - > fs_info - > btree_inode - > i_mapping , 1 ) ;
2008-04-30 21:59:35 +04:00
}
2008-05-09 19:52:25 +04:00
return ;
2007-05-02 23:53:43 +04:00
}
2007-10-16 00:17:34 +04:00
2008-05-12 20:59:19 +04:00
int btrfs_read_buffer ( struct extent_buffer * buf , u64 parent_transid )
2007-10-16 00:17:34 +04:00
{
2007-10-16 00:18:55 +04:00
struct btrfs_root * root = BTRFS_I ( buf - > first_page - > mapping - > host ) - > root ;
2008-04-10 00:28:12 +04:00
int ret ;
2008-05-12 20:59:19 +04:00
ret = btree_read_extent_buffer_pages ( root , buf , 0 , parent_transid ) ;
2009-01-06 05:25:51 +03:00
if ( ret = = 0 )
Btrfs: Change btree locking to use explicit blocking points
Most of the btrfs metadata operations can be protected by a spinlock,
but some operations still need to schedule.
So far, btrfs has been using a mutex along with a trylock loop,
most of the time it is able to avoid going for the full mutex, so
the trylock loop is a big performance gain.
This commit is step one for getting rid of the blocking locks entirely.
btrfs_tree_lock takes a spinlock, and the code explicitly switches
to a blocking lock when it starts an operation that can schedule.
We'll be able get rid of the blocking locks in smaller pieces over time.
Tracing allows us to find the most common cause of blocking, so we
can start with the hot spots first.
The basic idea is:
btrfs_tree_lock() returns with the spin lock held
btrfs_set_lock_blocking() sets the EXTENT_BUFFER_BLOCKING bit in
the extent buffer flags, and then drops the spin lock. The buffer is
still considered locked by all of the btrfs code.
If btrfs_tree_lock gets the spinlock but finds the blocking bit set, it drops
the spin lock and waits on a wait queue for the blocking bit to go away.
Much of the code that needs to set the blocking bit finishes without actually
blocking a good percentage of the time. So, an adaptive spin is still
used against the blocking bit to avoid very high context switch rates.
btrfs_clear_lock_blocking() clears the blocking bit and returns
with the spinlock held again.
btrfs_tree_unlock() can be called on either blocking or spinning locks,
it does the right thing based on the blocking bit.
ctree.c has a helper function to set/clear all the locked buffers in a
path as blocking.
Signed-off-by: Chris Mason <chris.mason@oracle.com>
2009-02-04 17:25:08 +03:00
set_bit ( EXTENT_BUFFER_UPTODATE , & buf - > bflags ) ;
2008-04-10 00:28:12 +04:00
return ret ;
2007-10-16 00:17:34 +04:00
}
2007-11-08 05:08:01 +03:00
2008-09-08 19:18:08 +04:00
int btree_lock_page_hook ( struct page * page )
{
struct inode * inode = page - > mapping - > host ;
2009-03-13 18:00:37 +03:00
struct btrfs_root * root = BTRFS_I ( inode ) - > root ;
2008-09-08 19:18:08 +04:00
struct extent_io_tree * io_tree = & BTRFS_I ( inode ) - > io_tree ;
struct extent_buffer * eb ;
unsigned long len ;
u64 bytenr = page_offset ( page ) ;
if ( page - > private = = EXTENT_PAGE_PRIVATE )
goto out ;
len = page - > private > > 2 ;
eb = find_extent_buffer ( io_tree , bytenr , len , GFP_NOFS ) ;
if ( ! eb )
goto out ;
btrfs_tree_lock ( eb ) ;
btrfs_set_header_flag ( eb , BTRFS_HEADER_FLAG_WRITTEN ) ;
2009-03-13 18:00:37 +03:00
if ( test_and_clear_bit ( EXTENT_BUFFER_DIRTY , & eb - > bflags ) ) {
spin_lock ( & root - > fs_info - > delalloc_lock ) ;
if ( root - > fs_info - > dirty_metadata_bytes > = eb - > len )
root - > fs_info - > dirty_metadata_bytes - = eb - > len ;
else
WARN_ON ( 1 ) ;
spin_unlock ( & root - > fs_info - > delalloc_lock ) ;
}
2008-09-08 19:18:08 +04:00
btrfs_tree_unlock ( eb ) ;
free_extent_buffer ( eb ) ;
out :
lock_page ( page ) ;
return 0 ;
}
2008-01-25 00:13:08 +03:00
static struct extent_io_ops btree_extent_io_ops = {
2008-09-08 19:18:08 +04:00
. write_cache_pages_lock_hook = btree_lock_page_hook ,
2008-04-10 00:28:12 +04:00
. readpage_end_io_hook = btree_readpage_end_io_hook ,
2008-03-24 22:01:56 +03:00
. submit_bio_hook = btree_submit_bio_hook ,
2008-03-24 22:02:07 +03:00
/* note we're sharing with inode.c for the merge bio hook */
. merge_bio_hook = btrfs_merge_bio_hook ,
2007-11-08 05:08:01 +03:00
} ;