2017-03-17 09:18:50 +03:00
// SPDX-License-Identifier: GPL-2.0
/*
* bcachefs journalling code , for btree insertions
*
* Copyright 2012 Google , Inc .
*/
# include "bcachefs.h"
2018-10-06 07:46:55 +03:00
# include "alloc_foreground.h"
2017-03-17 09:18:50 +03:00
# include "bkey_methods.h"
# include "btree_gc.h"
2021-01-23 01:56:34 +03:00
# include "btree_update.h"
2023-11-03 01:57:19 +03:00
# include "btree_write_buffer.h"
2017-03-17 09:18:50 +03:00
# include "buckets.h"
2021-04-03 23:24:13 +03:00
# include "error.h"
2017-03-17 09:18:50 +03:00
# include "journal.h"
# include "journal_io.h"
# include "journal_reclaim.h"
2022-03-11 00:43:52 +03:00
# include "journal_sb.h"
2017-03-17 09:18:50 +03:00
# include "journal_seq_blacklist.h"
# include "trace.h"
2022-03-15 04:48:42 +03:00
static const char * const bch2_journal_errors [ ] = {
2023-06-28 00:32:38 +03:00
# define x(n) #n,
2022-03-15 04:48:42 +03:00
JOURNAL_ERRORS ( )
2023-06-28 00:32:38 +03:00
# undef x
2022-03-15 04:48:42 +03:00
NULL
} ;
2024-01-31 19:28:13 +03:00
static inline bool journal_seq_unwritten ( struct journal * j , u64 seq )
{
return seq > j - > seq_ondisk ;
}
static bool __journal_entry_is_open ( union journal_res_state state )
{
return state . cur_entry_offset < JOURNAL_ENTRY_CLOSED_VAL ;
}
static inline unsigned nr_unwritten_journal_entries ( struct journal * j )
{
return atomic64_read ( & j - > seq ) - j - > seq_ondisk ;
}
static bool journal_entry_is_open ( struct journal * j )
{
return __journal_entry_is_open ( j - > reservations ) ;
}
2024-01-16 01:59:51 +03:00
static void bch2_journal_buf_to_text ( struct printbuf * out , struct journal * j , u64 seq )
{
union journal_res_state s = READ_ONCE ( j - > reservations ) ;
unsigned i = seq & JOURNAL_BUF_MASK ;
struct journal_buf * buf = j - > buf + i ;
prt_printf ( out , " seq: " ) ;
prt_tab ( out ) ;
prt_printf ( out , " %llu " , seq ) ;
prt_newline ( out ) ;
printbuf_indent_add ( out , 2 ) ;
prt_printf ( out , " refcount: " ) ;
prt_tab ( out ) ;
prt_printf ( out , " %u " , journal_state_count ( s , i ) ) ;
prt_newline ( out ) ;
prt_printf ( out , " size: " ) ;
prt_tab ( out ) ;
prt_human_readable_u64 ( out , vstruct_bytes ( buf - > data ) ) ;
prt_newline ( out ) ;
prt_printf ( out , " expires " ) ;
prt_tab ( out ) ;
prt_printf ( out , " %li jiffies " , buf - > expires - jiffies ) ;
prt_newline ( out ) ;
2024-01-31 22:26:15 +03:00
if ( buf - > write_done )
prt_printf ( out , " write done \n " ) ;
else if ( buf - > write_allocated )
prt_printf ( out , " write allocated \n " ) ;
else if ( buf - > write_started )
prt_printf ( out , " write started \n " ) ;
2024-01-16 01:59:51 +03:00
printbuf_indent_sub ( out , 2 ) ;
}
static void bch2_journal_bufs_to_text ( struct printbuf * out , struct journal * j )
{
if ( ! out - > nr_tabstops )
printbuf_tabstop_push ( out , 24 ) ;
for ( u64 seq = journal_last_unwritten_seq ( j ) ;
seq < = journal_cur_seq ( j ) ;
seq + + )
bch2_journal_buf_to_text ( out , j , seq ) ;
2024-01-31 19:28:13 +03:00
prt_printf ( out , " last buf %s \n " , journal_entry_is_open ( j ) ? " open " : " closed " ) ;
2017-03-17 09:18:50 +03:00
}
2020-11-15 00:04:30 +03:00
static inline struct journal_buf *
journal_seq_to_buf ( struct journal * j , u64 seq )
{
struct journal_buf * buf = NULL ;
EBUG_ON ( seq > journal_cur_seq ( j ) ) ;
if ( journal_seq_unwritten ( j , seq ) ) {
2020-11-14 02:36:33 +03:00
buf = j - > buf + ( seq & JOURNAL_BUF_MASK ) ;
2020-11-15 00:04:30 +03:00
EBUG_ON ( le64_to_cpu ( buf - > data - > seq ) ! = seq ) ;
}
return buf ;
}
2021-04-01 04:44:55 +03:00
static void journal_pin_list_init ( struct journal_entry_pin_list * p , int count )
2017-03-17 09:18:50 +03:00
{
2023-03-07 15:28:20 +03:00
unsigned i ;
2023-08-07 19:04:05 +03:00
2023-03-07 15:28:20 +03:00
for ( i = 0 ; i < ARRAY_SIZE ( p - > list ) ; i + + )
INIT_LIST_HEAD ( & p - > list [ i ] ) ;
2021-04-01 04:44:55 +03:00
INIT_LIST_HEAD ( & p - > flushed ) ;
atomic_set ( & p - > count , count ) ;
p - > devs . nr = 0 ;
}
2017-03-17 09:18:50 +03:00
2023-03-21 15:03:18 +03:00
/*
* Detect stuck journal conditions and trigger shutdown . Technically the journal
* can end up stuck for a variety of reasons , such as a blocked I / O , journal
* reservation lockup , etc . Since this is a fatal error with potentially
* unpredictable characteristics , we want to be fairly conservative before we
* decide to shut things down .
*
* Consider the journal stuck when it appears full with no ability to commit
* btree transactions , to discard journal buckets , nor acquire priority
* ( reserved watermark ) reservation .
*/
static inline bool
journal_error_check_stuck ( struct journal * j , int error , unsigned flags )
{
struct bch_fs * c = container_of ( j , struct bch_fs , journal ) ;
bool stuck = false ;
struct printbuf buf = PRINTBUF ;
if ( ! ( error = = JOURNAL_ERR_journal_full | |
error = = JOURNAL_ERR_journal_pin_full ) | |
nr_unwritten_journal_entries ( j ) | |
2023-06-28 00:32:38 +03:00
( flags & BCH_WATERMARK_MASK ) ! = BCH_WATERMARK_reclaim )
2023-03-21 15:03:18 +03:00
return stuck ;
spin_lock ( & j - > lock ) ;
if ( j - > can_discard ) {
spin_unlock ( & j - > lock ) ;
return stuck ;
}
stuck = true ;
/*
* The journal shutdown path will set - > err_seq , but do it here first to
* serialize against concurrent failures and avoid duplicate error
* reports .
*/
if ( j - > err_seq ) {
spin_unlock ( & j - > lock ) ;
return stuck ;
}
j - > err_seq = journal_cur_seq ( j ) ;
spin_unlock ( & j - > lock ) ;
bch_err ( c , " Journal stuck! Hava a pre-reservation but journal full (error %s) " ,
bch2_journal_errors [ error ] ) ;
bch2_journal_debug_to_text ( & buf , j ) ;
bch_err ( c , " %s " , buf . buf ) ;
printbuf_reset ( & buf ) ;
bch2_journal_pins_to_text ( & buf , j ) ;
bch_err ( c , " Journal pins: \n %s " , buf . buf ) ;
printbuf_exit ( & buf ) ;
bch2_fatal_error ( c ) ;
dump_stack ( ) ;
return stuck ;
}
2024-01-31 22:26:15 +03:00
void bch2_journal_do_writes ( struct journal * j )
{
for ( u64 seq = journal_last_unwritten_seq ( j ) ;
seq < = journal_cur_seq ( j ) ;
seq + + ) {
unsigned idx = seq & JOURNAL_BUF_MASK ;
struct journal_buf * w = j - > buf + idx ;
if ( w - > write_started & & ! w - > write_allocated )
break ;
if ( w - > write_started )
continue ;
if ( ! journal_state_count ( j - > reservations , idx ) ) {
w - > write_started = true ;
closure_call ( & w - > io , bch2_journal_write , j - > wq , NULL ) ;
}
break ;
}
}
bcachefs: fix race between journal entry close and pin set
bcachefs freeze testing via fstests generic/390 occasionally
reproduces the following BUG from bch2_fs_read_only():
BUG_ON(atomic_long_read(&c->btree_key_cache.nr_dirty));
This indicates that one or more dirty key cache keys still exist
after the attempt to flush and quiesce the fs. The sequence that
leads to this problem actually occurs on unfreeze (ro->rw), and
looks something like the following:
- Task A begins a transaction commit and acquires journal_res for
the current seq. This transaction intends to perform key cache
insertion.
- Task B begins a bch2_journal_flush() via bch2_sync_fs(). This ends
up in journal_entry_want_write(), which closes the current journal
entry and drops the reference to the pin list created on entry open.
The pin put pops the front of the journal via fast reclaim since the
reference count has dropped to 0.
- Task A attempts to set the journal pin for the associated cached
key, but bch2_journal_pin_set() skips the pin insert because the
seq of the transaction reservation is behind the front of the pin
list fifo.
The end result is that the pin associated with the cached key is not
added, which prevents a subsequent reclaim from processing the key
and thus leaves it dangling at freeze time. The fundamental cause of
this problem is that the front of the journal is allowed to pop
before a transaction with outstanding reservation on the associated
journal seq is able to add a pin. The count for the pin list
associated with the seq drops to zero and is prematurely reclaimed
as a result.
The logical fix for this problem lies in how the journal buffer is
managed in similar scenarios where the entry might have been closed
before a transaction with outstanding reservations happens to be
committed.
When a journal entry is opened, the current sequence number is
bumped, the associated pin list is initialized with a reference
count of 1, and the journal buffer reference count is bumped (via
journal_state_inc()). When a journal reservation is acquired, the
reservation also acquires a reference on the associated buffer. If
the journal entry is closed in the meantime, it drops both the pin
and buffer references held by the open entry, but the buffer still
has references held by outstanding reservation. After the associated
transaction commits, the reservation release drops the associated
buffer references and the buffer is written out once the reference
count has dropped to zero.
The fundamental problem here is that the lifecycle of the pin list
reference held by an open journal entry is too short to cover the
processing of transactions with outstanding reservations. The
simplest way to address this is to expand the pin list reference to
the lifecycle of the buffer vs. the shorter lifecycle of the open
journal entry. This ensures the pin list for a seq with outstanding
reservation cannot be popped and reclaimed before all outstanding
reservations have been released, even if the associated journal
entry has been closed for further reservations.
Move the pin put from journal entry close to where final processing
of the journal buffer occurs. Create a duplicate helper to cover the
case where the caller doesn't already hold the journal lock. This
allows generic/390 to pass reliably.
Signed-off-by: Brian Foster <bfoster@redhat.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2023-09-15 15:51:53 +03:00
/*
* Final processing when the last reference of a journal buffer has been
* dropped . Drop the pin list reference acquired at journal entry open and write
* the buffer , if requested .
*/
2024-01-31 22:26:15 +03:00
void bch2_journal_buf_put_final ( struct journal * j , u64 seq )
2019-02-19 01:39:42 +03:00
{
bcachefs: fix race between journal entry close and pin set
bcachefs freeze testing via fstests generic/390 occasionally
reproduces the following BUG from bch2_fs_read_only():
BUG_ON(atomic_long_read(&c->btree_key_cache.nr_dirty));
This indicates that one or more dirty key cache keys still exist
after the attempt to flush and quiesce the fs. The sequence that
leads to this problem actually occurs on unfreeze (ro->rw), and
looks something like the following:
- Task A begins a transaction commit and acquires journal_res for
the current seq. This transaction intends to perform key cache
insertion.
- Task B begins a bch2_journal_flush() via bch2_sync_fs(). This ends
up in journal_entry_want_write(), which closes the current journal
entry and drops the reference to the pin list created on entry open.
The pin put pops the front of the journal via fast reclaim since the
reference count has dropped to 0.
- Task A attempts to set the journal pin for the associated cached
key, but bch2_journal_pin_set() skips the pin insert because the
seq of the transaction reservation is behind the front of the pin
list fifo.
The end result is that the pin associated with the cached key is not
added, which prevents a subsequent reclaim from processing the key
and thus leaves it dangling at freeze time. The fundamental cause of
this problem is that the front of the journal is allowed to pop
before a transaction with outstanding reservation on the associated
journal seq is able to add a pin. The count for the pin list
associated with the seq drops to zero and is prematurely reclaimed
as a result.
The logical fix for this problem lies in how the journal buffer is
managed in similar scenarios where the entry might have been closed
before a transaction with outstanding reservations happens to be
committed.
When a journal entry is opened, the current sequence number is
bumped, the associated pin list is initialized with a reference
count of 1, and the journal buffer reference count is bumped (via
journal_state_inc()). When a journal reservation is acquired, the
reservation also acquires a reference on the associated buffer. If
the journal entry is closed in the meantime, it drops both the pin
and buffer references held by the open entry, but the buffer still
has references held by outstanding reservation. After the associated
transaction commits, the reservation release drops the associated
buffer references and the buffer is written out once the reference
count has dropped to zero.
The fundamental problem here is that the lifecycle of the pin list
reference held by an open journal entry is too short to cover the
processing of transactions with outstanding reservations. The
simplest way to address this is to expand the pin list reference to
the lifecycle of the buffer vs. the shorter lifecycle of the open
journal entry. This ensures the pin list for a seq with outstanding
reservation cannot be popped and reclaimed before all outstanding
reservations have been released, even if the associated journal
entry has been closed for further reservations.
Move the pin put from journal entry close to where final processing
of the journal buffer occurs. Create a duplicate helper to cover the
case where the caller doesn't already hold the journal lock. This
allows generic/390 to pass reliably.
Signed-off-by: Brian Foster <bfoster@redhat.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2023-09-15 15:51:53 +03:00
lockdep_assert_held ( & j - > lock ) ;
if ( __bch2_journal_pin_put ( j , seq ) )
bch2_journal_reclaim_fast ( j ) ;
2024-01-31 22:26:15 +03:00
bch2_journal_do_writes ( j ) ;
2019-02-19 01:39:42 +03:00
}
/*
* Returns true if journal entry is now closed :
2021-01-15 00:21:22 +03:00
*
* We don ' t close a journal_buf until the next journal_buf is finished writing ,
* and can be opened again - this also initializes the next journal_buf :
2019-02-19 01:39:42 +03:00
*/
2024-01-16 01:59:51 +03:00
static void __journal_entry_close ( struct journal * j , unsigned closed_val , bool trace )
2017-03-17 09:18:50 +03:00
{
struct bch_fs * c = container_of ( j , struct bch_fs , journal ) ;
2018-07-23 14:52:00 +03:00
struct journal_buf * buf = journal_cur_buf ( j ) ;
2017-03-17 09:18:50 +03:00
union journal_res_state old , new ;
u64 v = atomic64_read ( & j - > reservations . counter ) ;
2019-02-19 01:39:42 +03:00
unsigned sectors ;
2017-03-17 09:18:50 +03:00
2022-03-01 02:48:33 +03:00
BUG_ON ( closed_val ! = JOURNAL_ENTRY_CLOSED_VAL & &
closed_val ! = JOURNAL_ENTRY_ERROR_VAL ) ;
2017-03-17 09:18:50 +03:00
lockdep_assert_held ( & j - > lock ) ;
do {
old . v = new . v = v ;
2022-03-01 02:48:33 +03:00
new . cur_entry_offset = closed_val ;
2017-03-17 09:18:50 +03:00
2022-03-01 02:48:33 +03:00
if ( old . cur_entry_offset = = JOURNAL_ENTRY_ERROR_VAL | |
old . cur_entry_offset = = new . cur_entry_offset )
2022-03-01 03:29:19 +03:00
return ;
2017-03-17 09:18:50 +03:00
} while ( ( v = atomic64_cmpxchg ( & j - > reservations . counter ,
old . v , new . v ) ) ! = old . v ) ;
2022-03-01 02:48:33 +03:00
if ( ! __journal_entry_is_open ( old ) )
return ;
2021-01-15 00:21:22 +03:00
/* Close out old buffer: */
2017-03-17 09:18:50 +03:00
buf - > data - > u64s = cpu_to_le32 ( old . cur_entry_offset ) ;
2024-01-16 01:59:51 +03:00
if ( trace_journal_entry_close_enabled ( ) & & trace ) {
struct printbuf pbuf = PRINTBUF ;
pbuf . atomic + + ;
prt_str ( & pbuf , " entry size: " ) ;
prt_human_readable_u64 ( & pbuf , vstruct_bytes ( buf - > data ) ) ;
prt_newline ( & pbuf ) ;
2024-01-22 20:25:00 +03:00
bch2_prt_task_backtrace ( & pbuf , current , 1 , GFP_NOWAIT ) ;
2024-01-16 01:59:51 +03:00
trace_journal_entry_close ( c , pbuf . buf ) ;
printbuf_exit ( & pbuf ) ;
}
2023-11-17 04:41:10 +03:00
2019-02-19 01:39:42 +03:00
sectors = vstruct_blocks_plus ( buf - > data , c - > block_bits ,
buf - > u64s_reserved ) < < c - > block_bits ;
BUG_ON ( sectors > buf - > sectors ) ;
buf - > sectors = sectors ;
2017-03-17 09:18:50 +03:00
2018-07-17 19:19:14 +03:00
/*
* We have to set last_seq here , _before_ opening a new journal entry :
*
* A threads may replace an old pin with a new pin on their current
* journal reservation - the expectation being that the journal will
* contain either what the old pin protected or what the new pin
* protects .
*
* After the old pin is dropped journal_last_seq ( ) won ' t include the old
* pin , so we can only write the updated last_seq on the entry that
* contains whatever the new pin protects .
*
* Restated , we can _not_ update last_seq for a given entry if there
* could be a newer entry open with reservations / pins that have been
* taken against it .
*
* Hence , we want update / set last_seq on the current journal entry right
* before we open a new one :
*/
2021-05-08 06:32:26 +03:00
buf - > last_seq = journal_last_seq ( j ) ;
buf - > data - > last_seq = cpu_to_le64 ( buf - > last_seq ) ;
2022-03-01 03:17:27 +03:00
BUG_ON ( buf - > last_seq > le64_to_cpu ( buf - > data - > seq ) ) ;
2017-03-17 09:18:50 +03:00
cancel_delayed_work ( & j - > write_work ) ;
2019-02-21 21:33:21 +03:00
bch2_journal_space_available ( j ) ;
bcachefs: fix race between journal entry close and pin set
bcachefs freeze testing via fstests generic/390 occasionally
reproduces the following BUG from bch2_fs_read_only():
BUG_ON(atomic_long_read(&c->btree_key_cache.nr_dirty));
This indicates that one or more dirty key cache keys still exist
after the attempt to flush and quiesce the fs. The sequence that
leads to this problem actually occurs on unfreeze (ro->rw), and
looks something like the following:
- Task A begins a transaction commit and acquires journal_res for
the current seq. This transaction intends to perform key cache
insertion.
- Task B begins a bch2_journal_flush() via bch2_sync_fs(). This ends
up in journal_entry_want_write(), which closes the current journal
entry and drops the reference to the pin list created on entry open.
The pin put pops the front of the journal via fast reclaim since the
reference count has dropped to 0.
- Task A attempts to set the journal pin for the associated cached
key, but bch2_journal_pin_set() skips the pin insert because the
seq of the transaction reservation is behind the front of the pin
list fifo.
The end result is that the pin associated with the cached key is not
added, which prevents a subsequent reclaim from processing the key
and thus leaves it dangling at freeze time. The fundamental cause of
this problem is that the front of the journal is allowed to pop
before a transaction with outstanding reservation on the associated
journal seq is able to add a pin. The count for the pin list
associated with the seq drops to zero and is prematurely reclaimed
as a result.
The logical fix for this problem lies in how the journal buffer is
managed in similar scenarios where the entry might have been closed
before a transaction with outstanding reservations happens to be
committed.
When a journal entry is opened, the current sequence number is
bumped, the associated pin list is initialized with a reference
count of 1, and the journal buffer reference count is bumped (via
journal_state_inc()). When a journal reservation is acquired, the
reservation also acquires a reference on the associated buffer. If
the journal entry is closed in the meantime, it drops both the pin
and buffer references held by the open entry, but the buffer still
has references held by outstanding reservation. After the associated
transaction commits, the reservation release drops the associated
buffer references and the buffer is written out once the reference
count has dropped to zero.
The fundamental problem here is that the lifecycle of the pin list
reference held by an open journal entry is too short to cover the
processing of transactions with outstanding reservations. The
simplest way to address this is to expand the pin list reference to
the lifecycle of the buffer vs. the shorter lifecycle of the open
journal entry. This ensures the pin list for a seq with outstanding
reservation cannot be popped and reclaimed before all outstanding
reservations have been released, even if the associated journal
entry has been closed for further reservations.
Move the pin put from journal entry close to where final processing
of the journal buffer occurs. Create a duplicate helper to cover the
case where the caller doesn't already hold the journal lock. This
allows generic/390 to pass reliably.
Signed-off-by: Brian Foster <bfoster@redhat.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2023-09-15 15:51:53 +03:00
__bch2_journal_buf_put ( j , old . idx , le64_to_cpu ( buf - > data - > seq ) ) ;
2017-03-17 09:18:50 +03:00
}
2022-03-01 02:48:33 +03:00
void bch2_journal_halt ( struct journal * j )
{
spin_lock ( & j - > lock ) ;
2024-01-16 01:59:51 +03:00
__journal_entry_close ( j , JOURNAL_ENTRY_ERROR_VAL , true ) ;
2022-03-01 02:48:33 +03:00
if ( ! j - > err_seq )
j - > err_seq = journal_cur_seq ( j ) ;
2023-03-20 20:21:19 +03:00
journal_wake ( j ) ;
2022-03-01 02:48:33 +03:00
spin_unlock ( & j - > lock ) ;
}
2020-11-14 02:36:33 +03:00
static bool journal_entry_want_write ( struct journal * j )
{
2022-03-01 03:17:27 +03:00
bool ret = ! journal_entry_is_open ( j ) | |
2022-03-01 03:29:19 +03:00
journal_cur_seq ( j ) = = journal_last_unwritten_seq ( j ) ;
2022-02-25 18:28:20 +03:00
/* Don't close it yet if we already have a write in flight: */
2022-03-01 03:29:19 +03:00
if ( ret )
2024-01-16 01:59:51 +03:00
__journal_entry_close ( j , JOURNAL_ENTRY_CLOSED_VAL , true ) ;
2022-03-01 03:29:19 +03:00
else if ( nr_unwritten_journal_entries ( j ) ) {
2022-03-01 03:17:27 +03:00
struct journal_buf * buf = journal_cur_buf ( j ) ;
if ( ! buf - > flush_time ) {
buf - > flush_time = local_clock ( ) ? : 1 ;
buf - > expires = jiffies ;
}
}
2020-11-14 02:36:33 +03:00
return ret ;
}
2023-12-10 23:23:27 +03:00
bool bch2_journal_entry_close ( struct journal * j )
2017-03-17 09:18:50 +03:00
{
2019-02-19 01:39:42 +03:00
bool ret ;
2017-03-17 09:18:50 +03:00
2019-02-19 01:39:42 +03:00
spin_lock ( & j - > lock ) ;
2020-11-14 02:36:33 +03:00
ret = journal_entry_want_write ( j ) ;
2019-02-19 01:39:42 +03:00
spin_unlock ( & j - > lock ) ;
2017-03-17 09:18:50 +03:00
2019-02-19 01:39:42 +03:00
return ret ;
2017-03-17 09:18:50 +03:00
}
/*
* should _only_ called from journal_res_get ( ) - when we actually want a
* journal reservation - journal entry is open means journal is dirty :
*/
static int journal_entry_open ( struct journal * j )
{
2020-11-20 04:55:33 +03:00
struct bch_fs * c = container_of ( j , struct bch_fs , journal ) ;
2022-03-01 03:29:19 +03:00
struct journal_buf * buf = j - > buf +
( ( journal_cur_seq ( j ) + 1 ) & JOURNAL_BUF_MASK ) ;
2017-03-17 09:18:50 +03:00
union journal_res_state old , new ;
2019-02-21 21:33:21 +03:00
int u64s ;
2017-03-17 09:18:50 +03:00
u64 v ;
lockdep_assert_held ( & j - > lock ) ;
BUG_ON ( journal_entry_is_open ( j ) ) ;
2022-03-01 03:29:19 +03:00
BUG_ON ( BCH_SB_CLEAN ( c - > disk_sb . sb ) ) ;
2017-03-17 09:18:50 +03:00
2019-02-15 02:38:52 +03:00
if ( j - > blocked )
2022-03-15 04:48:42 +03:00
return JOURNAL_ERR_blocked ;
2019-02-15 02:38:52 +03:00
2019-02-21 21:33:21 +03:00
if ( j - > cur_entry_error )
return j - > cur_entry_error ;
2017-03-17 09:18:50 +03:00
2022-03-01 00:21:07 +03:00
if ( bch2_journal_error ( j ) )
2022-03-15 04:48:42 +03:00
return JOURNAL_ERR_insufficient_devices ; /* -EROFS */
2022-03-01 00:21:07 +03:00
2022-03-01 00:35:42 +03:00
if ( ! fifo_free ( & j - > pin ) )
2022-03-15 04:48:42 +03:00
return JOURNAL_ERR_journal_pin_full ;
2017-03-17 09:18:50 +03:00
2022-03-01 03:29:19 +03:00
if ( nr_unwritten_journal_entries ( j ) = = ARRAY_SIZE ( j - > buf ) )
2022-03-15 04:48:42 +03:00
return JOURNAL_ERR_max_in_flight ;
2022-03-01 03:29:19 +03:00
2022-03-01 00:35:42 +03:00
BUG_ON ( ! j - > cur_entry_sectors ) ;
2022-02-28 23:51:24 +03:00
2022-03-01 00:35:42 +03:00
buf - > expires =
( journal_cur_seq ( j ) = = j - > flushed_seq_ondisk
? jiffies
: j - > last_flush_write ) +
2022-02-25 18:28:20 +03:00
msecs_to_jiffies ( c - > opts . journal_flush_delay ) ;
2022-03-01 00:35:42 +03:00
2019-01-25 00:50:48 +03:00
buf - > u64s_reserved = j - > entry_u64s_reserved ;
2019-02-19 01:39:42 +03:00
buf - > disk_sectors = j - > cur_entry_sectors ;
buf - > sectors = min ( buf - > disk_sectors , buf - > buf_size > > 9 ) ;
2017-03-17 09:18:50 +03:00
2019-02-19 01:39:42 +03:00
u64s = ( int ) ( buf - > sectors < < 9 ) / sizeof ( u64 ) -
journal_entry_overhead ( j ) ;
2022-03-01 00:35:42 +03:00
u64s = clamp_t ( int , u64s , 0 , JOURNAL_ENTRY_CLOSED_VAL - 1 ) ;
2017-03-17 09:18:50 +03:00
2022-12-14 18:39:04 +03:00
if ( u64s < = ( ssize_t ) j - > early_journal_entries . nr )
2022-03-15 04:48:42 +03:00
return JOURNAL_ERR_journal_full ;
2017-03-17 09:18:50 +03:00
2022-03-07 01:20:39 +03:00
if ( fifo_empty ( & j - > pin ) & & j - > reclaim_thread )
wake_up_process ( j - > reclaim_thread ) ;
2022-03-01 00:35:42 +03:00
/*
* The fifo_push ( ) needs to happen at the same time as j - > seq is
* incremented for journal_last_seq ( ) to be calculated correctly
*/
atomic64_inc ( & j - > seq ) ;
journal_pin_list_init ( fifo_push_ref ( & j - > pin ) , 1 ) ;
2023-11-27 08:53:46 +03:00
BUG_ON ( j - > pin . back - 1 ! = atomic64_read ( & j - > seq ) ) ;
2022-03-01 00:35:42 +03:00
BUG_ON ( j - > buf + ( journal_cur_seq ( j ) & JOURNAL_BUF_MASK ) ! = buf ) ;
bkey_extent_init ( & buf - > key ) ;
2024-01-31 22:26:15 +03:00
buf - > noflush = false ;
buf - > must_flush = false ;
buf - > separate_flush = false ;
buf - > flush_time = 0 ;
2023-11-03 01:57:19 +03:00
buf - > need_flush_to_write_buffer = true ;
2024-01-31 22:26:15 +03:00
buf - > write_started = false ;
buf - > write_allocated = false ;
buf - > write_done = false ;
2022-03-01 00:35:42 +03:00
memset ( buf - > data , 0 , sizeof ( * buf - > data ) ) ;
buf - > data - > seq = cpu_to_le64 ( journal_cur_seq ( j ) ) ;
buf - > data - > u64s = 0 ;
2022-12-14 18:39:04 +03:00
if ( j - > early_journal_entries . nr ) {
memcpy ( buf - > data - > _data , j - > early_journal_entries . data ,
j - > early_journal_entries . nr * sizeof ( u64 ) ) ;
le32_add_cpu ( & buf - > data - > u64s , j - > early_journal_entries . nr ) ;
}
2017-03-17 09:18:50 +03:00
/*
* Must be set before marking the journal entry as open :
*/
j - > cur_entry_u64s = u64s ;
v = atomic64_read ( & j - > reservations . counter ) ;
do {
old . v = new . v = v ;
2022-03-01 00:21:07 +03:00
BUG_ON ( old . cur_entry_offset = = JOURNAL_ENTRY_ERROR_VAL ) ;
2022-03-01 03:29:19 +03:00
new . idx + + ;
2022-03-01 00:35:42 +03:00
BUG_ON ( journal_state_count ( new , new . idx ) ) ;
2022-03-01 03:29:19 +03:00
BUG_ON ( new . idx ! = ( journal_cur_seq ( j ) & JOURNAL_BUF_MASK ) ) ;
2017-03-17 09:18:50 +03:00
2019-02-19 01:39:42 +03:00
journal_state_inc ( & new ) ;
2022-12-14 18:39:04 +03:00
/* Handle any already added entries */
new . cur_entry_offset = le32_to_cpu ( buf - > data - > u64s ) ;
2017-03-17 09:18:50 +03:00
} while ( ( v = atomic64_cmpxchg ( & j - > reservations . counter ,
old . v , new . v ) ) ! = old . v ) ;
2024-01-31 19:24:37 +03:00
if ( nr_unwritten_journal_entries ( j ) = = 1 )
mod_delayed_work ( j - > wq ,
& j - > write_work ,
msecs_to_jiffies ( c - > opts . journal_flush_delay ) ) ;
2017-03-17 09:18:50 +03:00
journal_wake ( j ) ;
2022-12-14 18:39:04 +03:00
if ( j - > early_journal_entries . nr )
darray_exit ( & j - > early_journal_entries ) ;
2019-02-19 01:39:42 +03:00
return 0 ;
2017-03-17 09:18:50 +03:00
}
2019-02-15 02:38:52 +03:00
static bool journal_quiesced ( struct journal * j )
{
2022-03-01 02:48:33 +03:00
bool ret = atomic64_read ( & j - > seq ) = = j - > seq_ondisk ;
2019-02-15 02:38:52 +03:00
2019-02-19 01:39:42 +03:00
if ( ! ret )
2023-12-10 23:23:27 +03:00
bch2_journal_entry_close ( j ) ;
2019-02-15 02:38:52 +03:00
return ret ;
}
static void journal_quiesce ( struct journal * j )
{
wait_event ( j - > wait , journal_quiesced ( j ) ) ;
}
2017-03-17 09:18:50 +03:00
static void journal_write_work ( struct work_struct * work )
{
struct journal * j = container_of ( work , struct journal , write_work . work ) ;
2022-02-25 18:28:20 +03:00
spin_lock ( & j - > lock ) ;
2024-01-31 19:24:37 +03:00
if ( __journal_entry_is_open ( j - > reservations ) ) {
long delta = journal_cur_buf ( j ) - > expires - jiffies ;
2022-03-05 21:38:54 +03:00
2024-01-31 19:24:37 +03:00
if ( delta > 0 )
mod_delayed_work ( j - > wq , & j - > write_work , delta ) ;
else
__journal_entry_close ( j , JOURNAL_ENTRY_CLOSED_VAL , true ) ;
}
2022-02-25 18:28:20 +03:00
spin_unlock ( & j - > lock ) ;
2017-03-17 09:18:50 +03:00
}
static int __journal_res_get ( struct journal * j , struct journal_res * res ,
2018-11-19 05:35:59 +03:00
unsigned flags )
2017-03-17 09:18:50 +03:00
{
struct bch_fs * c = container_of ( j , struct bch_fs , journal ) ;
struct journal_buf * buf ;
2019-03-04 02:39:07 +03:00
bool can_discard ;
2017-03-17 09:18:50 +03:00
int ret ;
retry :
2018-11-19 05:35:59 +03:00
if ( journal_res_get_fast ( j , res , flags ) )
return 0 ;
2017-03-17 09:18:50 +03:00
2024-01-31 19:28:13 +03:00
if ( ( flags & BCH_WATERMARK_MASK ) < j - > watermark ) {
ret = JOURNAL_ERR_journal_full ;
can_discard = j - > can_discard ;
goto out ;
}
2019-02-19 01:39:42 +03:00
2024-01-31 19:28:13 +03:00
if ( j - > blocked )
return - BCH_ERR_journal_res_get_blocked ;
2019-02-19 01:39:42 +03:00
2024-01-31 19:28:13 +03:00
if ( bch2_journal_error ( j ) )
2023-03-20 20:21:19 +03:00
return - BCH_ERR_erofs_journal_err ;
2024-01-31 19:28:13 +03:00
if ( nr_unwritten_journal_entries ( j ) = = ARRAY_SIZE ( j - > buf ) & & ! journal_entry_is_open ( j ) ) {
ret = JOURNAL_ERR_max_in_flight ;
goto out ;
2023-03-20 20:21:19 +03:00
}
2024-01-31 19:28:13 +03:00
spin_lock ( & j - > lock ) ;
2017-03-17 09:18:50 +03:00
/*
* Recheck after taking the lock , so we don ' t race with another thread
2023-12-10 23:23:27 +03:00
* that just did journal_entry_open ( ) and call bch2_journal_entry_close ( )
2017-03-17 09:18:50 +03:00
* unnecessarily
*/
2018-11-19 05:35:59 +03:00
if ( journal_res_get_fast ( j , res , flags ) ) {
2024-01-31 19:28:13 +03:00
ret = 0 ;
2019-02-19 21:41:36 +03:00
goto unlock ;
}
2017-03-17 09:18:50 +03:00
/*
* If we couldn ' t get a reservation because the current buf filled up ,
* and we had room for a bigger entry on disk , signal that we want to
* realloc the journal bufs :
*/
buf = journal_cur_buf ( j ) ;
if ( journal_entry_is_open ( j ) & &
2019-02-19 01:39:42 +03:00
buf - > buf_size > > 9 < buf - > disk_sectors & &
buf - > buf_size < JOURNAL_ENTRY_SIZE_MAX )
j - > buf_size_want = max ( j - > buf_size_want , buf - > buf_size < < 1 ) ;
2017-03-17 09:18:50 +03:00
2024-01-16 01:59:51 +03:00
__journal_entry_close ( j , JOURNAL_ENTRY_CLOSED_VAL , false ) ;
2024-01-31 19:28:13 +03:00
ret = journal_entry_open ( j ) ? : JOURNAL_ERR_retry ;
2023-11-10 06:07:42 +03:00
unlock :
2019-03-04 02:39:07 +03:00
can_discard = j - > can_discard ;
2017-03-17 09:18:50 +03:00
spin_unlock ( & j - > lock ) ;
2024-01-31 19:28:13 +03:00
out :
if ( ret = = JOURNAL_ERR_retry )
2017-03-17 09:18:50 +03:00
goto retry ;
2024-01-31 19:28:13 +03:00
if ( ! ret )
return 0 ;
2023-03-21 15:03:18 +03:00
if ( journal_error_check_stuck ( j , ret , flags ) )
ret = - BCH_ERR_journal_res_get_blocked ;
2021-04-03 23:24:13 +03:00
2024-01-31 19:28:13 +03:00
if ( ret = = JOURNAL_ERR_max_in_flight & &
track_event_change ( & c - > times [ BCH_TIME_blocked_journal_max_in_flight ] ,
& j - > max_in_flight_start , true ) ) {
struct printbuf buf = PRINTBUF ;
prt_printf ( & buf , " seq %llu \n " , journal_cur_seq ( j ) ) ;
bch2_journal_bufs_to_text ( & buf , j ) ;
trace_journal_entry_full ( c , buf . buf ) ;
printbuf_exit ( & buf ) ;
count_event ( c , journal_entry_full ) ;
}
2020-11-18 21:21:59 +03:00
/*
* Journal is full - can ' t rely on reclaim from work item due to
* freezing :
*/
2022-03-15 04:48:42 +03:00
if ( ( ret = = JOURNAL_ERR_journal_full | |
ret = = JOURNAL_ERR_journal_pin_full ) & &
2020-11-18 21:21:59 +03:00
! ( flags & JOURNAL_RES_GET_NONBLOCK ) ) {
if ( can_discard ) {
bch2_journal_do_discards ( j ) ;
goto retry ;
2019-03-04 02:39:07 +03:00
}
2020-11-18 21:21:59 +03:00
if ( mutex_trylock ( & j - > reclaim_lock ) ) {
bch2_journal_reclaim ( j ) ;
mutex_unlock ( & j - > reclaim_lock ) ;
}
2019-02-19 01:39:42 +03:00
}
2017-03-17 09:18:50 +03:00
2022-12-13 23:17:40 +03:00
return ret = = JOURNAL_ERR_insufficient_devices
2023-07-08 00:09:26 +03:00
? - BCH_ERR_erofs_journal_err
2022-12-13 23:17:40 +03:00
: - BCH_ERR_journal_res_get_blocked ;
2017-03-17 09:18:50 +03:00
}
/*
* Essentially the entry function to the journaling code . When bcachefs is doing
* a btree insert , it calls this function to get the current journal write .
* Journal write is the structure used set up journal writes . The calling
* function will then add its keys to the structure , queuing them for the next
* write .
*
* To ensure forward progress , the current task must not be holding any
* btree node write locks .
*/
int bch2_journal_res_get_slowpath ( struct journal * j , struct journal_res * res ,
2018-11-19 05:35:59 +03:00
unsigned flags )
2017-03-17 09:18:50 +03:00
{
int ret ;
2019-02-21 21:33:21 +03:00
closure_wait_event ( & j - > async_wait ,
2023-08-07 19:04:05 +03:00
( ret = __journal_res_get ( j , res , flags ) ) ! = - BCH_ERR_journal_res_get_blocked | |
2018-11-19 05:35:59 +03:00
( flags & JOURNAL_RES_GET_NONBLOCK ) ) ;
return ret ;
2017-03-17 09:18:50 +03:00
}
2019-01-25 00:50:48 +03:00
/* journal_entry_res: */
void bch2_journal_entry_res_resize ( struct journal * j ,
struct journal_entry_res * res ,
unsigned new_u64s )
{
union journal_res_state state ;
int d = new_u64s - res - > u64s ;
spin_lock ( & j - > lock ) ;
j - > entry_u64s_reserved + = d ;
if ( d < = 0 )
2019-02-19 01:39:42 +03:00
goto out ;
2019-01-25 00:50:48 +03:00
2019-02-28 03:14:23 +03:00
j - > cur_entry_u64s = max_t ( int , 0 , j - > cur_entry_u64s - d ) ;
2019-01-25 00:50:48 +03:00
smp_mb ( ) ;
state = READ_ONCE ( j - > reservations ) ;
if ( state . cur_entry_offset < JOURNAL_ENTRY_CLOSED_VAL & &
state . cur_entry_offset > j - > cur_entry_u64s ) {
j - > cur_entry_u64s + = d ;
/*
* Not enough room in current journal entry , have to flush it :
*/
2024-01-16 01:59:51 +03:00
__journal_entry_close ( j , JOURNAL_ENTRY_CLOSED_VAL , true ) ;
2019-02-19 01:39:42 +03:00
} else {
journal_cur_buf ( j ) - > u64s_reserved + = d ;
2019-01-25 00:50:48 +03:00
}
out :
2019-02-19 01:39:42 +03:00
spin_unlock ( & j - > lock ) ;
2019-01-25 00:50:48 +03:00
res - > u64s + = d ;
}
/* journal flushing: */
2017-03-17 09:18:50 +03:00
/**
* bch2_journal_flush_seq_async - wait for a journal entry to be written
2023-09-13 01:41:22 +03:00
* @ j : journal object
* @ seq : seq to flush
* @ parent : closure object to wait with
* Returns : 1 if @ seq has already been flushed , 0 if @ seq is being flushed ,
* - EIO if @ seq will never be flushed
2017-03-17 09:18:50 +03:00
*
2023-09-13 01:41:22 +03:00
* Like bch2_journal_wait_on_seq , except that it triggers a write immediately if
2017-03-17 09:18:50 +03:00
* necessary
*/
2020-11-15 00:04:30 +03:00
int bch2_journal_flush_seq_async ( struct journal * j , u64 seq ,
2020-11-21 03:27:57 +03:00
struct closure * parent )
2017-03-17 09:18:50 +03:00
{
struct journal_buf * buf ;
2020-11-15 00:04:30 +03:00
int ret = 0 ;
2017-03-17 09:18:50 +03:00
bcachefs: Don't require flush/fua on every journal write
This patch adds a flag to journal entries which, if set, indicates that
they weren't done as flush/fua writes.
- non flush/fua journal writes don't update last_seq (i.e. they don't
free up space in the journal), thus the journal free space
calculations now check whether nonflush journal writes are currently
allowed (i.e. are we low on free space, or would doing a flush write
free up a lot of space in the journal)
- write_delay_ms, the user configurable option for when open journal
entries are automatically written, is now interpreted as the max
delay between flush journal writes (default 1 second).
- bch2_journal_flush_seq_async is changed to ensure a flush write >=
the requested sequence number has happened
- journal read/replay must now ignore, and blacklist, any journal
entries newer than the most recent flush entry in the journal. Also,
the way the read_entire_journal option is handled has been improved;
struct journal_replay now has an entry, 'ignore', for entries that
were read but should not be used.
- assorted refactoring and improvements related to journal read in
journal_io.c and recovery.c
Previously, we'd have to issue a flush/fua write every time we
accumulated a full journal entry - typically the bucket size. Now we
need to issue them much less frequently: when an fsync is requested, or
it's been more than write_delay_ms since the last flush, or when we need
to free up space in the journal. This is a significant performance
improvement on many write heavy workloads.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-11-14 17:59:58 +03:00
if ( seq < = j - > flushed_seq_ondisk )
2020-11-21 03:27:57 +03:00
return 1 ;
2017-03-17 09:18:50 +03:00
spin_lock ( & j - > lock ) ;
2020-11-21 03:27:57 +03:00
2021-11-14 01:53:55 +03:00
if ( WARN_ONCE ( seq > journal_cur_seq ( j ) ,
" requested to flush journal seq %llu, but currently at %llu " ,
seq , journal_cur_seq ( j ) ) )
goto out ;
2021-02-10 21:39:48 +03:00
2020-11-21 03:27:57 +03:00
/* Recheck under lock: */
2020-12-02 23:33:12 +03:00
if ( j - > err_seq & & seq > = j - > err_seq ) {
2020-11-15 00:04:30 +03:00
ret = - EIO ;
goto out ;
}
bcachefs: Don't require flush/fua on every journal write
This patch adds a flag to journal entries which, if set, indicates that
they weren't done as flush/fua writes.
- non flush/fua journal writes don't update last_seq (i.e. they don't
free up space in the journal), thus the journal free space
calculations now check whether nonflush journal writes are currently
allowed (i.e. are we low on free space, or would doing a flush write
free up a lot of space in the journal)
- write_delay_ms, the user configurable option for when open journal
entries are automatically written, is now interpreted as the max
delay between flush journal writes (default 1 second).
- bch2_journal_flush_seq_async is changed to ensure a flush write >=
the requested sequence number has happened
- journal read/replay must now ignore, and blacklist, any journal
entries newer than the most recent flush entry in the journal. Also,
the way the read_entire_journal option is handled has been improved;
struct journal_replay now has an entry, 'ignore', for entries that
were read but should not be used.
- assorted refactoring and improvements related to journal read in
journal_io.c and recovery.c
Previously, we'd have to issue a flush/fua write every time we
accumulated a full journal entry - typically the bucket size. Now we
need to issue them much less frequently: when an fsync is requested, or
it's been more than write_delay_ms since the last flush, or when we need
to free up space in the journal. This is a significant performance
improvement on many write heavy workloads.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-11-14 17:59:58 +03:00
if ( seq < = j - > flushed_seq_ondisk ) {
2020-11-15 00:04:30 +03:00
ret = 1 ;
goto out ;
}
2017-03-17 09:18:50 +03:00
bcachefs: Don't require flush/fua on every journal write
This patch adds a flag to journal entries which, if set, indicates that
they weren't done as flush/fua writes.
- non flush/fua journal writes don't update last_seq (i.e. they don't
free up space in the journal), thus the journal free space
calculations now check whether nonflush journal writes are currently
allowed (i.e. are we low on free space, or would doing a flush write
free up a lot of space in the journal)
- write_delay_ms, the user configurable option for when open journal
entries are automatically written, is now interpreted as the max
delay between flush journal writes (default 1 second).
- bch2_journal_flush_seq_async is changed to ensure a flush write >=
the requested sequence number has happened
- journal read/replay must now ignore, and blacklist, any journal
entries newer than the most recent flush entry in the journal. Also,
the way the read_entire_journal option is handled has been improved;
struct journal_replay now has an entry, 'ignore', for entries that
were read but should not be used.
- assorted refactoring and improvements related to journal read in
journal_io.c and recovery.c
Previously, we'd have to issue a flush/fua write every time we
accumulated a full journal entry - typically the bucket size. Now we
need to issue them much less frequently: when an fsync is requested, or
it's been more than write_delay_ms since the last flush, or when we need
to free up space in the journal. This is a significant performance
improvement on many write heavy workloads.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-11-14 17:59:58 +03:00
/* if seq was written, but not flushed - flush a newer one instead */
2022-03-01 03:17:27 +03:00
seq = max ( seq , journal_last_unwritten_seq ( j ) ) ;
bcachefs: Don't require flush/fua on every journal write
This patch adds a flag to journal entries which, if set, indicates that
they weren't done as flush/fua writes.
- non flush/fua journal writes don't update last_seq (i.e. they don't
free up space in the journal), thus the journal free space
calculations now check whether nonflush journal writes are currently
allowed (i.e. are we low on free space, or would doing a flush write
free up a lot of space in the journal)
- write_delay_ms, the user configurable option for when open journal
entries are automatically written, is now interpreted as the max
delay between flush journal writes (default 1 second).
- bch2_journal_flush_seq_async is changed to ensure a flush write >=
the requested sequence number has happened
- journal read/replay must now ignore, and blacklist, any journal
entries newer than the most recent flush entry in the journal. Also,
the way the read_entire_journal option is handled has been improved;
struct journal_replay now has an entry, 'ignore', for entries that
were read but should not be used.
- assorted refactoring and improvements related to journal read in
journal_io.c and recovery.c
Previously, we'd have to issue a flush/fua write every time we
accumulated a full journal entry - typically the bucket size. Now we
need to issue them much less frequently: when an fsync is requested, or
it's been more than write_delay_ms since the last flush, or when we need
to free up space in the journal. This is a significant performance
improvement on many write heavy workloads.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-11-14 17:59:58 +03:00
recheck_need_open :
2022-03-01 00:35:42 +03:00
if ( seq > journal_cur_seq ( j ) ) {
bcachefs: Don't require flush/fua on every journal write
This patch adds a flag to journal entries which, if set, indicates that
they weren't done as flush/fua writes.
- non flush/fua journal writes don't update last_seq (i.e. they don't
free up space in the journal), thus the journal free space
calculations now check whether nonflush journal writes are currently
allowed (i.e. are we low on free space, or would doing a flush write
free up a lot of space in the journal)
- write_delay_ms, the user configurable option for when open journal
entries are automatically written, is now interpreted as the max
delay between flush journal writes (default 1 second).
- bch2_journal_flush_seq_async is changed to ensure a flush write >=
the requested sequence number has happened
- journal read/replay must now ignore, and blacklist, any journal
entries newer than the most recent flush entry in the journal. Also,
the way the read_entire_journal option is handled has been improved;
struct journal_replay now has an entry, 'ignore', for entries that
were read but should not be used.
- assorted refactoring and improvements related to journal read in
journal_io.c and recovery.c
Previously, we'd have to issue a flush/fua write every time we
accumulated a full journal entry - typically the bucket size. Now we
need to issue them much less frequently: when an fsync is requested, or
it's been more than write_delay_ms since the last flush, or when we need
to free up space in the journal. This is a significant performance
improvement on many write heavy workloads.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-11-14 17:59:58 +03:00
struct journal_res res = { 0 } ;
2022-03-01 00:35:42 +03:00
if ( journal_entry_is_open ( j ) )
2024-01-16 01:59:51 +03:00
__journal_entry_close ( j , JOURNAL_ENTRY_CLOSED_VAL , true ) ;
2022-03-01 00:35:42 +03:00
bcachefs: Don't require flush/fua on every journal write
This patch adds a flag to journal entries which, if set, indicates that
they weren't done as flush/fua writes.
- non flush/fua journal writes don't update last_seq (i.e. they don't
free up space in the journal), thus the journal free space
calculations now check whether nonflush journal writes are currently
allowed (i.e. are we low on free space, or would doing a flush write
free up a lot of space in the journal)
- write_delay_ms, the user configurable option for when open journal
entries are automatically written, is now interpreted as the max
delay between flush journal writes (default 1 second).
- bch2_journal_flush_seq_async is changed to ensure a flush write >=
the requested sequence number has happened
- journal read/replay must now ignore, and blacklist, any journal
entries newer than the most recent flush entry in the journal. Also,
the way the read_entire_journal option is handled has been improved;
struct journal_replay now has an entry, 'ignore', for entries that
were read but should not be used.
- assorted refactoring and improvements related to journal read in
journal_io.c and recovery.c
Previously, we'd have to issue a flush/fua write every time we
accumulated a full journal entry - typically the bucket size. Now we
need to issue them much less frequently: when an fsync is requested, or
it's been more than write_delay_ms since the last flush, or when we need
to free up space in the journal. This is a significant performance
improvement on many write heavy workloads.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-11-14 17:59:58 +03:00
spin_unlock ( & j - > lock ) ;
ret = bch2_journal_res_get ( j , & res , jset_u64s ( 0 ) , 0 ) ;
if ( ret )
return ret ;
seq = res . seq ;
buf = j - > buf + ( seq & JOURNAL_BUF_MASK ) ;
buf - > must_flush = true ;
2022-02-25 18:28:20 +03:00
if ( ! buf - > flush_time ) {
buf - > flush_time = local_clock ( ) ? : 1 ;
buf - > expires = jiffies ;
}
bcachefs: Don't require flush/fua on every journal write
This patch adds a flag to journal entries which, if set, indicates that
they weren't done as flush/fua writes.
- non flush/fua journal writes don't update last_seq (i.e. they don't
free up space in the journal), thus the journal free space
calculations now check whether nonflush journal writes are currently
allowed (i.e. are we low on free space, or would doing a flush write
free up a lot of space in the journal)
- write_delay_ms, the user configurable option for when open journal
entries are automatically written, is now interpreted as the max
delay between flush journal writes (default 1 second).
- bch2_journal_flush_seq_async is changed to ensure a flush write >=
the requested sequence number has happened
- journal read/replay must now ignore, and blacklist, any journal
entries newer than the most recent flush entry in the journal. Also,
the way the read_entire_journal option is handled has been improved;
struct journal_replay now has an entry, 'ignore', for entries that
were read but should not be used.
- assorted refactoring and improvements related to journal read in
journal_io.c and recovery.c
Previously, we'd have to issue a flush/fua write every time we
accumulated a full journal entry - typically the bucket size. Now we
need to issue them much less frequently: when an fsync is requested, or
it's been more than write_delay_ms since the last flush, or when we need
to free up space in the journal. This is a significant performance
improvement on many write heavy workloads.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-11-14 17:59:58 +03:00
if ( parent & & ! closure_wait ( & buf - > wait , parent ) )
2017-03-17 09:18:50 +03:00
BUG ( ) ;
bcachefs: Don't require flush/fua on every journal write
This patch adds a flag to journal entries which, if set, indicates that
they weren't done as flush/fua writes.
- non flush/fua journal writes don't update last_seq (i.e. they don't
free up space in the journal), thus the journal free space
calculations now check whether nonflush journal writes are currently
allowed (i.e. are we low on free space, or would doing a flush write
free up a lot of space in the journal)
- write_delay_ms, the user configurable option for when open journal
entries are automatically written, is now interpreted as the max
delay between flush journal writes (default 1 second).
- bch2_journal_flush_seq_async is changed to ensure a flush write >=
the requested sequence number has happened
- journal read/replay must now ignore, and blacklist, any journal
entries newer than the most recent flush entry in the journal. Also,
the way the read_entire_journal option is handled has been improved;
struct journal_replay now has an entry, 'ignore', for entries that
were read but should not be used.
- assorted refactoring and improvements related to journal read in
journal_io.c and recovery.c
Previously, we'd have to issue a flush/fua write every time we
accumulated a full journal entry - typically the bucket size. Now we
need to issue them much less frequently: when an fsync is requested, or
it's been more than write_delay_ms since the last flush, or when we need
to free up space in the journal. This is a significant performance
improvement on many write heavy workloads.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-11-14 17:59:58 +03:00
bch2_journal_res_put ( j , & res ) ;
spin_lock ( & j - > lock ) ;
goto want_write ;
}
/*
* if write was kicked off without a flush , flush the next sequence
* number instead
*/
buf = journal_seq_to_buf ( j , seq ) ;
if ( buf - > noflush ) {
seq + + ;
goto recheck_need_open ;
}
buf - > must_flush = true ;
if ( parent & & ! closure_wait ( & buf - > wait , parent ) )
BUG ( ) ;
want_write :
2018-07-23 14:52:00 +03:00
if ( seq = = journal_cur_seq ( j ) )
2020-11-14 02:36:33 +03:00
journal_entry_want_write ( j ) ;
2020-11-15 00:04:30 +03:00
out :
2019-02-19 01:39:42 +03:00
spin_unlock ( & j - > lock ) ;
2017-03-17 09:18:50 +03:00
return ret ;
}
int bch2_journal_flush_seq ( struct journal * j , u64 seq )
{
u64 start_time = local_clock ( ) ;
int ret , ret2 ;
2021-12-10 23:41:38 +03:00
/*
* Don ' t update time_stats when @ seq is already flushed :
*/
if ( seq < = j - > flushed_seq_ondisk )
return 0 ;
2021-02-10 03:54:04 +03:00
ret = wait_event_interruptible ( j - > wait , ( ret2 = bch2_journal_flush_seq_async ( j , seq , NULL ) ) ) ;
2017-03-17 09:18:50 +03:00
2021-02-10 03:54:04 +03:00
if ( ! ret )
bch2_time_stats_update ( j - > flush_seq_time , start_time ) ;
2017-03-17 09:18:50 +03:00
return ret ? : ret2 < 0 ? ret2 : 0 ;
}
/*
* bch2_journal_flush_async - if there is an open journal entry , or a journal
* still being written , write it and wait for the write to complete
*/
void bch2_journal_flush_async ( struct journal * j , struct closure * parent )
{
2022-03-01 00:35:42 +03:00
bch2_journal_flush_seq_async ( j , atomic64_read ( & j - > seq ) , parent ) ;
2017-03-17 09:18:50 +03:00
}
int bch2_journal_flush ( struct journal * j )
{
2022-03-01 00:35:42 +03:00
return bch2_journal_flush_seq ( j , atomic64_read ( & j - > seq ) ) ;
2017-03-17 09:18:50 +03:00
}
2021-12-28 07:51:48 +03:00
/*
* bch2_journal_noflush_seq - tell the journal not to issue any flushes before
* @ seq
*/
bool bch2_journal_noflush_seq ( struct journal * j , u64 seq )
{
struct bch_fs * c = container_of ( j , struct bch_fs , journal ) ;
u64 unwritten_seq ;
bool ret = false ;
if ( ! ( c - > sb . features & ( 1ULL < < BCH_FEATURE_journal_no_flush ) ) )
return false ;
if ( seq < = c - > journal . flushed_seq_ondisk )
return false ;
spin_lock ( & j - > lock ) ;
if ( seq < = c - > journal . flushed_seq_ondisk )
goto out ;
2022-03-01 03:17:27 +03:00
for ( unwritten_seq = journal_last_unwritten_seq ( j ) ;
2021-12-28 07:51:48 +03:00
unwritten_seq < seq ;
unwritten_seq + + ) {
struct journal_buf * buf = journal_seq_to_buf ( j , unwritten_seq ) ;
/* journal write is already in flight, and was a flush write: */
2022-03-01 03:17:27 +03:00
if ( unwritten_seq = = journal_last_unwritten_seq ( j ) & & ! buf - > noflush )
2021-12-28 07:51:48 +03:00
goto out ;
buf - > noflush = true ;
}
ret = true ;
out :
spin_unlock ( & j - > lock ) ;
return ret ;
}
2022-03-10 22:25:16 +03:00
int bch2_journal_meta ( struct journal * j )
{
struct journal_buf * buf ;
struct journal_res res ;
int ret ;
memset ( & res , 0 , sizeof ( res ) ) ;
ret = bch2_journal_res_get ( j , & res , jset_u64s ( 0 ) , 0 ) ;
if ( ret )
return ret ;
buf = j - > buf + ( res . seq & JOURNAL_BUF_MASK ) ;
buf - > must_flush = true ;
if ( ! buf - > flush_time ) {
buf - > flush_time = local_clock ( ) ? : 1 ;
buf - > expires = jiffies ;
}
bch2_journal_res_put ( j , & res ) ;
return bch2_journal_flush_seq ( j , res . seq ) ;
}
2019-02-15 02:38:52 +03:00
/* block/unlock the journal: */
void bch2_journal_unblock ( struct journal * j )
{
spin_lock ( & j - > lock ) ;
j - > blocked - - ;
spin_unlock ( & j - > lock ) ;
journal_wake ( j ) ;
}
void bch2_journal_block ( struct journal * j )
{
spin_lock ( & j - > lock ) ;
j - > blocked + + ;
spin_unlock ( & j - > lock ) ;
journal_quiesce ( j ) ;
}
2023-11-03 01:57:19 +03:00
static struct journal_buf * __bch2_next_write_buffer_flush_journal_buf ( struct journal * j , u64 max_seq )
{
struct journal_buf * ret = NULL ;
mutex_lock ( & j - > buf_lock ) ;
spin_lock ( & j - > lock ) ;
max_seq = min ( max_seq , journal_cur_seq ( j ) ) ;
for ( u64 seq = journal_last_unwritten_seq ( j ) ;
seq < = max_seq ;
seq + + ) {
unsigned idx = seq & JOURNAL_BUF_MASK ;
struct journal_buf * buf = j - > buf + idx ;
if ( buf - > need_flush_to_write_buffer ) {
if ( seq = = journal_cur_seq ( j ) )
2024-01-16 01:59:51 +03:00
__journal_entry_close ( j , JOURNAL_ENTRY_CLOSED_VAL , true ) ;
2023-11-03 01:57:19 +03:00
union journal_res_state s ;
s . v = atomic64_read_acquire ( & j - > reservations . counter ) ;
ret = journal_state_count ( s , idx )
? ERR_PTR ( - EAGAIN )
: buf ;
break ;
}
}
spin_unlock ( & j - > lock ) ;
if ( IS_ERR_OR_NULL ( ret ) )
mutex_unlock ( & j - > buf_lock ) ;
return ret ;
}
struct journal_buf * bch2_next_write_buffer_flush_journal_buf ( struct journal * j , u64 max_seq )
{
struct journal_buf * ret ;
wait_event ( j - > wait , ( ret = __bch2_next_write_buffer_flush_journal_buf ( j , max_seq ) ) ! = ERR_PTR ( - EAGAIN ) ) ;
return ret ;
}
2017-03-17 09:18:50 +03:00
/* allocate journal on a device: */
static int __bch2_set_nr_journal_buckets ( struct bch_dev * ca , unsigned nr ,
bool new_fs , struct closure * cl )
{
struct bch_fs * c = ca - > fs ;
struct journal_device * ja = & ca - > journal ;
u64 * new_bucket_seq = NULL , * new_buckets = NULL ;
2022-03-11 00:43:52 +03:00
struct open_bucket * * ob = NULL ;
long * bu = NULL ;
2023-03-06 13:29:12 +03:00
unsigned i , pos , nr_got = 0 , nr_want = nr - ja - > nr ;
2017-03-17 09:18:50 +03:00
int ret = 0 ;
2023-03-06 13:29:12 +03:00
BUG_ON ( nr < = ja - > nr ) ;
2017-03-17 09:18:50 +03:00
2022-10-20 01:31:33 +03:00
bu = kcalloc ( nr_want , sizeof ( * bu ) , GFP_KERNEL ) ;
ob = kcalloc ( nr_want , sizeof ( * ob ) , GFP_KERNEL ) ;
new_buckets = kcalloc ( nr , sizeof ( u64 ) , GFP_KERNEL ) ;
new_bucket_seq = kcalloc ( nr , sizeof ( u64 ) , GFP_KERNEL ) ;
2022-03-11 00:43:52 +03:00
if ( ! bu | | ! ob | | ! new_buckets | | ! new_bucket_seq ) {
2023-03-14 22:35:57 +03:00
ret = - BCH_ERR_ENOMEM_set_nr_journal_buckets ;
2023-03-06 13:29:12 +03:00
goto err_free ;
2020-11-21 00:12:39 +03:00
}
2017-03-17 09:18:50 +03:00
2022-03-11 00:43:52 +03:00
for ( nr_got = 0 ; nr_got < nr_want ; nr_got + + ) {
if ( new_fs ) {
bu [ nr_got ] = bch2_bucket_alloc_new_fs ( ca ) ;
if ( bu [ nr_got ] < 0 ) {
2022-09-19 00:10:33 +03:00
ret = - BCH_ERR_ENOSPC_bucket_alloc ;
2022-03-11 00:43:52 +03:00
break ;
}
} else {
2023-06-25 02:30:10 +03:00
ob [ nr_got ] = bch2_bucket_alloc ( c , ca , BCH_WATERMARK_normal , cl ) ;
2022-12-13 23:17:40 +03:00
ret = PTR_ERR_OR_ZERO ( ob [ nr_got ] ) ;
if ( ret )
2022-03-11 00:43:52 +03:00
break ;
2023-03-06 13:29:12 +03:00
ret = bch2_trans_run ( c ,
2023-09-13 00:16:02 +03:00
bch2_trans_mark_metadata_bucket ( trans , ca ,
2023-03-06 13:29:12 +03:00
ob [ nr_got ] - > bucket , BCH_DATA_journal ,
ca - > mi . bucket_size ) ) ;
if ( ret ) {
bch2_open_bucket_put ( c , ob [ nr_got ] ) ;
2023-09-11 08:37:34 +03:00
bch_err_msg ( c , ret , " marking new journal buckets " ) ;
2023-03-06 13:29:12 +03:00
break ;
}
2022-03-11 00:43:52 +03:00
bu [ nr_got ] = ob [ nr_got ] - > bucket ;
}
2020-11-21 00:12:39 +03:00
}
2017-03-17 09:18:50 +03:00
2022-03-11 00:43:52 +03:00
if ( ! nr_got )
2023-03-06 13:29:12 +03:00
goto err_free ;
2022-03-11 00:43:52 +03:00
2023-03-06 13:29:12 +03:00
/* Don't return an error if we successfully allocated some buckets: */
ret = 0 ;
if ( c ) {
bch2_journal_flush_all_pins ( & c - > journal ) ;
bch2_journal_block ( & c - > journal ) ;
mutex_lock ( & c - > sb_lock ) ;
}
2017-03-17 09:18:50 +03:00
memcpy ( new_buckets , ja - > buckets , ja - > nr * sizeof ( u64 ) ) ;
memcpy ( new_bucket_seq , ja - > bucket_seq , ja - > nr * sizeof ( u64 ) ) ;
2023-03-06 13:29:12 +03:00
BUG_ON ( ja - > discard_idx > ja - > nr ) ;
pos = ja - > discard_idx ? : ja - > nr ;
memmove ( new_buckets + pos + nr_got ,
new_buckets + pos ,
sizeof ( new_buckets [ 0 ] ) * ( ja - > nr - pos ) ) ;
memmove ( new_bucket_seq + pos + nr_got ,
new_bucket_seq + pos ,
sizeof ( new_bucket_seq [ 0 ] ) * ( ja - > nr - pos ) ) ;
2017-03-17 09:18:50 +03:00
2022-03-11 00:43:52 +03:00
for ( i = 0 ; i < nr_got ; i + + ) {
2023-03-06 13:29:12 +03:00
new_buckets [ pos + i ] = bu [ i ] ;
new_bucket_seq [ pos + i ] = 0 ;
2022-03-11 00:43:52 +03:00
}
2017-03-17 09:18:50 +03:00
2023-03-06 13:29:12 +03:00
nr = ja - > nr + nr_got ;
2022-03-11 00:43:52 +03:00
2023-03-06 13:29:12 +03:00
ret = bch2_journal_buckets_to_sb ( c , ca , new_buckets , nr ) ;
if ( ret )
goto err_unblock ;
2021-05-17 06:46:08 +03:00
2023-03-06 13:29:12 +03:00
if ( ! new_fs )
2023-02-12 00:53:59 +03:00
bch2_write_super ( c ) ;
2023-03-06 13:29:12 +03:00
/* Commit: */
2022-03-11 00:43:52 +03:00
if ( c )
2023-03-06 13:29:12 +03:00
spin_lock ( & c - > journal . lock ) ;
2022-03-11 00:43:52 +03:00
2023-03-06 13:29:12 +03:00
swap ( new_buckets , ja - > buckets ) ;
swap ( new_bucket_seq , ja - > bucket_seq ) ;
ja - > nr = nr ;
if ( pos < = ja - > discard_idx )
ja - > discard_idx = ( ja - > discard_idx + nr_got ) % ja - > nr ;
if ( pos < = ja - > dirty_idx_ondisk )
ja - > dirty_idx_ondisk = ( ja - > dirty_idx_ondisk + nr_got ) % ja - > nr ;
if ( pos < = ja - > dirty_idx )
ja - > dirty_idx = ( ja - > dirty_idx + nr_got ) % ja - > nr ;
if ( pos < = ja - > cur_idx )
ja - > cur_idx = ( ja - > cur_idx + nr_got ) % ja - > nr ;
2022-03-11 00:43:52 +03:00
2023-02-12 00:53:59 +03:00
if ( c )
2023-03-06 13:29:12 +03:00
spin_unlock ( & c - > journal . lock ) ;
err_unblock :
if ( c ) {
bch2_journal_unblock ( & c - > journal ) ;
2023-02-12 00:53:59 +03:00
mutex_unlock ( & c - > sb_lock ) ;
2023-03-06 13:29:12 +03:00
}
2023-02-12 00:53:59 +03:00
2023-03-06 13:29:12 +03:00
if ( ret & & ! new_fs )
for ( i = 0 ; i < nr_got ; i + + )
bch2_trans_run ( c ,
2023-09-13 00:16:02 +03:00
bch2_trans_mark_metadata_bucket ( trans , ca ,
2023-03-06 13:29:12 +03:00
bu [ i ] , BCH_DATA_free , 0 ) ) ;
err_free :
if ( ! new_fs )
2022-03-11 00:43:52 +03:00
for ( i = 0 ; i < nr_got ; i + + )
bch2_open_bucket_put ( c , ob [ i ] ) ;
2017-03-17 09:18:50 +03:00
kfree ( new_bucket_seq ) ;
kfree ( new_buckets ) ;
2022-03-11 00:43:52 +03:00
kfree ( ob ) ;
kfree ( bu ) ;
2017-03-17 09:18:50 +03:00
return ret ;
}
/*
* Allocate more journal space at runtime - not currently making use if it , but
* the code works :
*/
int bch2_set_nr_journal_buckets ( struct bch_fs * c , struct bch_dev * ca ,
unsigned nr )
{
struct journal_device * ja = & ca - > journal ;
struct closure cl ;
2022-03-11 00:43:52 +03:00
int ret = 0 ;
2023-03-06 13:29:12 +03:00
closure_init_stack ( & cl ) ;
down_write ( & c - > state_lock ) ;
2022-03-11 00:43:52 +03:00
/* don't handle reducing nr of buckets yet: */
if ( nr < ja - > nr )
2023-03-06 13:29:12 +03:00
goto unlock ;
2017-03-17 09:18:50 +03:00
2023-03-06 13:29:12 +03:00
while ( ja - > nr < nr ) {
2023-09-13 01:41:22 +03:00
struct disk_reservation disk_res = { 0 , 0 , 0 } ;
2017-03-17 09:18:50 +03:00
/*
* note : journal buckets aren ' t really counted as _sectors_ used yet , so
* we don ' t need the disk reservation to avoid the BUG_ON ( ) in buckets . c
* when space used goes up without a reservation - but we do need the
* reservation to ensure we ' ll actually be able to allocate :
2023-02-12 00:53:59 +03:00
*
* XXX : that ' s not right , disk reservations only ensure a
* filesystem - wide allocation will succeed , this is a device
* specific allocation - we can hang here :
2017-03-17 09:18:50 +03:00
*/
2022-09-19 00:10:33 +03:00
ret = bch2_disk_reservation_get ( c , & disk_res ,
bucket_to_sector ( ca , nr - ja - > nr ) , 1 , 0 ) ;
2023-02-12 00:53:59 +03:00
if ( ret )
break ;
2017-03-17 09:18:50 +03:00
ret = __bch2_set_nr_journal_buckets ( ca , nr , false , & cl ) ;
bch2_disk_reservation_put ( c , & disk_res ) ;
2023-02-12 00:53:59 +03:00
closure_sync ( & cl ) ;
if ( ret & & ret ! = - BCH_ERR_bucket_alloc_blocked )
break ;
2022-03-11 00:43:52 +03:00
}
2017-03-17 09:18:50 +03:00
2023-12-17 06:43:41 +03:00
bch_err_fn ( c , ret ) ;
2023-03-06 13:29:12 +03:00
unlock :
up_write ( & c - > state_lock ) ;
2017-03-17 09:18:50 +03:00
return ret ;
}
int bch2_dev_journal_alloc ( struct bch_dev * ca )
{
unsigned nr ;
2023-06-20 20:49:25 +03:00
int ret ;
2017-03-17 09:18:50 +03:00
2023-06-20 20:49:25 +03:00
if ( dynamic_fault ( " bcachefs:add:journal_alloc " ) ) {
ret = - BCH_ERR_ENOMEM_set_nr_journal_buckets ;
goto err ;
}
2017-03-17 09:18:50 +03:00
2021-03-25 05:49:05 +03:00
/* 1/128th of the device by default: */
nr = ca - > mi . nbuckets > > 7 ;
2017-03-17 09:18:50 +03:00
/*
2021-03-25 05:49:05 +03:00
* clamp journal size to 8192 buckets or 8 GB ( in sectors ) , whichever
2017-03-17 09:18:50 +03:00
* is smaller :
*/
2021-03-25 05:49:05 +03:00
nr = clamp_t ( unsigned , nr ,
2017-03-17 09:18:50 +03:00
BCH_JOURNAL_BUCKETS_MIN ,
2021-03-25 05:49:05 +03:00
min ( 1 < < 13 ,
( 1 < < 24 ) / ca - > mi . bucket_size ) ) ;
2017-03-17 09:18:50 +03:00
2023-06-20 20:49:25 +03:00
ret = __bch2_set_nr_journal_buckets ( ca , nr , true , NULL ) ;
err :
2023-12-17 06:43:41 +03:00
bch_err_fn ( ca , ret ) ;
2023-06-20 20:49:25 +03:00
return ret ;
2017-03-17 09:18:50 +03:00
}
2023-10-21 20:54:39 +03:00
int bch2_fs_journal_alloc ( struct bch_fs * c )
{
2023-12-17 07:47:29 +03:00
for_each_online_member ( c , ca ) {
2023-10-21 20:54:39 +03:00
if ( ca - > journal . nr )
continue ;
int ret = bch2_dev_journal_alloc ( ca ) ;
if ( ret ) {
percpu_ref_put ( & ca - > io_ref ) ;
return ret ;
}
}
return 0 ;
}
2017-03-17 09:18:50 +03:00
/* startup/shutdown: */
static bool bch2_journal_writing_to_device ( struct journal * j , unsigned dev_idx )
{
2020-11-14 02:36:33 +03:00
bool ret = false ;
2022-03-01 03:17:27 +03:00
u64 seq ;
2017-03-17 09:18:50 +03:00
spin_lock ( & j - > lock ) ;
2022-03-01 03:17:27 +03:00
for ( seq = journal_last_unwritten_seq ( j ) ;
seq < = journal_cur_seq ( j ) & & ! ret ;
seq + + ) {
struct journal_buf * buf = journal_seq_to_buf ( j , seq ) ;
2017-03-17 09:18:50 +03:00
2023-03-11 00:28:37 +03:00
if ( bch2_bkey_has_device_c ( bkey_i_to_s_c ( & buf - > key ) , dev_idx ) )
2020-11-14 02:36:33 +03:00
ret = true ;
}
2017-03-17 09:18:50 +03:00
spin_unlock ( & j - > lock ) ;
return ret ;
}
void bch2_dev_journal_stop ( struct journal * j , struct bch_dev * ca )
{
wait_event ( j - > wait , ! bch2_journal_writing_to_device ( j , ca - > dev_idx ) ) ;
}
void bch2_fs_journal_stop ( struct journal * j )
{
2022-03-05 21:38:54 +03:00
bch2_journal_reclaim_stop ( j ) ;
2019-05-10 18:58:00 +03:00
bch2_journal_flush_all_pins ( j ) ;
2023-12-10 23:23:27 +03:00
wait_event ( j - > wait , bch2_journal_entry_close ( j ) ) ;
2017-03-17 09:18:50 +03:00
2020-10-25 08:08:28 +03:00
/*
* Always write a new journal entry , to make sure the clock hands are up
* to date ( and match the superblock )
*/
bch2_journal_meta ( j ) ;
2017-03-17 09:18:50 +03:00
2019-02-15 02:38:52 +03:00
journal_quiesce ( j ) ;
2018-07-23 14:52:00 +03:00
2017-03-17 09:18:50 +03:00
BUG_ON ( ! bch2_journal_error ( j ) & &
2021-01-22 03:15:49 +03:00
test_bit ( JOURNAL_REPLAY_DONE , & j - > flags ) & &
2022-03-01 00:35:42 +03:00
j - > last_empty_seq ! = journal_cur_seq ( j ) ) ;
2017-03-17 09:18:50 +03:00
cancel_delayed_work_sync ( & j - > write_work ) ;
}
2022-03-21 07:15:53 +03:00
int bch2_fs_journal_start ( struct journal * j , u64 cur_seq )
2017-03-17 09:18:50 +03:00
{
2018-07-22 05:57:20 +03:00
struct bch_fs * c = container_of ( j , struct bch_fs , journal ) ;
2019-04-05 04:53:12 +03:00
struct journal_entry_pin_list * p ;
2022-03-21 07:15:53 +03:00
struct journal_replay * i , * * _i ;
struct genradix_iter iter ;
bool had_entries = false ;
2019-04-05 04:53:12 +03:00
u64 last_seq = cur_seq , nr , seq ;
2022-03-21 07:15:53 +03:00
genradix_for_each_reverse ( & c - > journal_entries , iter , _i ) {
i = * _i ;
if ( ! i | | i - > ignore )
continue ;
last_seq = le64_to_cpu ( i - > j . last_seq ) ;
break ;
}
2019-04-05 04:53:12 +03:00
nr = cur_seq - last_seq ;
if ( nr + 1 > j - > pin . size ) {
free_fifo ( & j - > pin ) ;
init_fifo ( & j - > pin , roundup_pow_of_two ( nr + 1 ) , GFP_KERNEL ) ;
if ( ! j - > pin . data ) {
bch_err ( c , " error reallocating journal fifo (%llu open entries) " , nr ) ;
2023-03-14 22:35:57 +03:00
return - BCH_ERR_ENOMEM_journal_pin_fifo ;
2019-04-05 04:53:12 +03:00
}
}
2019-04-12 05:39:39 +03:00
j - > replay_journal_seq = last_seq ;
j - > replay_journal_seq_end = cur_seq ;
2019-04-05 04:53:12 +03:00
j - > last_seq_ondisk = last_seq ;
2021-12-21 00:55:49 +03:00
j - > flushed_seq_ondisk = cur_seq - 1 ;
2022-03-01 03:17:27 +03:00
j - > seq_ondisk = cur_seq - 1 ;
2019-04-05 04:53:12 +03:00
j - > pin . front = last_seq ;
j - > pin . back = cur_seq ;
atomic64_set ( & j - > seq , cur_seq - 1 ) ;
2021-04-01 04:44:55 +03:00
fifo_for_each_entry_ptr ( p , & j - > pin , seq )
journal_pin_list_init ( p , 1 ) ;
2019-04-05 04:53:12 +03:00
2022-03-21 07:15:53 +03:00
genradix_for_each ( & c - > journal_entries , iter , _i ) {
i = * _i ;
if ( ! i | | i - > ignore )
continue ;
2021-01-27 00:04:12 +03:00
2019-04-05 04:53:12 +03:00
seq = le64_to_cpu ( i - > j . seq ) ;
2020-06-14 01:43:14 +03:00
BUG_ON ( seq > = cur_seq ) ;
2019-04-05 04:53:12 +03:00
2020-06-14 01:43:14 +03:00
if ( seq < last_seq )
continue ;
2017-03-17 09:18:50 +03:00
2021-12-21 00:55:49 +03:00
if ( journal_entry_empty ( & i - > j ) )
j - > last_empty_seq = le64_to_cpu ( i - > j . seq ) ;
2021-01-27 00:04:12 +03:00
p = journal_seq_pin ( j , seq ) ;
p - > devs . nr = 0 ;
2024-01-27 08:05:03 +03:00
darray_for_each ( i - > ptrs , ptr )
bch2_dev_list_add_dev ( & p - > devs , ptr - > dev ) ;
2022-03-21 07:15:53 +03:00
had_entries = true ;
2019-04-05 04:53:12 +03:00
}
2017-03-17 09:18:50 +03:00
2022-03-21 07:15:53 +03:00
if ( ! had_entries )
2021-12-21 00:55:49 +03:00
j - > last_empty_seq = cur_seq ;
2017-03-17 09:18:50 +03:00
spin_lock ( & j - > lock ) ;
set_bit ( JOURNAL_STARTED , & j - > flags ) ;
bcachefs: Don't require flush/fua on every journal write
This patch adds a flag to journal entries which, if set, indicates that
they weren't done as flush/fua writes.
- non flush/fua journal writes don't update last_seq (i.e. they don't
free up space in the journal), thus the journal free space
calculations now check whether nonflush journal writes are currently
allowed (i.e. are we low on free space, or would doing a flush write
free up a lot of space in the journal)
- write_delay_ms, the user configurable option for when open journal
entries are automatically written, is now interpreted as the max
delay between flush journal writes (default 1 second).
- bch2_journal_flush_seq_async is changed to ensure a flush write >=
the requested sequence number has happened
- journal read/replay must now ignore, and blacklist, any journal
entries newer than the most recent flush entry in the journal. Also,
the way the read_entire_journal option is handled has been improved;
struct journal_replay now has an entry, 'ignore', for entries that
were read but should not be used.
- assorted refactoring and improvements related to journal read in
journal_io.c and recovery.c
Previously, we'd have to issue a flush/fua write every time we
accumulated a full journal entry - typically the bucket size. Now we
need to issue them much less frequently: when an fsync is requested, or
it's been more than write_delay_ms since the last flush, or when we need
to free up space in the journal. This is a significant performance
improvement on many write heavy workloads.
Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
2020-11-14 17:59:58 +03:00
j - > last_flush_write = jiffies ;
2017-03-17 09:18:50 +03:00
2022-03-01 03:29:19 +03:00
j - > reservations . idx = j - > reservations . unwritten_idx = journal_cur_seq ( j ) ;
j - > reservations . unwritten_idx + + ;
2017-03-17 09:18:50 +03:00
2018-07-22 05:57:20 +03:00
c - > last_bucket_seq_cleanup = journal_cur_seq ( j ) ;
2019-02-21 21:33:21 +03:00
bch2_journal_space_available ( j ) ;
2017-03-17 09:18:50 +03:00
spin_unlock ( & j - > lock ) ;
2021-06-21 23:30:52 +03:00
return bch2_journal_reclaim_start ( j ) ;
2017-03-17 09:18:50 +03:00
}
/* init/exit: */
void bch2_dev_journal_exit ( struct bch_dev * ca )
{
2024-01-31 21:20:28 +03:00
struct journal_device * ja = & ca - > journal ;
for ( unsigned i = 0 ; i < ARRAY_SIZE ( ja - > bio ) ; i + + ) {
kfree ( ja - > bio [ i ] ) ;
ja - > bio [ i ] = NULL ;
}
2017-03-17 09:18:50 +03:00
2024-01-31 21:20:28 +03:00
kfree ( ja - > buckets ) ;
kfree ( ja - > bucket_seq ) ;
ja - > buckets = NULL ;
ja - > bucket_seq = NULL ;
2017-03-17 09:18:50 +03:00
}
int bch2_dev_journal_init ( struct bch_dev * ca , struct bch_sb * sb )
{
struct journal_device * ja = & ca - > journal ;
struct bch_sb_field_journal * journal_buckets =
2023-09-27 00:49:34 +03:00
bch2_sb_field_get ( sb , journal ) ;
2022-03-11 00:43:52 +03:00
struct bch_sb_field_journal_v2 * journal_buckets_v2 =
2023-09-27 00:49:34 +03:00
bch2_sb_field_get ( sb , journal_v2 ) ;
2017-03-17 09:18:50 +03:00
2022-03-11 00:43:52 +03:00
ja - > nr = 0 ;
if ( journal_buckets_v2 ) {
unsigned nr = bch2_sb_field_journal_v2_nr_entries ( journal_buckets_v2 ) ;
2024-01-31 21:20:28 +03:00
for ( unsigned i = 0 ; i < nr ; i + + )
2022-03-11 00:43:52 +03:00
ja - > nr + = le64_to_cpu ( journal_buckets_v2 - > d [ i ] . nr ) ;
} else if ( journal_buckets ) {
ja - > nr = bch2_nr_journal_buckets ( journal_buckets ) ;
}
2017-03-17 09:18:50 +03:00
ja - > bucket_seq = kcalloc ( ja - > nr , sizeof ( u64 ) , GFP_KERNEL ) ;
if ( ! ja - > bucket_seq )
2023-03-14 22:35:57 +03:00
return - BCH_ERR_ENOMEM_dev_journal_init ;
2017-03-17 09:18:50 +03:00
2024-01-31 21:20:28 +03:00
unsigned nr_bvecs = DIV_ROUND_UP ( JOURNAL_ENTRY_SIZE_MAX , PAGE_SIZE ) ;
2017-03-17 09:18:50 +03:00
2024-01-31 21:20:28 +03:00
for ( unsigned i = 0 ; i < ARRAY_SIZE ( ja - > bio ) ; i + + ) {
2024-01-31 21:42:48 +03:00
ja - > bio [ i ] = kmalloc ( struct_size ( ja - > bio [ i ] , bio . bi_inline_vecs ,
nr_bvecs ) , GFP_KERNEL ) ;
2024-01-31 21:20:28 +03:00
if ( ! ja - > bio [ i ] )
return - BCH_ERR_ENOMEM_dev_journal_init ;
2024-01-31 21:42:48 +03:00
ja - > bio [ i ] - > ca = ca ;
ja - > bio [ i ] - > buf_idx = i ;
bio_init ( & ja - > bio [ i ] - > bio , NULL , ja - > bio [ i ] - > bio . bi_inline_vecs , nr_bvecs , 0 ) ;
2024-01-31 21:20:28 +03:00
}
2017-03-17 09:18:50 +03:00
ja - > buckets = kcalloc ( ja - > nr , sizeof ( u64 ) , GFP_KERNEL ) ;
if ( ! ja - > buckets )
2023-03-14 22:35:57 +03:00
return - BCH_ERR_ENOMEM_dev_journal_init ;
2017-03-17 09:18:50 +03:00
2022-03-11 00:43:52 +03:00
if ( journal_buckets_v2 ) {
unsigned nr = bch2_sb_field_journal_v2_nr_entries ( journal_buckets_v2 ) ;
2024-01-31 21:20:28 +03:00
unsigned dst = 0 ;
2022-03-11 00:43:52 +03:00
2024-01-31 21:20:28 +03:00
for ( unsigned i = 0 ; i < nr ; i + + )
for ( unsigned j = 0 ; j < le64_to_cpu ( journal_buckets_v2 - > d [ i ] . nr ) ; j + + )
2022-03-11 00:43:52 +03:00
ja - > buckets [ dst + + ] =
le64_to_cpu ( journal_buckets_v2 - > d [ i ] . start ) + j ;
} else if ( journal_buckets ) {
2024-01-31 21:20:28 +03:00
for ( unsigned i = 0 ; i < ja - > nr ; i + + )
2022-03-11 00:43:52 +03:00
ja - > buckets [ i ] = le64_to_cpu ( journal_buckets - > buckets [ i ] ) ;
}
2017-03-17 09:18:50 +03:00
return 0 ;
}
void bch2_fs_journal_exit ( struct journal * j )
{
2024-01-31 19:21:46 +03:00
if ( j - > wq )
destroy_workqueue ( j - > wq ) ;
2020-11-14 02:36:33 +03:00
2022-12-14 18:39:04 +03:00
darray_exit ( & j - > early_journal_entries ) ;
2024-01-31 19:21:46 +03:00
for ( unsigned i = 0 ; i < ARRAY_SIZE ( j - > buf ) ; i + + )
2024-02-01 14:35:46 +03:00
kvfree ( j - > buf [ i ] . data ) ;
2017-03-17 09:18:50 +03:00
free_fifo ( & j - > pin ) ;
}
int bch2_fs_journal_init ( struct journal * j )
{
static struct lock_class_key res_key ;
2023-11-03 04:06:52 +03:00
mutex_init ( & j - > buf_lock ) ;
2017-03-17 09:18:50 +03:00
spin_lock_init ( & j - > lock ) ;
spin_lock_init ( & j - > err_lock ) ;
init_waitqueue_head ( & j - > wait ) ;
INIT_DELAYED_WORK ( & j - > write_work , journal_write_work ) ;
2021-04-01 00:52:52 +03:00
init_waitqueue_head ( & j - > reclaim_wait ) ;
2018-07-17 19:19:14 +03:00
init_waitqueue_head ( & j - > pin_flush_wait ) ;
2017-03-17 09:18:50 +03:00
mutex_init ( & j - > reclaim_lock ) ;
2019-03-03 23:15:55 +03:00
mutex_init ( & j - > discard_lock ) ;
2017-03-17 09:18:50 +03:00
lockdep_init_map ( & j - > res_map , " journal res " , & res_key , 0 ) ;
atomic64_set ( & j - > reservations . counter ,
( ( union journal_res_state )
{ . cur_entry_offset = JOURNAL_ENTRY_CLOSED_VAL } ) . v ) ;
2023-07-07 11:38:29 +03:00
if ( ! ( init_fifo ( & j - > pin , JOURNAL_PIN , GFP_KERNEL ) ) )
return - BCH_ERR_ENOMEM_journal_pin_fifo ;
2017-03-17 09:18:50 +03:00
2024-01-31 19:21:46 +03:00
for ( unsigned i = 0 ; i < ARRAY_SIZE ( j - > buf ) ; i + + ) {
2020-11-14 02:36:33 +03:00
j - > buf [ i ] . buf_size = JOURNAL_ENTRY_SIZE_MIN ;
2024-02-01 14:35:46 +03:00
j - > buf [ i ] . data = kvmalloc ( j - > buf [ i ] . buf_size , GFP_KERNEL ) ;
2023-07-07 11:38:29 +03:00
if ( ! j - > buf [ i ] . data )
return - BCH_ERR_ENOMEM_journal_buf ;
2024-01-31 21:42:48 +03:00
j - > buf [ i ] . idx = i ;
2020-11-14 02:36:33 +03:00
}
2017-03-17 09:18:50 +03:00
j - > pin . front = j - > pin . back = 1 ;
2024-01-31 19:21:46 +03:00
j - > wq = alloc_workqueue ( " bcachefs_journal " ,
WQ_HIGHPRI | WQ_FREEZABLE | WQ_UNBOUND | WQ_MEM_RECLAIM , 512 ) ;
if ( ! j - > wq )
return - BCH_ERR_ENOMEM_fs_other_alloc ;
2023-07-07 11:38:29 +03:00
return 0 ;
2017-03-17 09:18:50 +03:00
}
/* debug: */
2020-12-06 00:25:05 +03:00
void __bch2_journal_debug_to_text ( struct printbuf * out , struct journal * j )
2017-03-17 09:18:50 +03:00
{
struct bch_fs * c = container_of ( j , struct bch_fs , journal ) ;
2019-02-26 22:28:08 +03:00
union journal_res_state s ;
2021-12-05 06:03:07 +03:00
unsigned long now = jiffies ;
2023-11-09 00:51:06 +03:00
u64 nr_writes = j - > nr_flush_writes + j - > nr_noflush_writes ;
2017-03-17 09:18:50 +03:00
2023-02-04 05:01:40 +03:00
if ( ! out - > nr_tabstops )
printbuf_tabstop_push ( out , 24 ) ;
2022-02-25 21:18:19 +03:00
out - > atomic + + ;
2017-03-17 09:18:50 +03:00
rcu_read_lock ( ) ;
2019-02-26 22:28:08 +03:00
s = READ_ONCE ( j - > reservations ) ;
2017-03-17 09:18:50 +03:00
2022-10-20 01:31:33 +03:00
prt_printf ( out , " dirty journal entries: \t %llu/%llu \n " , fifo_used ( & j - > pin ) , j - > pin . size ) ;
2023-02-04 05:01:40 +03:00
prt_printf ( out , " seq: \t \t \t %llu \n " , journal_cur_seq ( j ) ) ;
prt_printf ( out , " seq_ondisk: \t \t %llu \n " , j - > seq_ondisk ) ;
2023-11-09 00:51:06 +03:00
prt_printf ( out , " last_seq: \t \t %llu \n " , journal_last_seq ( j ) ) ;
2023-02-04 05:01:40 +03:00
prt_printf ( out , " last_seq_ondisk: \t %llu \n " , j - > last_seq_ondisk ) ;
2023-11-09 00:51:06 +03:00
prt_printf ( out , " flushed_seq_ondisk: \t %llu \n " , j - > flushed_seq_ondisk ) ;
prt_printf ( out , " watermark: \t \t %s \n " , bch2_watermarks [ j - > watermark ] ) ;
prt_printf ( out , " each entry reserved: \t %u \n " , j - > entry_u64s_reserved ) ;
2023-02-04 05:01:40 +03:00
prt_printf ( out , " nr flush writes: \t %llu \n " , j - > nr_flush_writes ) ;
2023-11-09 00:51:06 +03:00
prt_printf ( out , " nr noflush writes: \t %llu \n " , j - > nr_noflush_writes ) ;
prt_printf ( out , " average write size: \t " ) ;
prt_human_readable_u64 ( out , nr_writes ? div64_u64 ( j - > entry_bytes_written , nr_writes ) : 0 ) ;
prt_newline ( out ) ;
prt_printf ( out , " nr direct reclaim: \t %llu \n " , j - > nr_direct_reclaim ) ;
2023-02-04 05:01:40 +03:00
prt_printf ( out , " nr background reclaim: \t %llu \n " , j - > nr_background_reclaim ) ;
prt_printf ( out , " reclaim kicked: \t \t %u \n " , j - > reclaim_kicked ) ;
2023-11-09 00:51:06 +03:00
prt_printf ( out , " reclaim runs in: \t %u ms \n " , time_after ( j - > next_reclaim , now )
2021-12-05 06:03:07 +03:00
? jiffies_to_msecs ( j - > next_reclaim - jiffies ) : 0 ) ;
2023-11-09 00:51:06 +03:00
prt_printf ( out , " current entry sectors: \t %u \n " , j - > cur_entry_sectors ) ;
prt_printf ( out , " current entry error: \t %s \n " , bch2_journal_errors [ j - > cur_entry_error ] ) ;
2023-02-04 05:01:40 +03:00
prt_printf ( out , " current entry: \t \t " ) ;
2019-02-26 22:28:08 +03:00
switch ( s . cur_entry_offset ) {
case JOURNAL_ENTRY_ERROR_VAL :
2023-02-04 05:01:40 +03:00
prt_printf ( out , " error " ) ;
2019-02-26 22:28:08 +03:00
break ;
case JOURNAL_ENTRY_CLOSED_VAL :
2023-02-04 05:01:40 +03:00
prt_printf ( out , " closed " ) ;
2019-02-26 22:28:08 +03:00
break ;
default :
2023-02-04 05:01:40 +03:00
prt_printf ( out , " %u/%u " , s . cur_entry_offset , j - > cur_entry_u64s ) ;
2019-02-26 22:28:08 +03:00
break ;
}
2023-02-04 05:01:40 +03:00
prt_newline ( out ) ;
2024-01-16 01:59:51 +03:00
prt_printf ( out , " unwritten entries: " ) ;
prt_newline ( out ) ;
bch2_journal_bufs_to_text ( out , j ) ;
2019-02-26 22:28:08 +03:00
2023-02-04 05:01:40 +03:00
prt_printf ( out ,
2019-02-26 22:28:08 +03:00
" replay done: \t \t %i \n " ,
2018-11-09 09:24:07 +03:00
test_bit ( JOURNAL_REPLAY_DONE , & j - > flags ) ) ;
2017-03-17 09:18:50 +03:00
2023-02-04 05:01:40 +03:00
prt_printf ( out , " space: \n " ) ;
prt_printf ( out , " \t discarded \t %u:%u \n " ,
2020-11-14 20:29:21 +03:00
j - > space [ journal_space_discarded ] . next_entry ,
j - > space [ journal_space_discarded ] . total ) ;
2023-02-04 05:01:40 +03:00
prt_printf ( out , " \t clean ondisk \t %u:%u \n " ,
2020-11-14 20:29:21 +03:00
j - > space [ journal_space_clean_ondisk ] . next_entry ,
j - > space [ journal_space_clean_ondisk ] . total ) ;
2023-02-04 05:01:40 +03:00
prt_printf ( out , " \t clean \t \t %u:%u \n " ,
2020-11-14 20:29:21 +03:00
j - > space [ journal_space_clean ] . next_entry ,
j - > space [ journal_space_clean ] . total ) ;
2023-02-04 05:01:40 +03:00
prt_printf ( out , " \t total \t \t %u:%u \n " ,
2020-11-14 20:29:21 +03:00
j - > space [ journal_space_total ] . next_entry ,
j - > space [ journal_space_total ] . total ) ;
2023-12-17 10:34:05 +03:00
for_each_member_device_rcu ( c , ca , & c - > rw_devs [ BCH_DATA_journal ] ) {
2017-03-17 09:18:50 +03:00
struct journal_device * ja = & ca - > journal ;
2021-03-19 23:30:01 +03:00
if ( ! test_bit ( ca - > dev_idx , c - > rw_devs [ BCH_DATA_journal ] . d ) )
continue ;
2017-03-17 09:18:50 +03:00
if ( ! ja - > nr )
continue ;
2023-12-17 10:34:05 +03:00
prt_printf ( out , " dev %u: \n " , ca - > dev_idx ) ;
2023-02-04 05:01:40 +03:00
prt_printf ( out , " \t nr \t \t %u \n " , ja - > nr ) ;
prt_printf ( out , " \t bucket size \t %u \n " , ca - > mi . bucket_size ) ;
prt_printf ( out , " \t available \t %u:%u \n " , bch2_journal_dev_buckets_available ( j , ja , journal_space_discarded ) , ja - > sectors_free ) ;
prt_printf ( out , " \t discard_idx \t %u \n " , ja - > discard_idx ) ;
prt_printf ( out , " \t dirty_ondisk \t %u (seq %llu) \n " , ja - > dirty_idx_ondisk , ja - > bucket_seq [ ja - > dirty_idx_ondisk ] ) ;
prt_printf ( out , " \t dirty_idx \t %u (seq %llu) \n " , ja - > dirty_idx , ja - > bucket_seq [ ja - > dirty_idx ] ) ;
prt_printf ( out , " \t cur_idx \t \t %u (seq %llu) \n " , ja - > cur_idx , ja - > bucket_seq [ ja - > cur_idx ] ) ;
2017-03-17 09:18:50 +03:00
}
rcu_read_unlock ( ) ;
2022-02-25 21:18:19 +03:00
- - out - > atomic ;
2017-03-17 09:18:50 +03:00
}
2020-12-06 00:25:05 +03:00
void bch2_journal_debug_to_text ( struct printbuf * out , struct journal * j )
{
spin_lock ( & j - > lock ) ;
__bch2_journal_debug_to_text ( out , j ) ;
spin_unlock ( & j - > lock ) ;
}
2022-02-26 19:48:34 +03:00
bool bch2_journal_seq_pins_to_text ( struct printbuf * out , struct journal * j , u64 * seq )
2017-03-17 09:18:50 +03:00
{
struct journal_entry_pin_list * pin_list ;
struct journal_entry_pin * pin ;
spin_lock ( & j - > lock ) ;
2022-02-26 19:48:34 +03:00
* seq = max ( * seq , j - > pin . front ) ;
if ( * seq > = j - > pin . back ) {
spin_unlock ( & j - > lock ) ;
return true ;
}
2022-02-25 21:18:19 +03:00
out - > atomic + + ;
2022-02-26 19:48:34 +03:00
pin_list = journal_seq_pin ( j , * seq ) ;
2017-03-17 09:18:50 +03:00
2023-02-04 05:01:40 +03:00
prt_printf ( out , " %llu: count %u " , * seq , atomic_read ( & pin_list - > count ) ) ;
prt_newline ( out ) ;
printbuf_indent_add ( out , 2 ) ;
2022-02-25 02:19:32 +03:00
2024-01-31 21:42:48 +03:00
for ( unsigned i = 0 ; i < ARRAY_SIZE ( pin_list - > list ) ; i + + )
2023-03-07 15:28:20 +03:00
list_for_each_entry ( pin , & pin_list - > list [ i ] , list ) {
prt_printf ( out , " \t %px %ps " , pin , pin - > flush ) ;
prt_newline ( out ) ;
}
2017-03-17 09:18:50 +03:00
2022-02-26 19:48:34 +03:00
if ( ! list_empty ( & pin_list - > flushed ) ) {
2023-02-04 05:01:40 +03:00
prt_printf ( out , " flushed: " ) ;
prt_newline ( out ) ;
2022-02-26 19:48:34 +03:00
}
2017-03-17 09:18:50 +03:00
2022-02-26 19:48:34 +03:00
list_for_each_entry ( pin , & pin_list - > flushed , list ) {
2023-02-04 05:01:40 +03:00
prt_printf ( out , " \t %px %ps " , pin , pin - > flush ) ;
prt_newline ( out ) ;
2017-03-17 09:18:50 +03:00
}
2022-02-25 21:18:19 +03:00
2023-02-04 05:01:40 +03:00
printbuf_indent_sub ( out , 2 ) ;
2022-02-26 19:48:34 +03:00
2022-02-25 21:18:19 +03:00
- - out - > atomic ;
2017-03-17 09:18:50 +03:00
spin_unlock ( & j - > lock ) ;
2022-02-26 19:48:34 +03:00
return false ;
}
void bch2_journal_pins_to_text ( struct printbuf * out , struct journal * j )
{
u64 seq = 0 ;
while ( ! bch2_journal_seq_pins_to_text ( out , j , & seq ) )
seq + + ;
2017-03-17 09:18:50 +03:00
}