2018-06-06 05:42:14 +03:00
// SPDX-License-Identifier: GPL-2.0
2005-04-17 02:20:36 +04:00
/*
2005-11-02 06:59:41 +03:00
* Copyright ( c ) 2000 - 2005 Silicon Graphics , Inc .
* All Rights Reserved .
2005-04-17 02:20:36 +04:00
*/
# include "xfs.h"
# include "xfs_fs.h"
2013-10-23 03:51:50 +04:00
# include "xfs_shared.h"
2013-08-12 14:49:26 +04:00
# include "xfs_format.h"
2013-10-23 03:50:10 +04:00
# include "xfs_log_format.h"
# include "xfs_trans_resv.h"
2005-11-02 06:38:42 +03:00
# include "xfs_bit.h"
2005-04-17 02:20:36 +04:00
# include "xfs_sb.h"
# include "xfs_mount.h"
# include "xfs_inode.h"
2019-07-02 19:39:39 +03:00
# include "xfs_iwalk.h"
2013-10-23 03:50:10 +04:00
# include "xfs_quota.h"
2005-11-02 06:38:42 +03:00
# include "xfs_bmap.h"
2017-08-30 01:44:14 +03:00
# include "xfs_bmap_util.h"
2013-10-23 03:50:10 +04:00
# include "xfs_trans.h"
2005-04-17 02:20:36 +04:00
# include "xfs_trans_space.h"
# include "xfs_qm.h"
2009-12-15 02:14:59 +03:00
# include "xfs_trace.h"
2012-10-08 14:56:11 +04:00
# include "xfs_icache.h"
2019-11-02 19:40:53 +03:00
# include "xfs_error.h"
2021-06-02 03:48:24 +03:00
# include "xfs_ag.h"
2021-06-02 03:48:24 +03:00
# include "xfs_ialloc.h"
xfs: xfs_is_shutdown vs xlog_is_shutdown cage fight
I've been chasing a recent resurgence in generic/388 recovery
failure and/or corruption events. The events have largely been
uninitialised inode chunks being tripped over in log recovery
such as:
XFS (pmem1): User initiated shutdown received.
pmem1: writeback error on inode 12621949, offset 1019904, sector 12968096
XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xa3/0xf0 (fs/xfs/xfs_fsops.c:500). Shutting down filesystem.
XFS (pmem1): Please unmount the filesystem and rectify the problem(s)
XFS (pmem1): Unmounting Filesystem
XFS (pmem1): Mounting V5 Filesystem
XFS (pmem1): Starting recovery (logdev: internal)
XFS (pmem1): bad inode magic/vsn daddr 8723584 #0 (magic=1818)
XFS (pmem1): Metadata corruption detected at xfs_inode_buf_verify+0x180/0x190, xfs_inode block 0x851c80 xfs_inode_buf_verify
XFS (pmem1): Unmount and run xfs_repair
XFS (pmem1): First 128 bytes of corrupted metadata buffer:
00000000: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000010: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000020: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000030: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000040: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000050: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000060: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000070: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
XFS (pmem1): metadata I/O error in "xlog_recover_items_pass2+0x52/0xc0" at daddr 0x851c80 len 32 error 117
XFS (pmem1): log mount/recovery failed: error -117
XFS (pmem1): log mount failed
There have been isolated random other issues, too - xfs_repair fails
because it finds some corruption in symlink blocks, rmap
inconsistencies, etc - but they are nowhere near as common as the
uninitialised inode chunk failure.
The problem has clearly happened at runtime before recovery has run;
I can see the ICREATE log item in the log shortly before the
actively recovered range of the log. This means the ICREATE was
definitely created and written to the log, but for some reason the
tail of the log has been moved past the ordered buffer log item that
tracks INODE_ALLOC buffers and, supposedly, prevents the tail of the
log moving past the ICREATE log item before the inode chunk buffer
is written to disk.
Tracing the fsstress processes that are running when the filesystem
shut down immediately pin-pointed the problem:
user shutdown marks xfs_mount as shutdown
godown-213341 [008] 6398.022871: console: [ 6397.915392] XFS (pmem1): User initiated shutdown received.
.....
aild tries to push ordered inode cluster buffer
xfsaild/pmem1-213314 [001] 6398.022974: xfs_buf_trylock: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 16 pincount 0 lock 0 flags DONE|INODES|PAGES caller xfs_inode_item_push+0x8e
xfsaild/pmem1-213314 [001] 6398.022976: xfs_ilock_nowait: dev 259:1 ino 0x851c80 flags ILOCK_SHARED caller xfs_iflush_cluster+0xae
xfs_iflush_cluster() checks xfs_is_shutdown(), returns true,
calls xfs_iflush_abort() to kill writeback of the inode.
Inode is removed from AIL, drops cluster buffer reference.
xfsaild/pmem1-213314 [001] 6398.022977: xfs_ail_delete: dev 259:1 lip 0xffff88880247ed80 old lsn 7/20344 new lsn 7/21000 type XFS_LI_INODE flags IN_AIL
xfsaild/pmem1-213314 [001] 6398.022978: xfs_buf_rele: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 17 pincount 0 lock 0 flags DONE|INODES|PAGES caller xfs_iflush_abort+0xd7
.....
All inodes on cluster buffer are aborted, then the cluster buffer
itself is aborted and removed from the AIL *without writeback*:
xfsaild/pmem1-213314 [001] 6398.023011: xfs_buf_error_relse: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 2 pincount 0 lock 0 flags ASYNC|DONE|STALE|INODES|PAGES caller xfs_buf_ioend_fail+0x33
xfsaild/pmem1-213314 [001] 6398.023012: xfs_ail_delete: dev 259:1 lip 0xffff8888053efde8 old lsn 7/20344 new lsn 7/20344 type XFS_LI_BUF flags IN_AIL
The inode buffer was at 7/20344 when it was removed from the AIL.
xfsaild/pmem1-213314 [001] 6398.023012: xfs_buf_item_relse: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 2 pincount 0 lock 0 flags ASYNC|DONE|STALE|INODES|PAGES caller xfs_buf_item_done+0x31
xfsaild/pmem1-213314 [001] 6398.023012: xfs_buf_rele: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 2 pincount 0 lock 0 flags ASYNC|DONE|STALE|INODES|PAGES caller xfs_buf_item_relse+0x39
.....
Userspace is still running, doing stuff. an fsstress process runs
syncfs() or sync() and we end up in sync_fs_one_sb() which issues
a log force. This pushes on the CIL:
fsstress-213322 [001] 6398.024430: xfs_fs_sync_fs: dev 259:1 m_features 0x20000000019ff6e9 opstate (clean|shutdown|inodegc|blockgc) s_flags 0x70810000 caller sync_fs_one_sb+0x26
fsstress-213322 [001] 6398.024430: xfs_log_force: dev 259:1 lsn 0x0 caller xfs_fs_sync_fs+0x82
fsstress-213322 [001] 6398.024430: xfs_log_force: dev 259:1 lsn 0x5f caller xfs_log_force+0x7c
<...>-194402 [001] 6398.024467: kmem_alloc: size 176 flags 0x14 caller xlog_cil_push_work+0x9f
And the CIL fills up iclogs with pending changes. This picks up
the current tail from the AIL:
<...>-194402 [001] 6398.024497: xlog_iclog_get_space: dev 259:1 state XLOG_STATE_ACTIVE refcnt 1 offset 0 lsn 0x0 flags caller xlog_write+0x149
<...>-194402 [001] 6398.024498: xlog_iclog_switch: dev 259:1 state XLOG_STATE_ACTIVE refcnt 1 offset 0 lsn 0x700005408 flags caller xlog_state_get_iclog_space+0x37e
<...>-194402 [001] 6398.024521: xlog_iclog_release: dev 259:1 state XLOG_STATE_WANT_SYNC refcnt 1 offset 32256 lsn 0x700005408 flags caller xlog_write+0x5f9
<...>-194402 [001] 6398.024522: xfs_log_assign_tail_lsn: dev 259:1 new tail lsn 7/21000, old lsn 7/20344, last sync 7/21448
And it moves the tail of the log to 7/21000 from 7/20344. This
*moves the tail of the log beyond the ICREATE transaction* that was
at 7/20344 and pinned by the inode cluster buffer that was cancelled
above.
....
godown-213341 [008] 6398.027005: xfs_force_shutdown: dev 259:1 tag logerror flags log_io|force_umount file fs/xfs/xfs_fsops.c line_num 500
godown-213341 [008] 6398.027022: console: [ 6397.915406] pmem1: writeback error on inode 12621949, offset 1019904, sector 12968096
godown-213341 [008] 6398.030551: console: [ 6397.919546] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xa3/0xf0 (fs/
And finally the log itself is now shutdown, stopping all further
writes to the log. But this is too late to prevent the corruption
that moving the tail of the log forwards after we start cancelling
writeback causes.
The fundamental problem here is that we are using the wrong shutdown
checks for log items. We've long conflated mount shutdown with log
shutdown state, and I started separating that recently with the
atomic shutdown state changes in commit b36d4651e165 ("xfs: make
forced shutdown processing atomic"). The changes in that commit
series are directly responsible for being able to diagnose this
issue because it clearly separated mount shutdown from log shutdown.
Essentially, once we start cancelling writeback of log items and
removing them from the AIL because the filesystem is shut down, we
*cannot* update the journal because we may have cancelled the items
that pin the tail of the log. That moves the tail of the log
forwards without having written the metadata back, hence we have
corrupt in memory state and writing to the journal propagates that
to the on-disk state.
What commit b36d4651e165 makes clear is that log item state needs to
change relative to log shutdown, not mount shutdown. IOWs, anything
that aborts metadata writeback needs to check log shutdown state
because log items directly affect log consistency. Having them check
mount shutdown state introduces the above race condition where we
cancel metadata writeback before the log shuts down.
To fix this, this patch works through all log items and converts
shutdown checks to use xlog_is_shutdown() rather than
xfs_is_shutdown(), so that we don't start aborting metadata
writeback before we shut off journal writes.
AFAICT, this race condition is a zero day IO error handling bug in
XFS that dates back to the introduction of XLOG_IO_ERROR,
XLOG_STATE_IOERROR and XFS_FORCED_SHUTDOWN back in January 1997.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Darrick J. Wong <djwong@kernel.org>
Signed-off-by: Darrick J. Wong <djwong@kernel.org>
2022-03-17 19:09:13 +03:00
# include "xfs_log_priv.h"
2005-04-17 02:20:36 +04:00
/*
* The global quota manager . There is only one of these for the entire
* system , _not_ one per file system . XQM keeps track of the overall
* quota functionality , including maintaining the freelist and hash
* tables of dquots .
*/
2019-11-13 04:04:26 +03:00
STATIC int xfs_qm_init_quotainos ( struct xfs_mount * mp ) ;
STATIC int xfs_qm_init_quotainfo ( struct xfs_mount * mp ) ;
2005-04-17 02:20:36 +04:00
2019-11-13 04:04:26 +03:00
STATIC void xfs_qm_destroy_quotainos ( struct xfs_quotainfo * qi ) ;
2013-08-28 04:18:07 +04:00
STATIC void xfs_qm_dqfree_one ( struct xfs_dquot * dqp ) ;
2012-03-14 20:53:34 +04:00
/*
* We use the batch lookup interface to iterate over the dquots as it
* currently is the only interface into the radix tree code that allows
* fuzzy lookups instead of exact matches . Holding the lock over multiple
* operations is fine as all callers are used either during mount / umount
* or quotaoff .
*/
# define XFS_DQ_LOOKUP_BATCH 32
STATIC int
xfs_qm_dquot_walk (
struct xfs_mount * mp ,
2020-07-16 03:53:43 +03:00
xfs_dqtype_t type ,
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
int ( * execute ) ( struct xfs_dquot * dqp , void * data ) ,
void * data )
2012-03-14 20:53:34 +04:00
{
struct xfs_quotainfo * qi = mp - > m_quotainfo ;
2013-06-28 02:25:05 +04:00
struct radix_tree_root * tree = xfs_dquot_tree ( qi , type ) ;
2012-03-14 20:53:34 +04:00
uint32_t next_index ;
int last_error = 0 ;
int skipped ;
int nr_found ;
restart :
skipped = 0 ;
next_index = 0 ;
nr_found = 0 ;
while ( 1 ) {
struct xfs_dquot * batch [ XFS_DQ_LOOKUP_BATCH ] ;
2023-01-03 20:54:28 +03:00
int error ;
2012-03-14 20:53:34 +04:00
int i ;
mutex_lock ( & qi - > qi_tree_lock ) ;
nr_found = radix_tree_gang_lookup ( tree , ( void * * ) batch ,
next_index , XFS_DQ_LOOKUP_BATCH ) ;
if ( ! nr_found ) {
mutex_unlock ( & qi - > qi_tree_lock ) ;
break ;
}
for ( i = 0 ; i < nr_found ; i + + ) {
struct xfs_dquot * dqp = batch [ i ] ;
2020-07-14 20:37:30 +03:00
next_index = dqp - > q_id + 1 ;
2012-03-14 20:53:34 +04:00
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
error = execute ( batch [ i ] , data ) ;
2014-06-25 08:58:08 +04:00
if ( error = = - EAGAIN ) {
2012-03-14 20:53:34 +04:00
skipped + + ;
continue ;
}
2014-06-25 08:58:08 +04:00
if ( error & & last_error ! = - EFSCORRUPTED )
2012-03-14 20:53:34 +04:00
last_error = error ;
}
mutex_unlock ( & qi - > qi_tree_lock ) ;
/* bail out if the filesystem is corrupted. */
2014-06-25 08:58:08 +04:00
if ( last_error = = - EFSCORRUPTED ) {
2012-03-14 20:53:34 +04:00
skipped = 0 ;
break ;
}
2017-07-24 18:33:25 +03:00
/* we're done if id overflows back to zero */
if ( ! next_index )
break ;
2012-03-14 20:53:34 +04:00
}
if ( skipped ) {
delay ( 1 ) ;
goto restart ;
}
return last_error ;
}
/*
* Purge a dquot from all tracking data structures and free it .
*/
STATIC int
xfs_qm_dqpurge (
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
struct xfs_dquot * dqp ,
void * data )
2012-03-14 20:53:34 +04:00
{
xfs: xfs_is_shutdown vs xlog_is_shutdown cage fight
I've been chasing a recent resurgence in generic/388 recovery
failure and/or corruption events. The events have largely been
uninitialised inode chunks being tripped over in log recovery
such as:
XFS (pmem1): User initiated shutdown received.
pmem1: writeback error on inode 12621949, offset 1019904, sector 12968096
XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xa3/0xf0 (fs/xfs/xfs_fsops.c:500). Shutting down filesystem.
XFS (pmem1): Please unmount the filesystem and rectify the problem(s)
XFS (pmem1): Unmounting Filesystem
XFS (pmem1): Mounting V5 Filesystem
XFS (pmem1): Starting recovery (logdev: internal)
XFS (pmem1): bad inode magic/vsn daddr 8723584 #0 (magic=1818)
XFS (pmem1): Metadata corruption detected at xfs_inode_buf_verify+0x180/0x190, xfs_inode block 0x851c80 xfs_inode_buf_verify
XFS (pmem1): Unmount and run xfs_repair
XFS (pmem1): First 128 bytes of corrupted metadata buffer:
00000000: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000010: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000020: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000030: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000040: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000050: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000060: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000070: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
XFS (pmem1): metadata I/O error in "xlog_recover_items_pass2+0x52/0xc0" at daddr 0x851c80 len 32 error 117
XFS (pmem1): log mount/recovery failed: error -117
XFS (pmem1): log mount failed
There have been isolated random other issues, too - xfs_repair fails
because it finds some corruption in symlink blocks, rmap
inconsistencies, etc - but they are nowhere near as common as the
uninitialised inode chunk failure.
The problem has clearly happened at runtime before recovery has run;
I can see the ICREATE log item in the log shortly before the
actively recovered range of the log. This means the ICREATE was
definitely created and written to the log, but for some reason the
tail of the log has been moved past the ordered buffer log item that
tracks INODE_ALLOC buffers and, supposedly, prevents the tail of the
log moving past the ICREATE log item before the inode chunk buffer
is written to disk.
Tracing the fsstress processes that are running when the filesystem
shut down immediately pin-pointed the problem:
user shutdown marks xfs_mount as shutdown
godown-213341 [008] 6398.022871: console: [ 6397.915392] XFS (pmem1): User initiated shutdown received.
.....
aild tries to push ordered inode cluster buffer
xfsaild/pmem1-213314 [001] 6398.022974: xfs_buf_trylock: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 16 pincount 0 lock 0 flags DONE|INODES|PAGES caller xfs_inode_item_push+0x8e
xfsaild/pmem1-213314 [001] 6398.022976: xfs_ilock_nowait: dev 259:1 ino 0x851c80 flags ILOCK_SHARED caller xfs_iflush_cluster+0xae
xfs_iflush_cluster() checks xfs_is_shutdown(), returns true,
calls xfs_iflush_abort() to kill writeback of the inode.
Inode is removed from AIL, drops cluster buffer reference.
xfsaild/pmem1-213314 [001] 6398.022977: xfs_ail_delete: dev 259:1 lip 0xffff88880247ed80 old lsn 7/20344 new lsn 7/21000 type XFS_LI_INODE flags IN_AIL
xfsaild/pmem1-213314 [001] 6398.022978: xfs_buf_rele: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 17 pincount 0 lock 0 flags DONE|INODES|PAGES caller xfs_iflush_abort+0xd7
.....
All inodes on cluster buffer are aborted, then the cluster buffer
itself is aborted and removed from the AIL *without writeback*:
xfsaild/pmem1-213314 [001] 6398.023011: xfs_buf_error_relse: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 2 pincount 0 lock 0 flags ASYNC|DONE|STALE|INODES|PAGES caller xfs_buf_ioend_fail+0x33
xfsaild/pmem1-213314 [001] 6398.023012: xfs_ail_delete: dev 259:1 lip 0xffff8888053efde8 old lsn 7/20344 new lsn 7/20344 type XFS_LI_BUF flags IN_AIL
The inode buffer was at 7/20344 when it was removed from the AIL.
xfsaild/pmem1-213314 [001] 6398.023012: xfs_buf_item_relse: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 2 pincount 0 lock 0 flags ASYNC|DONE|STALE|INODES|PAGES caller xfs_buf_item_done+0x31
xfsaild/pmem1-213314 [001] 6398.023012: xfs_buf_rele: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 2 pincount 0 lock 0 flags ASYNC|DONE|STALE|INODES|PAGES caller xfs_buf_item_relse+0x39
.....
Userspace is still running, doing stuff. an fsstress process runs
syncfs() or sync() and we end up in sync_fs_one_sb() which issues
a log force. This pushes on the CIL:
fsstress-213322 [001] 6398.024430: xfs_fs_sync_fs: dev 259:1 m_features 0x20000000019ff6e9 opstate (clean|shutdown|inodegc|blockgc) s_flags 0x70810000 caller sync_fs_one_sb+0x26
fsstress-213322 [001] 6398.024430: xfs_log_force: dev 259:1 lsn 0x0 caller xfs_fs_sync_fs+0x82
fsstress-213322 [001] 6398.024430: xfs_log_force: dev 259:1 lsn 0x5f caller xfs_log_force+0x7c
<...>-194402 [001] 6398.024467: kmem_alloc: size 176 flags 0x14 caller xlog_cil_push_work+0x9f
And the CIL fills up iclogs with pending changes. This picks up
the current tail from the AIL:
<...>-194402 [001] 6398.024497: xlog_iclog_get_space: dev 259:1 state XLOG_STATE_ACTIVE refcnt 1 offset 0 lsn 0x0 flags caller xlog_write+0x149
<...>-194402 [001] 6398.024498: xlog_iclog_switch: dev 259:1 state XLOG_STATE_ACTIVE refcnt 1 offset 0 lsn 0x700005408 flags caller xlog_state_get_iclog_space+0x37e
<...>-194402 [001] 6398.024521: xlog_iclog_release: dev 259:1 state XLOG_STATE_WANT_SYNC refcnt 1 offset 32256 lsn 0x700005408 flags caller xlog_write+0x5f9
<...>-194402 [001] 6398.024522: xfs_log_assign_tail_lsn: dev 259:1 new tail lsn 7/21000, old lsn 7/20344, last sync 7/21448
And it moves the tail of the log to 7/21000 from 7/20344. This
*moves the tail of the log beyond the ICREATE transaction* that was
at 7/20344 and pinned by the inode cluster buffer that was cancelled
above.
....
godown-213341 [008] 6398.027005: xfs_force_shutdown: dev 259:1 tag logerror flags log_io|force_umount file fs/xfs/xfs_fsops.c line_num 500
godown-213341 [008] 6398.027022: console: [ 6397.915406] pmem1: writeback error on inode 12621949, offset 1019904, sector 12968096
godown-213341 [008] 6398.030551: console: [ 6397.919546] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xa3/0xf0 (fs/
And finally the log itself is now shutdown, stopping all further
writes to the log. But this is too late to prevent the corruption
that moving the tail of the log forwards after we start cancelling
writeback causes.
The fundamental problem here is that we are using the wrong shutdown
checks for log items. We've long conflated mount shutdown with log
shutdown state, and I started separating that recently with the
atomic shutdown state changes in commit b36d4651e165 ("xfs: make
forced shutdown processing atomic"). The changes in that commit
series are directly responsible for being able to diagnose this
issue because it clearly separated mount shutdown from log shutdown.
Essentially, once we start cancelling writeback of log items and
removing them from the AIL because the filesystem is shut down, we
*cannot* update the journal because we may have cancelled the items
that pin the tail of the log. That moves the tail of the log
forwards without having written the metadata back, hence we have
corrupt in memory state and writing to the journal propagates that
to the on-disk state.
What commit b36d4651e165 makes clear is that log item state needs to
change relative to log shutdown, not mount shutdown. IOWs, anything
that aborts metadata writeback needs to check log shutdown state
because log items directly affect log consistency. Having them check
mount shutdown state introduces the above race condition where we
cancel metadata writeback before the log shuts down.
To fix this, this patch works through all log items and converts
shutdown checks to use xlog_is_shutdown() rather than
xfs_is_shutdown(), so that we don't start aborting metadata
writeback before we shut off journal writes.
AFAICT, this race condition is a zero day IO error handling bug in
XFS that dates back to the introduction of XLOG_IO_ERROR,
XLOG_STATE_IOERROR and XFS_FORCED_SHUTDOWN back in January 1997.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Darrick J. Wong <djwong@kernel.org>
Signed-off-by: Darrick J. Wong <djwong@kernel.org>
2022-03-17 19:09:13 +03:00
struct xfs_quotainfo * qi = dqp - > q_mount - > m_quotainfo ;
2020-03-27 18:29:45 +03:00
int error = - EAGAIN ;
2012-03-14 20:53:34 +04:00
xfs_dqlock ( dqp ) ;
2020-07-14 20:37:13 +03:00
if ( ( dqp - > q_flags & XFS_DQFLAG_FREEING ) | | dqp - > q_nrefs ! = 0 )
2020-03-27 18:29:45 +03:00
goto out_unlock ;
2012-03-14 20:53:34 +04:00
2020-07-14 20:37:13 +03:00
dqp - > q_flags | = XFS_DQFLAG_FREEING ;
2012-03-14 20:53:34 +04:00
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
xfs_dqflock ( dqp ) ;
2012-03-14 20:53:34 +04:00
/*
* If we are turning this type of quotas off , we don ' t care
* about the dirty metadata sitting in this dquot . OTOH , if
* we ' re unmounting , we do care , so we flush it and wait .
*/
if ( XFS_DQ_IS_DIRTY ( dqp ) ) {
2012-04-23 09:58:37 +04:00
struct xfs_buf * bp = NULL ;
2012-03-14 20:53:34 +04:00
/*
* We don ' t care about getting disk errors here . We need
* to purge this dquot anyway , so we go ahead regardless .
*/
2012-04-23 09:58:37 +04:00
error = xfs_qm_dqflush ( dqp , & bp ) ;
2018-05-05 01:30:20 +03:00
if ( ! error ) {
2012-04-23 09:58:37 +04:00
error = xfs_bwrite ( bp ) ;
xfs_buf_relse ( bp ) ;
2020-03-27 18:29:45 +03:00
} else if ( error = = - EAGAIN ) {
2020-07-14 20:37:13 +03:00
dqp - > q_flags & = ~ XFS_DQFLAG_FREEING ;
2020-03-27 18:29:45 +03:00
goto out_unlock ;
2012-04-23 09:58:37 +04:00
}
2012-03-14 20:53:34 +04:00
xfs_dqflock ( dqp ) ;
}
ASSERT ( atomic_read ( & dqp - > q_pincount ) = = 0 ) ;
xfs: xfs_is_shutdown vs xlog_is_shutdown cage fight
I've been chasing a recent resurgence in generic/388 recovery
failure and/or corruption events. The events have largely been
uninitialised inode chunks being tripped over in log recovery
such as:
XFS (pmem1): User initiated shutdown received.
pmem1: writeback error on inode 12621949, offset 1019904, sector 12968096
XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xa3/0xf0 (fs/xfs/xfs_fsops.c:500). Shutting down filesystem.
XFS (pmem1): Please unmount the filesystem and rectify the problem(s)
XFS (pmem1): Unmounting Filesystem
XFS (pmem1): Mounting V5 Filesystem
XFS (pmem1): Starting recovery (logdev: internal)
XFS (pmem1): bad inode magic/vsn daddr 8723584 #0 (magic=1818)
XFS (pmem1): Metadata corruption detected at xfs_inode_buf_verify+0x180/0x190, xfs_inode block 0x851c80 xfs_inode_buf_verify
XFS (pmem1): Unmount and run xfs_repair
XFS (pmem1): First 128 bytes of corrupted metadata buffer:
00000000: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000010: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000020: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000030: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000040: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000050: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000060: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000070: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
XFS (pmem1): metadata I/O error in "xlog_recover_items_pass2+0x52/0xc0" at daddr 0x851c80 len 32 error 117
XFS (pmem1): log mount/recovery failed: error -117
XFS (pmem1): log mount failed
There have been isolated random other issues, too - xfs_repair fails
because it finds some corruption in symlink blocks, rmap
inconsistencies, etc - but they are nowhere near as common as the
uninitialised inode chunk failure.
The problem has clearly happened at runtime before recovery has run;
I can see the ICREATE log item in the log shortly before the
actively recovered range of the log. This means the ICREATE was
definitely created and written to the log, but for some reason the
tail of the log has been moved past the ordered buffer log item that
tracks INODE_ALLOC buffers and, supposedly, prevents the tail of the
log moving past the ICREATE log item before the inode chunk buffer
is written to disk.
Tracing the fsstress processes that are running when the filesystem
shut down immediately pin-pointed the problem:
user shutdown marks xfs_mount as shutdown
godown-213341 [008] 6398.022871: console: [ 6397.915392] XFS (pmem1): User initiated shutdown received.
.....
aild tries to push ordered inode cluster buffer
xfsaild/pmem1-213314 [001] 6398.022974: xfs_buf_trylock: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 16 pincount 0 lock 0 flags DONE|INODES|PAGES caller xfs_inode_item_push+0x8e
xfsaild/pmem1-213314 [001] 6398.022976: xfs_ilock_nowait: dev 259:1 ino 0x851c80 flags ILOCK_SHARED caller xfs_iflush_cluster+0xae
xfs_iflush_cluster() checks xfs_is_shutdown(), returns true,
calls xfs_iflush_abort() to kill writeback of the inode.
Inode is removed from AIL, drops cluster buffer reference.
xfsaild/pmem1-213314 [001] 6398.022977: xfs_ail_delete: dev 259:1 lip 0xffff88880247ed80 old lsn 7/20344 new lsn 7/21000 type XFS_LI_INODE flags IN_AIL
xfsaild/pmem1-213314 [001] 6398.022978: xfs_buf_rele: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 17 pincount 0 lock 0 flags DONE|INODES|PAGES caller xfs_iflush_abort+0xd7
.....
All inodes on cluster buffer are aborted, then the cluster buffer
itself is aborted and removed from the AIL *without writeback*:
xfsaild/pmem1-213314 [001] 6398.023011: xfs_buf_error_relse: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 2 pincount 0 lock 0 flags ASYNC|DONE|STALE|INODES|PAGES caller xfs_buf_ioend_fail+0x33
xfsaild/pmem1-213314 [001] 6398.023012: xfs_ail_delete: dev 259:1 lip 0xffff8888053efde8 old lsn 7/20344 new lsn 7/20344 type XFS_LI_BUF flags IN_AIL
The inode buffer was at 7/20344 when it was removed from the AIL.
xfsaild/pmem1-213314 [001] 6398.023012: xfs_buf_item_relse: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 2 pincount 0 lock 0 flags ASYNC|DONE|STALE|INODES|PAGES caller xfs_buf_item_done+0x31
xfsaild/pmem1-213314 [001] 6398.023012: xfs_buf_rele: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 2 pincount 0 lock 0 flags ASYNC|DONE|STALE|INODES|PAGES caller xfs_buf_item_relse+0x39
.....
Userspace is still running, doing stuff. an fsstress process runs
syncfs() or sync() and we end up in sync_fs_one_sb() which issues
a log force. This pushes on the CIL:
fsstress-213322 [001] 6398.024430: xfs_fs_sync_fs: dev 259:1 m_features 0x20000000019ff6e9 opstate (clean|shutdown|inodegc|blockgc) s_flags 0x70810000 caller sync_fs_one_sb+0x26
fsstress-213322 [001] 6398.024430: xfs_log_force: dev 259:1 lsn 0x0 caller xfs_fs_sync_fs+0x82
fsstress-213322 [001] 6398.024430: xfs_log_force: dev 259:1 lsn 0x5f caller xfs_log_force+0x7c
<...>-194402 [001] 6398.024467: kmem_alloc: size 176 flags 0x14 caller xlog_cil_push_work+0x9f
And the CIL fills up iclogs with pending changes. This picks up
the current tail from the AIL:
<...>-194402 [001] 6398.024497: xlog_iclog_get_space: dev 259:1 state XLOG_STATE_ACTIVE refcnt 1 offset 0 lsn 0x0 flags caller xlog_write+0x149
<...>-194402 [001] 6398.024498: xlog_iclog_switch: dev 259:1 state XLOG_STATE_ACTIVE refcnt 1 offset 0 lsn 0x700005408 flags caller xlog_state_get_iclog_space+0x37e
<...>-194402 [001] 6398.024521: xlog_iclog_release: dev 259:1 state XLOG_STATE_WANT_SYNC refcnt 1 offset 32256 lsn 0x700005408 flags caller xlog_write+0x5f9
<...>-194402 [001] 6398.024522: xfs_log_assign_tail_lsn: dev 259:1 new tail lsn 7/21000, old lsn 7/20344, last sync 7/21448
And it moves the tail of the log to 7/21000 from 7/20344. This
*moves the tail of the log beyond the ICREATE transaction* that was
at 7/20344 and pinned by the inode cluster buffer that was cancelled
above.
....
godown-213341 [008] 6398.027005: xfs_force_shutdown: dev 259:1 tag logerror flags log_io|force_umount file fs/xfs/xfs_fsops.c line_num 500
godown-213341 [008] 6398.027022: console: [ 6397.915406] pmem1: writeback error on inode 12621949, offset 1019904, sector 12968096
godown-213341 [008] 6398.030551: console: [ 6397.919546] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xa3/0xf0 (fs/
And finally the log itself is now shutdown, stopping all further
writes to the log. But this is too late to prevent the corruption
that moving the tail of the log forwards after we start cancelling
writeback causes.
The fundamental problem here is that we are using the wrong shutdown
checks for log items. We've long conflated mount shutdown with log
shutdown state, and I started separating that recently with the
atomic shutdown state changes in commit b36d4651e165 ("xfs: make
forced shutdown processing atomic"). The changes in that commit
series are directly responsible for being able to diagnose this
issue because it clearly separated mount shutdown from log shutdown.
Essentially, once we start cancelling writeback of log items and
removing them from the AIL because the filesystem is shut down, we
*cannot* update the journal because we may have cancelled the items
that pin the tail of the log. That moves the tail of the log
forwards without having written the metadata back, hence we have
corrupt in memory state and writing to the journal propagates that
to the on-disk state.
What commit b36d4651e165 makes clear is that log item state needs to
change relative to log shutdown, not mount shutdown. IOWs, anything
that aborts metadata writeback needs to check log shutdown state
because log items directly affect log consistency. Having them check
mount shutdown state introduces the above race condition where we
cancel metadata writeback before the log shuts down.
To fix this, this patch works through all log items and converts
shutdown checks to use xlog_is_shutdown() rather than
xfs_is_shutdown(), so that we don't start aborting metadata
writeback before we shut off journal writes.
AFAICT, this race condition is a zero day IO error handling bug in
XFS that dates back to the introduction of XLOG_IO_ERROR,
XLOG_STATE_IOERROR and XFS_FORCED_SHUTDOWN back in January 1997.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Darrick J. Wong <djwong@kernel.org>
Signed-off-by: Darrick J. Wong <djwong@kernel.org>
2022-03-17 19:09:13 +03:00
ASSERT ( xlog_is_shutdown ( dqp - > q_logitem . qli_item . li_log ) | |
2018-05-09 17:47:34 +03:00
! test_bit ( XFS_LI_IN_AIL , & dqp - > q_logitem . qli_item . li_flags ) ) ;
2012-03-14 20:53:34 +04:00
xfs_dqfunlock ( dqp ) ;
xfs_dqunlock ( dqp ) ;
2020-07-14 20:37:32 +03:00
radix_tree_delete ( xfs_dquot_tree ( qi , xfs_dquot_type ( dqp ) ) , dqp - > q_id ) ;
2012-03-14 20:53:34 +04:00
qi - > qi_dquots - - ;
/*
* We move dquots to the freelist as soon as their reference count
* hits zero , so it really should be on the freelist here .
*/
ASSERT ( ! list_empty ( & dqp - > q_lru ) ) ;
2013-08-28 04:18:07 +04:00
list_lru_del ( & qi - > qi_lru , & dqp - > q_lru ) ;
xfs: xfs_is_shutdown vs xlog_is_shutdown cage fight
I've been chasing a recent resurgence in generic/388 recovery
failure and/or corruption events. The events have largely been
uninitialised inode chunks being tripped over in log recovery
such as:
XFS (pmem1): User initiated shutdown received.
pmem1: writeback error on inode 12621949, offset 1019904, sector 12968096
XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xa3/0xf0 (fs/xfs/xfs_fsops.c:500). Shutting down filesystem.
XFS (pmem1): Please unmount the filesystem and rectify the problem(s)
XFS (pmem1): Unmounting Filesystem
XFS (pmem1): Mounting V5 Filesystem
XFS (pmem1): Starting recovery (logdev: internal)
XFS (pmem1): bad inode magic/vsn daddr 8723584 #0 (magic=1818)
XFS (pmem1): Metadata corruption detected at xfs_inode_buf_verify+0x180/0x190, xfs_inode block 0x851c80 xfs_inode_buf_verify
XFS (pmem1): Unmount and run xfs_repair
XFS (pmem1): First 128 bytes of corrupted metadata buffer:
00000000: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000010: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000020: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000030: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000040: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000050: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000060: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
00000070: 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 18 ................
XFS (pmem1): metadata I/O error in "xlog_recover_items_pass2+0x52/0xc0" at daddr 0x851c80 len 32 error 117
XFS (pmem1): log mount/recovery failed: error -117
XFS (pmem1): log mount failed
There have been isolated random other issues, too - xfs_repair fails
because it finds some corruption in symlink blocks, rmap
inconsistencies, etc - but they are nowhere near as common as the
uninitialised inode chunk failure.
The problem has clearly happened at runtime before recovery has run;
I can see the ICREATE log item in the log shortly before the
actively recovered range of the log. This means the ICREATE was
definitely created and written to the log, but for some reason the
tail of the log has been moved past the ordered buffer log item that
tracks INODE_ALLOC buffers and, supposedly, prevents the tail of the
log moving past the ICREATE log item before the inode chunk buffer
is written to disk.
Tracing the fsstress processes that are running when the filesystem
shut down immediately pin-pointed the problem:
user shutdown marks xfs_mount as shutdown
godown-213341 [008] 6398.022871: console: [ 6397.915392] XFS (pmem1): User initiated shutdown received.
.....
aild tries to push ordered inode cluster buffer
xfsaild/pmem1-213314 [001] 6398.022974: xfs_buf_trylock: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 16 pincount 0 lock 0 flags DONE|INODES|PAGES caller xfs_inode_item_push+0x8e
xfsaild/pmem1-213314 [001] 6398.022976: xfs_ilock_nowait: dev 259:1 ino 0x851c80 flags ILOCK_SHARED caller xfs_iflush_cluster+0xae
xfs_iflush_cluster() checks xfs_is_shutdown(), returns true,
calls xfs_iflush_abort() to kill writeback of the inode.
Inode is removed from AIL, drops cluster buffer reference.
xfsaild/pmem1-213314 [001] 6398.022977: xfs_ail_delete: dev 259:1 lip 0xffff88880247ed80 old lsn 7/20344 new lsn 7/21000 type XFS_LI_INODE flags IN_AIL
xfsaild/pmem1-213314 [001] 6398.022978: xfs_buf_rele: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 17 pincount 0 lock 0 flags DONE|INODES|PAGES caller xfs_iflush_abort+0xd7
.....
All inodes on cluster buffer are aborted, then the cluster buffer
itself is aborted and removed from the AIL *without writeback*:
xfsaild/pmem1-213314 [001] 6398.023011: xfs_buf_error_relse: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 2 pincount 0 lock 0 flags ASYNC|DONE|STALE|INODES|PAGES caller xfs_buf_ioend_fail+0x33
xfsaild/pmem1-213314 [001] 6398.023012: xfs_ail_delete: dev 259:1 lip 0xffff8888053efde8 old lsn 7/20344 new lsn 7/20344 type XFS_LI_BUF flags IN_AIL
The inode buffer was at 7/20344 when it was removed from the AIL.
xfsaild/pmem1-213314 [001] 6398.023012: xfs_buf_item_relse: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 2 pincount 0 lock 0 flags ASYNC|DONE|STALE|INODES|PAGES caller xfs_buf_item_done+0x31
xfsaild/pmem1-213314 [001] 6398.023012: xfs_buf_rele: dev 259:1 daddr 0x851c80 bbcount 0x20 hold 2 pincount 0 lock 0 flags ASYNC|DONE|STALE|INODES|PAGES caller xfs_buf_item_relse+0x39
.....
Userspace is still running, doing stuff. an fsstress process runs
syncfs() or sync() and we end up in sync_fs_one_sb() which issues
a log force. This pushes on the CIL:
fsstress-213322 [001] 6398.024430: xfs_fs_sync_fs: dev 259:1 m_features 0x20000000019ff6e9 opstate (clean|shutdown|inodegc|blockgc) s_flags 0x70810000 caller sync_fs_one_sb+0x26
fsstress-213322 [001] 6398.024430: xfs_log_force: dev 259:1 lsn 0x0 caller xfs_fs_sync_fs+0x82
fsstress-213322 [001] 6398.024430: xfs_log_force: dev 259:1 lsn 0x5f caller xfs_log_force+0x7c
<...>-194402 [001] 6398.024467: kmem_alloc: size 176 flags 0x14 caller xlog_cil_push_work+0x9f
And the CIL fills up iclogs with pending changes. This picks up
the current tail from the AIL:
<...>-194402 [001] 6398.024497: xlog_iclog_get_space: dev 259:1 state XLOG_STATE_ACTIVE refcnt 1 offset 0 lsn 0x0 flags caller xlog_write+0x149
<...>-194402 [001] 6398.024498: xlog_iclog_switch: dev 259:1 state XLOG_STATE_ACTIVE refcnt 1 offset 0 lsn 0x700005408 flags caller xlog_state_get_iclog_space+0x37e
<...>-194402 [001] 6398.024521: xlog_iclog_release: dev 259:1 state XLOG_STATE_WANT_SYNC refcnt 1 offset 32256 lsn 0x700005408 flags caller xlog_write+0x5f9
<...>-194402 [001] 6398.024522: xfs_log_assign_tail_lsn: dev 259:1 new tail lsn 7/21000, old lsn 7/20344, last sync 7/21448
And it moves the tail of the log to 7/21000 from 7/20344. This
*moves the tail of the log beyond the ICREATE transaction* that was
at 7/20344 and pinned by the inode cluster buffer that was cancelled
above.
....
godown-213341 [008] 6398.027005: xfs_force_shutdown: dev 259:1 tag logerror flags log_io|force_umount file fs/xfs/xfs_fsops.c line_num 500
godown-213341 [008] 6398.027022: console: [ 6397.915406] pmem1: writeback error on inode 12621949, offset 1019904, sector 12968096
godown-213341 [008] 6398.030551: console: [ 6397.919546] XFS (pmem1): Log I/O Error (0x6) detected at xfs_fs_goingdown+0xa3/0xf0 (fs/
And finally the log itself is now shutdown, stopping all further
writes to the log. But this is too late to prevent the corruption
that moving the tail of the log forwards after we start cancelling
writeback causes.
The fundamental problem here is that we are using the wrong shutdown
checks for log items. We've long conflated mount shutdown with log
shutdown state, and I started separating that recently with the
atomic shutdown state changes in commit b36d4651e165 ("xfs: make
forced shutdown processing atomic"). The changes in that commit
series are directly responsible for being able to diagnose this
issue because it clearly separated mount shutdown from log shutdown.
Essentially, once we start cancelling writeback of log items and
removing them from the AIL because the filesystem is shut down, we
*cannot* update the journal because we may have cancelled the items
that pin the tail of the log. That moves the tail of the log
forwards without having written the metadata back, hence we have
corrupt in memory state and writing to the journal propagates that
to the on-disk state.
What commit b36d4651e165 makes clear is that log item state needs to
change relative to log shutdown, not mount shutdown. IOWs, anything
that aborts metadata writeback needs to check log shutdown state
because log items directly affect log consistency. Having them check
mount shutdown state introduces the above race condition where we
cancel metadata writeback before the log shuts down.
To fix this, this patch works through all log items and converts
shutdown checks to use xlog_is_shutdown() rather than
xfs_is_shutdown(), so that we don't start aborting metadata
writeback before we shut off journal writes.
AFAICT, this race condition is a zero day IO error handling bug in
XFS that dates back to the introduction of XLOG_IO_ERROR,
XLOG_STATE_IOERROR and XFS_FORCED_SHUTDOWN back in January 1997.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Darrick J. Wong <djwong@kernel.org>
Signed-off-by: Darrick J. Wong <djwong@kernel.org>
2022-03-17 19:09:13 +03:00
XFS_STATS_DEC ( dqp - > q_mount , xs_qm_dquot_unused ) ;
2012-03-14 20:53:34 +04:00
xfs_qm_dqdestroy ( dqp ) ;
xfs: fix infinite loop by detaching the group/project hints from user dquot
xfs_quota(8) will hang up if trying to turn group/project quota off
before the user quota is off, this could be 100% reproduced by:
# mount -ouquota,gquota /dev/sda7 /xfs
# mkdir /xfs/test
# xfs_quota -xc 'off -g' /xfs <-- hangs up
# echo w > /proc/sysrq-trigger
# dmesg
SysRq : Show Blocked State
task PC stack pid father
xfs_quota D 0000000000000000 0 27574 2551 0x00000000
[snip]
Call Trace:
[<ffffffff81aaa21d>] schedule+0xad/0xc0
[<ffffffff81aa327e>] schedule_timeout+0x35e/0x3c0
[<ffffffff8114b506>] ? mark_held_locks+0x176/0x1c0
[<ffffffff810ad6c0>] ? call_timer_fn+0x2c0/0x2c0
[<ffffffffa0c25380>] ? xfs_qm_shrink_count+0x30/0x30 [xfs]
[<ffffffff81aa3306>] schedule_timeout_uninterruptible+0x26/0x30
[<ffffffffa0c26155>] xfs_qm_dquot_walk+0x235/0x260 [xfs]
[<ffffffffa0c059d8>] ? xfs_perag_get+0x1d8/0x2d0 [xfs]
[<ffffffffa0c05805>] ? xfs_perag_get+0x5/0x2d0 [xfs]
[<ffffffffa0b7707e>] ? xfs_inode_ag_iterator+0xae/0xf0 [xfs]
[<ffffffffa0c22280>] ? xfs_trans_free_dqinfo+0x50/0x50 [xfs]
[<ffffffffa0b7709f>] ? xfs_inode_ag_iterator+0xcf/0xf0 [xfs]
[<ffffffffa0c261e6>] xfs_qm_dqpurge_all+0x66/0xb0 [xfs]
[<ffffffffa0c2497a>] xfs_qm_scall_quotaoff+0x20a/0x5f0 [xfs]
[<ffffffffa0c2b8f6>] xfs_fs_set_xstate+0x136/0x180 [xfs]
[<ffffffff8136cf7a>] do_quotactl+0x53a/0x6b0
[<ffffffff812fba4b>] ? iput+0x5b/0x90
[<ffffffff8136d257>] SyS_quotactl+0x167/0x1d0
[<ffffffff814cf2ee>] ? trace_hardirqs_on_thunk+0x3a/0x3f
[<ffffffff81abcd19>] system_call_fastpath+0x16/0x1b
It's fine if we turn user quota off at first, then turn off other
kind of quotas if they are enabled since the group/project dquot
refcount is decreased to zero once the user quota if off. Otherwise,
those dquots refcount is non-zero due to the user dquot might refer
to them as hint(s). Hence, above operation cause an infinite loop
at xfs_qm_dquot_walk() while trying to purge dquot cache.
This problem has been around since Linux 3.4, it was introduced by:
[ b84a3a9675 xfs: remove the per-filesystem list of dquots ]
Originally we will release the group dquot pointers because the user
dquots maybe carrying around as a hint via xfs_qm_detach_gdquots().
However, with above change, there is no such work to be done before
purging group/project dquot cache.
In order to solve this problem, this patch introduces a special routine
xfs_qm_dqpurge_hints(), and it would release the group/project dquot
pointers the user dquots maybe carrying around as a hint, and then it
will proceed to purge the user dquot cache if requested.
Cc: stable@vger.kernel.org
Signed-off-by: Jie Liu <jeff.liu@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2013-11-26 17:38:49 +04:00
return 0 ;
2020-03-27 18:29:45 +03:00
out_unlock :
xfs_dqunlock ( dqp ) ;
return error ;
xfs: fix infinite loop by detaching the group/project hints from user dquot
xfs_quota(8) will hang up if trying to turn group/project quota off
before the user quota is off, this could be 100% reproduced by:
# mount -ouquota,gquota /dev/sda7 /xfs
# mkdir /xfs/test
# xfs_quota -xc 'off -g' /xfs <-- hangs up
# echo w > /proc/sysrq-trigger
# dmesg
SysRq : Show Blocked State
task PC stack pid father
xfs_quota D 0000000000000000 0 27574 2551 0x00000000
[snip]
Call Trace:
[<ffffffff81aaa21d>] schedule+0xad/0xc0
[<ffffffff81aa327e>] schedule_timeout+0x35e/0x3c0
[<ffffffff8114b506>] ? mark_held_locks+0x176/0x1c0
[<ffffffff810ad6c0>] ? call_timer_fn+0x2c0/0x2c0
[<ffffffffa0c25380>] ? xfs_qm_shrink_count+0x30/0x30 [xfs]
[<ffffffff81aa3306>] schedule_timeout_uninterruptible+0x26/0x30
[<ffffffffa0c26155>] xfs_qm_dquot_walk+0x235/0x260 [xfs]
[<ffffffffa0c059d8>] ? xfs_perag_get+0x1d8/0x2d0 [xfs]
[<ffffffffa0c05805>] ? xfs_perag_get+0x5/0x2d0 [xfs]
[<ffffffffa0b7707e>] ? xfs_inode_ag_iterator+0xae/0xf0 [xfs]
[<ffffffffa0c22280>] ? xfs_trans_free_dqinfo+0x50/0x50 [xfs]
[<ffffffffa0b7709f>] ? xfs_inode_ag_iterator+0xcf/0xf0 [xfs]
[<ffffffffa0c261e6>] xfs_qm_dqpurge_all+0x66/0xb0 [xfs]
[<ffffffffa0c2497a>] xfs_qm_scall_quotaoff+0x20a/0x5f0 [xfs]
[<ffffffffa0c2b8f6>] xfs_fs_set_xstate+0x136/0x180 [xfs]
[<ffffffff8136cf7a>] do_quotactl+0x53a/0x6b0
[<ffffffff812fba4b>] ? iput+0x5b/0x90
[<ffffffff8136d257>] SyS_quotactl+0x167/0x1d0
[<ffffffff814cf2ee>] ? trace_hardirqs_on_thunk+0x3a/0x3f
[<ffffffff81abcd19>] system_call_fastpath+0x16/0x1b
It's fine if we turn user quota off at first, then turn off other
kind of quotas if they are enabled since the group/project dquot
refcount is decreased to zero once the user quota if off. Otherwise,
those dquots refcount is non-zero due to the user dquot might refer
to them as hint(s). Hence, above operation cause an infinite loop
at xfs_qm_dquot_walk() while trying to purge dquot cache.
This problem has been around since Linux 3.4, it was introduced by:
[ b84a3a9675 xfs: remove the per-filesystem list of dquots ]
Originally we will release the group dquot pointers because the user
dquots maybe carrying around as a hint via xfs_qm_detach_gdquots().
However, with above change, there is no such work to be done before
purging group/project dquot cache.
In order to solve this problem, this patch introduces a special routine
xfs_qm_dqpurge_hints(), and it would release the group/project dquot
pointers the user dquots maybe carrying around as a hint, and then it
will proceed to purge the user dquot cache if requested.
Cc: stable@vger.kernel.org
Signed-off-by: Jie Liu <jeff.liu@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2013-11-26 17:38:49 +04:00
}
2012-03-14 20:53:34 +04:00
/*
* Purge the dquot cache .
*/
2021-08-06 21:05:36 +03:00
static void
2012-03-14 20:53:34 +04:00
xfs_qm_dqpurge_all (
2021-08-06 21:05:36 +03:00
struct xfs_mount * mp )
2012-03-14 20:53:34 +04:00
{
2021-08-06 21:05:36 +03:00
xfs_qm_dquot_walk ( mp , XFS_DQTYPE_USER , xfs_qm_dqpurge , NULL ) ;
xfs_qm_dquot_walk ( mp , XFS_DQTYPE_GROUP , xfs_qm_dqpurge , NULL ) ;
xfs_qm_dquot_walk ( mp , XFS_DQTYPE_PROJ , xfs_qm_dqpurge , NULL ) ;
2012-03-14 20:53:34 +04:00
}
2005-04-17 02:20:36 +04:00
/*
* Just destroy the quotainfo structure .
*/
void
2009-06-08 17:33:32 +04:00
xfs_qm_unmount (
struct xfs_mount * mp )
2005-04-17 02:20:36 +04:00
{
2009-06-08 17:33:32 +04:00
if ( mp - > m_quotainfo ) {
2021-08-06 21:05:36 +03:00
xfs_qm_dqpurge_all ( mp ) ;
2005-04-17 02:20:36 +04:00
xfs_qm_destroy_quotainfo ( mp ) ;
2009-06-08 17:33:32 +04:00
}
2005-04-17 02:20:36 +04:00
}
/*
* Called from the vfsops layer .
*/
2008-12-03 14:20:36 +03:00
void
2005-04-17 02:20:36 +04:00
xfs_qm_unmount_quotas (
xfs_mount_t * mp )
{
/*
* Release the dquots that root inode , et al might be holding ,
* before we flush quotas and blow away the quotainfo structure .
*/
ASSERT ( mp - > m_rootip ) ;
xfs_qm_dqdetach ( mp - > m_rootip ) ;
if ( mp - > m_rbmip )
xfs_qm_dqdetach ( mp - > m_rbmip ) ;
if ( mp - > m_rsumip )
xfs_qm_dqdetach ( mp - > m_rsumip ) ;
/*
2008-12-03 14:20:36 +03:00
* Release the quota inodes .
2005-04-17 02:20:36 +04:00
*/
if ( mp - > m_quotainfo ) {
2008-12-03 14:20:36 +03:00
if ( mp - > m_quotainfo - > qi_uquotaip ) {
2018-07-25 22:52:32 +03:00
xfs_irele ( mp - > m_quotainfo - > qi_uquotaip ) ;
2008-12-03 14:20:36 +03:00
mp - > m_quotainfo - > qi_uquotaip = NULL ;
2005-04-17 02:20:36 +04:00
}
2008-12-03 14:20:36 +03:00
if ( mp - > m_quotainfo - > qi_gquotaip ) {
2018-07-25 22:52:32 +03:00
xfs_irele ( mp - > m_quotainfo - > qi_gquotaip ) ;
2008-12-03 14:20:36 +03:00
mp - > m_quotainfo - > qi_gquotaip = NULL ;
2005-04-17 02:20:36 +04:00
}
2013-07-11 09:00:40 +04:00
if ( mp - > m_quotainfo - > qi_pquotaip ) {
2018-07-25 22:52:32 +03:00
xfs_irele ( mp - > m_quotainfo - > qi_pquotaip ) ;
2013-07-11 09:00:40 +04:00
mp - > m_quotainfo - > qi_pquotaip = NULL ;
}
2005-04-17 02:20:36 +04:00
}
}
STATIC int
xfs_qm_dqattach_one (
2019-11-13 04:04:02 +03:00
struct xfs_inode * ip ,
2020-07-16 03:53:43 +03:00
xfs_dqtype_t type ,
2019-11-13 04:04:02 +03:00
bool doalloc ,
struct xfs_dquot * * IO_idqpp )
2005-04-17 02:20:36 +04:00
{
2019-11-13 04:04:02 +03:00
struct xfs_dquot * dqp ;
int error ;
2005-04-17 02:20:36 +04:00
2008-04-22 11:34:00 +04:00
ASSERT ( xfs_isilocked ( ip , XFS_ILOCK_EXCL ) ) ;
2005-04-17 02:20:36 +04:00
error = 0 ;
2009-02-08 23:51:42 +03:00
2005-04-17 02:20:36 +04:00
/*
xfs: remove dquot hints
group and project quota hints are currently stored on the user
dquot. If we are attaching quotas to the inode, then the group and
project dquots are stored as hints on the user dquot to save having
to look them up again later.
The thing is, the hints are not used for that inode for the rest of
the life of the inode - the dquots are attached directly to the
inode itself - so the only time the hints are used is when an inode
first has dquots attached.
When the hints on the user dquot don't match the dquots being
attache dto the inode, they are then removed and replaced with the
new hints. If a user is concurrently modifying files in different
group and/or project contexts, then this leads to thrashing of the
hints attached to user dquot.
If user quotas are not enabled, then hints are never even used.
So, if the hints are used to avoid the cost of the lookup, is the
cost of the lookup significant enough to justify the hint
infrstructure? Maybe it was once, when there was a global quota
manager shared between all XFS filesystems and was hash table based.
However, lookups are now much simpler, requiring only a single lock and
radix tree lookup local to the filesystem and no hash or LRU
manipulations to be made. Hence the cost of lookup is much lower
than when hints were implemented. Turns out that benchmarks show
that, too, with thir being no differnce in performance when doing
file creation workloads as a single user with user, group and
project quotas enabled - the hints do not make the code go any
faster. In fact, removing the hints shows a 2-3% reduction in the
time it takes to create 50 million inodes....
So, let's just get rid of the hints and the complexity around them.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2014-05-05 11:30:15 +04:00
* See if we already have it in the inode itself . IO_idqpp is & i_udquot
* or & i_gdquot . This made the code look weird , but made the logic a lot
* simpler .
2005-04-17 02:20:36 +04:00
*/
2009-02-08 23:51:42 +03:00
dqp = * IO_idqpp ;
if ( dqp ) {
2009-12-15 02:14:59 +03:00
trace_xfs_dqattach_found ( dqp ) ;
2009-02-08 23:51:42 +03:00
return 0 ;
2005-04-17 02:20:36 +04:00
}
/*
xfs: remove dquot hints
group and project quota hints are currently stored on the user
dquot. If we are attaching quotas to the inode, then the group and
project dquots are stored as hints on the user dquot to save having
to look them up again later.
The thing is, the hints are not used for that inode for the rest of
the life of the inode - the dquots are attached directly to the
inode itself - so the only time the hints are used is when an inode
first has dquots attached.
When the hints on the user dquot don't match the dquots being
attache dto the inode, they are then removed and replaced with the
new hints. If a user is concurrently modifying files in different
group and/or project contexts, then this leads to thrashing of the
hints attached to user dquot.
If user quotas are not enabled, then hints are never even used.
So, if the hints are used to avoid the cost of the lookup, is the
cost of the lookup significant enough to justify the hint
infrstructure? Maybe it was once, when there was a global quota
manager shared between all XFS filesystems and was hash table based.
However, lookups are now much simpler, requiring only a single lock and
radix tree lookup local to the filesystem and no hash or LRU
manipulations to be made. Hence the cost of lookup is much lower
than when hints were implemented. Turns out that benchmarks show
that, too, with thir being no differnce in performance when doing
file creation workloads as a single user with user, group and
project quotas enabled - the hints do not make the code go any
faster. In fact, removing the hints shows a 2-3% reduction in the
time it takes to create 50 million inodes....
So, let's just get rid of the hints and the complexity around them.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2014-05-05 11:30:15 +04:00
* Find the dquot from somewhere . This bumps the reference count of
* dquot and returns it locked . This can return ENOENT if dquot didn ' t
* exist on disk and we didn ' t ask it to allocate ; ESRCH if quotas got
* turned off suddenly .
2005-04-17 02:20:36 +04:00
*/
2018-05-05 01:30:22 +03:00
error = xfs_qm_dqget_inode ( ip , type , doalloc , & dqp ) ;
2009-02-08 23:51:42 +03:00
if ( error )
return error ;
2005-04-17 02:20:36 +04:00
2009-12-15 02:14:59 +03:00
trace_xfs_dqattach_get ( dqp ) ;
2009-02-08 23:51:42 +03:00
2005-04-17 02:20:36 +04:00
/*
* dqget may have dropped and re - acquired the ilock , but it guarantees
* that the dquot returned is the one that should go in the inode .
*/
* IO_idqpp = dqp ;
2009-02-08 23:51:42 +03:00
xfs_dqunlock ( dqp ) ;
return 0 ;
2005-04-17 02:20:36 +04:00
}
2012-03-27 18:34:46 +04:00
static bool
xfs_qm_need_dqattach (
struct xfs_inode * ip )
{
struct xfs_mount * mp = ip - > i_mount ;
if ( ! XFS_IS_QUOTA_ON ( mp ) )
return false ;
if ( ! XFS_NOT_DQATTACHED ( mp , ip ) )
return false ;
2013-06-28 02:25:04 +04:00
if ( xfs_is_quota_inode ( & mp - > m_sb , ip - > i_ino ) )
2012-03-27 18:34:46 +04:00
return false ;
return true ;
}
2005-04-17 02:20:36 +04:00
/*
2005-06-21 09:38:48 +04:00
* Given a locked inode , attach dquot ( s ) to it , taking U / G / P - QUOTAON
* into account .
2018-05-05 01:30:24 +03:00
* If @ doalloc is true , the dquot ( s ) will be allocated if needed .
2005-04-17 02:20:36 +04:00
* Inode may get unlocked and relocked in here , and the caller must deal with
* the consequences .
*/
int
2009-06-08 17:33:32 +04:00
xfs_qm_dqattach_locked (
2005-04-17 02:20:36 +04:00
xfs_inode_t * ip ,
2018-05-05 01:30:22 +03:00
bool doalloc )
2005-04-17 02:20:36 +04:00
{
xfs_mount_t * mp = ip - > i_mount ;
int error = 0 ;
2012-03-27 18:34:46 +04:00
if ( ! xfs_qm_need_dqattach ( ip ) )
2006-01-15 04:37:08 +03:00
return 0 ;
2005-04-17 02:20:36 +04:00
2009-06-08 17:33:32 +04:00
ASSERT ( xfs_isilocked ( ip , XFS_ILOCK_EXCL ) ) ;
2005-04-17 02:20:36 +04:00
xfs: remove dquot hints
group and project quota hints are currently stored on the user
dquot. If we are attaching quotas to the inode, then the group and
project dquots are stored as hints on the user dquot to save having
to look them up again later.
The thing is, the hints are not used for that inode for the rest of
the life of the inode - the dquots are attached directly to the
inode itself - so the only time the hints are used is when an inode
first has dquots attached.
When the hints on the user dquot don't match the dquots being
attache dto the inode, they are then removed and replaced with the
new hints. If a user is concurrently modifying files in different
group and/or project contexts, then this leads to thrashing of the
hints attached to user dquot.
If user quotas are not enabled, then hints are never even used.
So, if the hints are used to avoid the cost of the lookup, is the
cost of the lookup significant enough to justify the hint
infrstructure? Maybe it was once, when there was a global quota
manager shared between all XFS filesystems and was hash table based.
However, lookups are now much simpler, requiring only a single lock and
radix tree lookup local to the filesystem and no hash or LRU
manipulations to be made. Hence the cost of lookup is much lower
than when hints were implemented. Turns out that benchmarks show
that, too, with thir being no differnce in performance when doing
file creation workloads as a single user with user, group and
project quotas enabled - the hints do not make the code go any
faster. In fact, removing the hints shows a 2-3% reduction in the
time it takes to create 50 million inodes....
So, let's just get rid of the hints and the complexity around them.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2014-05-05 11:30:15 +04:00
if ( XFS_IS_UQUOTA_ON ( mp ) & & ! ip - > i_udquot ) {
2020-09-22 19:19:18 +03:00
error = xfs_qm_dqattach_one ( ip , XFS_DQTYPE_USER ,
doalloc , & ip - > i_udquot ) ;
2005-04-17 02:20:36 +04:00
if ( error )
goto done ;
xfs: remove dquot hints
group and project quota hints are currently stored on the user
dquot. If we are attaching quotas to the inode, then the group and
project dquots are stored as hints on the user dquot to save having
to look them up again later.
The thing is, the hints are not used for that inode for the rest of
the life of the inode - the dquots are attached directly to the
inode itself - so the only time the hints are used is when an inode
first has dquots attached.
When the hints on the user dquot don't match the dquots being
attache dto the inode, they are then removed and replaced with the
new hints. If a user is concurrently modifying files in different
group and/or project contexts, then this leads to thrashing of the
hints attached to user dquot.
If user quotas are not enabled, then hints are never even used.
So, if the hints are used to avoid the cost of the lookup, is the
cost of the lookup significant enough to justify the hint
infrstructure? Maybe it was once, when there was a global quota
manager shared between all XFS filesystems and was hash table based.
However, lookups are now much simpler, requiring only a single lock and
radix tree lookup local to the filesystem and no hash or LRU
manipulations to be made. Hence the cost of lookup is much lower
than when hints were implemented. Turns out that benchmarks show
that, too, with thir being no differnce in performance when doing
file creation workloads as a single user with user, group and
project quotas enabled - the hints do not make the code go any
faster. In fact, removing the hints shows a 2-3% reduction in the
time it takes to create 50 million inodes....
So, let's just get rid of the hints and the complexity around them.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2014-05-05 11:30:15 +04:00
ASSERT ( ip - > i_udquot ) ;
2005-04-17 02:20:36 +04:00
}
2008-04-22 11:34:00 +04:00
xfs: remove dquot hints
group and project quota hints are currently stored on the user
dquot. If we are attaching quotas to the inode, then the group and
project dquots are stored as hints on the user dquot to save having
to look them up again later.
The thing is, the hints are not used for that inode for the rest of
the life of the inode - the dquots are attached directly to the
inode itself - so the only time the hints are used is when an inode
first has dquots attached.
When the hints on the user dquot don't match the dquots being
attache dto the inode, they are then removed and replaced with the
new hints. If a user is concurrently modifying files in different
group and/or project contexts, then this leads to thrashing of the
hints attached to user dquot.
If user quotas are not enabled, then hints are never even used.
So, if the hints are used to avoid the cost of the lookup, is the
cost of the lookup significant enough to justify the hint
infrstructure? Maybe it was once, when there was a global quota
manager shared between all XFS filesystems and was hash table based.
However, lookups are now much simpler, requiring only a single lock and
radix tree lookup local to the filesystem and no hash or LRU
manipulations to be made. Hence the cost of lookup is much lower
than when hints were implemented. Turns out that benchmarks show
that, too, with thir being no differnce in performance when doing
file creation workloads as a single user with user, group and
project quotas enabled - the hints do not make the code go any
faster. In fact, removing the hints shows a 2-3% reduction in the
time it takes to create 50 million inodes....
So, let's just get rid of the hints and the complexity around them.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2014-05-05 11:30:15 +04:00
if ( XFS_IS_GQUOTA_ON ( mp ) & & ! ip - > i_gdquot ) {
2020-09-22 19:19:18 +03:00
error = xfs_qm_dqattach_one ( ip , XFS_DQTYPE_GROUP ,
doalloc , & ip - > i_gdquot ) ;
2005-04-17 02:20:36 +04:00
if ( error )
goto done ;
xfs: remove dquot hints
group and project quota hints are currently stored on the user
dquot. If we are attaching quotas to the inode, then the group and
project dquots are stored as hints on the user dquot to save having
to look them up again later.
The thing is, the hints are not used for that inode for the rest of
the life of the inode - the dquots are attached directly to the
inode itself - so the only time the hints are used is when an inode
first has dquots attached.
When the hints on the user dquot don't match the dquots being
attache dto the inode, they are then removed and replaced with the
new hints. If a user is concurrently modifying files in different
group and/or project contexts, then this leads to thrashing of the
hints attached to user dquot.
If user quotas are not enabled, then hints are never even used.
So, if the hints are used to avoid the cost of the lookup, is the
cost of the lookup significant enough to justify the hint
infrstructure? Maybe it was once, when there was a global quota
manager shared between all XFS filesystems and was hash table based.
However, lookups are now much simpler, requiring only a single lock and
radix tree lookup local to the filesystem and no hash or LRU
manipulations to be made. Hence the cost of lookup is much lower
than when hints were implemented. Turns out that benchmarks show
that, too, with thir being no differnce in performance when doing
file creation workloads as a single user with user, group and
project quotas enabled - the hints do not make the code go any
faster. In fact, removing the hints shows a 2-3% reduction in the
time it takes to create 50 million inodes....
So, let's just get rid of the hints and the complexity around them.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2014-05-05 11:30:15 +04:00
ASSERT ( ip - > i_gdquot ) ;
2005-04-17 02:20:36 +04:00
}
xfs: remove dquot hints
group and project quota hints are currently stored on the user
dquot. If we are attaching quotas to the inode, then the group and
project dquots are stored as hints on the user dquot to save having
to look them up again later.
The thing is, the hints are not used for that inode for the rest of
the life of the inode - the dquots are attached directly to the
inode itself - so the only time the hints are used is when an inode
first has dquots attached.
When the hints on the user dquot don't match the dquots being
attache dto the inode, they are then removed and replaced with the
new hints. If a user is concurrently modifying files in different
group and/or project contexts, then this leads to thrashing of the
hints attached to user dquot.
If user quotas are not enabled, then hints are never even used.
So, if the hints are used to avoid the cost of the lookup, is the
cost of the lookup significant enough to justify the hint
infrstructure? Maybe it was once, when there was a global quota
manager shared between all XFS filesystems and was hash table based.
However, lookups are now much simpler, requiring only a single lock and
radix tree lookup local to the filesystem and no hash or LRU
manipulations to be made. Hence the cost of lookup is much lower
than when hints were implemented. Turns out that benchmarks show
that, too, with thir being no differnce in performance when doing
file creation workloads as a single user with user, group and
project quotas enabled - the hints do not make the code go any
faster. In fact, removing the hints shows a 2-3% reduction in the
time it takes to create 50 million inodes....
So, let's just get rid of the hints and the complexity around them.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2014-05-05 11:30:15 +04:00
if ( XFS_IS_PQUOTA_ON ( mp ) & & ! ip - > i_pdquot ) {
2020-09-22 19:19:18 +03:00
error = xfs_qm_dqattach_one ( ip , XFS_DQTYPE_PROJ ,
2018-05-05 01:30:22 +03:00
doalloc , & ip - > i_pdquot ) ;
2013-07-11 09:00:40 +04:00
if ( error )
goto done ;
xfs: remove dquot hints
group and project quota hints are currently stored on the user
dquot. If we are attaching quotas to the inode, then the group and
project dquots are stored as hints on the user dquot to save having
to look them up again later.
The thing is, the hints are not used for that inode for the rest of
the life of the inode - the dquots are attached directly to the
inode itself - so the only time the hints are used is when an inode
first has dquots attached.
When the hints on the user dquot don't match the dquots being
attache dto the inode, they are then removed and replaced with the
new hints. If a user is concurrently modifying files in different
group and/or project contexts, then this leads to thrashing of the
hints attached to user dquot.
If user quotas are not enabled, then hints are never even used.
So, if the hints are used to avoid the cost of the lookup, is the
cost of the lookup significant enough to justify the hint
infrstructure? Maybe it was once, when there was a global quota
manager shared between all XFS filesystems and was hash table based.
However, lookups are now much simpler, requiring only a single lock and
radix tree lookup local to the filesystem and no hash or LRU
manipulations to be made. Hence the cost of lookup is much lower
than when hints were implemented. Turns out that benchmarks show
that, too, with thir being no differnce in performance when doing
file creation workloads as a single user with user, group and
project quotas enabled - the hints do not make the code go any
faster. In fact, removing the hints shows a 2-3% reduction in the
time it takes to create 50 million inodes....
So, let's just get rid of the hints and the complexity around them.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2014-05-05 11:30:15 +04:00
ASSERT ( ip - > i_pdquot ) ;
2013-07-11 09:00:40 +04:00
}
xfs: remove dquot hints
group and project quota hints are currently stored on the user
dquot. If we are attaching quotas to the inode, then the group and
project dquots are stored as hints on the user dquot to save having
to look them up again later.
The thing is, the hints are not used for that inode for the rest of
the life of the inode - the dquots are attached directly to the
inode itself - so the only time the hints are used is when an inode
first has dquots attached.
When the hints on the user dquot don't match the dquots being
attache dto the inode, they are then removed and replaced with the
new hints. If a user is concurrently modifying files in different
group and/or project contexts, then this leads to thrashing of the
hints attached to user dquot.
If user quotas are not enabled, then hints are never even used.
So, if the hints are used to avoid the cost of the lookup, is the
cost of the lookup significant enough to justify the hint
infrstructure? Maybe it was once, when there was a global quota
manager shared between all XFS filesystems and was hash table based.
However, lookups are now much simpler, requiring only a single lock and
radix tree lookup local to the filesystem and no hash or LRU
manipulations to be made. Hence the cost of lookup is much lower
than when hints were implemented. Turns out that benchmarks show
that, too, with thir being no differnce in performance when doing
file creation workloads as a single user with user, group and
project quotas enabled - the hints do not make the code go any
faster. In fact, removing the hints shows a 2-3% reduction in the
time it takes to create 50 million inodes....
So, let's just get rid of the hints and the complexity around them.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2014-05-05 11:30:15 +04:00
done :
2005-04-17 02:20:36 +04:00
/*
xfs: remove dquot hints
group and project quota hints are currently stored on the user
dquot. If we are attaching quotas to the inode, then the group and
project dquots are stored as hints on the user dquot to save having
to look them up again later.
The thing is, the hints are not used for that inode for the rest of
the life of the inode - the dquots are attached directly to the
inode itself - so the only time the hints are used is when an inode
first has dquots attached.
When the hints on the user dquot don't match the dquots being
attache dto the inode, they are then removed and replaced with the
new hints. If a user is concurrently modifying files in different
group and/or project contexts, then this leads to thrashing of the
hints attached to user dquot.
If user quotas are not enabled, then hints are never even used.
So, if the hints are used to avoid the cost of the lookup, is the
cost of the lookup significant enough to justify the hint
infrstructure? Maybe it was once, when there was a global quota
manager shared between all XFS filesystems and was hash table based.
However, lookups are now much simpler, requiring only a single lock and
radix tree lookup local to the filesystem and no hash or LRU
manipulations to be made. Hence the cost of lookup is much lower
than when hints were implemented. Turns out that benchmarks show
that, too, with thir being no differnce in performance when doing
file creation workloads as a single user with user, group and
project quotas enabled - the hints do not make the code go any
faster. In fact, removing the hints shows a 2-3% reduction in the
time it takes to create 50 million inodes....
So, let's just get rid of the hints and the complexity around them.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2014-05-05 11:30:15 +04:00
* Don ' t worry about the dquots that we may have attached before any
* error - they ' ll get detached later if it has not already been done .
2005-04-17 02:20:36 +04:00
*/
2009-06-08 17:33:32 +04:00
ASSERT ( xfs_isilocked ( ip , XFS_ILOCK_EXCL ) ) ;
return error ;
}
2005-04-17 02:20:36 +04:00
2009-06-08 17:33:32 +04:00
int
xfs_qm_dqattach (
2018-05-05 01:30:21 +03:00
struct xfs_inode * ip )
2009-06-08 17:33:32 +04:00
{
int error ;
2012-03-27 18:34:46 +04:00
if ( ! xfs_qm_need_dqattach ( ip ) )
return 0 ;
2009-06-08 17:33:32 +04:00
xfs_ilock ( ip , XFS_ILOCK_EXCL ) ;
2018-05-05 01:30:22 +03:00
error = xfs_qm_dqattach_locked ( ip , false ) ;
2009-06-08 17:33:32 +04:00
xfs_iunlock ( ip , XFS_ILOCK_EXCL ) ;
2005-04-17 02:20:36 +04:00
2006-01-15 04:37:08 +03:00
return error ;
2005-04-17 02:20:36 +04:00
}
/*
* Release dquots ( and their references ) if any .
* The inode should be locked EXCL except when this ' s called by
* xfs_ireclaim .
*/
void
xfs_qm_dqdetach (
xfs_inode_t * ip )
{
2013-07-11 09:00:40 +04:00
if ( ! ( ip - > i_udquot | | ip - > i_gdquot | | ip - > i_pdquot ) )
2005-04-17 02:20:36 +04:00
return ;
2009-12-15 02:14:59 +03:00
trace_xfs_dquot_dqdetach ( ip ) ;
2013-06-28 02:25:04 +04:00
ASSERT ( ! xfs_is_quota_inode ( & ip - > i_mount - > m_sb , ip - > i_ino ) ) ;
2005-04-17 02:20:36 +04:00
if ( ip - > i_udquot ) {
xfs_qm_dqrele ( ip - > i_udquot ) ;
ip - > i_udquot = NULL ;
}
if ( ip - > i_gdquot ) {
xfs_qm_dqrele ( ip - > i_gdquot ) ;
ip - > i_gdquot = NULL ;
}
2013-07-11 09:00:40 +04:00
if ( ip - > i_pdquot ) {
xfs_qm_dqrele ( ip - > i_pdquot ) ;
ip - > i_pdquot = NULL ;
}
2005-04-17 02:20:36 +04:00
}
2013-08-28 04:18:07 +04:00
struct xfs_qm_isolate {
struct list_head buffers ;
struct list_head dispose ;
} ;
static enum lru_status
xfs_qm_dquot_isolate (
struct list_head * item ,
2015-02-13 01:59:35 +03:00
struct list_lru_one * lru ,
2013-08-28 04:18:07 +04:00
spinlock_t * lru_lock ,
void * arg )
2014-09-29 04:43:40 +04:00
__releases ( lru_lock ) __acquires ( lru_lock )
2013-08-28 04:18:07 +04:00
{
struct xfs_dquot * dqp = container_of ( item ,
struct xfs_dquot , q_lru ) ;
struct xfs_qm_isolate * isol = arg ;
if ( ! xfs_dqlock_nowait ( dqp ) )
goto out_miss_busy ;
2022-12-08 19:29:22 +03:00
/*
* If something else is freeing this dquot and hasn ' t yet removed it
* from the LRU , leave it for the freeing task to complete the freeing
* process rather than risk it being free from under us here .
*/
if ( dqp - > q_flags & XFS_DQFLAG_FREEING )
goto out_miss_unlock ;
2013-08-28 04:18:07 +04:00
/*
* This dquot has acquired a reference in the meantime remove it from
* the freelist and try again .
*/
if ( dqp - > q_nrefs ) {
xfs_dqunlock ( dqp ) ;
2015-10-12 10:21:22 +03:00
XFS_STATS_INC ( dqp - > q_mount , xs_qm_dqwants ) ;
2013-08-28 04:18:07 +04:00
trace_xfs_dqreclaim_want ( dqp ) ;
2015-02-13 01:59:35 +03:00
list_lru_isolate ( lru , & dqp - > q_lru ) ;
2015-10-12 10:21:22 +03:00
XFS_STATS_DEC ( dqp - > q_mount , xs_qm_dquot_unused ) ;
2013-08-28 04:18:08 +04:00
return LRU_REMOVED ;
2013-08-28 04:18:07 +04:00
}
/*
* If the dquot is dirty , flush it . If it ' s already being flushed , just
* skip it so there is time for the IO to complete before we try to
* reclaim it again on the next LRU pass .
*/
2022-12-08 19:29:22 +03:00
if ( ! xfs_dqflock_nowait ( dqp ) )
goto out_miss_unlock ;
2013-08-28 04:18:07 +04:00
if ( XFS_DQ_IS_DIRTY ( dqp ) ) {
struct xfs_buf * bp = NULL ;
int error ;
trace_xfs_dqreclaim_dirty ( dqp ) ;
/* we have to drop the LRU lock to flush the dquot */
spin_unlock ( lru_lock ) ;
error = xfs_qm_dqflush ( dqp , & bp ) ;
2018-05-05 01:30:20 +03:00
if ( error )
2013-08-28 04:18:07 +04:00
goto out_unlock_dirty ;
xfs_buf_delwri_queue ( bp , & isol - > buffers ) ;
xfs_buf_relse ( bp ) ;
goto out_unlock_dirty ;
}
xfs_dqfunlock ( dqp ) ;
/*
* Prevent lookups now that we are past the point of no return .
*/
2020-07-14 20:37:13 +03:00
dqp - > q_flags | = XFS_DQFLAG_FREEING ;
2013-08-28 04:18:07 +04:00
xfs_dqunlock ( dqp ) ;
ASSERT ( dqp - > q_nrefs = = 0 ) ;
2015-02-13 01:59:35 +03:00
list_lru_isolate_move ( lru , & dqp - > q_lru , & isol - > dispose ) ;
2015-10-12 10:21:22 +03:00
XFS_STATS_DEC ( dqp - > q_mount , xs_qm_dquot_unused ) ;
2013-08-28 04:18:07 +04:00
trace_xfs_dqreclaim_done ( dqp ) ;
2015-10-12 10:21:22 +03:00
XFS_STATS_INC ( dqp - > q_mount , xs_qm_dqreclaims ) ;
2013-08-28 04:18:08 +04:00
return LRU_REMOVED ;
2013-08-28 04:18:07 +04:00
2022-12-08 19:29:22 +03:00
out_miss_unlock :
xfs_dqunlock ( dqp ) ;
2013-08-28 04:18:07 +04:00
out_miss_busy :
trace_xfs_dqreclaim_busy ( dqp ) ;
2015-10-12 10:21:22 +03:00
XFS_STATS_INC ( dqp - > q_mount , xs_qm_dqreclaim_misses ) ;
2013-08-28 04:18:08 +04:00
return LRU_SKIP ;
2013-08-28 04:18:07 +04:00
out_unlock_dirty :
trace_xfs_dqreclaim_busy ( dqp ) ;
2015-10-12 10:21:22 +03:00
XFS_STATS_INC ( dqp - > q_mount , xs_qm_dqreclaim_misses ) ;
2013-08-28 04:18:08 +04:00
xfs_dqunlock ( dqp ) ;
spin_lock ( lru_lock ) ;
return LRU_RETRY ;
2013-08-28 04:18:07 +04:00
}
2013-08-28 04:18:08 +04:00
static unsigned long
2013-08-28 04:18:07 +04:00
xfs_qm_shrink_scan (
struct shrinker * shrink ,
struct shrink_control * sc )
{
2023-09-11 12:44:35 +03:00
struct xfs_quotainfo * qi = shrink - > private_data ;
2013-08-28 04:18:07 +04:00
struct xfs_qm_isolate isol ;
2013-08-28 04:18:08 +04:00
unsigned long freed ;
2013-08-28 04:18:07 +04:00
int error ;
2015-11-07 03:28:21 +03:00
if ( ( sc - > gfp_mask & ( __GFP_FS | __GFP_DIRECT_RECLAIM ) ) ! = ( __GFP_FS | __GFP_DIRECT_RECLAIM ) )
2013-08-28 04:18:07 +04:00
return 0 ;
INIT_LIST_HEAD ( & isol . buffers ) ;
INIT_LIST_HEAD ( & isol . dispose ) ;
list_lru: introduce list_lru_shrink_{count,walk}
Kmem accounting of memcg is unusable now, because it lacks slab shrinker
support. That means when we hit the limit we will get ENOMEM w/o any
chance to recover. What we should do then is to call shrink_slab, which
would reclaim old inode/dentry caches from this cgroup. This is what
this patch set is intended to do.
Basically, it does two things. First, it introduces the notion of
per-memcg slab shrinker. A shrinker that wants to reclaim objects per
cgroup should mark itself as SHRINKER_MEMCG_AWARE. Then it will be
passed the memory cgroup to scan from in shrink_control->memcg. For
such shrinkers shrink_slab iterates over the whole cgroup subtree under
the target cgroup and calls the shrinker for each kmem-active memory
cgroup.
Secondly, this patch set makes the list_lru structure per-memcg. It's
done transparently to list_lru users - everything they have to do is to
tell list_lru_init that they want memcg-aware list_lru. Then the
list_lru will automatically distribute objects among per-memcg lists
basing on which cgroup the object is accounted to. This way to make FS
shrinkers (icache, dcache) memcg-aware we only need to make them use
memcg-aware list_lru, and this is what this patch set does.
As before, this patch set only enables per-memcg kmem reclaim when the
pressure goes from memory.limit, not from memory.kmem.limit. Handling
memory.kmem.limit is going to be tricky due to GFP_NOFS allocations, and
it is still unclear whether we will have this knob in the unified
hierarchy.
This patch (of 9):
NUMA aware slab shrinkers use the list_lru structure to distribute
objects coming from different NUMA nodes to different lists. Whenever
such a shrinker needs to count or scan objects from a particular node,
it issues commands like this:
count = list_lru_count_node(lru, sc->nid);
freed = list_lru_walk_node(lru, sc->nid, isolate_func,
isolate_arg, &sc->nr_to_scan);
where sc is an instance of the shrink_control structure passed to it
from vmscan.
To simplify this, let's add special list_lru functions to be used by
shrinkers, list_lru_shrink_count() and list_lru_shrink_walk(), which
consolidate the nid and nr_to_scan arguments in the shrink_control
structure.
This will also allow us to avoid patching shrinkers that use list_lru
when we make shrink_slab() per-memcg - all we will have to do is extend
the shrink_control structure to include the target memcg and make
list_lru_shrink_{count,walk} handle this appropriately.
Signed-off-by: Vladimir Davydov <vdavydov@parallels.com>
Suggested-by: Dave Chinner <david@fromorbit.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Michal Hocko <mhocko@suse.cz>
Cc: Greg Thelen <gthelen@google.com>
Cc: Glauber Costa <glommer@gmail.com>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Christoph Lameter <cl@linux.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: David Rientjes <rientjes@google.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Tejun Heo <tj@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2015-02-13 01:58:47 +03:00
freed = list_lru_shrink_walk ( & qi - > qi_lru , sc ,
xfs_qm_dquot_isolate , & isol ) ;
2013-08-28 04:18:07 +04:00
error = xfs_buf_delwri_submit ( & isol . buffers ) ;
if ( error )
xfs_warn ( NULL , " %s: dquot reclaim failed " , __func__ ) ;
while ( ! list_empty ( & isol . dispose ) ) {
struct xfs_dquot * dqp ;
dqp = list_first_entry ( & isol . dispose , struct xfs_dquot , q_lru ) ;
list_del_init ( & dqp - > q_lru ) ;
xfs_qm_dqfree_one ( dqp ) ;
}
return freed ;
}
2013-08-28 04:18:08 +04:00
static unsigned long
2013-08-28 04:18:07 +04:00
xfs_qm_shrink_count (
struct shrinker * shrink ,
struct shrink_control * sc )
{
2023-09-11 12:44:35 +03:00
struct xfs_quotainfo * qi = shrink - > private_data ;
2013-08-28 04:18:07 +04:00
list_lru: introduce list_lru_shrink_{count,walk}
Kmem accounting of memcg is unusable now, because it lacks slab shrinker
support. That means when we hit the limit we will get ENOMEM w/o any
chance to recover. What we should do then is to call shrink_slab, which
would reclaim old inode/dentry caches from this cgroup. This is what
this patch set is intended to do.
Basically, it does two things. First, it introduces the notion of
per-memcg slab shrinker. A shrinker that wants to reclaim objects per
cgroup should mark itself as SHRINKER_MEMCG_AWARE. Then it will be
passed the memory cgroup to scan from in shrink_control->memcg. For
such shrinkers shrink_slab iterates over the whole cgroup subtree under
the target cgroup and calls the shrinker for each kmem-active memory
cgroup.
Secondly, this patch set makes the list_lru structure per-memcg. It's
done transparently to list_lru users - everything they have to do is to
tell list_lru_init that they want memcg-aware list_lru. Then the
list_lru will automatically distribute objects among per-memcg lists
basing on which cgroup the object is accounted to. This way to make FS
shrinkers (icache, dcache) memcg-aware we only need to make them use
memcg-aware list_lru, and this is what this patch set does.
As before, this patch set only enables per-memcg kmem reclaim when the
pressure goes from memory.limit, not from memory.kmem.limit. Handling
memory.kmem.limit is going to be tricky due to GFP_NOFS allocations, and
it is still unclear whether we will have this knob in the unified
hierarchy.
This patch (of 9):
NUMA aware slab shrinkers use the list_lru structure to distribute
objects coming from different NUMA nodes to different lists. Whenever
such a shrinker needs to count or scan objects from a particular node,
it issues commands like this:
count = list_lru_count_node(lru, sc->nid);
freed = list_lru_walk_node(lru, sc->nid, isolate_func,
isolate_arg, &sc->nr_to_scan);
where sc is an instance of the shrink_control structure passed to it
from vmscan.
To simplify this, let's add special list_lru functions to be used by
shrinkers, list_lru_shrink_count() and list_lru_shrink_walk(), which
consolidate the nid and nr_to_scan arguments in the shrink_control
structure.
This will also allow us to avoid patching shrinkers that use list_lru
when we make shrink_slab() per-memcg - all we will have to do is extend
the shrink_control structure to include the target memcg and make
list_lru_shrink_{count,walk} handle this appropriately.
Signed-off-by: Vladimir Davydov <vdavydov@parallels.com>
Suggested-by: Dave Chinner <david@fromorbit.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Michal Hocko <mhocko@suse.cz>
Cc: Greg Thelen <gthelen@google.com>
Cc: Glauber Costa <glommer@gmail.com>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Christoph Lameter <cl@linux.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: David Rientjes <rientjes@google.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Tejun Heo <tj@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2015-02-13 01:58:47 +03:00
return list_lru_shrink_count ( & qi - > qi_lru , sc ) ;
2013-08-28 04:18:07 +04:00
}
2016-02-08 03:27:55 +03:00
STATIC void
xfs_qm_set_defquota (
2019-11-13 04:04:26 +03:00
struct xfs_mount * mp ,
2020-07-16 03:53:43 +03:00
xfs_dqtype_t type ,
2019-11-13 04:04:26 +03:00
struct xfs_quotainfo * qinf )
2016-02-08 03:27:55 +03:00
{
2019-11-13 04:04:02 +03:00
struct xfs_dquot * dqp ;
struct xfs_def_quota * defq ;
2016-02-08 03:27:55 +03:00
int error ;
2018-05-05 01:30:23 +03:00
error = xfs_qm_dqget_uncached ( mp , 0 , type , & dqp ) ;
2018-05-14 16:34:32 +03:00
if ( error )
return ;
2016-02-08 03:27:55 +03:00
2020-05-21 23:07:00 +03:00
defq = xfs_get_defquota ( qinf , xfs_dquot_type ( dqp ) ) ;
2016-02-08 03:27:55 +03:00
2018-05-14 16:34:32 +03:00
/*
* Timers and warnings have been already set , let ' s just set the
* default limits for this quota type
*/
2020-07-14 20:37:32 +03:00
defq - > blk . hard = dqp - > q_blk . hardlimit ;
defq - > blk . soft = dqp - > q_blk . softlimit ;
defq - > ino . hard = dqp - > q_ino . hardlimit ;
defq - > ino . soft = dqp - > q_ino . softlimit ;
defq - > rtb . hard = dqp - > q_rtb . hardlimit ;
defq - > rtb . soft = dqp - > q_rtb . softlimit ;
2018-05-14 16:34:32 +03:00
xfs_qm_dqdestroy ( dqp ) ;
}
/* Initialize quota time limits from the root dquot. */
static void
xfs_qm_init_timelimits (
struct xfs_mount * mp ,
2020-07-16 03:53:43 +03:00
xfs_dqtype_t type )
2018-05-14 16:34:32 +03:00
{
2020-05-21 23:07:01 +03:00
struct xfs_quotainfo * qinf = mp - > m_quotainfo ;
struct xfs_def_quota * defq ;
2018-05-14 16:34:32 +03:00
struct xfs_dquot * dqp ;
int error ;
2020-05-21 23:07:01 +03:00
defq = xfs_get_defquota ( qinf , type ) ;
2020-07-14 20:37:32 +03:00
defq - > blk . time = XFS_QM_BTIMELIMIT ;
defq - > ino . time = XFS_QM_ITIMELIMIT ;
defq - > rtb . time = XFS_QM_RTBTIMELIMIT ;
2018-05-14 16:34:32 +03:00
/*
* We try to get the limits from the superuser ' s limits fields .
* This is quite hacky , but it is standard quota practice .
*
* Since we may not have done a quotacheck by this point , just read
* the dquot without attaching it to any hashtables or lists .
*/
error = xfs_qm_dqget_uncached ( mp , 0 , type , & dqp ) ;
if ( error )
return ;
/*
* The warnings and timers set the grace period given to
* a user or group before he or she can not perform any
* more writing . If it is zero , a default is used .
*/
2020-07-14 20:37:32 +03:00
if ( dqp - > q_blk . timer )
2020-07-14 20:37:32 +03:00
defq - > blk . time = dqp - > q_blk . timer ;
2020-07-14 20:37:32 +03:00
if ( dqp - > q_ino . timer )
2020-07-14 20:37:32 +03:00
defq - > ino . time = dqp - > q_ino . timer ;
2020-07-14 20:37:32 +03:00
if ( dqp - > q_rtb . timer )
2020-07-14 20:37:32 +03:00
defq - > rtb . time = dqp - > q_rtb . timer ;
2018-05-14 16:34:32 +03:00
xfs_qm_dqdestroy ( dqp ) ;
2016-02-08 03:27:55 +03:00
}
2005-04-17 02:20:36 +04:00
/*
* This initializes all the quota information that ' s kept in the
* mount structure
*/
2005-06-21 09:36:52 +04:00
STATIC int
2005-04-17 02:20:36 +04:00
xfs_qm_init_quotainfo (
2018-05-05 01:30:23 +03:00
struct xfs_mount * mp )
2005-04-17 02:20:36 +04:00
{
2018-05-05 01:30:23 +03:00
struct xfs_quotainfo * qinf ;
int error ;
2005-04-17 02:20:36 +04:00
2021-08-06 21:05:37 +03:00
ASSERT ( XFS_IS_QUOTA_ON ( mp ) ) ;
2005-04-17 02:20:36 +04:00
2019-11-13 04:04:26 +03:00
qinf = mp - > m_quotainfo = kmem_zalloc ( sizeof ( struct xfs_quotainfo ) , 0 ) ;
2005-04-17 02:20:36 +04:00
2014-06-25 08:58:08 +04:00
error = list_lru_init ( & qinf - > qi_lru ) ;
2014-05-15 03:23:24 +04:00
if ( error )
goto out_free_qinf ;
2013-08-28 04:18:18 +04:00
2005-04-17 02:20:36 +04:00
/*
* See if quotainodes are setup , and if not , allocate them ,
* and change the superblock accordingly .
*/
2014-05-15 03:23:24 +04:00
error = xfs_qm_init_quotainos ( mp ) ;
if ( error )
goto out_free_lru ;
2005-04-17 02:20:36 +04:00
2012-03-13 12:52:35 +04:00
INIT_RADIX_TREE ( & qinf - > qi_uquota_tree , GFP_NOFS ) ;
INIT_RADIX_TREE ( & qinf - > qi_gquota_tree , GFP_NOFS ) ;
2013-07-11 09:00:40 +04:00
INIT_RADIX_TREE ( & qinf - > qi_pquota_tree , GFP_NOFS ) ;
2012-03-13 12:52:35 +04:00
mutex_init ( & qinf - > qi_tree_lock ) ;
2005-04-17 02:20:36 +04:00
/* mutex used to serialize quotaoffs */
2006-01-10 02:59:21 +03:00
mutex_init ( & qinf - > qi_quotaofflock ) ;
2005-04-17 02:20:36 +04:00
/* Precalc some constants */
qinf - > qi_dqchunklen = XFS_FSB_TO_BB ( mp , XFS_DQUOT_CLUSTER_SIZE_FSB ) ;
2014-04-14 13:03:34 +04:00
qinf - > qi_dqperchunk = xfs_calc_dquots_per_chunk ( qinf - > qi_dqchunklen ) ;
2021-08-19 04:46:37 +03:00
if ( xfs_has_bigtime ( mp ) ) {
2020-08-17 19:59:51 +03:00
qinf - > qi_expiry_min =
xfs_dq_bigtime_to_unix ( XFS_DQ_BIGTIME_EXPIRY_MIN ) ;
qinf - > qi_expiry_max =
xfs_dq_bigtime_to_unix ( XFS_DQ_BIGTIME_EXPIRY_MAX ) ;
} else {
qinf - > qi_expiry_min = XFS_DQ_LEGACY_EXPIRY_MIN ;
qinf - > qi_expiry_max = XFS_DQ_LEGACY_EXPIRY_MAX ;
}
2020-08-24 21:58:01 +03:00
trace_xfs_quota_expiry_range ( mp , qinf - > qi_expiry_min ,
qinf - > qi_expiry_max ) ;
2005-04-17 02:20:36 +04:00
mp - > m_qflags | = ( mp - > m_sb . sb_qflags & XFS_ALL_QUOTA_CHKD ) ;
2020-07-16 03:42:36 +03:00
xfs_qm_init_timelimits ( mp , XFS_DQTYPE_USER ) ;
xfs_qm_init_timelimits ( mp , XFS_DQTYPE_GROUP ) ;
xfs_qm_init_timelimits ( mp , XFS_DQTYPE_PROJ ) ;
2005-04-17 02:20:36 +04:00
2021-08-06 21:05:37 +03:00
if ( XFS_IS_UQUOTA_ON ( mp ) )
2020-07-16 03:42:36 +03:00
xfs_qm_set_defquota ( mp , XFS_DQTYPE_USER , qinf ) ;
2021-08-06 21:05:37 +03:00
if ( XFS_IS_GQUOTA_ON ( mp ) )
2020-07-16 03:42:36 +03:00
xfs_qm_set_defquota ( mp , XFS_DQTYPE_GROUP , qinf ) ;
2021-08-06 21:05:37 +03:00
if ( XFS_IS_PQUOTA_ON ( mp ) )
2020-07-16 03:42:36 +03:00
xfs_qm_set_defquota ( mp , XFS_DQTYPE_PROJ , qinf ) ;
2016-02-08 03:27:55 +03:00
2023-09-11 12:44:35 +03:00
qinf - > qi_shrinker = shrinker_alloc ( SHRINKER_NUMA_AWARE , " xfs-qm:%s " ,
mp - > m_super - > s_id ) ;
if ( ! qinf - > qi_shrinker ) {
error = - ENOMEM ;
2017-12-22 00:18:26 +03:00
goto out_free_inos ;
2023-09-11 12:44:35 +03:00
}
qinf - > qi_shrinker - > count_objects = xfs_qm_shrink_count ;
qinf - > qi_shrinker - > scan_objects = xfs_qm_shrink_scan ;
qinf - > qi_shrinker - > private_data = qinf ;
shrinker_register ( qinf - > qi_shrinker ) ;
2017-12-22 00:18:26 +03:00
2006-01-15 04:37:08 +03:00
return 0 ;
2014-05-15 03:23:24 +04:00
2017-12-22 00:18:26 +03:00
out_free_inos :
mutex_destroy ( & qinf - > qi_quotaofflock ) ;
mutex_destroy ( & qinf - > qi_tree_lock ) ;
xfs_qm_destroy_quotainos ( qinf ) ;
2014-05-15 03:23:24 +04:00
out_free_lru :
list_lru_destroy ( & qinf - > qi_lru ) ;
out_free_qinf :
kmem_free ( qinf ) ;
mp - > m_quotainfo = NULL ;
return error ;
2005-04-17 02:20:36 +04:00
}
/*
* Gets called when unmounting a filesystem or when all quotas get
* turned off .
* This purges the quota inodes , destroys locks and frees itself .
*/
void
xfs_qm_destroy_quotainfo (
2019-11-13 04:04:26 +03:00
struct xfs_mount * mp )
2005-04-17 02:20:36 +04:00
{
2019-11-13 04:04:26 +03:00
struct xfs_quotainfo * qi ;
2005-04-17 02:20:36 +04:00
qi = mp - > m_quotainfo ;
ASSERT ( qi ! = NULL ) ;
2023-09-11 12:44:35 +03:00
shrinker_free ( qi - > qi_shrinker ) ;
2013-08-28 04:18:18 +04:00
list_lru_destroy ( & qi - > qi_lru ) ;
2017-12-22 00:18:26 +03:00
xfs_qm_destroy_quotainos ( qi ) ;
2017-12-22 00:18:26 +03:00
mutex_destroy ( & qi - > qi_tree_lock ) ;
2005-04-17 02:20:36 +04:00
mutex_destroy ( & qi - > qi_quotaofflock ) ;
2008-05-19 10:31:57 +04:00
kmem_free ( qi ) ;
2005-04-17 02:20:36 +04:00
mp - > m_quotainfo = NULL ;
}
/*
* Create an inode and return with a reference already taken , but unlocked
* This is how we create quota inodes
*/
STATIC int
xfs_qm_qino_alloc (
2020-12-09 21:05:15 +03:00
struct xfs_mount * mp ,
struct xfs_inode * * ipp ,
unsigned int flags )
2005-04-17 02:20:36 +04:00
{
2020-12-09 21:05:15 +03:00
struct xfs_trans * tp ;
int error ;
bool need_alloc = true ;
2005-04-17 02:20:36 +04:00
2020-12-09 21:05:15 +03:00
* ipp = NULL ;
2013-07-20 02:36:02 +04:00
/*
* With superblock that doesn ' t have separate pquotino , we
* share an inode between gquota and pquota . If the on - disk
* superblock has GQUOTA and the filesystem is now mounted
* with PQUOTA , just use sb_gquotino for sb_pquotino and
* vice - versa .
*/
2021-08-19 04:46:37 +03:00
if ( ! xfs_has_pquotino ( mp ) & &
2013-07-20 02:36:02 +04:00
( flags & ( XFS_QMOPT_PQUOTA | XFS_QMOPT_GQUOTA ) ) ) {
xfs_ino_t ino = NULLFSINO ;
if ( ( flags & XFS_QMOPT_PQUOTA ) & &
( mp - > m_sb . sb_gquotino ! = NULLFSINO ) ) {
ino = mp - > m_sb . sb_gquotino ;
2019-11-11 23:53:22 +03:00
if ( XFS_IS_CORRUPT ( mp ,
mp - > m_sb . sb_pquotino ! = NULLFSINO ) )
2019-11-02 19:40:53 +03:00
return - EFSCORRUPTED ;
2013-07-20 02:36:02 +04:00
} else if ( ( flags & XFS_QMOPT_GQUOTA ) & &
( mp - > m_sb . sb_pquotino ! = NULLFSINO ) ) {
ino = mp - > m_sb . sb_pquotino ;
2019-11-11 23:53:22 +03:00
if ( XFS_IS_CORRUPT ( mp ,
mp - > m_sb . sb_gquotino ! = NULLFSINO ) )
2019-11-02 19:40:53 +03:00
return - EFSCORRUPTED ;
2013-07-20 02:36:02 +04:00
}
if ( ino ! = NULLFSINO ) {
2020-12-09 21:05:15 +03:00
error = xfs_iget ( mp , NULL , ino , 0 , 0 , ipp ) ;
2013-07-20 02:36:02 +04:00
if ( error )
return error ;
mp - > m_sb . sb_gquotino = NULLFSINO ;
mp - > m_sb . sb_pquotino = NULLFSINO ;
2015-02-23 14:38:08 +03:00
need_alloc = false ;
2013-07-20 02:36:02 +04:00
}
}
2016-04-06 02:19:55 +03:00
error = xfs_trans_alloc ( mp , & M_RES ( mp ) - > tr_create ,
2020-04-23 07:54:28 +03:00
need_alloc ? XFS_QM_QINOCREATE_SPACE_RES ( mp ) : 0 ,
0 , 0 , & tp ) ;
2016-04-06 02:19:55 +03:00
if ( error )
2006-01-15 04:37:08 +03:00
return error ;
2005-04-17 02:20:36 +04:00
2015-02-23 14:38:08 +03:00
if ( need_alloc ) {
2021-06-02 03:48:24 +03:00
xfs_ino_t ino ;
error = xfs_dialloc ( & tp , 0 , S_IFREG , & ino ) ;
if ( ! error )
2023-01-13 14:49:25 +03:00
error = xfs_init_new_inode ( & nop_mnt_idmap , tp , NULL , ino ,
2021-06-02 03:48:24 +03:00
S_IFREG , 1 , 0 , 0 , false , ipp ) ;
2013-07-20 02:36:02 +04:00
if ( error ) {
2015-06-04 06:47:56 +03:00
xfs_trans_cancel ( tp ) ;
2013-07-20 02:36:02 +04:00
return error ;
}
2005-04-17 02:20:36 +04:00
}
/*
* Make the changes in the superblock , and log those too .
* sbfields arg may contain fields other than * QUOTINO ;
* VERSIONNUM for example .
*/
2007-10-11 11:42:32 +04:00
spin_lock ( & mp - > m_sb_lock ) ;
2005-04-17 02:20:36 +04:00
if ( flags & XFS_QMOPT_SBVERSION ) {
2021-08-19 04:46:37 +03:00
ASSERT ( ! xfs_has_quota ( mp ) ) ;
2005-04-17 02:20:36 +04:00
2021-08-19 04:46:37 +03:00
xfs_add_quota ( mp ) ;
2005-04-17 02:20:36 +04:00
mp - > m_sb . sb_uquotino = NULLFSINO ;
mp - > m_sb . sb_gquotino = NULLFSINO ;
2013-07-20 02:36:02 +04:00
mp - > m_sb . sb_pquotino = NULLFSINO ;
2005-04-17 02:20:36 +04:00
2013-07-20 02:36:02 +04:00
/* qflags will get updated fully _after_ quotacheck */
mp - > m_sb . sb_qflags = mp - > m_qflags & XFS_ALL_QUOTA_ACCT ;
2005-04-17 02:20:36 +04:00
}
if ( flags & XFS_QMOPT_UQUOTA )
2020-12-09 21:05:15 +03:00
mp - > m_sb . sb_uquotino = ( * ipp ) - > i_ino ;
2013-07-20 02:36:02 +04:00
else if ( flags & XFS_QMOPT_GQUOTA )
2020-12-09 21:05:15 +03:00
mp - > m_sb . sb_gquotino = ( * ipp ) - > i_ino ;
2013-07-20 02:36:02 +04:00
else
2020-12-09 21:05:15 +03:00
mp - > m_sb . sb_pquotino = ( * ipp ) - > i_ino ;
2007-10-11 11:42:32 +04:00
spin_unlock ( & mp - > m_sb_lock ) ;
2015-01-22 01:10:31 +03:00
xfs_log_sb ( tp ) ;
2005-04-17 02:20:36 +04:00
2015-06-04 06:48:08 +03:00
error = xfs_trans_commit ( tp ) ;
2015-02-23 14:38:08 +03:00
if ( error ) {
2021-08-19 04:46:53 +03:00
ASSERT ( xfs_is_shutdown ( mp ) ) ;
2011-03-07 02:05:35 +03:00
xfs_alert ( mp , " %s failed (error %d)! " , __func__ , error ) ;
2005-04-17 02:20:36 +04:00
}
2015-02-23 14:38:08 +03:00
if ( need_alloc )
2020-12-09 21:05:15 +03:00
xfs_finish_inode_setup ( * ipp ) ;
2015-02-23 14:38:08 +03:00
return error ;
2005-04-17 02:20:36 +04:00
}
2008-04-10 06:20:10 +04:00
STATIC void
2005-04-17 02:20:36 +04:00
xfs_qm_reset_dqcounts (
2020-07-16 03:53:43 +03:00
struct xfs_mount * mp ,
struct xfs_buf * bp ,
xfs_dqid_t id ,
xfs_dqtype_t type )
2005-04-17 02:20:36 +04:00
{
xfs: rework dquot CRCs
Calculating dquot CRCs when the backing buffer is written back just
doesn't work reliably. There are several places which manipulate
dquots directly in the buffers, and they don't calculate CRCs
appropriately, nor do they always set the buffer up to calculate
CRCs appropriately.
Firstly, if we log a dquot buffer (e.g. during allocation) it gets
logged without valid CRC, and so on recovery we end up with a dquot
that is not valid.
Secondly, if we recover/repair a dquot, we don't have a verifier
attached to the buffer and hence CRCs are not calculated on the way
down to disk.
Thirdly, calculating the CRC after we've changed the contents means
that if we re-read the dquot from the buffer, we cannot verify the
contents of the dquot are valid, as the CRC is invalid.
So, to avoid all the dquot CRC errors that are being detected by the
read verifier, change to using the same model as for inodes. That
is, dquot CRCs are calculated and written to the backing buffer at
the time the dquot is flushed to the backing buffer. If we modify
the dquot directly in the backing buffer, calculate the CRC
immediately after the modification is complete. Hence the dquot in
the on-disk buffer should always have a valid CRC.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Ben Myers <bpm@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2013-06-03 09:28:46 +04:00
struct xfs_dqblk * dqb ;
2005-04-17 02:20:36 +04:00
int j ;
2009-12-15 02:14:59 +03:00
trace_xfs_reset_dqcounts ( bp , _RET_IP_ ) ;
2005-04-17 02:20:36 +04:00
/*
* Reset all counters and timers . They ' ll be
* started afresh by xfs_qm_quotacheck .
*/
# ifdef DEBUG
2017-04-19 22:55:57 +03:00
j = ( int ) XFS_FSB_TO_B ( mp , XFS_DQUOT_CLUSTER_SIZE_FSB ) /
2021-10-12 02:12:37 +03:00
sizeof ( struct xfs_dqblk ) ;
2010-04-20 11:01:30 +04:00
ASSERT ( mp - > m_quotainfo - > qi_dqperchunk = = j ) ;
2005-04-17 02:20:36 +04:00
# endif
xfs: rework dquot CRCs
Calculating dquot CRCs when the backing buffer is written back just
doesn't work reliably. There are several places which manipulate
dquots directly in the buffers, and they don't calculate CRCs
appropriately, nor do they always set the buffer up to calculate
CRCs appropriately.
Firstly, if we log a dquot buffer (e.g. during allocation) it gets
logged without valid CRC, and so on recovery we end up with a dquot
that is not valid.
Secondly, if we recover/repair a dquot, we don't have a verifier
attached to the buffer and hence CRCs are not calculated on the way
down to disk.
Thirdly, calculating the CRC after we've changed the contents means
that if we re-read the dquot from the buffer, we cannot verify the
contents of the dquot are valid, as the CRC is invalid.
So, to avoid all the dquot CRC errors that are being detected by the
read verifier, change to using the same model as for inodes. That
is, dquot CRCs are calculated and written to the backing buffer at
the time the dquot is flushed to the backing buffer. If we modify
the dquot directly in the backing buffer, calculate the CRC
immediately after the modification is complete. Hence the dquot in
the on-disk buffer should always have a valid CRC.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Ben Myers <bpm@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2013-06-03 09:28:46 +04:00
dqb = bp - > b_addr ;
2010-04-20 11:01:30 +04:00
for ( j = 0 ; j < mp - > m_quotainfo - > qi_dqperchunk ; j + + ) {
xfs: rework dquot CRCs
Calculating dquot CRCs when the backing buffer is written back just
doesn't work reliably. There are several places which manipulate
dquots directly in the buffers, and they don't calculate CRCs
appropriately, nor do they always set the buffer up to calculate
CRCs appropriately.
Firstly, if we log a dquot buffer (e.g. during allocation) it gets
logged without valid CRC, and so on recovery we end up with a dquot
that is not valid.
Secondly, if we recover/repair a dquot, we don't have a verifier
attached to the buffer and hence CRCs are not calculated on the way
down to disk.
Thirdly, calculating the CRC after we've changed the contents means
that if we re-read the dquot from the buffer, we cannot verify the
contents of the dquot are valid, as the CRC is invalid.
So, to avoid all the dquot CRC errors that are being detected by the
read verifier, change to using the same model as for inodes. That
is, dquot CRCs are calculated and written to the backing buffer at
the time the dquot is flushed to the backing buffer. If we modify
the dquot directly in the backing buffer, calculate the CRC
immediately after the modification is complete. Hence the dquot in
the on-disk buffer should always have a valid CRC.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Ben Myers <bpm@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2013-06-03 09:28:46 +04:00
struct xfs_disk_dquot * ddq ;
ddq = ( struct xfs_disk_dquot * ) & dqb [ j ] ;
2005-04-17 02:20:36 +04:00
/*
* Do a sanity check , and if needed , repair the dqblk . Don ' t
* output any warnings because it ' s perfectly possible to
2018-01-08 21:51:25 +03:00
* find uninitialised dquot blks . See comment in
* xfs_dquot_verify .
2005-04-17 02:20:36 +04:00
*/
2020-07-16 03:41:24 +03:00
if ( xfs_dqblk_verify ( mp , & dqb [ j ] , id + j ) | |
2020-07-16 04:05:39 +03:00
( dqb [ j ] . dd_diskdq . d_type & XFS_DQTYPE_REC_MASK ) ! = type )
2018-05-07 19:20:17 +03:00
xfs_dqblk_repair ( mp , & dqb [ j ] , id + j , type ) ;
2018-01-08 21:51:24 +03:00
2015-02-23 14:34:17 +03:00
/*
* Reset type in case we are reusing group quota file for
* project quotas or vice versa
*/
2020-07-16 04:05:39 +03:00
ddq - > d_type = type ;
2005-11-02 07:01:12 +03:00
ddq - > d_bcount = 0 ;
ddq - > d_icount = 0 ;
ddq - > d_rtbcount = 0 ;
2020-03-25 06:12:53 +03:00
/*
* dquot id 0 stores the default grace period and the maximum
* warning limit that were set by the administrator , so we
* should not reset them .
*/
if ( ddq - > d_id ! = 0 ) {
ddq - > d_btimer = 0 ;
ddq - > d_itimer = 0 ;
ddq - > d_rtbtimer = 0 ;
ddq - > d_bwarns = 0 ;
ddq - > d_iwarns = 0 ;
ddq - > d_rtbwarns = 0 ;
2021-08-19 04:46:37 +03:00
if ( xfs_has_bigtime ( mp ) )
2020-08-17 19:59:51 +03:00
ddq - > d_type | = XFS_DQTYPE_BIGTIME ;
2020-03-25 06:12:53 +03:00
}
xfs: rework dquot CRCs
Calculating dquot CRCs when the backing buffer is written back just
doesn't work reliably. There are several places which manipulate
dquots directly in the buffers, and they don't calculate CRCs
appropriately, nor do they always set the buffer up to calculate
CRCs appropriately.
Firstly, if we log a dquot buffer (e.g. during allocation) it gets
logged without valid CRC, and so on recovery we end up with a dquot
that is not valid.
Secondly, if we recover/repair a dquot, we don't have a verifier
attached to the buffer and hence CRCs are not calculated on the way
down to disk.
Thirdly, calculating the CRC after we've changed the contents means
that if we re-read the dquot from the buffer, we cannot verify the
contents of the dquot are valid, as the CRC is invalid.
So, to avoid all the dquot CRC errors that are being detected by the
read verifier, change to using the same model as for inodes. That
is, dquot CRCs are calculated and written to the backing buffer at
the time the dquot is flushed to the backing buffer. If we modify
the dquot directly in the backing buffer, calculate the CRC
immediately after the modification is complete. Hence the dquot in
the on-disk buffer should always have a valid CRC.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Ben Myers <bpm@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2013-06-03 09:28:46 +04:00
2021-08-19 04:46:37 +03:00
if ( xfs_has_crc ( mp ) ) {
xfs: rework dquot CRCs
Calculating dquot CRCs when the backing buffer is written back just
doesn't work reliably. There are several places which manipulate
dquots directly in the buffers, and they don't calculate CRCs
appropriately, nor do they always set the buffer up to calculate
CRCs appropriately.
Firstly, if we log a dquot buffer (e.g. during allocation) it gets
logged without valid CRC, and so on recovery we end up with a dquot
that is not valid.
Secondly, if we recover/repair a dquot, we don't have a verifier
attached to the buffer and hence CRCs are not calculated on the way
down to disk.
Thirdly, calculating the CRC after we've changed the contents means
that if we re-read the dquot from the buffer, we cannot verify the
contents of the dquot are valid, as the CRC is invalid.
So, to avoid all the dquot CRC errors that are being detected by the
read verifier, change to using the same model as for inodes. That
is, dquot CRCs are calculated and written to the backing buffer at
the time the dquot is flushed to the backing buffer. If we modify
the dquot directly in the backing buffer, calculate the CRC
immediately after the modification is complete. Hence the dquot in
the on-disk buffer should always have a valid CRC.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Ben Myers <bpm@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2013-06-03 09:28:46 +04:00
xfs_update_cksum ( ( char * ) & dqb [ j ] ,
sizeof ( struct xfs_dqblk ) ,
XFS_DQUOT_CRC_OFF ) ;
}
2005-04-17 02:20:36 +04:00
}
}
STATIC int
2018-05-05 01:31:20 +03:00
xfs_qm_reset_dqcounts_all (
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
struct xfs_mount * mp ,
xfs_dqid_t firstid ,
xfs_fsblock_t bno ,
xfs_filblks_t blkcnt ,
2020-07-16 03:53:43 +03:00
xfs_dqtype_t type ,
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
struct list_head * buffer_list )
2005-04-17 02:20:36 +04:00
{
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
struct xfs_buf * bp ;
2020-07-14 20:37:12 +03:00
int error = 0 ;
2005-04-17 02:20:36 +04:00
ASSERT ( blkcnt > 0 ) ;
/*
* Blkcnt arg can be a very big number , and might even be
* larger than the log itself . So , we have to break it up into
* manageable - sized transactions .
* Note that we don ' t start a permanent transaction here ; we might
* not be able to get a log reservation for the whole thing up front ,
* and we don ' t really care to either , because we just discard
* everything if we were to crash in the middle of this loop .
*/
while ( blkcnt - - ) {
error = xfs_trans_read_buf ( mp , NULL , mp - > m_ddev_targp ,
XFS_FSB_TO_DADDR ( mp , bno ) ,
2012-11-14 10:50:13 +04:00
mp - > m_quotainfo - > qi_dqchunklen , 0 , & bp ,
2012-11-14 10:54:40 +04:00
& xfs_dquot_buf_ops ) ;
2005-04-17 02:20:36 +04:00
2013-04-03 09:11:16 +04:00
/*
xfs: rework dquot CRCs
Calculating dquot CRCs when the backing buffer is written back just
doesn't work reliably. There are several places which manipulate
dquots directly in the buffers, and they don't calculate CRCs
appropriately, nor do they always set the buffer up to calculate
CRCs appropriately.
Firstly, if we log a dquot buffer (e.g. during allocation) it gets
logged without valid CRC, and so on recovery we end up with a dquot
that is not valid.
Secondly, if we recover/repair a dquot, we don't have a verifier
attached to the buffer and hence CRCs are not calculated on the way
down to disk.
Thirdly, calculating the CRC after we've changed the contents means
that if we re-read the dquot from the buffer, we cannot verify the
contents of the dquot are valid, as the CRC is invalid.
So, to avoid all the dquot CRC errors that are being detected by the
read verifier, change to using the same model as for inodes. That
is, dquot CRCs are calculated and written to the backing buffer at
the time the dquot is flushed to the backing buffer. If we modify
the dquot directly in the backing buffer, calculate the CRC
immediately after the modification is complete. Hence the dquot in
the on-disk buffer should always have a valid CRC.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Ben Myers <bpm@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2013-06-03 09:28:46 +04:00
* CRC and validation errors will return a EFSCORRUPTED here . If
* this occurs , re - read without CRC validation so that we can
* repair the damage via xfs_qm_reset_dqcounts ( ) . This process
* will leave a trace in the log indicating corruption has
* been detected .
2013-04-03 09:11:16 +04:00
*/
2014-06-25 08:58:08 +04:00
if ( error = = - EFSCORRUPTED ) {
xfs: rework dquot CRCs
Calculating dquot CRCs when the backing buffer is written back just
doesn't work reliably. There are several places which manipulate
dquots directly in the buffers, and they don't calculate CRCs
appropriately, nor do they always set the buffer up to calculate
CRCs appropriately.
Firstly, if we log a dquot buffer (e.g. during allocation) it gets
logged without valid CRC, and so on recovery we end up with a dquot
that is not valid.
Secondly, if we recover/repair a dquot, we don't have a verifier
attached to the buffer and hence CRCs are not calculated on the way
down to disk.
Thirdly, calculating the CRC after we've changed the contents means
that if we re-read the dquot from the buffer, we cannot verify the
contents of the dquot are valid, as the CRC is invalid.
So, to avoid all the dquot CRC errors that are being detected by the
read verifier, change to using the same model as for inodes. That
is, dquot CRCs are calculated and written to the backing buffer at
the time the dquot is flushed to the backing buffer. If we modify
the dquot directly in the backing buffer, calculate the CRC
immediately after the modification is complete. Hence the dquot in
the on-disk buffer should always have a valid CRC.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Ben Myers <bpm@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2013-06-03 09:28:46 +04:00
error = xfs_trans_read_buf ( mp , NULL , mp - > m_ddev_targp ,
XFS_FSB_TO_DADDR ( mp , bno ) ,
mp - > m_quotainfo - > qi_dqchunklen , 0 , & bp ,
NULL ) ;
}
if ( error )
break ;
2014-08-04 06:43:26 +04:00
/*
* A corrupt buffer might not have a verifier attached , so
* make sure we have the correct one attached before writeback
* occurs .
*/
bp - > b_ops = & xfs_dquot_buf_ops ;
2008-04-10 06:20:10 +04:00
xfs_qm_reset_dqcounts ( mp , bp , firstid , type ) ;
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
xfs_buf_delwri_queue ( bp , buffer_list ) ;
2011-08-23 12:28:06 +04:00
xfs_buf_relse ( bp ) ;
xfs: rework dquot CRCs
Calculating dquot CRCs when the backing buffer is written back just
doesn't work reliably. There are several places which manipulate
dquots directly in the buffers, and they don't calculate CRCs
appropriately, nor do they always set the buffer up to calculate
CRCs appropriately.
Firstly, if we log a dquot buffer (e.g. during allocation) it gets
logged without valid CRC, and so on recovery we end up with a dquot
that is not valid.
Secondly, if we recover/repair a dquot, we don't have a verifier
attached to the buffer and hence CRCs are not calculated on the way
down to disk.
Thirdly, calculating the CRC after we've changed the contents means
that if we re-read the dquot from the buffer, we cannot verify the
contents of the dquot are valid, as the CRC is invalid.
So, to avoid all the dquot CRC errors that are being detected by the
read verifier, change to using the same model as for inodes. That
is, dquot CRCs are calculated and written to the backing buffer at
the time the dquot is flushed to the backing buffer. If we modify
the dquot directly in the backing buffer, calculate the CRC
immediately after the modification is complete. Hence the dquot in
the on-disk buffer should always have a valid CRC.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Ben Myers <bpm@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2013-06-03 09:28:46 +04:00
/* goto the next block. */
2005-04-17 02:20:36 +04:00
bno + + ;
2010-04-20 11:01:30 +04:00
firstid + = mp - > m_quotainfo - > qi_dqperchunk ;
2005-04-17 02:20:36 +04:00
}
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
2006-01-15 04:37:08 +03:00
return error ;
2005-04-17 02:20:36 +04:00
}
/*
2018-05-05 01:31:20 +03:00
* Iterate over all allocated dquot blocks in this quota inode , zeroing all
* counters for every chunk of dquots that we find .
2005-04-17 02:20:36 +04:00
*/
STATIC int
2018-05-05 01:31:20 +03:00
xfs_qm_reset_dqcounts_buf (
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
struct xfs_mount * mp ,
struct xfs_inode * qip ,
2020-07-16 03:53:43 +03:00
xfs_dqtype_t type ,
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
struct list_head * buffer_list )
2005-04-17 02:20:36 +04:00
{
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
struct xfs_bmbt_irec * map ;
2005-04-17 02:20:36 +04:00
int i , nmaps ; /* number of map entries */
int error ; /* return value */
xfs_fileoff_t lblkno ;
xfs_filblks_t maxlblkcnt ;
xfs_dqid_t firstid ;
xfs_fsblock_t rablkno ;
xfs_filblks_t rablkcnt ;
error = 0 ;
/*
2006-03-29 02:55:14 +04:00
* This looks racy , but we can ' t keep an inode lock across a
2005-04-17 02:20:36 +04:00
* trans_reserve . But , this gets called during quotacheck , and that
* happens only at mount time which is single threaded .
*/
2021-03-29 21:11:40 +03:00
if ( qip - > i_nblocks = = 0 )
2006-01-15 04:37:08 +03:00
return 0 ;
2005-04-17 02:20:36 +04:00
2019-08-26 22:06:22 +03:00
map = kmem_alloc ( XFS_DQITER_MAP_SIZE * sizeof ( * map ) , 0 ) ;
2005-04-17 02:20:36 +04:00
lblkno = 0 ;
2012-06-08 09:44:54 +04:00
maxlblkcnt = XFS_B_TO_FSB ( mp , mp - > m_super - > s_maxbytes ) ;
2005-04-17 02:20:36 +04:00
do {
2013-12-07 00:30:14 +04:00
uint lock_mode ;
2005-04-17 02:20:36 +04:00
nmaps = XFS_DQITER_MAP_SIZE ;
/*
* We aren ' t changing the inode itself . Just changing
* some of its data . No new blocks are added here , and
* the inode is never added to the transaction .
*/
2013-12-07 00:30:14 +04:00
lock_mode = xfs_ilock_data_map_shared ( qip ) ;
2011-09-19 00:40:45 +04:00
error = xfs_bmapi_read ( qip , lblkno , maxlblkcnt - lblkno ,
map , & nmaps , 0 ) ;
2013-12-07 00:30:14 +04:00
xfs_iunlock ( qip , lock_mode ) ;
2005-04-17 02:20:36 +04:00
if ( error )
break ;
ASSERT ( nmaps < = XFS_DQITER_MAP_SIZE ) ;
for ( i = 0 ; i < nmaps ; i + + ) {
ASSERT ( map [ i ] . br_startblock ! = DELAYSTARTBLOCK ) ;
ASSERT ( map [ i ] . br_blockcount ) ;
lblkno + = map [ i ] . br_blockcount ;
if ( map [ i ] . br_startblock = = HOLESTARTBLOCK )
continue ;
firstid = ( xfs_dqid_t ) map [ i ] . br_startoff *
2010-04-20 11:01:30 +04:00
mp - > m_quotainfo - > qi_dqperchunk ;
2005-04-17 02:20:36 +04:00
/*
* Do a read - ahead on the next extent .
*/
if ( ( i + 1 < nmaps ) & &
( map [ i + 1 ] . br_startblock ! = HOLESTARTBLOCK ) ) {
rablkcnt = map [ i + 1 ] . br_blockcount ;
rablkno = map [ i + 1 ] . br_startblock ;
while ( rablkcnt - - ) {
2010-10-06 22:41:18 +04:00
xfs_buf_readahead ( mp - > m_ddev_targp ,
2005-04-17 02:20:36 +04:00
XFS_FSB_TO_DADDR ( mp , rablkno ) ,
2012-11-12 15:54:01 +04:00
mp - > m_quotainfo - > qi_dqchunklen ,
2014-08-04 06:43:26 +04:00
& xfs_dquot_buf_ops ) ;
2005-04-17 02:20:36 +04:00
rablkno + + ;
}
}
/*
* Iterate thru all the blks in the extent and
* reset the counters of all the dquots inside them .
*/
2018-05-05 01:31:20 +03:00
error = xfs_qm_reset_dqcounts_all ( mp , firstid ,
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
map [ i ] . br_startblock ,
map [ i ] . br_blockcount ,
2020-07-14 20:37:12 +03:00
type , buffer_list ) ;
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
if ( error )
goto out ;
2005-04-17 02:20:36 +04:00
}
} while ( nmaps > 0 ) ;
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
out :
2008-05-19 10:31:57 +04:00
kmem_free ( map ) ;
2006-01-15 04:37:08 +03:00
return error ;
2005-04-17 02:20:36 +04:00
}
/*
* Called by dqusage_adjust in doing a quotacheck .
2010-09-06 05:44:22 +04:00
*
* Given the inode , and a dquot id this updates both the incore dqout as well
* as the buffer copy . This is so that once the quotacheck is done , we can
* just log all the buffers , as opposed to logging numerous updates to
* individual dquots .
2005-04-17 02:20:36 +04:00
*/
2010-09-06 05:44:22 +04:00
STATIC int
2005-04-17 02:20:36 +04:00
xfs_qm_quotacheck_dqadjust (
2010-09-06 05:44:22 +04:00
struct xfs_inode * ip ,
2020-07-16 03:53:43 +03:00
xfs_dqtype_t type ,
2005-04-17 02:20:36 +04:00
xfs_qcnt_t nblks ,
xfs_qcnt_t rtblks )
{
2010-09-06 05:44:22 +04:00
struct xfs_mount * mp = ip - > i_mount ;
struct xfs_dquot * dqp ;
2018-05-05 01:30:22 +03:00
xfs_dqid_t id ;
2010-09-06 05:44:22 +04:00
int error ;
2018-05-05 01:30:22 +03:00
id = xfs_qm_id_for_quotatype ( ip , type ) ;
2018-05-05 01:30:24 +03:00
error = xfs_qm_dqget ( mp , id , type , true , & dqp ) ;
2010-09-06 05:44:22 +04:00
if ( error ) {
/*
* Shouldn ' t be able to turn off quotas here .
*/
2014-06-25 08:58:08 +04:00
ASSERT ( error ! = - ESRCH ) ;
ASSERT ( error ! = - ENOENT ) ;
2010-09-06 05:44:22 +04:00
return error ;
}
2009-12-15 02:14:59 +03:00
trace_xfs_dqadjust ( dqp ) ;
2005-04-17 02:20:36 +04:00
/*
* Adjust the inode count and the block count to reflect this inode ' s
* resource usage .
*/
2020-07-14 20:37:31 +03:00
dqp - > q_ino . count + + ;
2020-07-14 20:37:30 +03:00
dqp - > q_ino . reserved + + ;
2005-04-17 02:20:36 +04:00
if ( nblks ) {
2020-07-14 20:37:31 +03:00
dqp - > q_blk . count + = nblks ;
2020-07-14 20:37:30 +03:00
dqp - > q_blk . reserved + = nblks ;
2005-04-17 02:20:36 +04:00
}
if ( rtblks ) {
2020-07-14 20:37:31 +03:00
dqp - > q_rtb . count + = rtblks ;
2020-07-14 20:37:30 +03:00
dqp - > q_rtb . reserved + = rtblks ;
2005-04-17 02:20:36 +04:00
}
/*
* Set default limits , adjust timers ( since we changed usages )
2010-04-20 11:01:53 +04:00
*
* There are no timers for the default values set in the root dquot .
2005-04-17 02:20:36 +04:00
*/
2020-07-14 20:37:30 +03:00
if ( dqp - > q_id ) {
2020-07-14 20:37:33 +03:00
xfs_qm_adjust_dqlimits ( dqp ) ;
xfs_qm_adjust_dqtimers ( dqp ) ;
2005-04-17 02:20:36 +04:00
}
2020-07-14 20:37:13 +03:00
dqp - > q_flags | = XFS_DQFLAG_DIRTY ;
2010-09-06 05:44:22 +04:00
xfs_qm_dqput ( dqp ) ;
return 0 ;
2005-04-17 02:20:36 +04:00
}
/*
* callback routine supplied to bulkstat ( ) . Given an inumber , find its
* dquots and update them to account for resources taken by that inode .
*/
/* ARGSUSED */
STATIC int
xfs_qm_dqusage_adjust (
2019-07-02 19:39:39 +03:00
struct xfs_mount * mp ,
struct xfs_trans * tp ,
xfs_ino_t ino ,
void * data )
2005-04-17 02:20:36 +04:00
{
2019-07-02 19:39:39 +03:00
struct xfs_inode * ip ;
xfs_qcnt_t nblks ;
xfs_filblks_t rtblks = 0 ; /* total rt blks */
int error ;
2005-04-17 02:20:36 +04:00
2021-08-06 21:05:37 +03:00
ASSERT ( XFS_IS_QUOTA_ON ( mp ) ) ;
2005-04-17 02:20:36 +04:00
/*
* rootino must have its resources accounted for , not so with the quota
* inodes .
*/
2019-07-02 19:39:39 +03:00
if ( xfs_is_quota_inode ( & mp - > m_sb , ino ) )
return 0 ;
2005-04-17 02:20:36 +04:00
/*
2018-05-05 01:30:22 +03:00
* We don ' t _need_ to take the ilock EXCL here because quotacheck runs
* at mount time and therefore nobody will be racing chown / chproj .
2005-04-17 02:20:36 +04:00
*/
2019-07-02 19:39:39 +03:00
error = xfs_iget ( mp , tp , ino , XFS_IGET_DONTCACHE , 0 , & ip ) ;
if ( error = = - EINVAL | | error = = - ENOENT )
return 0 ;
if ( error )
2006-01-15 04:37:08 +03:00
return error ;
2005-04-17 02:20:36 +04:00
2023-09-24 18:35:53 +03:00
/*
* Reload the incore unlinked list to avoid failure in inodegc .
* Use an unlocked check here because unrecovered unlinked inodes
* should be somewhat rare .
*/
if ( xfs_inode_unlinked_incomplete ( ip ) ) {
error = xfs_inode_reload_unlinked ( ip ) ;
if ( error ) {
xfs_force_shutdown ( mp , SHUTDOWN_CORRUPT_INCORE ) ;
goto error0 ;
}
}
2023-09-11 18:39:08 +03:00
2010-09-06 05:44:22 +04:00
ASSERT ( ip - > i_delayed_blks = = 0 ) ;
2005-04-17 02:20:36 +04:00
2010-09-06 05:44:22 +04:00
if ( XFS_IS_REALTIME_INODE ( ip ) ) {
2022-07-09 20:56:05 +03:00
struct xfs_ifork * ifp = xfs_ifork_ptr ( ip , XFS_DATA_FORK ) ;
2017-08-30 01:44:14 +03:00
2021-04-13 21:15:09 +03:00
error = xfs_iread_extents ( tp , ip , XFS_DATA_FORK ) ;
if ( error )
goto error0 ;
2017-08-30 01:44:14 +03:00
xfs_bmap_count_leaves ( ifp , & rtblks ) ;
2005-04-17 02:20:36 +04:00
}
2021-03-29 21:11:40 +03:00
nblks = ( xfs_qcnt_t ) ip - > i_nblocks - rtblks ;
2023-09-11 18:39:08 +03:00
xfs_iflags_clear ( ip , XFS_IQUOTAUNCHECKED ) ;
2005-04-17 02:20:36 +04:00
/*
* Add the ( disk blocks and inode ) resources occupied by this
* inode to its dquots . We do this adjustment in the incore dquot ,
* and also copy the changes to its buffer .
* We don ' t care about putting these changes in a transaction
* envelope because if we crash in the middle of a ' quotacheck '
* we have to start from the beginning anyway .
* Once we ' re done , we ' ll log all the dquot bufs .
*
2006-03-29 02:55:14 +04:00
* The * QUOTA_ON checks below may look pretty racy , but quotachecks
2005-04-17 02:20:36 +04:00
* and quotaoffs don ' t race . ( Quotachecks happen at mount time only ) .
*/
if ( XFS_IS_UQUOTA_ON ( mp ) ) {
2020-07-16 03:42:36 +03:00
error = xfs_qm_quotacheck_dqadjust ( ip , XFS_DQTYPE_USER , nblks ,
2018-05-05 01:30:22 +03:00
rtblks ) ;
2010-09-06 05:44:22 +04:00
if ( error )
goto error0 ;
2005-04-17 02:20:36 +04:00
}
2010-09-06 05:44:22 +04:00
if ( XFS_IS_GQUOTA_ON ( mp ) ) {
2020-07-16 03:42:36 +03:00
error = xfs_qm_quotacheck_dqadjust ( ip , XFS_DQTYPE_GROUP , nblks ,
2018-05-05 01:30:22 +03:00
rtblks ) ;
2010-09-06 05:44:22 +04:00
if ( error )
goto error0 ;
2005-04-17 02:20:36 +04:00
}
2010-09-06 05:44:22 +04:00
if ( XFS_IS_PQUOTA_ON ( mp ) ) {
2020-07-16 03:42:36 +03:00
error = xfs_qm_quotacheck_dqadjust ( ip , XFS_DQTYPE_PROJ , nblks ,
2018-05-05 01:30:22 +03:00
rtblks ) ;
2010-09-06 05:44:22 +04:00
if ( error )
goto error0 ;
}
error0 :
2018-07-25 22:52:32 +03:00
xfs_irele ( ip ) ;
2010-09-06 05:44:22 +04:00
return error ;
2005-04-17 02:20:36 +04:00
}
2012-03-14 20:53:34 +04:00
STATIC int
xfs_qm_flush_one (
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
struct xfs_dquot * dqp ,
void * data )
2012-03-14 20:53:34 +04:00
{
2017-06-15 07:21:45 +03:00
struct xfs_mount * mp = dqp - > q_mount ;
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
struct list_head * buffer_list = data ;
2012-04-23 09:58:37 +04:00
struct xfs_buf * bp = NULL ;
2012-03-14 20:53:34 +04:00
int error = 0 ;
xfs_dqlock ( dqp ) ;
2020-07-14 20:37:13 +03:00
if ( dqp - > q_flags & XFS_DQFLAG_FREEING )
2012-03-14 20:53:34 +04:00
goto out_unlock ;
if ( ! XFS_DQ_IS_DIRTY ( dqp ) )
goto out_unlock ;
2017-06-15 07:21:45 +03:00
/*
* The only way the dquot is already flush locked by the time quotacheck
* gets here is if reclaim flushed it before the dqadjust walk dirtied
* it for the final time . Quotacheck collects all dquot bufs in the
* local delwri queue before dquots are dirtied , so reclaim can ' t have
* possibly queued it for I / O . The only way out is to push the buffer to
* cycle the flush lock .
*/
if ( ! xfs_dqflock_nowait ( dqp ) ) {
/* buf is pinned in-core by delwri list */
2022-07-07 15:05:18 +03:00
error = xfs_buf_incore ( mp - > m_ddev_targp , dqp - > q_blkno ,
mp - > m_quotainfo - > qi_dqchunklen , 0 , & bp ) ;
if ( error )
2017-06-15 07:21:45 +03:00
goto out_unlock ;
2022-07-07 15:05:18 +03:00
xfs: fix intermittent hang during quotacheck
Every now and then, I see the following hang during mount time
quotacheck when running fstests. Turning on KASAN seems to make it
happen somewhat more frequently. I've edited the backtrace for brevity.
XFS (sdd): Quotacheck needed: Please wait.
XFS: Assertion failed: bp->b_flags & _XBF_DELWRI_Q, file: fs/xfs/xfs_buf.c, line: 2411
------------[ cut here ]------------
WARNING: CPU: 0 PID: 1831409 at fs/xfs/xfs_message.c:104 assfail+0x46/0x4a [xfs]
CPU: 0 PID: 1831409 Comm: mount Tainted: G W 5.19.0-rc6-xfsx #rc6 09911566947b9f737b036b4af85e399e4b9aef64
Hardware name: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.15.0-1 04/01/2014
RIP: 0010:assfail+0x46/0x4a [xfs]
Code: a0 8f 41 a0 e8 45 fe ff ff 8a 1d 2c 36 10 00 80 fb 01 76 0f 0f b6 f3 48 c7 c7 c0 f0 4f a0 e8 10 f0 02 e1 80 e3 01 74 02 0f 0b <0f> 0b 5b c3 48 8d 45 10 48 89 e2 4c 89 e6 48 89 1c 24 48 89 44 24
RSP: 0018:ffffc900078c7b30 EFLAGS: 00010246
RAX: 0000000000000000 RBX: ffff8880099ac000 RCX: 000000007fffffff
RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffffffffa0418fa0
RBP: ffff8880197bc1c0 R08: 0000000000000000 R09: 000000000000000a
R10: 000000000000000a R11: f000000000000000 R12: ffffc900078c7d20
R13: 00000000fffffff5 R14: ffffc900078c7d20 R15: 0000000000000000
FS: 00007f0449903800(0000) GS:ffff88803ec00000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: 00005610ada631f0 CR3: 0000000014dd8002 CR4: 00000000001706f0
Call Trace:
<TASK>
xfs_buf_delwri_pushbuf+0x150/0x160 [xfs 4561f5b32c9bfb874ec98d58d0719464e1f87368]
xfs_qm_flush_one+0xd6/0x130 [xfs 4561f5b32c9bfb874ec98d58d0719464e1f87368]
xfs_qm_dquot_walk.isra.0+0x109/0x1e0 [xfs 4561f5b32c9bfb874ec98d58d0719464e1f87368]
xfs_qm_quotacheck+0x319/0x490 [xfs 4561f5b32c9bfb874ec98d58d0719464e1f87368]
xfs_qm_mount_quotas+0x65/0x2c0 [xfs 4561f5b32c9bfb874ec98d58d0719464e1f87368]
xfs_mountfs+0x6b5/0xab0 [xfs 4561f5b32c9bfb874ec98d58d0719464e1f87368]
xfs_fs_fill_super+0x781/0x990 [xfs 4561f5b32c9bfb874ec98d58d0719464e1f87368]
get_tree_bdev+0x175/0x280
vfs_get_tree+0x1a/0x80
path_mount+0x6f5/0xaa0
__x64_sys_mount+0x103/0x140
do_syscall_64+0x2b/0x80
entry_SYSCALL_64_after_hwframe+0x46/0xb0
I /think/ this can happen if xfs_qm_flush_one is racing with
xfs_qm_dquot_isolate (i.e. dquot reclaim) when the second function has
taken the dquot flush lock but xfs_qm_dqflush hasn't yet locked the
dquot buffer, let alone queued it to the delwri list. In this case,
flush_one will fail to get the dquot flush lock, but it can lock the
incore buffer, but xfs_buf_delwri_pushbuf will then trip over this
ASSERT, which checks that the buffer isn't on a delwri list. The hang
results because the _delwri_submit_buffers ignores non DELWRI_Q buffers,
which means that xfs_buf_iowait waits forever for an IO that has not yet
been scheduled.
AFAICT, a reasonable solution here is to detect a dquot buffer that is
not on a DELWRI list, drop it, and return -EAGAIN to try the flush
again. It's not /that/ big of a deal if quotacheck writes the dquot
buffer repeatedly before we even set QUOTA_CHKD.
Signed-off-by: Darrick J. Wong <djwong@kernel.org>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
2022-08-04 03:33:00 +03:00
if ( ! ( bp - > b_flags & _XBF_DELWRI_Q ) ) {
error = - EAGAIN ;
xfs_buf_relse ( bp ) ;
goto out_unlock ;
}
2017-06-15 07:21:45 +03:00
xfs_buf_unlock ( bp ) ;
xfs_buf_delwri_pushbuf ( bp , buffer_list ) ;
xfs_buf_rele ( bp ) ;
error = - EAGAIN ;
goto out_unlock ;
}
2012-04-23 09:58:37 +04:00
error = xfs_qm_dqflush ( dqp , & bp ) ;
if ( error )
goto out_unlock ;
2012-03-14 20:53:34 +04:00
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
xfs_buf_delwri_queue ( bp , buffer_list ) ;
2012-04-23 09:58:37 +04:00
xfs_buf_relse ( bp ) ;
2012-03-14 20:53:34 +04:00
out_unlock :
xfs_dqunlock ( dqp ) ;
return error ;
}
2005-04-17 02:20:36 +04:00
/*
* Walk thru all the filesystem inodes and construct a consistent view
* of the disk quota world . If the quotacheck fails , disable quotas .
*/
2014-07-24 14:49:57 +04:00
STATIC int
2005-04-17 02:20:36 +04:00
xfs_qm_quotacheck (
xfs_mount_t * mp )
{
2019-07-02 19:39:39 +03:00
int error , error2 ;
2013-06-28 02:25:07 +04:00
uint flags ;
LIST_HEAD ( buffer_list ) ;
struct xfs_inode * uip = mp - > m_quotainfo - > qi_uquotaip ;
struct xfs_inode * gip = mp - > m_quotainfo - > qi_gquotaip ;
2013-07-11 09:00:40 +04:00
struct xfs_inode * pip = mp - > m_quotainfo - > qi_pquotaip ;
2005-04-17 02:20:36 +04:00
flags = 0 ;
2013-07-11 09:00:40 +04:00
ASSERT ( uip | | gip | | pip ) ;
2021-08-06 21:05:37 +03:00
ASSERT ( XFS_IS_QUOTA_ON ( mp ) ) ;
2005-04-17 02:20:36 +04:00
2011-03-07 02:08:35 +03:00
xfs_notice ( mp , " Quotacheck needed: Please wait. " ) ;
2005-04-17 02:20:36 +04:00
/*
2005-06-21 09:38:48 +04:00
* First we go thru all the dquots on disk , USR and GRP / PRJ , and reset
2005-04-17 02:20:36 +04:00
* their counters to zero . We need a clean slate .
* We don ' t log our changes till later .
*/
2010-04-20 11:01:30 +04:00
if ( uip ) {
2020-07-16 03:42:36 +03:00
error = xfs_qm_reset_dqcounts_buf ( mp , uip , XFS_DQTYPE_USER ,
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
& buffer_list ) ;
2010-04-20 11:01:30 +04:00
if ( error )
2005-04-17 02:20:36 +04:00
goto error_return ;
flags | = XFS_UQUOTA_CHKD ;
}
2010-04-20 11:01:30 +04:00
if ( gip ) {
2020-07-16 03:42:36 +03:00
error = xfs_qm_reset_dqcounts_buf ( mp , gip , XFS_DQTYPE_GROUP ,
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
& buffer_list ) ;
2010-04-20 11:01:30 +04:00
if ( error )
2005-04-17 02:20:36 +04:00
goto error_return ;
2013-07-11 09:00:40 +04:00
flags | = XFS_GQUOTA_CHKD ;
}
if ( pip ) {
2020-07-16 03:42:36 +03:00
error = xfs_qm_reset_dqcounts_buf ( mp , pip , XFS_DQTYPE_PROJ ,
2013-07-11 09:00:40 +04:00
& buffer_list ) ;
if ( error )
goto error_return ;
flags | = XFS_PQUOTA_CHKD ;
2005-04-17 02:20:36 +04:00
}
2023-09-11 18:39:08 +03:00
xfs_set_quotacheck_running ( mp ) ;
2019-07-04 06:36:28 +03:00
error = xfs_iwalk_threaded ( mp , 0 , 0 , xfs_qm_dqusage_adjust , 0 , true ,
NULL ) ;
2023-09-11 18:39:08 +03:00
xfs_clear_quotacheck_running ( mp ) ;
2023-03-06 02:13:22 +03:00
/*
* On error , the inode walk may have partially populated the dquot
* caches . We must purge them before disabling quota and tearing down
* the quotainfo , or else the dquots will leak .
*/
if ( error )
goto error_purge ;
2005-04-17 02:20:36 +04:00
2008-04-10 06:20:17 +04:00
/*
2012-03-14 20:53:34 +04:00
* We ' ve made all the changes that we need to make incore . Flush them
* down to disk buffers if everything was updated successfully .
2008-04-10 06:20:17 +04:00
*/
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
if ( XFS_IS_UQUOTA_ON ( mp ) ) {
2020-07-16 03:42:36 +03:00
error = xfs_qm_dquot_walk ( mp , XFS_DQTYPE_USER , xfs_qm_flush_one ,
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
& buffer_list ) ;
}
2012-03-14 20:53:34 +04:00
if ( XFS_IS_GQUOTA_ON ( mp ) ) {
2020-07-16 03:42:36 +03:00
error2 = xfs_qm_dquot_walk ( mp , XFS_DQTYPE_GROUP , xfs_qm_flush_one ,
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
& buffer_list ) ;
2012-03-14 20:53:34 +04:00
if ( ! error )
error = error2 ;
}
if ( XFS_IS_PQUOTA_ON ( mp ) ) {
2020-07-16 03:42:36 +03:00
error2 = xfs_qm_dquot_walk ( mp , XFS_DQTYPE_PROJ , xfs_qm_flush_one ,
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
& buffer_list ) ;
2012-03-14 20:53:34 +04:00
if ( ! error )
error = error2 ;
}
2008-04-10 06:20:17 +04:00
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
error2 = xfs_buf_delwri_submit ( & buffer_list ) ;
if ( ! error )
error = error2 ;
2005-04-17 02:20:36 +04:00
/*
* We can get this error if we couldn ' t do a dquot allocation inside
* xfs_qm_dqusage_adjust ( via bulkstat ) . We don ' t care about the
* dirty dquots that might be cached , we just want to get rid of them
* and turn quotaoff . The dquots won ' t be attached to any of the inodes
* at this point ( because we intentionally didn ' t in dqget_noattach ) .
*/
2023-03-06 02:13:22 +03:00
if ( error )
goto error_purge ;
2005-04-17 02:20:36 +04:00
/*
* If one type of quotas is off , then it will lose its
* quotachecked status , since we won ' t be doing accounting for
* that type anymore .
*/
2012-01-23 21:31:43 +04:00
mp - > m_qflags & = ~ XFS_ALL_QUOTA_CHKD ;
2005-04-17 02:20:36 +04:00
mp - > m_qflags | = flags ;
2023-03-06 02:13:22 +03:00
error_return :
2017-04-21 22:40:44 +03:00
xfs_buf_delwri_cancel ( & buffer_list ) ;
xfs: on-stack delayed write buffer lists
Queue delwri buffers on a local on-stack list instead of a per-buftarg one,
and write back the buffers per-process instead of by waking up xfsbufd.
This is now easily doable given that we have very few places left that write
delwri buffers:
- log recovery:
Only done at mount time, and already forcing out the buffers
synchronously using xfs_flush_buftarg
- quotacheck:
Same story.
- dquot reclaim:
Writes out dirty dquots on the LRU under memory pressure. We might
want to look into doing more of this via xfsaild, but it's already
more optimal than the synchronous inode reclaim that writes each
buffer synchronously.
- xfsaild:
This is the main beneficiary of the change. By keeping a local list
of buffers to write we reduce latency of writing out buffers, and
more importably we can remove all the delwri list promotions which
were hitting the buffer cache hard under sustained metadata loads.
The implementation is very straight forward - xfs_buf_delwri_queue now gets
a new list_head pointer that it adds the delwri buffers to, and all callers
need to eventually submit the list using xfs_buf_delwi_submit or
xfs_buf_delwi_submit_nowait. Buffers that already are on a delwri list are
skipped in xfs_buf_delwri_queue, assuming they already are on another delwri
list. The biggest change to pass down the buffer list was done to the AIL
pushing. Now that we operate on buffers the trylock, push and pushbuf log
item methods are merged into a single push routine, which tries to lock the
item, and if possible add the buffer that needs writeback to the buffer list.
This leads to much simpler code than the previous split but requires the
individual IOP_PUSH instances to unlock and reacquire the AIL around calls
to blocking routines.
Given that xfsailds now also handle writing out buffers, the conditions for
log forcing and the sleep times needed some small changes. The most
important one is that we consider an AIL busy as long we still have buffers
to push, and the other one is that we do increment the pushed LSN for
buffers that are under flushing at this moment, but still count them towards
the stuck items for restart purposes. Without this we could hammer on stuck
items without ever forcing the log and not make progress under heavy random
delete workloads on fast flash storage devices.
[ Dave Chinner:
- rebase on previous patches.
- improved comments for XBF_DELWRI_Q handling
- fix XBF_ASYNC handling in queue submission (test 106 failure)
- rename delwri submit function buffer list parameters for clarity
- xfs_efd_item_push() should return XFS_ITEM_PINNED ]
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-04-23 09:58:39 +04:00
2005-04-17 02:20:36 +04:00
if ( error ) {
2011-03-07 02:08:35 +03:00
xfs_warn ( mp ,
" Quotacheck: Unsuccessful (Error %d): Disabling quotas. " ,
error ) ;
2005-04-17 02:20:36 +04:00
/*
* We must turn off quotas .
*/
ASSERT ( mp - > m_quotainfo ! = NULL ) ;
xfs_qm_destroy_quotainfo ( mp ) ;
2008-04-10 06:20:38 +04:00
if ( xfs_mount_reset_sbqflags ( mp ) ) {
2011-03-07 02:08:35 +03:00
xfs_warn ( mp ,
" Quotacheck: Failed to reset quota flags. " ) ;
2008-04-10 06:20:38 +04:00
}
2011-03-07 02:08:35 +03:00
} else
xfs_notice ( mp , " Quotacheck: Done. " ) ;
2014-06-22 09:03:54 +04:00
return error ;
2023-03-06 02:13:22 +03:00
error_purge :
/*
* On error , we may have inodes queued for inactivation . This may try
* to attach dquots to the inode before running cleanup operations on
* the inode and this can race with the xfs_qm_destroy_quotainfo ( ) call
* below that frees mp - > m_quotainfo . To avoid this race , flush all the
* pending inodegc operations before we purge the dquots from memory ,
* ensuring that background inactivation is idle whilst we turn off
* quotas .
*/
xfs_inodegc_flush ( mp ) ;
xfs_qm_dqpurge_all ( mp ) ;
goto error_return ;
2005-04-17 02:20:36 +04:00
}
2014-07-24 14:49:57 +04:00
/*
* This is called from xfs_mountfs to start quotas and initialize all
* necessary data structures like quotainfo . This is also responsible for
* running a quotacheck as necessary . We are guaranteed that the superblock
* is consistently read in at this point .
*
* If we fail here , the mount will continue with quota turned off . We don ' t
* need to inidicate success or failure at all .
*/
void
xfs_qm_mount_quotas (
struct xfs_mount * mp )
{
int error = 0 ;
uint sbf ;
/*
* If quotas on realtime volumes is not supported , we disable
* quotas immediately .
*/
if ( mp - > m_sb . sb_rextents ) {
xfs_notice ( mp , " Cannot turn on quotas for realtime filesystem " ) ;
mp - > m_qflags = 0 ;
goto write_changes ;
}
2021-08-06 21:05:37 +03:00
ASSERT ( XFS_IS_QUOTA_ON ( mp ) ) ;
2014-07-24 14:49:57 +04:00
/*
* Allocate the quotainfo structure inside the mount struct , and
* create quotainode ( s ) , and change / rev superblock if necessary .
*/
error = xfs_qm_init_quotainfo ( mp ) ;
if ( error ) {
/*
* We must turn off quotas .
*/
ASSERT ( mp - > m_quotainfo = = NULL ) ;
mp - > m_qflags = 0 ;
goto write_changes ;
}
/*
* If any of the quotas are not consistent , do a quotacheck .
*/
if ( XFS_QM_NEED_QUOTACHECK ( mp ) ) {
error = xfs_qm_quotacheck ( mp ) ;
if ( error ) {
/* Quotacheck failed and disabled quotas. */
return ;
}
}
/*
* If one type of quotas is off , then it will lose its
* quotachecked status , since we won ' t be doing accounting for
* that type anymore .
*/
if ( ! XFS_IS_UQUOTA_ON ( mp ) )
mp - > m_qflags & = ~ XFS_UQUOTA_CHKD ;
if ( ! XFS_IS_GQUOTA_ON ( mp ) )
mp - > m_qflags & = ~ XFS_GQUOTA_CHKD ;
if ( ! XFS_IS_PQUOTA_ON ( mp ) )
mp - > m_qflags & = ~ XFS_PQUOTA_CHKD ;
write_changes :
/*
* We actually don ' t have to acquire the m_sb_lock at all .
* This can only be called from mount , and that ' s single threaded . XXX
*/
spin_lock ( & mp - > m_sb_lock ) ;
sbf = mp - > m_sb . sb_qflags ;
mp - > m_sb . sb_qflags = mp - > m_qflags & XFS_MOUNT_QUOTA_ALL ;
spin_unlock ( & mp - > m_sb_lock ) ;
if ( sbf ! = ( mp - > m_qflags & XFS_MOUNT_QUOTA_ALL ) ) {
2015-01-22 01:10:31 +03:00
if ( xfs_sync_sb ( mp , false ) ) {
2014-07-24 14:49:57 +04:00
/*
* We could only have been turning quotas off .
* We aren ' t in very good shape actually because
* the incore structures are convinced that quotas are
* off , but the on disk superblock doesn ' t know that !
*/
2021-08-06 21:05:37 +03:00
ASSERT ( ! ( XFS_IS_QUOTA_ON ( mp ) ) ) ;
2014-07-24 14:49:57 +04:00
xfs_alert ( mp , " %s: Superblock update failed! " ,
__func__ ) ;
}
}
if ( error ) {
xfs_warn ( mp , " Failed to initialize disk quotas. " ) ;
return ;
}
}
2005-04-17 02:20:36 +04:00
/*
* This is called after the superblock has been read in and we ' re ready to
* iget the quota inodes .
*/
STATIC int
xfs_qm_init_quotainos (
xfs_mount_t * mp )
{
2013-06-28 02:25:07 +04:00
struct xfs_inode * uip = NULL ;
struct xfs_inode * gip = NULL ;
2013-07-11 09:00:40 +04:00
struct xfs_inode * pip = NULL ;
2013-06-28 02:25:07 +04:00
int error ;
uint flags = 0 ;
2005-04-17 02:20:36 +04:00
ASSERT ( mp - > m_quotainfo ) ;
/*
* Get the uquota and gquota inodes
*/
2021-08-19 04:46:37 +03:00
if ( xfs_has_quota ( mp ) ) {
2005-04-17 02:20:36 +04:00
if ( XFS_IS_UQUOTA_ON ( mp ) & &
mp - > m_sb . sb_uquotino ! = NULLFSINO ) {
ASSERT ( mp - > m_sb . sb_uquotino > 0 ) ;
2013-06-28 02:25:07 +04:00
error = xfs_iget ( mp , NULL , mp - > m_sb . sb_uquotino ,
0 , 0 , & uip ) ;
if ( error )
2014-06-22 09:04:54 +04:00
return error ;
2005-04-17 02:20:36 +04:00
}
2013-07-11 09:00:40 +04:00
if ( XFS_IS_GQUOTA_ON ( mp ) & &
2005-04-17 02:20:36 +04:00
mp - > m_sb . sb_gquotino ! = NULLFSINO ) {
ASSERT ( mp - > m_sb . sb_gquotino > 0 ) ;
2013-06-28 02:25:07 +04:00
error = xfs_iget ( mp , NULL , mp - > m_sb . sb_gquotino ,
0 , 0 , & gip ) ;
if ( error )
goto error_rele ;
2005-04-17 02:20:36 +04:00
}
2013-07-11 09:00:40 +04:00
if ( XFS_IS_PQUOTA_ON ( mp ) & &
2013-07-20 02:36:02 +04:00
mp - > m_sb . sb_pquotino ! = NULLFSINO ) {
ASSERT ( mp - > m_sb . sb_pquotino > 0 ) ;
error = xfs_iget ( mp , NULL , mp - > m_sb . sb_pquotino ,
2013-07-11 09:00:40 +04:00
0 , 0 , & pip ) ;
if ( error )
goto error_rele ;
}
2005-04-17 02:20:36 +04:00
} else {
flags | = XFS_QMOPT_SBVERSION ;
}
/*
2013-07-11 09:00:40 +04:00
* Create the three inodes , if they don ' t exist already . The changes
2005-04-17 02:20:36 +04:00
* made above will get added to a transaction and logged in one of
* the qino_alloc calls below . If the device is readonly ,
* temporarily switch to read - write to do this .
*/
if ( XFS_IS_UQUOTA_ON ( mp ) & & uip = = NULL ) {
2013-06-28 02:25:07 +04:00
error = xfs_qm_qino_alloc ( mp , & uip ,
flags | XFS_QMOPT_UQUOTA ) ;
if ( error )
goto error_rele ;
2005-04-17 02:20:36 +04:00
flags & = ~ XFS_QMOPT_SBVERSION ;
}
2013-07-11 09:00:40 +04:00
if ( XFS_IS_GQUOTA_ON ( mp ) & & gip = = NULL ) {
2005-06-21 09:38:48 +04:00
error = xfs_qm_qino_alloc ( mp , & gip ,
2013-07-11 09:00:40 +04:00
flags | XFS_QMOPT_GQUOTA ) ;
if ( error )
goto error_rele ;
flags & = ~ XFS_QMOPT_SBVERSION ;
}
if ( XFS_IS_PQUOTA_ON ( mp ) & & pip = = NULL ) {
error = xfs_qm_qino_alloc ( mp , & pip ,
flags | XFS_QMOPT_PQUOTA ) ;
2013-06-28 02:25:07 +04:00
if ( error )
goto error_rele ;
2005-04-17 02:20:36 +04:00
}
2010-04-20 11:01:30 +04:00
mp - > m_quotainfo - > qi_uquotaip = uip ;
mp - > m_quotainfo - > qi_gquotaip = gip ;
2013-07-11 09:00:40 +04:00
mp - > m_quotainfo - > qi_pquotaip = pip ;
2005-04-17 02:20:36 +04:00
2006-01-15 04:37:08 +03:00
return 0 ;
2013-06-28 02:25:07 +04:00
error_rele :
if ( uip )
2018-07-25 22:52:32 +03:00
xfs_irele ( uip ) ;
2013-06-28 02:25:07 +04:00
if ( gip )
2018-07-25 22:52:32 +03:00
xfs_irele ( gip ) ;
2013-07-11 09:00:40 +04:00
if ( pip )
2018-07-25 22:52:32 +03:00
xfs_irele ( pip ) ;
2014-06-22 09:04:54 +04:00
return error ;
2005-04-17 02:20:36 +04:00
}
2017-12-22 00:18:26 +03:00
STATIC void
xfs_qm_destroy_quotainos (
2019-11-13 04:04:26 +03:00
struct xfs_quotainfo * qi )
2017-12-22 00:18:26 +03:00
{
if ( qi - > qi_uquotaip ) {
2018-07-25 22:52:32 +03:00
xfs_irele ( qi - > qi_uquotaip ) ;
2017-12-22 00:18:26 +03:00
qi - > qi_uquotaip = NULL ; /* paranoia */
}
if ( qi - > qi_gquotaip ) {
2018-07-25 22:52:32 +03:00
xfs_irele ( qi - > qi_gquotaip ) ;
2017-12-22 00:18:26 +03:00
qi - > qi_gquotaip = NULL ;
}
if ( qi - > qi_pquotaip ) {
2018-07-25 22:52:32 +03:00
xfs_irele ( qi - > qi_pquotaip ) ;
2017-12-22 00:18:26 +03:00
qi - > qi_pquotaip = NULL ;
}
}
2012-02-01 17:57:20 +04:00
STATIC void
xfs_qm_dqfree_one (
struct xfs_dquot * dqp )
{
struct xfs_mount * mp = dqp - > q_mount ;
struct xfs_quotainfo * qi = mp - > m_quotainfo ;
2005-04-17 02:20:36 +04:00
2012-03-13 12:52:35 +04:00
mutex_lock ( & qi - > qi_tree_lock ) ;
2020-07-14 20:37:32 +03:00
radix_tree_delete ( xfs_dquot_tree ( qi , xfs_dquot_type ( dqp ) ) , dqp - > q_id ) ;
2010-04-13 09:06:50 +04:00
2012-02-01 17:57:20 +04:00
qi - > qi_dquots - - ;
2012-03-14 20:53:34 +04:00
mutex_unlock ( & qi - > qi_tree_lock ) ;
2012-02-01 17:57:20 +04:00
xfs_qm_dqdestroy ( dqp ) ;
}
2005-04-17 02:20:36 +04:00
/* --------------- utility functions for vnodeops ---------------- */
/*
2010-10-06 22:41:17 +04:00
* Given an inode , a uid , gid and prid make sure that we have
2005-04-17 02:20:36 +04:00
* allocated relevant dquot ( s ) on disk , and that we won ' t exceed inode
* quotas by creating this file .
* This also attaches dquot ( s ) to the given inode after locking it ,
* and returns the dquots corresponding to the uid and / or gid .
*
* in : inode ( unlocked )
* out : udquot , gdquot with references taken and unlocked
*/
int
xfs_qm_vop_dqalloc (
2009-06-08 17:33:32 +04:00
struct xfs_inode * ip ,
2020-02-21 19:31:27 +03:00
kuid_t uid ,
kgid_t gid ,
2009-06-08 17:33:32 +04:00
prid_t prid ,
uint flags ,
struct xfs_dquot * * O_udqpp ,
2013-07-11 09:00:40 +04:00
struct xfs_dquot * * O_gdqpp ,
struct xfs_dquot * * O_pdqpp )
2005-04-17 02:20:36 +04:00
{
2009-06-08 17:33:32 +04:00
struct xfs_mount * mp = ip - > i_mount ;
2020-02-21 19:31:27 +03:00
struct inode * inode = VFS_I ( ip ) ;
2020-02-21 19:31:27 +03:00
struct user_namespace * user_ns = inode - > i_sb - > s_user_ns ;
2013-06-28 02:25:07 +04:00
struct xfs_dquot * uq = NULL ;
struct xfs_dquot * gq = NULL ;
2013-07-11 09:00:40 +04:00
struct xfs_dquot * pq = NULL ;
2009-06-08 17:33:32 +04:00
int error ;
uint lockflags ;
2005-04-17 02:20:36 +04:00
2021-08-06 21:05:37 +03:00
if ( ! XFS_IS_QUOTA_ON ( mp ) )
2005-04-17 02:20:36 +04:00
return 0 ;
lockflags = XFS_ILOCK_EXCL ;
xfs_ilock ( ip , lockflags ) ;
2007-08-30 11:21:12 +04:00
if ( ( flags & XFS_QMOPT_INHERIT ) & & XFS_INHERIT_GID ( ip ) )
2020-02-21 19:31:27 +03:00
gid = inode - > i_gid ;
2005-04-17 02:20:36 +04:00
/*
* Attach the dquot ( s ) to this inode , doing a dquot allocation
* if necessary . The dquot ( s ) will not be locked .
*/
if ( XFS_NOT_DQATTACHED ( mp , ip ) ) {
2018-05-05 01:30:22 +03:00
error = xfs_qm_dqattach_locked ( ip , true ) ;
2009-06-08 17:33:32 +04:00
if ( error ) {
2005-04-17 02:20:36 +04:00
xfs_iunlock ( ip , lockflags ) ;
2006-01-15 04:37:08 +03:00
return error ;
2005-04-17 02:20:36 +04:00
}
}
2005-06-21 09:38:48 +04:00
if ( ( flags & XFS_QMOPT_UQUOTA ) & & XFS_IS_UQUOTA_ON ( mp ) ) {
2020-10-07 03:50:14 +03:00
ASSERT ( O_udqpp ) ;
2020-02-21 19:31:27 +03:00
if ( ! uid_eq ( inode - > i_uid , uid ) ) {
2005-04-17 02:20:36 +04:00
/*
* What we need is the dquot that has this uid , and
* if we send the inode to dqget , the uid of the inode
* takes priority over what ' s sent in the uid argument .
* We must unlock inode here before calling dqget if
* we ' re not sending the inode , because otherwise
* we ' ll deadlock by doing trans_reserve while
* holding ilock .
*/
xfs_iunlock ( ip , lockflags ) ;
2020-02-21 19:31:27 +03:00
error = xfs_qm_dqget ( mp , from_kuid ( user_ns , uid ) ,
2020-07-16 03:42:36 +03:00
XFS_DQTYPE_USER , true , & uq ) ;
2013-06-28 02:25:07 +04:00
if ( error ) {
2014-06-25 08:58:08 +04:00
ASSERT ( error ! = - ENOENT ) ;
2006-01-15 04:37:08 +03:00
return error ;
2005-04-17 02:20:36 +04:00
}
/*
* Get the ilock in the right order .
*/
xfs_dqunlock ( uq ) ;
lockflags = XFS_ILOCK_SHARED ;
xfs_ilock ( ip , lockflags ) ;
} else {
/*
* Take an extra reference , because we ' ll return
* this to caller
*/
ASSERT ( ip - > i_udquot ) ;
2011-12-07 01:58:22 +04:00
uq = xfs_qm_dqhold ( ip - > i_udquot ) ;
2005-04-17 02:20:36 +04:00
}
}
2005-06-21 09:38:48 +04:00
if ( ( flags & XFS_QMOPT_GQUOTA ) & & XFS_IS_GQUOTA_ON ( mp ) ) {
2020-10-07 03:50:14 +03:00
ASSERT ( O_gdqpp ) ;
2020-02-21 19:31:27 +03:00
if ( ! gid_eq ( inode - > i_gid , gid ) ) {
2005-04-17 02:20:36 +04:00
xfs_iunlock ( ip , lockflags ) ;
2020-02-21 19:31:27 +03:00
error = xfs_qm_dqget ( mp , from_kgid ( user_ns , gid ) ,
2020-07-16 03:42:36 +03:00
XFS_DQTYPE_GROUP , true , & gq ) ;
2013-06-28 02:25:07 +04:00
if ( error ) {
2014-06-25 08:58:08 +04:00
ASSERT ( error ! = - ENOENT ) ;
2013-06-28 02:25:07 +04:00
goto error_rele ;
2005-04-17 02:20:36 +04:00
}
xfs_dqunlock ( gq ) ;
lockflags = XFS_ILOCK_SHARED ;
xfs_ilock ( ip , lockflags ) ;
} else {
ASSERT ( ip - > i_gdquot ) ;
2011-12-07 01:58:22 +04:00
gq = xfs_qm_dqhold ( ip - > i_gdquot ) ;
2005-04-17 02:20:36 +04:00
}
2013-07-11 09:00:40 +04:00
}
if ( ( flags & XFS_QMOPT_PQUOTA ) & & XFS_IS_PQUOTA_ON ( mp ) ) {
2020-10-07 03:50:14 +03:00
ASSERT ( O_pdqpp ) ;
2021-03-29 21:11:39 +03:00
if ( ip - > i_projid ! = prid ) {
2005-06-21 09:38:48 +04:00
xfs_iunlock ( ip , lockflags ) ;
2020-09-17 00:31:55 +03:00
error = xfs_qm_dqget ( mp , prid ,
2020-07-16 03:42:36 +03:00
XFS_DQTYPE_PROJ , true , & pq ) ;
2013-06-28 02:25:07 +04:00
if ( error ) {
2014-06-25 08:58:08 +04:00
ASSERT ( error ! = - ENOENT ) ;
2013-06-28 02:25:07 +04:00
goto error_rele ;
2005-06-21 09:38:48 +04:00
}
2013-07-11 09:00:40 +04:00
xfs_dqunlock ( pq ) ;
2005-06-21 09:38:48 +04:00
lockflags = XFS_ILOCK_SHARED ;
xfs_ilock ( ip , lockflags ) ;
} else {
2013-07-11 09:00:40 +04:00
ASSERT ( ip - > i_pdquot ) ;
pq = xfs_qm_dqhold ( ip - > i_pdquot ) ;
2005-06-21 09:38:48 +04:00
}
2005-04-17 02:20:36 +04:00
}
2020-04-23 07:54:27 +03:00
trace_xfs_dquot_dqalloc ( ip ) ;
2005-04-17 02:20:36 +04:00
xfs_iunlock ( ip , lockflags ) ;
if ( O_udqpp )
* O_udqpp = uq ;
2014-12-01 00:24:20 +03:00
else
2005-04-17 02:20:36 +04:00
xfs_qm_dqrele ( uq ) ;
if ( O_gdqpp )
* O_gdqpp = gq ;
2014-12-01 00:24:20 +03:00
else
2005-04-17 02:20:36 +04:00
xfs_qm_dqrele ( gq ) ;
2013-07-11 09:00:40 +04:00
if ( O_pdqpp )
* O_pdqpp = pq ;
2014-12-01 00:24:20 +03:00
else
2013-07-11 09:00:40 +04:00
xfs_qm_dqrele ( pq ) ;
2006-01-15 04:37:08 +03:00
return 0 ;
2013-06-28 02:25:07 +04:00
error_rele :
2014-12-01 00:24:20 +03:00
xfs_qm_dqrele ( gq ) ;
xfs_qm_dqrele ( uq ) ;
2013-06-28 02:25:07 +04:00
return error ;
2005-04-17 02:20:36 +04:00
}
/*
* Actually transfer ownership , and do dquot modifications .
* These were already reserved .
*/
2019-11-13 04:04:02 +03:00
struct xfs_dquot *
2005-04-17 02:20:36 +04:00
xfs_qm_vop_chown (
2019-11-13 04:04:02 +03:00
struct xfs_trans * tp ,
struct xfs_inode * ip ,
struct xfs_dquot * * IO_olddq ,
struct xfs_dquot * newdq )
2005-04-17 02:20:36 +04:00
{
2019-11-13 04:04:02 +03:00
struct xfs_dquot * prevdq ;
2005-06-21 09:48:47 +04:00
uint bfield = XFS_IS_REALTIME_INODE ( ip ) ?
XFS_TRANS_DQ_RTBCOUNT : XFS_TRANS_DQ_BCOUNT ;
2009-06-08 17:33:32 +04:00
2008-04-22 11:34:00 +04:00
ASSERT ( xfs_isilocked ( ip , XFS_ILOCK_EXCL ) ) ;
2021-08-06 21:05:37 +03:00
ASSERT ( XFS_IS_QUOTA_ON ( ip - > i_mount ) ) ;
2005-04-17 02:20:36 +04:00
/* old dquot */
prevdq = * IO_olddq ;
ASSERT ( prevdq ) ;
ASSERT ( prevdq ! = newdq ) ;
2021-03-29 21:11:40 +03:00
xfs_trans_mod_dquot ( tp , prevdq , bfield , - ( ip - > i_nblocks ) ) ;
2005-06-21 09:48:47 +04:00
xfs_trans_mod_dquot ( tp , prevdq , XFS_TRANS_DQ_ICOUNT , - 1 ) ;
2005-04-17 02:20:36 +04:00
/* the sparkling new dquot */
2021-03-29 21:11:40 +03:00
xfs_trans_mod_dquot ( tp , newdq , bfield , ip - > i_nblocks ) ;
2005-06-21 09:48:47 +04:00
xfs_trans_mod_dquot ( tp , newdq , XFS_TRANS_DQ_ICOUNT , 1 ) ;
2005-04-17 02:20:36 +04:00
2021-01-30 06:06:10 +03:00
/*
* Back when we made quota reservations for the chown , we reserved the
* ondisk blocks + delalloc blocks with the new dquot . Now that we ' ve
* switched the dquots , decrease the new dquot ' s block reservation
* ( having already bumped up the real counter ) so that we don ' t have
* any reservation to give back when we commit .
*/
xfs_trans_mod_dquot ( tp , newdq , XFS_TRANS_DQ_RES_BLKS ,
- ip - > i_delayed_blks ) ;
/*
* Give the incore reservation for delalloc blocks back to the old
* dquot . We don ' t normally handle delalloc quota reservations
* transactionally , so just lock the dquot and subtract from the
* reservation . Dirty the transaction because it ' s too late to turn
* back now .
*/
tp - > t_flags | = XFS_TRANS_DIRTY ;
xfs_dqlock ( prevdq ) ;
ASSERT ( prevdq - > q_blk . reserved > = ip - > i_delayed_blks ) ;
prevdq - > q_blk . reserved - = ip - > i_delayed_blks ;
xfs_dqunlock ( prevdq ) ;
2005-04-17 02:20:36 +04:00
/*
2011-12-07 01:58:22 +04:00
* Take an extra reference , because the inode is going to keep
* this dquot pointer even after the trans_commit .
2005-04-17 02:20:36 +04:00
*/
2011-12-07 01:58:22 +04:00
* IO_olddq = xfs_qm_dqhold ( newdq ) ;
2005-04-17 02:20:36 +04:00
2006-01-15 04:37:08 +03:00
return prevdq ;
2005-04-17 02:20:36 +04:00
}
int
xfs_qm_vop_rename_dqattach (
2009-06-08 17:33:32 +04:00
struct xfs_inode * * i_tab )
2005-04-17 02:20:36 +04:00
{
2009-06-08 17:33:32 +04:00
struct xfs_mount * mp = i_tab [ 0 ] - > i_mount ;
int i ;
2005-04-17 02:20:36 +04:00
2021-08-06 21:05:37 +03:00
if ( ! XFS_IS_QUOTA_ON ( mp ) )
2006-01-15 04:37:08 +03:00
return 0 ;
2005-04-17 02:20:36 +04:00
2009-06-08 17:33:32 +04:00
for ( i = 0 ; ( i < 4 & & i_tab [ i ] ) ; i + + ) {
struct xfs_inode * ip = i_tab [ i ] ;
int error ;
2005-04-17 02:20:36 +04:00
/*
* Watch out for duplicate entries in the table .
*/
2009-06-08 17:33:32 +04:00
if ( i = = 0 | | ip ! = i_tab [ i - 1 ] ) {
if ( XFS_NOT_DQATTACHED ( mp , ip ) ) {
2018-05-05 01:30:21 +03:00
error = xfs_qm_dqattach ( ip ) ;
2005-04-17 02:20:36 +04:00
if ( error )
2006-01-15 04:37:08 +03:00
return error ;
2005-04-17 02:20:36 +04:00
}
}
}
2006-01-15 04:37:08 +03:00
return 0 ;
2005-04-17 02:20:36 +04:00
}
void
2009-06-08 17:33:32 +04:00
xfs_qm_vop_create_dqattach (
struct xfs_trans * tp ,
struct xfs_inode * ip ,
struct xfs_dquot * udqp ,
2013-07-11 09:00:40 +04:00
struct xfs_dquot * gdqp ,
struct xfs_dquot * pdqp )
2005-04-17 02:20:36 +04:00
{
2009-06-08 17:33:32 +04:00
struct xfs_mount * mp = tp - > t_mountp ;
2021-08-06 21:05:37 +03:00
if ( ! XFS_IS_QUOTA_ON ( mp ) )
2005-04-17 02:20:36 +04:00
return ;
2008-04-22 11:34:00 +04:00
ASSERT ( xfs_isilocked ( ip , XFS_ILOCK_EXCL ) ) ;
2005-04-17 02:20:36 +04:00
2013-11-26 17:38:54 +04:00
if ( udqp & & XFS_IS_UQUOTA_ON ( mp ) ) {
2005-04-17 02:20:36 +04:00
ASSERT ( ip - > i_udquot = = NULL ) ;
2020-07-14 20:37:30 +03:00
ASSERT ( i_uid_read ( VFS_I ( ip ) ) = = udqp - > q_id ) ;
2011-12-07 01:58:22 +04:00
ip - > i_udquot = xfs_qm_dqhold ( udqp ) ;
2005-04-17 02:20:36 +04:00
xfs_trans_mod_dquot ( tp , udqp , XFS_TRANS_DQ_ICOUNT , 1 ) ;
}
2013-11-26 17:38:54 +04:00
if ( gdqp & & XFS_IS_GQUOTA_ON ( mp ) ) {
2005-04-17 02:20:36 +04:00
ASSERT ( ip - > i_gdquot = = NULL ) ;
2020-07-14 20:37:30 +03:00
ASSERT ( i_gid_read ( VFS_I ( ip ) ) = = gdqp - > q_id ) ;
2020-02-21 19:31:27 +03:00
2011-12-07 01:58:22 +04:00
ip - > i_gdquot = xfs_qm_dqhold ( gdqp ) ;
2005-04-17 02:20:36 +04:00
xfs_trans_mod_dquot ( tp , gdqp , XFS_TRANS_DQ_ICOUNT , 1 ) ;
}
2013-11-26 17:38:54 +04:00
if ( pdqp & & XFS_IS_PQUOTA_ON ( mp ) ) {
2013-07-11 09:00:40 +04:00
ASSERT ( ip - > i_pdquot = = NULL ) ;
2021-03-29 21:11:39 +03:00
ASSERT ( ip - > i_projid = = pdqp - > q_id ) ;
2013-07-11 09:00:40 +04:00
ip - > i_pdquot = xfs_qm_dqhold ( pdqp ) ;
xfs_trans_mod_dquot ( tp , pdqp , XFS_TRANS_DQ_ICOUNT , 1 ) ;
}
2005-04-17 02:20:36 +04:00
}
2021-08-06 21:05:40 +03:00
/* Decide if this inode's dquot is near an enforcement boundary. */
bool
xfs_inode_near_dquot_enforcement (
struct xfs_inode * ip ,
xfs_dqtype_t type )
{
struct xfs_dquot * dqp ;
int64_t freesp ;
/* We only care for quotas that are enabled and enforced. */
dqp = xfs_inode_dquot ( ip , type ) ;
if ( ! dqp | | ! xfs_dquot_is_enforced ( dqp ) )
return false ;
if ( xfs_dquot_res_over_limits ( & dqp - > q_ino ) | |
xfs_dquot_res_over_limits ( & dqp - > q_rtb ) )
return true ;
/* For space on the data device, check the various thresholds. */
if ( ! dqp - > q_prealloc_hi_wmark )
return false ;
if ( dqp - > q_blk . reserved < dqp - > q_prealloc_lo_wmark )
return false ;
if ( dqp - > q_blk . reserved > = dqp - > q_prealloc_hi_wmark )
return true ;
freesp = dqp - > q_prealloc_hi_wmark - dqp - > q_blk . reserved ;
if ( freesp < dqp - > q_low_space [ XFS_QLOWSP_5_PCNT ] )
return true ;
return false ;
}