License cleanup: add SPDX GPL-2.0 license identifier to files with no license
Many source files in the tree are missing licensing information, which
makes it harder for compliance tools to determine the correct license.
By default all files without license information are under the default
license of the kernel, which is GPL version 2.
Update the files which contain no license information with the 'GPL-2.0'
SPDX license identifier. The SPDX identifier is a legally binding
shorthand, which can be used instead of the full boiler plate text.
This patch is based on work done by Thomas Gleixner and Kate Stewart and
Philippe Ombredanne.
How this work was done:
Patches were generated and checked against linux-4.14-rc6 for a subset of
the use cases:
- file had no licensing information it it.
- file was a */uapi/* one with no licensing information in it,
- file was a */uapi/* one with existing licensing information,
Further patches will be generated in subsequent months to fix up cases
where non-standard license headers were used, and references to license
had to be inferred by heuristics based on keywords.
The analysis to determine which SPDX License Identifier to be applied to
a file was done in a spreadsheet of side by side results from of the
output of two independent scanners (ScanCode & Windriver) producing SPDX
tag:value files created by Philippe Ombredanne. Philippe prepared the
base worksheet, and did an initial spot review of a few 1000 files.
The 4.13 kernel was the starting point of the analysis with 60,537 files
assessed. Kate Stewart did a file by file comparison of the scanner
results in the spreadsheet to determine which SPDX license identifier(s)
to be applied to the file. She confirmed any determination that was not
immediately clear with lawyers working with the Linux Foundation.
Criteria used to select files for SPDX license identifier tagging was:
- Files considered eligible had to be source code files.
- Make and config files were included as candidates if they contained >5
lines of source
- File already had some variant of a license header in it (even if <5
lines).
All documentation files were explicitly excluded.
The following heuristics were used to determine which SPDX license
identifiers to apply.
- when both scanners couldn't find any license traces, file was
considered to have no license information in it, and the top level
COPYING file license applied.
For non */uapi/* files that summary was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 11139
and resulted in the first patch in this series.
If that file was a */uapi/* path one, it was "GPL-2.0 WITH
Linux-syscall-note" otherwise it was "GPL-2.0". Results of that was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 WITH Linux-syscall-note 930
and resulted in the second patch in this series.
- if a file had some form of licensing information in it, and was one
of the */uapi/* ones, it was denoted with the Linux-syscall-note if
any GPL family license was found in the file or had no licensing in
it (per prior point). Results summary:
SPDX license identifier # files
---------------------------------------------------|------
GPL-2.0 WITH Linux-syscall-note 270
GPL-2.0+ WITH Linux-syscall-note 169
((GPL-2.0 WITH Linux-syscall-note) OR BSD-2-Clause) 21
((GPL-2.0 WITH Linux-syscall-note) OR BSD-3-Clause) 17
LGPL-2.1+ WITH Linux-syscall-note 15
GPL-1.0+ WITH Linux-syscall-note 14
((GPL-2.0+ WITH Linux-syscall-note) OR BSD-3-Clause) 5
LGPL-2.0+ WITH Linux-syscall-note 4
LGPL-2.1 WITH Linux-syscall-note 3
((GPL-2.0 WITH Linux-syscall-note) OR MIT) 3
((GPL-2.0 WITH Linux-syscall-note) AND MIT) 1
and that resulted in the third patch in this series.
- when the two scanners agreed on the detected license(s), that became
the concluded license(s).
- when there was disagreement between the two scanners (one detected a
license but the other didn't, or they both detected different
licenses) a manual inspection of the file occurred.
- In most cases a manual inspection of the information in the file
resulted in a clear resolution of the license that should apply (and
which scanner probably needed to revisit its heuristics).
- When it was not immediately clear, the license identifier was
confirmed with lawyers working with the Linux Foundation.
- If there was any question as to the appropriate license identifier,
the file was flagged for further research and to be revisited later
in time.
In total, over 70 hours of logged manual review was done on the
spreadsheet to determine the SPDX license identifiers to apply to the
source files by Kate, Philippe, Thomas and, in some cases, confirmation
by lawyers working with the Linux Foundation.
Kate also obtained a third independent scan of the 4.13 code base from
FOSSology, and compared selected files where the other two scanners
disagreed against that SPDX file, to see if there was new insights. The
Windriver scanner is based on an older version of FOSSology in part, so
they are related.
Thomas did random spot checks in about 500 files from the spreadsheets
for the uapi headers and agreed with SPDX license identifier in the
files he inspected. For the non-uapi files Thomas did random spot checks
in about 15000 files.
In initial set of patches against 4.14-rc6, 3 files were found to have
copy/paste license identifier errors, and have been fixed to reflect the
correct identifier.
Additionally Philippe spent 10 hours this week doing a detailed manual
inspection and review of the 12,461 patched files from the initial patch
version early this week with:
- a full scancode scan run, collecting the matched texts, detected
license ids and scores
- reviewing anything where there was a license detected (about 500+
files) to ensure that the applied SPDX license was correct
- reviewing anything where there was no detection but the patch license
was not GPL-2.0 WITH Linux-syscall-note to ensure that the applied
SPDX license was correct
This produced a worksheet with 20 files needing minor correction. This
worksheet was then exported into 3 different .csv files for the
different types of files to be modified.
These .csv files were then reviewed by Greg. Thomas wrote a script to
parse the csv files and add the proper SPDX tag to the file, in the
format that the file expected. This script was further refined by Greg
based on the output to detect more types of files automatically and to
distinguish between header and source .c files (which need different
comment types.) Finally Greg ran the script using the .csv files to
generate the patches.
Reviewed-by: Kate Stewart <kstewart@linuxfoundation.org>
Reviewed-by: Philippe Ombredanne <pombredanne@nexb.com>
Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
2017-11-01 17:07:57 +03:00
// SPDX-License-Identifier: GPL-2.0
2008-01-29 16:04:06 +03:00
/*
* Functions related to segment and merge handling
*/
# include <linux/kernel.h>
# include <linux/module.h>
# include <linux/bio.h>
# include <linux/blkdev.h>
2021-09-20 15:33:27 +03:00
# include <linux/blk-integrity.h>
2008-01-29 16:04:06 +03:00
# include <linux/scatterlist.h>
2021-11-23 21:53:12 +03:00
# include <linux/part_stat.h>
2022-03-15 03:30:11 +03:00
# include <linux/blk-cgroup.h>
2008-01-29 16:04:06 +03:00
2015-12-03 17:32:30 +03:00
# include <trace/events/block.h>
2008-01-29 16:04:06 +03:00
# include "blk.h"
2021-11-23 21:53:08 +03:00
# include "blk-mq-sched.h"
2020-08-28 05:52:54 +03:00
# include "blk-rq-qos.h"
2021-10-05 18:11:56 +03:00
# include "blk-throttle.h"
2008-01-29 16:04:06 +03:00
2021-10-12 19:18:03 +03:00
static inline void bio_get_first_bvec ( struct bio * bio , struct bio_vec * bv )
{
* bv = mp_bvec_iter_bvec ( bio - > bi_io_vec , bio - > bi_iter ) ;
}
static inline void bio_get_last_bvec ( struct bio * bio , struct bio_vec * bv )
{
struct bvec_iter iter = bio - > bi_iter ;
int idx ;
bio_get_first_bvec ( bio , bv ) ;
if ( bv - > bv_len = = bio - > bi_iter . bi_size )
return ; /* this bio only has a single bvec */
bio_advance_iter ( bio , & iter , iter . bi_size ) ;
if ( ! iter . bi_bvec_done )
idx = iter . bi_idx - 1 ;
else /* in the middle of bvec */
idx = iter . bi_idx ;
* bv = bio - > bi_io_vec [ idx ] ;
/*
* iter . bi_bvec_done records actual length of the last bvec
* if this bio ends in the middle of one io vector
*/
if ( iter . bi_bvec_done )
bv - > bv_len = iter . bi_bvec_done ;
}
2018-09-24 10:43:48 +03:00
static inline bool bio_will_gap ( struct request_queue * q ,
struct request * prev_rq , struct bio * prev , struct bio * next )
{
struct bio_vec pb , nb ;
if ( ! bio_has_data ( prev ) | | ! queue_virt_boundary ( q ) )
return false ;
/*
* Don ' t merge if the 1 st bio starts with non - zero offset , otherwise it
* is quite difficult to respect the sg gap limit . We work hard to
* merge a huge number of small single bios in case of mkfs .
*/
if ( prev_rq )
bio_get_first_bvec ( prev_rq - > bio , & pb ) ;
else
bio_get_first_bvec ( prev , & pb ) ;
2018-11-07 16:58:14 +03:00
if ( pb . bv_offset & queue_virt_boundary ( q ) )
2018-09-24 10:43:48 +03:00
return true ;
/*
* We don ' t need to worry about the situation that the merged segment
* ends in unaligned virt boundary :
*
* - if ' pb ' ends aligned , the merged segment ends aligned
* - if ' pb ' ends unaligned , the next bio must include
* one single bvec of ' nb ' , otherwise the ' nb ' can ' t
* merge with ' pb '
*/
bio_get_last_bvec ( prev , & pb ) ;
bio_get_first_bvec ( next , & nb ) ;
2019-05-21 10:01:42 +03:00
if ( biovec_phys_mergeable ( q , & pb , & nb ) )
2018-09-24 10:43:48 +03:00
return false ;
return __bvec_gap_to_prev ( q , & pb , nb . bv_offset ) ;
}
static inline bool req_gap_back_merge ( struct request * req , struct bio * bio )
{
return bio_will_gap ( req - > q , req , req - > biotail , bio ) ;
}
static inline bool req_gap_front_merge ( struct request * req , struct bio * bio )
{
return bio_will_gap ( req - > q , NULL , bio , req - > bio ) ;
}
2022-07-27 19:22:55 +03:00
static struct bio * bio_split_discard ( struct bio * bio , struct request_queue * q ,
unsigned * nsegs , struct bio_set * bs )
2015-04-24 08:37:18 +03:00
{
unsigned int max_discard_sectors , granularity ;
int alignment ;
sector_t tmp ;
unsigned split_sectors ;
2015-10-20 18:13:52 +03:00
* nsegs = 1 ;
2015-04-24 08:37:18 +03:00
/* Zero-sector (unknown) and one-sector granularities are the same. */
granularity = max ( q - > limits . discard_granularity > > 9 , 1U ) ;
2018-10-29 15:57:17 +03:00
max_discard_sectors = min ( q - > limits . max_discard_sectors ,
bio_allowed_max_sectors ( q ) ) ;
2015-04-24 08:37:18 +03:00
max_discard_sectors - = max_discard_sectors % granularity ;
if ( unlikely ( ! max_discard_sectors ) ) {
/* XXX: warn */
return NULL ;
}
if ( bio_sectors ( bio ) < = max_discard_sectors )
return NULL ;
split_sectors = max_discard_sectors ;
/*
* If the next starting sector would be misaligned , stop the discard at
* the previous aligned sector .
*/
alignment = ( q - > limits . discard_alignment > > 9 ) % granularity ;
tmp = bio - > bi_iter . bi_sector + split_sectors - alignment ;
tmp = sector_div ( tmp , granularity ) ;
if ( split_sectors > tmp )
split_sectors - = tmp ;
return bio_split ( bio , split_sectors , GFP_NOIO , bs ) ;
}
2022-07-27 19:22:55 +03:00
static struct bio * bio_split_write_zeroes ( struct bio * bio ,
struct request_queue * q , unsigned * nsegs , struct bio_set * bs )
2017-04-05 20:21:01 +03:00
{
2019-07-03 15:24:35 +03:00
* nsegs = 0 ;
2017-04-05 20:21:01 +03:00
if ( ! q - > limits . max_write_zeroes_sectors )
return NULL ;
if ( bio_sectors ( bio ) < = q - > limits . max_write_zeroes_sectors )
return NULL ;
return bio_split ( bio , q - > limits . max_write_zeroes_sectors , GFP_NOIO , bs ) ;
}
2019-08-02 01:50:44 +03:00
/*
* Return the maximum number of sectors from the start of a bio that may be
* submitted as a single request to a block device . If enough sectors remain ,
* align the end to the physical block size . Otherwise align the end to the
* logical block size . This approach minimizes the number of non - aligned
* requests that are submitted to a block device if the start of a bio is not
* aligned to a physical block boundary .
*/
2022-07-27 19:22:55 +03:00
static inline unsigned get_max_io_size ( struct bio * bio ,
struct request_queue * q )
2016-01-23 03:05:33 +03:00
{
2019-08-02 01:50:44 +03:00
unsigned pbs = queue_physical_block_size ( q ) > > SECTOR_SHIFT ;
unsigned lbs = queue_logical_block_size ( q ) > > SECTOR_SHIFT ;
2022-06-14 12:09:33 +03:00
unsigned max_sectors = queue_max_sectors ( q ) , start , end ;
2016-01-23 03:05:33 +03:00
2022-06-14 12:09:33 +03:00
if ( q - > limits . chunk_sectors ) {
max_sectors = min ( max_sectors ,
blk_chunk_sectors_left ( bio - > bi_iter . bi_sector ,
q - > limits . chunk_sectors ) ) ;
}
2016-01-23 03:05:33 +03:00
2022-06-14 12:09:32 +03:00
start = bio - > bi_iter . bi_sector & ( pbs - 1 ) ;
end = ( start + max_sectors ) & ~ ( pbs - 1 ) ;
if ( end > start )
return end - start ;
return max_sectors & ~ ( lbs - 1 ) ;
2016-01-23 03:05:33 +03:00
}
2019-12-29 05:32:30 +03:00
static inline unsigned get_max_segment_size ( const struct request_queue * q ,
struct page * start_page ,
unsigned long offset )
2019-02-15 14:13:12 +03:00
{
unsigned long mask = queue_segment_boundary ( q ) ;
2019-12-29 05:32:30 +03:00
offset = mask & ( page_to_phys ( start_page ) + offset ) ;
2020-01-11 15:57:43 +03:00
/*
* overflow may be triggered in case of zero page physical address
* on 32 bit arch , use queue ' s max segment size when that happens .
*/
return min_not_zero ( mask - offset + 1 ,
( unsigned long ) queue_max_segment_size ( q ) ) ;
2019-02-15 14:13:12 +03:00
}
2019-08-02 01:50:43 +03:00
/**
* bvec_split_segs - verify whether or not a bvec should be split in the middle
* @ q : [ in ] request queue associated with the bio associated with @ bv
* @ bv : [ in ] bvec to examine
* @ nsegs : [ in , out ] Number of segments in the bio being built . Incremented
* by the number of segments from @ bv that may be appended to that
* bio without exceeding @ max_segs
2022-06-10 22:58:25 +03:00
* @ bytes : [ in , out ] Number of bytes in the bio being built . Incremented
* by the number of bytes from @ bv that may be appended to that
* bio without exceeding @ max_bytes
2019-08-02 01:50:43 +03:00
* @ max_segs : [ in ] upper bound for * @ nsegs
2022-06-10 22:58:25 +03:00
* @ max_bytes : [ in ] upper bound for * @ bytes
2019-08-02 01:50:43 +03:00
*
* When splitting a bio , it can happen that a bvec is encountered that is too
* big to fit in a single segment and hence that it has to be split in the
* middle . This function verifies whether or not that should happen . The value
* % true is returned if and only if appending the entire @ bv to a bio with
* * @ nsegs segments and * @ sectors sectors would make that bio unacceptable for
* the block driver .
2019-02-15 14:13:12 +03:00
*/
2019-08-02 01:50:40 +03:00
static bool bvec_split_segs ( const struct request_queue * q ,
const struct bio_vec * bv , unsigned * nsegs ,
2022-06-10 22:58:25 +03:00
unsigned * bytes , unsigned max_segs ,
unsigned max_bytes )
2019-02-15 14:13:12 +03:00
{
2022-06-10 22:58:25 +03:00
unsigned max_len = min ( max_bytes , UINT_MAX ) - * bytes ;
2019-08-02 01:50:43 +03:00
unsigned len = min ( bv - > bv_len , max_len ) ;
2019-02-15 14:13:12 +03:00
unsigned total_len = 0 ;
2019-08-02 01:50:42 +03:00
unsigned seg_size = 0 ;
2019-02-15 14:13:12 +03:00
2019-08-02 01:50:42 +03:00
while ( len & & * nsegs < max_segs ) {
2019-12-29 05:32:30 +03:00
seg_size = get_max_segment_size ( q , bv - > bv_page ,
bv - > bv_offset + total_len ) ;
2019-02-15 14:13:12 +03:00
seg_size = min ( seg_size , len ) ;
2019-08-02 01:50:42 +03:00
( * nsegs ) + + ;
2019-02-15 14:13:12 +03:00
total_len + = seg_size ;
len - = seg_size ;
if ( ( bv - > bv_offset + total_len ) & queue_virt_boundary ( q ) )
break ;
}
2022-06-10 22:58:25 +03:00
* bytes + = total_len ;
2019-02-15 14:13:12 +03:00
2019-08-02 01:50:43 +03:00
/* tell the caller to split the bvec if it is too big to fit */
return len > 0 | | bv - > bv_len > max_len ;
2019-02-15 14:13:12 +03:00
}
2019-08-02 01:50:41 +03:00
/**
2022-07-27 19:22:55 +03:00
* bio_split_rw - split a bio in two bios
2019-08-02 01:50:41 +03:00
* @ bio : [ in ] bio to be split
2022-07-27 19:22:55 +03:00
* @ q : [ in ] request queue pointer
2019-08-02 01:50:41 +03:00
* @ segs : [ out ] number of segments in the bio with the first half of the sectors
2022-07-27 19:22:55 +03:00
* @ bs : [ in ] bio set to allocate the clone from
2019-08-02 01:50:41 +03:00
*
* Clone @ bio , update the bi_iter of the clone to represent the first sectors
* of @ bio and update @ bio - > bi_iter to represent the remaining sectors . The
* following is guaranteed for the cloned bio :
2022-07-27 19:22:55 +03:00
* - That it has at most get_max_io_size ( @ bio , @ q ) sectors .
2019-08-02 01:50:41 +03:00
* - That it has at most queue_max_segments ( @ q ) segments .
*
* Except for discard requests the cloned bio will point at the bi_io_vec of
* the original bio . It is the responsibility of the caller to ensure that the
* original bio is not freed before the cloned bio . The caller is also
* responsible for ensuring that @ bs is only destroyed after processing of the
* split bio has finished .
*/
2022-07-27 19:22:55 +03:00
static struct bio * bio_split_rw ( struct bio * bio , struct request_queue * q ,
unsigned * segs , struct bio_set * bs )
2015-04-24 08:37:18 +03:00
{
2015-09-03 01:46:02 +03:00
struct bio_vec bv , bvprv , * bvprvp = NULL ;
2015-04-24 08:37:18 +03:00
struct bvec_iter iter ;
2022-06-10 22:58:25 +03:00
unsigned nsegs = 0 , bytes = 0 ;
2022-07-27 19:22:55 +03:00
const unsigned max_bytes = get_max_io_size ( bio , q ) < < 9 ;
2019-03-03 16:17:48 +03:00
const unsigned max_segs = queue_max_segments ( q ) ;
2015-04-24 08:37:18 +03:00
2019-02-15 14:13:12 +03:00
bio_for_each_bvec ( bv , bio , iter ) {
2015-04-24 08:37:18 +03:00
/*
* If the queue doesn ' t support SG gaps and adding this
* offset would create a gap , disallow it .
*/
2015-09-03 01:46:02 +03:00
if ( bvprvp & & bvec_gap_to_prev ( q , bvprvp , bv . bv_offset ) )
2015-04-24 08:37:18 +03:00
goto split ;
2019-08-02 01:50:43 +03:00
if ( nsegs < max_segs & &
2022-06-10 22:58:25 +03:00
bytes + bv . bv_len < = max_bytes & &
2019-08-02 01:50:43 +03:00
bv . bv_offset + bv . bv_len < = PAGE_SIZE ) {
nsegs + + ;
2022-06-10 22:58:25 +03:00
bytes + = bv . bv_len ;
} else if ( bvec_split_segs ( q , & bv , & nsegs , & bytes , max_segs ,
max_bytes ) ) {
2017-12-18 15:22:16 +03:00
goto split ;
2016-01-13 01:08:39 +03:00
}
2015-04-24 08:37:18 +03:00
bvprv = bv ;
2015-11-24 05:35:29 +03:00
bvprvp = & bvprv ;
2015-04-24 08:37:18 +03:00
}
2019-06-06 13:29:03 +03:00
* segs = nsegs ;
return NULL ;
2015-04-24 08:37:18 +03:00
split :
2015-10-20 18:13:52 +03:00
* segs = nsegs ;
block: disable iopoll for split bio
iopoll is initially for small size, latency sensitive IO. It doesn't
work well for big IO, especially when it needs to be split to multiple
bios. In this case, the returned cookie of __submit_bio_noacct_mq() is
indeed the cookie of the last split bio. The completion of *this* last
split bio done by iopoll doesn't mean the whole original bio has
completed. Callers of iopoll still need to wait for completion of other
split bios.
Besides bio splitting may cause more trouble for iopoll which isn't
supposed to be used in case of big IO.
iopoll for split bio may cause potential race if CPU migration happens
during bio submission. Since the returned cookie is that of the last
split bio, polling on the corresponding hardware queue doesn't help
complete other split bios, if these split bios are enqueued into
different hardware queues. Since interrupts are disabled for polling
queues, the completion of these other split bios depends on timeout
mechanism, thus causing a potential hang.
iopoll for split bio may also cause hang for sync polling. Currently
both the blkdev and iomap-based fs (ext4/xfs, etc) support sync polling
in direct IO routine. These routines will submit bio without REQ_NOWAIT
flag set, and then start sync polling in current process context. The
process may hang in blk_mq_get_tag() if the submitted bio has to be
split into multiple bios and can rapidly exhaust the queue depth. The
process are waiting for the completion of the previously allocated
requests, which should be reaped by the following polling, and thus
causing a deadlock.
To avoid these subtle trouble described above, just disable iopoll for
split bio and return BLK_QC_T_NONE in this case. The side effect is that
non-HIPRI IO also returns BLK_QC_T_NONE now. It should be acceptable
since the returned cookie is never used for non-HIPRI IO.
Suggested-by: Ming Lei <ming.lei@redhat.com>
Signed-off-by: Jeffle Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Ming Lei <ming.lei@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
2020-11-26 12:18:52 +03:00
2022-06-10 22:58:25 +03:00
/*
* Individual bvecs might not be logical block aligned . Round down the
* split size so that each bio is properly block size aligned , even if
* we do not use the full hardware limits .
*/
bytes = ALIGN_DOWN ( bytes , queue_logical_block_size ( q ) ) ;
block: disable iopoll for split bio
iopoll is initially for small size, latency sensitive IO. It doesn't
work well for big IO, especially when it needs to be split to multiple
bios. In this case, the returned cookie of __submit_bio_noacct_mq() is
indeed the cookie of the last split bio. The completion of *this* last
split bio done by iopoll doesn't mean the whole original bio has
completed. Callers of iopoll still need to wait for completion of other
split bios.
Besides bio splitting may cause more trouble for iopoll which isn't
supposed to be used in case of big IO.
iopoll for split bio may cause potential race if CPU migration happens
during bio submission. Since the returned cookie is that of the last
split bio, polling on the corresponding hardware queue doesn't help
complete other split bios, if these split bios are enqueued into
different hardware queues. Since interrupts are disabled for polling
queues, the completion of these other split bios depends on timeout
mechanism, thus causing a potential hang.
iopoll for split bio may also cause hang for sync polling. Currently
both the blkdev and iomap-based fs (ext4/xfs, etc) support sync polling
in direct IO routine. These routines will submit bio without REQ_NOWAIT
flag set, and then start sync polling in current process context. The
process may hang in blk_mq_get_tag() if the submitted bio has to be
split into multiple bios and can rapidly exhaust the queue depth. The
process are waiting for the completion of the previously allocated
requests, which should be reaped by the following polling, and thus
causing a deadlock.
To avoid these subtle trouble described above, just disable iopoll for
split bio and return BLK_QC_T_NONE in this case. The side effect is that
non-HIPRI IO also returns BLK_QC_T_NONE now. It should be acceptable
since the returned cookie is never used for non-HIPRI IO.
Suggested-by: Ming Lei <ming.lei@redhat.com>
Signed-off-by: Jeffle Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Ming Lei <ming.lei@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
2020-11-26 12:18:52 +03:00
/*
* Bio splitting may cause subtle trouble such as hang when doing sync
* iopoll in direct IO routine . Given performance gain of iopoll for
* big IO can be trival , disable iopoll when split needed .
*/
2021-10-12 14:12:21 +03:00
bio_clear_polled ( bio ) ;
2022-06-10 22:58:25 +03:00
return bio_split ( bio , bytes > > SECTOR_SHIFT , GFP_NOIO , bs ) ;
2015-04-24 08:37:18 +03:00
}
2019-08-02 01:50:41 +03:00
/**
2022-07-27 19:22:55 +03:00
* __bio_split_to_limits - split a bio to fit the queue limits
* @ bio : bio to be split
* @ q : request_queue new bio is being queued at
* @ nr_segs : returns the number of segments in the returned bio
*
* Check if @ bio needs splitting based on the queue limits , and if so split off
* a bio fitting the limits from the beginning of @ bio and return it . @ bio is
* shortened to the remainder and re - submitted .
2019-08-02 01:50:41 +03:00
*
2022-07-27 19:22:55 +03:00
* The split bio is allocated from @ q - > bio_split , which is provided by the
* block layer .
2019-08-02 01:50:41 +03:00
*/
2022-07-27 19:22:55 +03:00
struct bio * __bio_split_to_limits ( struct bio * bio , struct request_queue * q ,
2021-10-13 21:43:41 +03:00
unsigned int * nr_segs )
2015-04-24 08:37:18 +03:00
{
2022-07-27 19:22:55 +03:00
struct bio * split ;
2015-04-24 08:37:18 +03:00
2022-07-27 19:22:55 +03:00
switch ( bio_op ( bio ) ) {
2016-08-16 10:59:35 +03:00
case REQ_OP_DISCARD :
case REQ_OP_SECURE_ERASE :
2022-07-27 19:22:55 +03:00
split = bio_split_discard ( bio , q , nr_segs , & q - > bio_split ) ;
2016-08-16 10:59:35 +03:00
break ;
2016-11-30 23:28:59 +03:00
case REQ_OP_WRITE_ZEROES :
2022-07-27 19:22:55 +03:00
split = bio_split_write_zeroes ( bio , q , nr_segs , & q - > bio_split ) ;
2016-11-30 23:28:59 +03:00
break ;
2016-08-16 10:59:35 +03:00
default :
2022-07-27 19:22:55 +03:00
split = bio_split_rw ( bio , q , nr_segs , & q - > bio_split ) ;
2016-08-16 10:59:35 +03:00
break ;
}
2015-10-20 18:13:52 +03:00
2015-04-24 08:37:18 +03:00
if ( split ) {
2015-10-20 18:13:53 +03:00
/* there isn't chance to merge the splitted bio */
2016-08-06 00:35:16 +03:00
split - > bi_opf | = REQ_NOMERGE ;
2015-10-20 18:13:53 +03:00
2022-07-13 17:02:26 +03:00
blkcg_bio_issue_init ( split ) ;
2022-07-27 19:22:55 +03:00
bio_chain ( split , bio ) ;
trace_block_split ( split , bio - > bi_iter . bi_sector ) ;
submit_bio_noacct ( bio ) ;
return split ;
2015-04-24 08:37:18 +03:00
}
2022-07-27 19:22:55 +03:00
return bio ;
2015-04-24 08:37:18 +03:00
}
2019-06-06 13:29:01 +03:00
2019-08-02 01:50:41 +03:00
/**
2022-07-27 19:22:55 +03:00
* bio_split_to_limits - split a bio to fit the queue limits
* @ bio : bio to be split
*
* Check if @ bio needs splitting based on the queue limits of @ bio - > bi_bdev , and
* if so split off a bio fitting the limits from the beginning of @ bio and
* return it . @ bio is shortened to the remainder and re - submitted .
2019-08-02 01:50:41 +03:00
*
2022-07-27 19:22:55 +03:00
* The split bio is allocated from @ q - > bio_split , which is provided by the
* block layer .
2019-08-02 01:50:41 +03:00
*/
2022-07-27 19:22:55 +03:00
struct bio * bio_split_to_limits ( struct bio * bio )
2019-06-06 13:29:01 +03:00
{
2022-07-27 19:22:55 +03:00
struct request_queue * q = bdev_get_queue ( bio - > bi_bdev ) ;
2019-06-06 13:29:01 +03:00
unsigned int nr_segs ;
2022-07-27 19:22:55 +03:00
if ( bio_may_exceed_limits ( bio , q ) )
return __bio_split_to_limits ( bio , q , & nr_segs ) ;
return bio ;
2019-06-06 13:29:01 +03:00
}
2022-07-27 19:22:55 +03:00
EXPORT_SYMBOL ( bio_split_to_limits ) ;
2015-04-24 08:37:18 +03:00
2019-06-06 13:29:02 +03:00
unsigned int blk_recalc_rq_segments ( struct request * rq )
2008-01-29 16:04:06 +03:00
{
2019-05-21 10:01:43 +03:00
unsigned int nr_phys_segs = 0 ;
2022-06-10 22:58:25 +03:00
unsigned int bytes = 0 ;
2019-06-06 13:29:02 +03:00
struct req_iterator iter ;
2019-05-21 10:01:43 +03:00
struct bio_vec bv ;
2008-01-29 16:04:06 +03:00
2019-06-06 13:29:02 +03:00
if ( ! rq - > bio )
2009-02-23 11:03:10 +03:00
return 0 ;
2008-01-29 16:04:06 +03:00
2019-06-06 13:29:02 +03:00
switch ( bio_op ( rq - > bio ) ) {
2016-11-30 23:28:59 +03:00
case REQ_OP_DISCARD :
case REQ_OP_SECURE_ERASE :
2021-02-11 17:38:07 +03:00
if ( queue_max_discard_segments ( rq - > q ) > 1 ) {
struct bio * bio = rq - > bio ;
for_each_bio ( bio )
nr_phys_segs + + ;
return nr_phys_segs ;
}
return 1 ;
2016-11-30 23:28:59 +03:00
case REQ_OP_WRITE_ZEROES :
2016-12-09 01:20:32 +03:00
return 0 ;
2022-07-14 21:06:30 +03:00
default :
break ;
2016-11-30 23:28:59 +03:00
}
2014-02-08 00:53:46 +04:00
2019-06-06 13:29:02 +03:00
rq_for_each_bvec ( bv , rq , iter )
2022-06-10 22:58:25 +03:00
bvec_split_segs ( rq - > q , & bv , & nr_phys_segs , & bytes ,
2019-08-02 01:50:43 +03:00
UINT_MAX , UINT_MAX ) ;
2009-02-23 11:03:10 +03:00
return nr_phys_segs ;
}
2019-02-27 15:40:11 +03:00
static inline struct scatterlist * blk_next_sg ( struct scatterlist * * sg ,
2019-02-15 14:13:13 +03:00
struct scatterlist * sglist )
{
if ( ! * sg )
return sglist ;
/*
* If the driver previously mapped a shorter list , we could see a
* termination bit prematurely unless it fully inits the sg table
* on each mapping . We KNOW that there must be more entries here
* or the driver would be buggy , so force clear the termination bit
* to avoid doing a full sg_init_table ( ) in drivers for each command .
*/
sg_unmark_end ( * sg ) ;
return sg_next ( * sg ) ;
}
static unsigned blk_bvec_map_sg ( struct request_queue * q ,
struct bio_vec * bvec , struct scatterlist * sglist ,
struct scatterlist * * sg )
{
unsigned nbytes = bvec - > bv_len ;
2019-04-11 09:23:27 +03:00
unsigned nsegs = 0 , total = 0 ;
2019-02-15 14:13:13 +03:00
while ( nbytes > 0 ) {
2019-04-11 09:23:27 +03:00
unsigned offset = bvec - > bv_offset + total ;
2019-12-29 05:32:30 +03:00
unsigned len = min ( get_max_segment_size ( q , bvec - > bv_page ,
offset ) , nbytes ) ;
2019-04-19 09:56:24 +03:00
struct page * page = bvec - > bv_page ;
/*
* Unfortunately a fair number of drivers barf on scatterlists
* that have an offset larger than PAGE_SIZE , despite other
* subsystems dealing with that invariant just fine . For now
* stick to the legacy format where we never present those from
* the block layer , but the code below should be removed once
* these offenders ( mostly MMC / SD drivers ) are fixed .
*/
page + = ( offset > > PAGE_SHIFT ) ;
offset & = ~ PAGE_MASK ;
2019-02-15 14:13:13 +03:00
* sg = blk_next_sg ( sg , sglist ) ;
2019-04-19 09:56:24 +03:00
sg_set_page ( * sg , page , len , offset ) ;
2019-02-15 14:13:13 +03:00
2019-04-11 09:23:27 +03:00
total + = len ;
nbytes - = len ;
2019-02-15 14:13:13 +03:00
nsegs + + ;
}
return nsegs ;
}
2019-03-17 13:01:11 +03:00
static inline int __blk_bvec_map_sg ( struct bio_vec bv ,
struct scatterlist * sglist , struct scatterlist * * sg )
{
* sg = blk_next_sg ( sg , sglist ) ;
sg_set_page ( * sg , bv . bv_page , bv . bv_len , bv . bv_offset ) ;
return 1 ;
}
2019-03-17 13:01:12 +03:00
/* only try to merge bvecs into one sg if they are from two bios */
static inline bool
__blk_segment_map_sg_merge ( struct request_queue * q , struct bio_vec * bvec ,
struct bio_vec * bvprv , struct scatterlist * * sg )
2012-08-03 01:42:03 +04:00
{
int nbytes = bvec - > bv_len ;
2019-03-17 13:01:12 +03:00
if ( ! * sg )
return false ;
2012-08-03 01:42:03 +04:00
2019-03-17 13:01:12 +03:00
if ( ( * sg ) - > length + nbytes > queue_max_segment_size ( q ) )
return false ;
if ( ! biovec_phys_mergeable ( q , bvprv , bvec ) )
return false ;
( * sg ) - > length + = nbytes ;
return true ;
2012-08-03 01:42:03 +04:00
}
2014-02-08 00:53:46 +04:00
static int __blk_bios_map_sg ( struct request_queue * q , struct bio * bio ,
struct scatterlist * sglist ,
struct scatterlist * * sg )
2008-01-29 16:04:06 +03:00
{
treewide: Remove uninitialized_var() usage
Using uninitialized_var() is dangerous as it papers over real bugs[1]
(or can in the future), and suppresses unrelated compiler warnings
(e.g. "unused variable"). If the compiler thinks it is uninitialized,
either simply initialize the variable or make compiler changes.
In preparation for removing[2] the[3] macro[4], remove all remaining
needless uses with the following script:
git grep '\buninitialized_var\b' | cut -d: -f1 | sort -u | \
xargs perl -pi -e \
's/\buninitialized_var\(([^\)]+)\)/\1/g;
s:\s*/\* (GCC be quiet|to make compiler happy) \*/$::g;'
drivers/video/fbdev/riva/riva_hw.c was manually tweaked to avoid
pathological white-space.
No outstanding warnings were found building allmodconfig with GCC 9.3.0
for x86_64, i386, arm64, arm, powerpc, powerpc64le, s390x, mips, sparc64,
alpha, and m68k.
[1] https://lore.kernel.org/lkml/20200603174714.192027-1-glider@google.com/
[2] https://lore.kernel.org/lkml/CA+55aFw+Vbj0i=1TGqCR5vQkCzWJ0QxK6CernOU6eedsudAixw@mail.gmail.com/
[3] https://lore.kernel.org/lkml/CA+55aFwgbgqhbp1fkxvRKEpzyR5J8n1vKT1VZdz9knmPuXhOeg@mail.gmail.com/
[4] https://lore.kernel.org/lkml/CA+55aFz2500WfbKXAx8s67wrm9=yVJu65TpLgN_ybYNv0VEOKA@mail.gmail.com/
Reviewed-by: Leon Romanovsky <leonro@mellanox.com> # drivers/infiniband and mlx4/mlx5
Acked-by: Jason Gunthorpe <jgg@mellanox.com> # IB
Acked-by: Kalle Valo <kvalo@codeaurora.org> # wireless drivers
Reviewed-by: Chao Yu <yuchao0@huawei.com> # erofs
Signed-off-by: Kees Cook <keescook@chromium.org>
2020-06-03 23:09:38 +03:00
struct bio_vec bvec , bvprv = { NULL } ;
2014-02-08 00:53:46 +04:00
struct bvec_iter iter ;
2018-12-13 18:17:10 +03:00
int nsegs = 0 ;
2019-03-17 13:01:12 +03:00
bool new_bio = false ;
2014-02-08 00:53:46 +04:00
2019-03-17 13:01:12 +03:00
for_each_bio ( bio ) {
bio_for_each_bvec ( bvec , bio , iter ) {
/*
* Only try to merge bvecs from two bios given we
* have done bio internal merge when adding pages
* to bio
*/
if ( new_bio & &
__blk_segment_map_sg_merge ( q , & bvec , & bvprv , sg ) )
goto next_bvec ;
if ( bvec . bv_offset + bvec . bv_len < = PAGE_SIZE )
nsegs + = __blk_bvec_map_sg ( bvec , sglist , sg ) ;
else
nsegs + = blk_bvec_map_sg ( q , & bvec , sglist , sg ) ;
next_bvec :
new_bio = false ;
}
2019-04-02 05:26:44 +03:00
if ( likely ( bio - > bi_iter . bi_size ) ) {
bvprv = bvec ;
new_bio = true ;
}
2019-03-17 13:01:12 +03:00
}
2008-01-29 16:04:06 +03:00
2014-02-08 00:53:46 +04:00
return nsegs ;
}
/*
* map a request to scatterlist , return number of sg entries setup . Caller
* must make sure sg can hold rq - > nr_phys_segments entries
*/
2020-04-14 10:42:22 +03:00
int __blk_rq_map_sg ( struct request_queue * q , struct request * rq ,
struct scatterlist * sglist , struct scatterlist * * last_sg )
2014-02-08 00:53:46 +04:00
{
int nsegs = 0 ;
2016-12-09 01:20:32 +03:00
if ( rq - > rq_flags & RQF_SPECIAL_PAYLOAD )
2020-04-14 10:42:22 +03:00
nsegs = __blk_bvec_map_sg ( rq - > special_vec , sglist , last_sg ) ;
2016-12-09 01:20:32 +03:00
else if ( rq - > bio )
2020-04-14 10:42:22 +03:00
nsegs = __blk_bios_map_sg ( q , rq - > bio , sglist , last_sg ) ;
2008-04-11 14:56:52 +04:00
2020-04-14 10:42:22 +03:00
if ( * last_sg )
sg_mark_end ( * last_sg ) ;
2008-01-29 16:04:06 +03:00
2015-11-24 05:35:31 +03:00
/*
* Something must have been wrong if the figured number of
* segment is bigger than number of req ' s physical segments
*/
2016-12-09 01:20:32 +03:00
WARN_ON ( nsegs > blk_rq_nr_phys_segments ( rq ) ) ;
2015-11-24 05:35:31 +03:00
2008-01-29 16:04:06 +03:00
return nsegs ;
}
2020-04-14 10:42:22 +03:00
EXPORT_SYMBOL ( __blk_rq_map_sg ) ;
2008-01-29 16:04:06 +03:00
2020-08-17 12:52:39 +03:00
static inline unsigned int blk_rq_get_max_segments ( struct request * rq )
{
if ( req_op ( rq ) = = REQ_OP_DISCARD )
return queue_max_discard_segments ( rq - > q ) ;
return queue_max_segments ( rq - > q ) ;
}
2021-09-20 15:33:26 +03:00
static inline unsigned int blk_rq_get_max_sectors ( struct request * rq ,
sector_t offset )
{
struct request_queue * q = rq - > q ;
2022-06-14 12:09:31 +03:00
unsigned int max_sectors ;
2021-09-20 15:33:26 +03:00
if ( blk_rq_is_passthrough ( rq ) )
return q - > limits . max_hw_sectors ;
2022-06-14 12:09:31 +03:00
max_sectors = blk_queue_get_max_sectors ( q , req_op ( rq ) ) ;
2021-09-20 15:33:26 +03:00
if ( ! q - > limits . chunk_sectors | |
req_op ( rq ) = = REQ_OP_DISCARD | |
req_op ( rq ) = = REQ_OP_SECURE_ERASE )
2022-06-14 12:09:31 +03:00
return max_sectors ;
return min ( max_sectors ,
blk_chunk_sectors_left ( offset , q - > limits . chunk_sectors ) ) ;
2021-09-20 15:33:26 +03:00
}
2019-06-06 13:29:01 +03:00
static inline int ll_new_hw_segment ( struct request * req , struct bio * bio ,
unsigned int nr_phys_segs )
2008-01-29 16:04:06 +03:00
{
2022-03-15 03:30:11 +03:00
if ( ! blk_cgroup_mergeable ( req , bio ) )
goto no_merge ;
2021-06-28 05:33:12 +03:00
if ( blk_integrity_merge_bio ( req - > q , req , bio ) = = false )
2010-09-10 22:50:10 +04:00
goto no_merge ;
2021-06-28 05:33:12 +03:00
/* discard request merge won't add new segment */
if ( req_op ( req ) = = REQ_OP_DISCARD )
return 1 ;
if ( req - > nr_phys_segments + nr_phys_segs > blk_rq_get_max_segments ( req ) )
2010-09-10 22:50:10 +04:00
goto no_merge ;
2008-01-29 16:04:06 +03:00
/*
* This will form the start of a new hw segment . Bump both
* counters .
*/
req - > nr_phys_segments + = nr_phys_segs ;
return 1 ;
2010-09-10 22:50:10 +04:00
no_merge :
2019-06-06 13:29:01 +03:00
req_set_nomerge ( req - > q , req ) ;
2010-09-10 22:50:10 +04:00
return 0 ;
2008-01-29 16:04:06 +03:00
}
2019-06-06 13:29:01 +03:00
int ll_back_merge_fn ( struct request * req , struct bio * bio , unsigned int nr_segs )
2008-01-29 16:04:06 +03:00
{
2015-09-03 19:28:20 +03:00
if ( req_gap_back_merge ( req , bio ) )
return 0 ;
2015-09-11 18:03:04 +03:00
if ( blk_integrity_rq ( req ) & &
integrity_req_gap_back_merge ( req , bio ) )
return 0 ;
block: Inline encryption support for blk-mq
We must have some way of letting a storage device driver know what
encryption context it should use for en/decrypting a request. However,
it's the upper layers (like the filesystem/fscrypt) that know about and
manages encryption contexts. As such, when the upper layer submits a bio
to the block layer, and this bio eventually reaches a device driver with
support for inline encryption, the device driver will need to have been
told the encryption context for that bio.
We want to communicate the encryption context from the upper layer to the
storage device along with the bio, when the bio is submitted to the block
layer. To do this, we add a struct bio_crypt_ctx to struct bio, which can
represent an encryption context (note that we can't use the bi_private
field in struct bio to do this because that field does not function to pass
information across layers in the storage stack). We also introduce various
functions to manipulate the bio_crypt_ctx and make the bio/request merging
logic aware of the bio_crypt_ctx.
We also make changes to blk-mq to make it handle bios with encryption
contexts. blk-mq can merge many bios into the same request. These bios need
to have contiguous data unit numbers (the necessary changes to blk-merge
are also made to ensure this) - as such, it suffices to keep the data unit
number of just the first bio, since that's all a storage driver needs to
infer the data unit number to use for each data block in each bio in a
request. blk-mq keeps track of the encryption context to be used for all
the bios in a request with the request's rq_crypt_ctx. When the first bio
is added to an empty request, blk-mq will program the encryption context
of that bio into the request_queue's keyslot manager, and store the
returned keyslot in the request's rq_crypt_ctx. All the functions to
operate on encryption contexts are in blk-crypto.c.
Upper layers only need to call bio_crypt_set_ctx with the encryption key,
algorithm and data_unit_num; they don't have to worry about getting a
keyslot for each encryption context, as blk-mq/blk-crypto handles that.
Blk-crypto also makes it possible for request-based layered devices like
dm-rq to make use of inline encryption hardware by cloning the
rq_crypt_ctx and programming a keyslot in the new request_queue when
necessary.
Note that any user of the block layer can submit bios with an
encryption context, such as filesystems, device-mapper targets, etc.
Signed-off-by: Satya Tangirala <satyat@google.com>
Reviewed-by: Eric Biggers <ebiggers@google.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
2020-05-14 03:37:18 +03:00
if ( ! bio_crypt_ctx_back_mergeable ( req , bio ) )
return 0 ;
2012-09-18 20:19:26 +04:00
if ( blk_rq_sectors ( req ) + bio_sectors ( bio ) >
2016-07-21 06:40:47 +03:00
blk_rq_get_max_sectors ( req , blk_rq_pos ( req ) ) ) {
2019-06-06 13:29:01 +03:00
req_set_nomerge ( req - > q , req ) ;
2008-01-29 16:04:06 +03:00
return 0 ;
}
2019-06-06 13:29:01 +03:00
return ll_new_hw_segment ( req , bio , nr_segs ) ;
2008-01-29 16:04:06 +03:00
}
2020-10-06 10:07:19 +03:00
static int ll_front_merge_fn ( struct request * req , struct bio * bio ,
unsigned int nr_segs )
2008-01-29 16:04:06 +03:00
{
2015-09-03 19:28:20 +03:00
if ( req_gap_front_merge ( req , bio ) )
return 0 ;
2015-09-11 18:03:04 +03:00
if ( blk_integrity_rq ( req ) & &
integrity_req_gap_front_merge ( req , bio ) )
return 0 ;
block: Inline encryption support for blk-mq
We must have some way of letting a storage device driver know what
encryption context it should use for en/decrypting a request. However,
it's the upper layers (like the filesystem/fscrypt) that know about and
manages encryption contexts. As such, when the upper layer submits a bio
to the block layer, and this bio eventually reaches a device driver with
support for inline encryption, the device driver will need to have been
told the encryption context for that bio.
We want to communicate the encryption context from the upper layer to the
storage device along with the bio, when the bio is submitted to the block
layer. To do this, we add a struct bio_crypt_ctx to struct bio, which can
represent an encryption context (note that we can't use the bi_private
field in struct bio to do this because that field does not function to pass
information across layers in the storage stack). We also introduce various
functions to manipulate the bio_crypt_ctx and make the bio/request merging
logic aware of the bio_crypt_ctx.
We also make changes to blk-mq to make it handle bios with encryption
contexts. blk-mq can merge many bios into the same request. These bios need
to have contiguous data unit numbers (the necessary changes to blk-merge
are also made to ensure this) - as such, it suffices to keep the data unit
number of just the first bio, since that's all a storage driver needs to
infer the data unit number to use for each data block in each bio in a
request. blk-mq keeps track of the encryption context to be used for all
the bios in a request with the request's rq_crypt_ctx. When the first bio
is added to an empty request, blk-mq will program the encryption context
of that bio into the request_queue's keyslot manager, and store the
returned keyslot in the request's rq_crypt_ctx. All the functions to
operate on encryption contexts are in blk-crypto.c.
Upper layers only need to call bio_crypt_set_ctx with the encryption key,
algorithm and data_unit_num; they don't have to worry about getting a
keyslot for each encryption context, as blk-mq/blk-crypto handles that.
Blk-crypto also makes it possible for request-based layered devices like
dm-rq to make use of inline encryption hardware by cloning the
rq_crypt_ctx and programming a keyslot in the new request_queue when
necessary.
Note that any user of the block layer can submit bios with an
encryption context, such as filesystems, device-mapper targets, etc.
Signed-off-by: Satya Tangirala <satyat@google.com>
Reviewed-by: Eric Biggers <ebiggers@google.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
2020-05-14 03:37:18 +03:00
if ( ! bio_crypt_ctx_front_mergeable ( req , bio ) )
return 0 ;
2012-09-18 20:19:26 +04:00
if ( blk_rq_sectors ( req ) + bio_sectors ( bio ) >
2016-07-21 06:40:47 +03:00
blk_rq_get_max_sectors ( req , bio - > bi_iter . bi_sector ) ) {
2019-06-06 13:29:01 +03:00
req_set_nomerge ( req - > q , req ) ;
2008-01-29 16:04:06 +03:00
return 0 ;
}
2019-06-06 13:29:01 +03:00
return ll_new_hw_segment ( req , bio , nr_segs ) ;
2008-01-29 16:04:06 +03:00
}
2018-02-02 00:01:02 +03:00
static bool req_attempt_discard_merge ( struct request_queue * q , struct request * req ,
struct request * next )
{
unsigned short segments = blk_rq_nr_discard_segments ( req ) ;
if ( segments > = queue_max_discard_segments ( q ) )
goto no_merge ;
if ( blk_rq_sectors ( req ) + bio_sectors ( next - > bio ) >
blk_rq_get_max_sectors ( req , blk_rq_pos ( req ) ) )
goto no_merge ;
req - > nr_phys_segments = segments + blk_rq_nr_discard_segments ( next ) ;
return true ;
no_merge :
req_set_nomerge ( q , req ) ;
return false ;
}
2008-01-29 16:04:06 +03:00
static int ll_merge_requests_fn ( struct request_queue * q , struct request * req ,
struct request * next )
{
int total_phys_segments ;
2015-09-03 19:28:20 +03:00
if ( req_gap_back_merge ( req , next - > bio ) )
2015-02-11 18:20:13 +03:00
return 0 ;
2008-01-29 16:04:06 +03:00
/*
* Will it become too large ?
*/
2012-09-18 20:19:26 +04:00
if ( ( blk_rq_sectors ( req ) + blk_rq_sectors ( next ) ) >
2016-07-21 06:40:47 +03:00
blk_rq_get_max_sectors ( req , blk_rq_pos ( req ) ) )
2008-01-29 16:04:06 +03:00
return 0 ;
total_phys_segments = req - > nr_phys_segments + next - > nr_phys_segments ;
2020-08-17 12:52:39 +03:00
if ( total_phys_segments > blk_rq_get_max_segments ( req ) )
2008-01-29 16:04:06 +03:00
return 0 ;
2022-03-15 03:30:11 +03:00
if ( ! blk_cgroup_mergeable ( req , next - > bio ) )
return 0 ;
2014-09-27 03:20:06 +04:00
if ( blk_integrity_merge_rq ( q , req , next ) = = false )
2010-09-10 22:50:10 +04:00
return 0 ;
block: Inline encryption support for blk-mq
We must have some way of letting a storage device driver know what
encryption context it should use for en/decrypting a request. However,
it's the upper layers (like the filesystem/fscrypt) that know about and
manages encryption contexts. As such, when the upper layer submits a bio
to the block layer, and this bio eventually reaches a device driver with
support for inline encryption, the device driver will need to have been
told the encryption context for that bio.
We want to communicate the encryption context from the upper layer to the
storage device along with the bio, when the bio is submitted to the block
layer. To do this, we add a struct bio_crypt_ctx to struct bio, which can
represent an encryption context (note that we can't use the bi_private
field in struct bio to do this because that field does not function to pass
information across layers in the storage stack). We also introduce various
functions to manipulate the bio_crypt_ctx and make the bio/request merging
logic aware of the bio_crypt_ctx.
We also make changes to blk-mq to make it handle bios with encryption
contexts. blk-mq can merge many bios into the same request. These bios need
to have contiguous data unit numbers (the necessary changes to blk-merge
are also made to ensure this) - as such, it suffices to keep the data unit
number of just the first bio, since that's all a storage driver needs to
infer the data unit number to use for each data block in each bio in a
request. blk-mq keeps track of the encryption context to be used for all
the bios in a request with the request's rq_crypt_ctx. When the first bio
is added to an empty request, blk-mq will program the encryption context
of that bio into the request_queue's keyslot manager, and store the
returned keyslot in the request's rq_crypt_ctx. All the functions to
operate on encryption contexts are in blk-crypto.c.
Upper layers only need to call bio_crypt_set_ctx with the encryption key,
algorithm and data_unit_num; they don't have to worry about getting a
keyslot for each encryption context, as blk-mq/blk-crypto handles that.
Blk-crypto also makes it possible for request-based layered devices like
dm-rq to make use of inline encryption hardware by cloning the
rq_crypt_ctx and programming a keyslot in the new request_queue when
necessary.
Note that any user of the block layer can submit bios with an
encryption context, such as filesystems, device-mapper targets, etc.
Signed-off-by: Satya Tangirala <satyat@google.com>
Reviewed-by: Eric Biggers <ebiggers@google.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
2020-05-14 03:37:18 +03:00
if ( ! bio_crypt_ctx_merge_rq ( req , next ) )
return 0 ;
2008-01-29 16:04:06 +03:00
/* Merge is OK... */
req - > nr_phys_segments = total_phys_segments ;
return 1 ;
}
2009-07-03 12:48:17 +04:00
/**
* blk_rq_set_mixed_merge - mark a request as mixed merge
* @ rq : request to mark as mixed merge
*
* Description :
* @ rq is about to be mixed merged . Make sure the attributes
* which can be mixed are set in each bio and mark @ rq as mixed
* merged .
*/
void blk_rq_set_mixed_merge ( struct request * rq )
{
2022-07-14 21:06:32 +03:00
blk_opf_t ff = rq - > cmd_flags & REQ_FAILFAST_MASK ;
2009-07-03 12:48:17 +04:00
struct bio * bio ;
2016-10-20 16:12:13 +03:00
if ( rq - > rq_flags & RQF_MIXED_MERGE )
2009-07-03 12:48:17 +04:00
return ;
/*
* @ rq will no longer represent mixable attributes for all the
* contained bios . It will just track those of the first one .
* Distributes the attributs to each bio .
*/
for ( bio = rq - > bio ; bio ; bio = bio - > bi_next ) {
2016-08-06 00:35:16 +03:00
WARN_ON_ONCE ( ( bio - > bi_opf & REQ_FAILFAST_MASK ) & &
( bio - > bi_opf & REQ_FAILFAST_MASK ) ! = ff ) ;
bio - > bi_opf | = ff ;
2009-07-03 12:48:17 +04:00
}
2016-10-20 16:12:13 +03:00
rq - > rq_flags | = RQF_MIXED_MERGE ;
2009-07-03 12:48:17 +04:00
}
2020-05-27 08:24:15 +03:00
static void blk_account_io_merge_request ( struct request * req )
2009-03-27 12:31:51 +03:00
{
if ( blk_do_io_stat ( req ) ) {
2018-12-06 19:41:18 +03:00
part_stat_lock ( ) ;
2020-05-27 08:24:15 +03:00
part_stat_inc ( req - > part , merges [ op_stat_group ( req_op ( req ) ) ] ) ;
2009-03-27 12:31:51 +03:00
part_stat_unlock ( ) ;
}
}
2020-05-27 08:24:15 +03:00
2018-11-15 04:19:46 +03:00
static enum elv_merge blk_try_req_merge ( struct request * req ,
struct request * next )
2018-10-27 14:52:14 +03:00
{
if ( blk_discard_mergable ( req ) )
return ELEVATOR_DISCARD_MERGE ;
else if ( blk_rq_pos ( req ) + blk_rq_sectors ( req ) = = blk_rq_pos ( next ) )
return ELEVATOR_BACK_MERGE ;
return ELEVATOR_NO_MERGE ;
}
2009-03-27 12:31:51 +03:00
2008-01-29 16:04:06 +03:00
/*
2017-02-02 18:54:40 +03:00
* For non - mq , this has to be called with the request spinlock acquired .
* For mq with scheduling , the appropriate queue wide lock should be held .
2008-01-29 16:04:06 +03:00
*/
2017-02-02 18:54:40 +03:00
static struct request * attempt_merge ( struct request_queue * q ,
struct request * req , struct request * next )
2008-01-29 16:04:06 +03:00
{
if ( ! rq_mergeable ( req ) | | ! rq_mergeable ( next ) )
2017-02-02 18:54:40 +03:00
return NULL ;
2008-01-29 16:04:06 +03:00
2016-06-09 17:00:36 +03:00
if ( req_op ( req ) ! = req_op ( next ) )
2017-02-02 18:54:40 +03:00
return NULL ;
2012-09-18 20:19:26 +04:00
2021-11-26 15:17:59 +03:00
if ( rq_data_dir ( req ) ! = rq_data_dir ( next ) )
2017-02-02 18:54:40 +03:00
return NULL ;
2008-01-29 16:04:06 +03:00
2018-11-20 04:52:37 +03:00
if ( req - > ioprio ! = next - > ioprio )
return NULL ;
2008-01-29 16:04:06 +03:00
/*
* If we are allowed to merge , then append bio list
* from next to rq and release next . merge_requests_fn
* will have updated segment counts , update sector
2018-02-02 00:01:02 +03:00
* counts here . Handle DISCARDs separately , as they
* have separate settings .
2008-01-29 16:04:06 +03:00
*/
2018-10-27 14:52:14 +03:00
switch ( blk_try_req_merge ( req , next ) ) {
case ELEVATOR_DISCARD_MERGE :
2018-02-02 00:01:02 +03:00
if ( ! req_attempt_discard_merge ( q , req , next ) )
return NULL ;
2018-10-27 14:52:14 +03:00
break ;
case ELEVATOR_BACK_MERGE :
if ( ! ll_merge_requests_fn ( q , req , next ) )
return NULL ;
break ;
default :
2017-02-02 18:54:40 +03:00
return NULL ;
2018-10-27 14:52:14 +03:00
}
2008-01-29 16:04:06 +03:00
2009-07-03 12:48:17 +04:00
/*
* If failfast settings disagree or any of the two is already
* a mixed merge , mark both as mixed before proceeding . This
* makes sure that all involved bios have mixable attributes
* set properly .
*/
2016-10-20 16:12:13 +03:00
if ( ( ( req - > rq_flags | next - > rq_flags ) & RQF_MIXED_MERGE ) | |
2009-07-03 12:48:17 +04:00
( req - > cmd_flags & REQ_FAILFAST_MASK ) ! =
( next - > cmd_flags & REQ_FAILFAST_MASK ) ) {
blk_rq_set_mixed_merge ( req ) ;
blk_rq_set_mixed_merge ( next ) ;
}
2008-01-29 16:04:06 +03:00
/*
block: consolidate struct request timestamp fields
Currently, struct request has four timestamp fields:
- A start time, set at get_request time, in jiffies, used for iostats
- An I/O start time, set at start_request time, in ktime nanoseconds,
used for blk-stats (i.e., wbt, kyber, hybrid polling)
- Another start time and another I/O start time, used for cfq and bfq
These can all be consolidated into one start time and one I/O start
time, both in ktime nanoseconds, shaving off up to 16 bytes from struct
request depending on the kernel config.
Signed-off-by: Omar Sandoval <osandov@fb.com>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
2018-05-09 12:08:53 +03:00
* At this point we have either done a back merge or front merge . We
* need the smaller start_time_ns of the merged requests to be the
* current request for accounting purposes .
2008-01-29 16:04:06 +03:00
*/
block: consolidate struct request timestamp fields
Currently, struct request has four timestamp fields:
- A start time, set at get_request time, in jiffies, used for iostats
- An I/O start time, set at start_request time, in ktime nanoseconds,
used for blk-stats (i.e., wbt, kyber, hybrid polling)
- Another start time and another I/O start time, used for cfq and bfq
These can all be consolidated into one start time and one I/O start
time, both in ktime nanoseconds, shaving off up to 16 bytes from struct
request depending on the kernel config.
Signed-off-by: Omar Sandoval <osandov@fb.com>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
2018-05-09 12:08:53 +03:00
if ( next - > start_time_ns < req - > start_time_ns )
req - > start_time_ns = next - > start_time_ns ;
2008-01-29 16:04:06 +03:00
req - > biotail - > bi_next = next - > bio ;
req - > biotail = next - > biotail ;
2009-05-07 17:24:44 +04:00
req - > __data_len + = blk_rq_bytes ( next ) ;
2008-01-29 16:04:06 +03:00
2018-11-30 19:38:18 +03:00
if ( ! blk_discard_mergable ( req ) )
2018-02-02 00:01:02 +03:00
elv_merge_requests ( q , req , next ) ;
2008-01-29 16:04:06 +03:00
2009-04-22 16:01:49 +04:00
/*
* ' next ' is going away , so update stats accordingly
*/
2020-05-27 08:24:15 +03:00
blk_account_io_merge_request ( next ) ;
2008-01-29 16:04:06 +03:00
2020-12-03 19:21:39 +03:00
trace_block_rq_merge ( next ) ;
2020-06-17 16:58:23 +03:00
2017-02-03 19:48:28 +03:00
/*
* ownership of bio passed from next to req , return ' next ' for
* the caller to free
*/
2009-03-24 14:35:07 +03:00
next - > bio = NULL ;
2017-02-02 18:54:40 +03:00
return next ;
2008-01-29 16:04:06 +03:00
}
2020-10-06 10:07:19 +03:00
static struct request * attempt_back_merge ( struct request_queue * q ,
struct request * rq )
2008-01-29 16:04:06 +03:00
{
struct request * next = elv_latter_request ( q , rq ) ;
if ( next )
return attempt_merge ( q , rq , next ) ;
2017-02-02 18:54:40 +03:00
return NULL ;
2008-01-29 16:04:06 +03:00
}
2020-10-06 10:07:19 +03:00
static struct request * attempt_front_merge ( struct request_queue * q ,
struct request * rq )
2008-01-29 16:04:06 +03:00
{
struct request * prev = elv_former_request ( q , rq ) ;
if ( prev )
return attempt_merge ( q , prev , rq ) ;
2017-02-02 18:54:40 +03:00
return NULL ;
2008-01-29 16:04:06 +03:00
}
2011-03-21 12:14:27 +03:00
2021-06-23 12:36:34 +03:00
/*
* Try to merge ' next ' into ' rq ' . Return true if the merge happened , false
* otherwise . The caller is responsible for freeing ' next ' if the merge
* happened .
*/
bool blk_attempt_req_merge ( struct request_queue * q , struct request * rq ,
struct request * next )
2011-03-21 12:14:27 +03:00
{
2021-06-23 12:36:34 +03:00
return attempt_merge ( q , rq , next ) ;
2011-03-21 12:14:27 +03:00
}
2012-02-08 12:19:38 +04:00
bool blk_rq_merge_ok ( struct request * rq , struct bio * bio )
{
2012-09-18 20:19:25 +04:00
if ( ! rq_mergeable ( rq ) | | ! bio_mergeable ( bio ) )
2012-02-08 12:19:38 +04:00
return false ;
2016-06-09 17:00:36 +03:00
if ( req_op ( rq ) ! = bio_op ( bio ) )
2012-09-18 20:19:26 +04:00
return false ;
2012-02-08 12:19:38 +04:00
/* different data direction or already started, don't merge */
if ( bio_data_dir ( bio ) ! = rq_data_dir ( rq ) )
return false ;
2022-03-15 03:30:11 +03:00
/* don't merge across cgroup boundaries */
if ( ! blk_cgroup_mergeable ( rq , bio ) )
return false ;
2012-02-08 12:19:38 +04:00
/* only merge integrity protected bio into ditto rq */
2014-09-27 03:20:06 +04:00
if ( blk_integrity_merge_bio ( rq - > q , rq , bio ) = = false )
2012-02-08 12:19:38 +04:00
return false ;
block: Inline encryption support for blk-mq
We must have some way of letting a storage device driver know what
encryption context it should use for en/decrypting a request. However,
it's the upper layers (like the filesystem/fscrypt) that know about and
manages encryption contexts. As such, when the upper layer submits a bio
to the block layer, and this bio eventually reaches a device driver with
support for inline encryption, the device driver will need to have been
told the encryption context for that bio.
We want to communicate the encryption context from the upper layer to the
storage device along with the bio, when the bio is submitted to the block
layer. To do this, we add a struct bio_crypt_ctx to struct bio, which can
represent an encryption context (note that we can't use the bi_private
field in struct bio to do this because that field does not function to pass
information across layers in the storage stack). We also introduce various
functions to manipulate the bio_crypt_ctx and make the bio/request merging
logic aware of the bio_crypt_ctx.
We also make changes to blk-mq to make it handle bios with encryption
contexts. blk-mq can merge many bios into the same request. These bios need
to have contiguous data unit numbers (the necessary changes to blk-merge
are also made to ensure this) - as such, it suffices to keep the data unit
number of just the first bio, since that's all a storage driver needs to
infer the data unit number to use for each data block in each bio in a
request. blk-mq keeps track of the encryption context to be used for all
the bios in a request with the request's rq_crypt_ctx. When the first bio
is added to an empty request, blk-mq will program the encryption context
of that bio into the request_queue's keyslot manager, and store the
returned keyslot in the request's rq_crypt_ctx. All the functions to
operate on encryption contexts are in blk-crypto.c.
Upper layers only need to call bio_crypt_set_ctx with the encryption key,
algorithm and data_unit_num; they don't have to worry about getting a
keyslot for each encryption context, as blk-mq/blk-crypto handles that.
Blk-crypto also makes it possible for request-based layered devices like
dm-rq to make use of inline encryption hardware by cloning the
rq_crypt_ctx and programming a keyslot in the new request_queue when
necessary.
Note that any user of the block layer can submit bios with an
encryption context, such as filesystems, device-mapper targets, etc.
Signed-off-by: Satya Tangirala <satyat@google.com>
Reviewed-by: Eric Biggers <ebiggers@google.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
2020-05-14 03:37:18 +03:00
/* Only merge if the crypt contexts are compatible */
if ( ! bio_crypt_rq_ctx_compatible ( rq , bio ) )
return false ;
2018-11-20 04:52:37 +03:00
if ( rq - > ioprio ! = bio_prio ( bio ) )
return false ;
2012-02-08 12:19:38 +04:00
return true ;
}
2017-02-08 16:46:48 +03:00
enum elv_merge blk_try_merge ( struct request * rq , struct bio * bio )
2012-02-08 12:19:38 +04:00
{
2018-10-27 14:52:14 +03:00
if ( blk_discard_mergable ( rq ) )
2017-02-08 16:46:49 +03:00
return ELEVATOR_DISCARD_MERGE ;
else if ( blk_rq_pos ( rq ) + blk_rq_sectors ( rq ) = = bio - > bi_iter . bi_sector )
2012-02-08 12:19:38 +04:00
return ELEVATOR_BACK_MERGE ;
2013-10-12 02:44:27 +04:00
else if ( blk_rq_pos ( rq ) - bio_sectors ( bio ) = = bio - > bi_iter . bi_sector )
2012-02-08 12:19:38 +04:00
return ELEVATOR_FRONT_MERGE ;
return ELEVATOR_NO_MERGE ;
}
2020-08-28 05:52:54 +03:00
static void blk_account_io_merge_bio ( struct request * req )
{
if ( ! blk_do_io_stat ( req ) )
return ;
part_stat_lock ( ) ;
part_stat_inc ( req - > part , merges [ op_stat_group ( req_op ( req ) ) ] ) ;
part_stat_unlock ( ) ;
}
2020-10-06 10:07:19 +03:00
enum bio_merge_status {
BIO_MERGE_OK ,
BIO_MERGE_NONE ,
BIO_MERGE_FAILED ,
} ;
static enum bio_merge_status bio_attempt_back_merge ( struct request * req ,
struct bio * bio , unsigned int nr_segs )
2020-08-28 05:52:54 +03:00
{
2022-07-14 21:06:32 +03:00
const blk_opf_t ff = bio - > bi_opf & REQ_FAILFAST_MASK ;
2020-08-28 05:52:54 +03:00
if ( ! ll_back_merge_fn ( req , bio , nr_segs ) )
2020-08-28 05:52:56 +03:00
return BIO_MERGE_FAILED ;
2020-08-28 05:52:54 +03:00
2020-12-03 19:21:36 +03:00
trace_block_bio_backmerge ( bio ) ;
2020-08-28 05:52:54 +03:00
rq_qos_merge ( req - > q , req , bio ) ;
if ( ( req - > cmd_flags & REQ_FAILFAST_MASK ) ! = ff )
blk_rq_set_mixed_merge ( req ) ;
req - > biotail - > bi_next = bio ;
req - > biotail = bio ;
req - > __data_len + = bio - > bi_iter . bi_size ;
bio_crypt_free_ctx ( bio ) ;
blk_account_io_merge_bio ( req ) ;
2020-08-28 05:52:56 +03:00
return BIO_MERGE_OK ;
2020-08-28 05:52:54 +03:00
}
2020-10-06 10:07:19 +03:00
static enum bio_merge_status bio_attempt_front_merge ( struct request * req ,
struct bio * bio , unsigned int nr_segs )
2020-08-28 05:52:54 +03:00
{
2022-07-14 21:06:32 +03:00
const blk_opf_t ff = bio - > bi_opf & REQ_FAILFAST_MASK ;
2020-08-28 05:52:54 +03:00
if ( ! ll_front_merge_fn ( req , bio , nr_segs ) )
2020-08-28 05:52:56 +03:00
return BIO_MERGE_FAILED ;
2020-08-28 05:52:54 +03:00
2020-12-03 19:21:36 +03:00
trace_block_bio_frontmerge ( bio ) ;
2020-08-28 05:52:54 +03:00
rq_qos_merge ( req - > q , req , bio ) ;
if ( ( req - > cmd_flags & REQ_FAILFAST_MASK ) ! = ff )
blk_rq_set_mixed_merge ( req ) ;
bio - > bi_next = req - > bio ;
req - > bio = bio ;
req - > __sector = bio - > bi_iter . bi_sector ;
req - > __data_len + = bio - > bi_iter . bi_size ;
bio_crypt_do_front_merge ( req , bio ) ;
blk_account_io_merge_bio ( req ) ;
2020-08-28 05:52:56 +03:00
return BIO_MERGE_OK ;
2020-08-28 05:52:54 +03:00
}
2020-10-06 10:07:19 +03:00
static enum bio_merge_status bio_attempt_discard_merge ( struct request_queue * q ,
struct request * req , struct bio * bio )
2020-08-28 05:52:54 +03:00
{
unsigned short segments = blk_rq_nr_discard_segments ( req ) ;
if ( segments > = queue_max_discard_segments ( q ) )
goto no_merge ;
if ( blk_rq_sectors ( req ) + bio_sectors ( bio ) >
blk_rq_get_max_sectors ( req , blk_rq_pos ( req ) ) )
goto no_merge ;
rq_qos_merge ( q , req , bio ) ;
req - > biotail - > bi_next = bio ;
req - > biotail = bio ;
req - > __data_len + = bio - > bi_iter . bi_size ;
req - > nr_phys_segments = segments + 1 ;
blk_account_io_merge_bio ( req ) ;
2020-08-28 05:52:56 +03:00
return BIO_MERGE_OK ;
2020-08-28 05:52:54 +03:00
no_merge :
req_set_nomerge ( q , req ) ;
2020-08-28 05:52:56 +03:00
return BIO_MERGE_FAILED ;
}
static enum bio_merge_status blk_attempt_bio_merge ( struct request_queue * q ,
struct request * rq ,
struct bio * bio ,
unsigned int nr_segs ,
bool sched_allow_merge )
{
if ( ! blk_rq_merge_ok ( rq , bio ) )
return BIO_MERGE_NONE ;
switch ( blk_try_merge ( rq , bio ) ) {
case ELEVATOR_BACK_MERGE :
2020-09-02 04:45:25 +03:00
if ( ! sched_allow_merge | | blk_mq_sched_allow_merge ( q , rq , bio ) )
2020-08-28 05:52:56 +03:00
return bio_attempt_back_merge ( rq , bio , nr_segs ) ;
break ;
case ELEVATOR_FRONT_MERGE :
2020-09-02 04:45:25 +03:00
if ( ! sched_allow_merge | | blk_mq_sched_allow_merge ( q , rq , bio ) )
2020-08-28 05:52:56 +03:00
return bio_attempt_front_merge ( rq , bio , nr_segs ) ;
break ;
case ELEVATOR_DISCARD_MERGE :
return bio_attempt_discard_merge ( q , rq , bio ) ;
default :
return BIO_MERGE_NONE ;
}
return BIO_MERGE_FAILED ;
2020-08-28 05:52:54 +03:00
}
/**
* blk_attempt_plug_merge - try to merge with % current ' s plugged list
* @ q : request_queue new bio is being queued at
* @ bio : new bio being queued
* @ nr_segs : number of segments in @ bio
2021-10-18 19:07:09 +03:00
* from the passed in @ q already in the plug list
2020-08-28 05:52:54 +03:00
*
block: only check previous entry for plug merge attempt
Currently we scan the entire plug list, which is potentially very
expensive. In an IOPS bound workload, we can drive about 5.6M IOPS with
merging enabled, and profiling shows that the plug merge check is the
(by far) most expensive thing we're doing:
Overhead Command Shared Object Symbol
+ 20.89% io_uring [kernel.vmlinux] [k] blk_attempt_plug_merge
+ 4.98% io_uring [kernel.vmlinux] [k] io_submit_sqes
+ 4.78% io_uring [kernel.vmlinux] [k] blkdev_direct_IO
+ 4.61% io_uring [kernel.vmlinux] [k] blk_mq_submit_bio
Instead of browsing the whole list, just check the previously inserted
entry. That is enough for a naive merge check and will catch most cases,
and for devices that need full merging, the IO scheduler attached to
such devices will do that anyway. The plug merge is meant to be an
inexpensive check to avoid getting a request, but if we repeatedly
scan the list for every single insert, it is very much not a cheap
check.
With this patch, the workload instead runs at ~7.0M IOPS, providing
a 25% improvement. Disabling merging entirely yields another 5%
improvement.
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
2021-10-14 16:24:07 +03:00
* Determine whether @ bio being queued on @ q can be merged with the previous
* request on % current ' s plugged list . Returns % true if merge was successful ,
2020-08-28 05:52:54 +03:00
* otherwise % false .
*
* Plugging coalesces IOs from the same issuer for the same purpose without
* going through @ q - > queue_lock . As such it ' s more of an issuing mechanism
* than scheduling , and the request , while may have elvpriv data , is not
* added on the elevator at this point . In addition , we don ' t have
* reliable access to the elevator outside queue lock . Only check basic
* merging parameters without querying the elevator .
*
* Caller must ensure ! blk_queue_nomerges ( q ) beforehand .
*/
bool blk_attempt_plug_merge ( struct request_queue * q , struct bio * bio ,
2021-11-23 19:04:41 +03:00
unsigned int nr_segs )
2020-08-28 05:52:54 +03:00
{
struct blk_plug * plug ;
struct request * rq ;
2022-07-06 10:03:38 +03:00
plug = blk_mq_plug ( bio ) ;
2021-10-18 19:12:12 +03:00
if ( ! plug | | rq_list_empty ( plug - > mq_list ) )
2020-08-28 05:52:54 +03:00
return false ;
2022-03-11 20:21:43 +03:00
rq_list_for_each ( & plug - > mq_list , rq ) {
if ( rq - > q = = q ) {
if ( blk_attempt_bio_merge ( q , rq , bio , nr_segs , false ) = =
BIO_MERGE_OK )
return true ;
break ;
}
/*
* Only keep iterating plug list for merges if we have multiple
* queues
*/
if ( ! plug - > multiple_queues )
break ;
2020-08-28 05:52:54 +03:00
}
return false ;
}
2020-08-28 05:52:55 +03:00
/*
* Iterate list of requests and see if we can merge this bio with any
* of them .
*/
bool blk_bio_list_merge ( struct request_queue * q , struct list_head * list ,
struct bio * bio , unsigned int nr_segs )
{
struct request * rq ;
int checked = 8 ;
list_for_each_entry_reverse ( rq , list , queuelist ) {
if ( ! checked - - )
break ;
2020-08-28 05:52:56 +03:00
switch ( blk_attempt_bio_merge ( q , rq , bio , nr_segs , true ) ) {
case BIO_MERGE_NONE :
2020-08-28 05:52:55 +03:00
continue ;
2020-08-28 05:52:56 +03:00
case BIO_MERGE_OK :
return true ;
case BIO_MERGE_FAILED :
return false ;
2020-08-28 05:52:55 +03:00
}
}
return false ;
}
EXPORT_SYMBOL_GPL ( blk_bio_list_merge ) ;
2020-10-06 10:07:19 +03:00
bool blk_mq_sched_try_merge ( struct request_queue * q , struct bio * bio ,
unsigned int nr_segs , struct request * * merged_request )
{
struct request * rq ;
switch ( elv_merge ( q , & rq , bio ) ) {
case ELEVATOR_BACK_MERGE :
if ( ! blk_mq_sched_allow_merge ( q , rq , bio ) )
return false ;
if ( bio_attempt_back_merge ( rq , bio , nr_segs ) ! = BIO_MERGE_OK )
return false ;
* merged_request = attempt_back_merge ( q , rq ) ;
if ( ! * merged_request )
elv_merged_request ( q , rq , ELEVATOR_BACK_MERGE ) ;
return true ;
case ELEVATOR_FRONT_MERGE :
if ( ! blk_mq_sched_allow_merge ( q , rq , bio ) )
return false ;
if ( bio_attempt_front_merge ( rq , bio , nr_segs ) ! = BIO_MERGE_OK )
return false ;
* merged_request = attempt_front_merge ( q , rq ) ;
if ( ! * merged_request )
elv_merged_request ( q , rq , ELEVATOR_FRONT_MERGE ) ;
return true ;
case ELEVATOR_DISCARD_MERGE :
return bio_attempt_discard_merge ( q , rq , bio ) = = BIO_MERGE_OK ;
default :
return false ;
}
}
EXPORT_SYMBOL_GPL ( blk_mq_sched_try_merge ) ;