License cleanup: add SPDX GPL-2.0 license identifier to files with no license
Many source files in the tree are missing licensing information, which
makes it harder for compliance tools to determine the correct license.
By default all files without license information are under the default
license of the kernel, which is GPL version 2.
Update the files which contain no license information with the 'GPL-2.0'
SPDX license identifier. The SPDX identifier is a legally binding
shorthand, which can be used instead of the full boiler plate text.
This patch is based on work done by Thomas Gleixner and Kate Stewart and
Philippe Ombredanne.
How this work was done:
Patches were generated and checked against linux-4.14-rc6 for a subset of
the use cases:
- file had no licensing information it it.
- file was a */uapi/* one with no licensing information in it,
- file was a */uapi/* one with existing licensing information,
Further patches will be generated in subsequent months to fix up cases
where non-standard license headers were used, and references to license
had to be inferred by heuristics based on keywords.
The analysis to determine which SPDX License Identifier to be applied to
a file was done in a spreadsheet of side by side results from of the
output of two independent scanners (ScanCode & Windriver) producing SPDX
tag:value files created by Philippe Ombredanne. Philippe prepared the
base worksheet, and did an initial spot review of a few 1000 files.
The 4.13 kernel was the starting point of the analysis with 60,537 files
assessed. Kate Stewart did a file by file comparison of the scanner
results in the spreadsheet to determine which SPDX license identifier(s)
to be applied to the file. She confirmed any determination that was not
immediately clear with lawyers working with the Linux Foundation.
Criteria used to select files for SPDX license identifier tagging was:
- Files considered eligible had to be source code files.
- Make and config files were included as candidates if they contained >5
lines of source
- File already had some variant of a license header in it (even if <5
lines).
All documentation files were explicitly excluded.
The following heuristics were used to determine which SPDX license
identifiers to apply.
- when both scanners couldn't find any license traces, file was
considered to have no license information in it, and the top level
COPYING file license applied.
For non */uapi/* files that summary was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 11139
and resulted in the first patch in this series.
If that file was a */uapi/* path one, it was "GPL-2.0 WITH
Linux-syscall-note" otherwise it was "GPL-2.0". Results of that was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 WITH Linux-syscall-note 930
and resulted in the second patch in this series.
- if a file had some form of licensing information in it, and was one
of the */uapi/* ones, it was denoted with the Linux-syscall-note if
any GPL family license was found in the file or had no licensing in
it (per prior point). Results summary:
SPDX license identifier # files
---------------------------------------------------|------
GPL-2.0 WITH Linux-syscall-note 270
GPL-2.0+ WITH Linux-syscall-note 169
((GPL-2.0 WITH Linux-syscall-note) OR BSD-2-Clause) 21
((GPL-2.0 WITH Linux-syscall-note) OR BSD-3-Clause) 17
LGPL-2.1+ WITH Linux-syscall-note 15
GPL-1.0+ WITH Linux-syscall-note 14
((GPL-2.0+ WITH Linux-syscall-note) OR BSD-3-Clause) 5
LGPL-2.0+ WITH Linux-syscall-note 4
LGPL-2.1 WITH Linux-syscall-note 3
((GPL-2.0 WITH Linux-syscall-note) OR MIT) 3
((GPL-2.0 WITH Linux-syscall-note) AND MIT) 1
and that resulted in the third patch in this series.
- when the two scanners agreed on the detected license(s), that became
the concluded license(s).
- when there was disagreement between the two scanners (one detected a
license but the other didn't, or they both detected different
licenses) a manual inspection of the file occurred.
- In most cases a manual inspection of the information in the file
resulted in a clear resolution of the license that should apply (and
which scanner probably needed to revisit its heuristics).
- When it was not immediately clear, the license identifier was
confirmed with lawyers working with the Linux Foundation.
- If there was any question as to the appropriate license identifier,
the file was flagged for further research and to be revisited later
in time.
In total, over 70 hours of logged manual review was done on the
spreadsheet to determine the SPDX license identifiers to apply to the
source files by Kate, Philippe, Thomas and, in some cases, confirmation
by lawyers working with the Linux Foundation.
Kate also obtained a third independent scan of the 4.13 code base from
FOSSology, and compared selected files where the other two scanners
disagreed against that SPDX file, to see if there was new insights. The
Windriver scanner is based on an older version of FOSSology in part, so
they are related.
Thomas did random spot checks in about 500 files from the spreadsheets
for the uapi headers and agreed with SPDX license identifier in the
files he inspected. For the non-uapi files Thomas did random spot checks
in about 15000 files.
In initial set of patches against 4.14-rc6, 3 files were found to have
copy/paste license identifier errors, and have been fixed to reflect the
correct identifier.
Additionally Philippe spent 10 hours this week doing a detailed manual
inspection and review of the 12,461 patched files from the initial patch
version early this week with:
- a full scancode scan run, collecting the matched texts, detected
license ids and scores
- reviewing anything where there was a license detected (about 500+
files) to ensure that the applied SPDX license was correct
- reviewing anything where there was no detection but the patch license
was not GPL-2.0 WITH Linux-syscall-note to ensure that the applied
SPDX license was correct
This produced a worksheet with 20 files needing minor correction. This
worksheet was then exported into 3 different .csv files for the
different types of files to be modified.
These .csv files were then reviewed by Greg. Thomas wrote a script to
parse the csv files and add the proper SPDX tag to the file, in the
format that the file expected. This script was further refined by Greg
based on the output to detect more types of files automatically and to
distinguish between header and source .c files (which need different
comment types.) Finally Greg ran the script using the .csv files to
generate the patches.
Reviewed-by: Kate Stewart <kstewart@linuxfoundation.org>
Reviewed-by: Philippe Ombredanne <pombredanne@nexb.com>
Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
2017-11-01 15:07:57 +01:00
|
|
|
// SPDX-License-Identifier: GPL-2.0
|
2008-01-29 14:04:06 +01:00
|
|
|
/*
|
|
|
|
* Functions related to segment and merge handling
|
|
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <linux/module.h>
|
|
|
|
#include <linux/bio.h>
|
|
|
|
#include <linux/blkdev.h>
|
|
|
|
#include <linux/scatterlist.h>
|
|
|
|
|
2015-12-03 15:32:30 +01:00
|
|
|
#include <trace/events/block.h>
|
|
|
|
|
2008-01-29 14:04:06 +01:00
|
|
|
#include "blk.h"
|
|
|
|
|
2018-09-24 09:43:48 +02:00
|
|
|
static inline bool bio_will_gap(struct request_queue *q,
|
|
|
|
struct request *prev_rq, struct bio *prev, struct bio *next)
|
|
|
|
{
|
|
|
|
struct bio_vec pb, nb;
|
|
|
|
|
|
|
|
if (!bio_has_data(prev) || !queue_virt_boundary(q))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Don't merge if the 1st bio starts with non-zero offset, otherwise it
|
|
|
|
* is quite difficult to respect the sg gap limit. We work hard to
|
|
|
|
* merge a huge number of small single bios in case of mkfs.
|
|
|
|
*/
|
|
|
|
if (prev_rq)
|
|
|
|
bio_get_first_bvec(prev_rq->bio, &pb);
|
|
|
|
else
|
|
|
|
bio_get_first_bvec(prev, &pb);
|
2018-11-07 14:58:14 +01:00
|
|
|
if (pb.bv_offset & queue_virt_boundary(q))
|
2018-09-24 09:43:48 +02:00
|
|
|
return true;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We don't need to worry about the situation that the merged segment
|
|
|
|
* ends in unaligned virt boundary:
|
|
|
|
*
|
|
|
|
* - if 'pb' ends aligned, the merged segment ends aligned
|
|
|
|
* - if 'pb' ends unaligned, the next bio must include
|
|
|
|
* one single bvec of 'nb', otherwise the 'nb' can't
|
|
|
|
* merge with 'pb'
|
|
|
|
*/
|
|
|
|
bio_get_last_bvec(prev, &pb);
|
|
|
|
bio_get_first_bvec(next, &nb);
|
2019-05-21 09:01:42 +02:00
|
|
|
if (biovec_phys_mergeable(q, &pb, &nb))
|
2018-09-24 09:43:48 +02:00
|
|
|
return false;
|
|
|
|
return __bvec_gap_to_prev(q, &pb, nb.bv_offset);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline bool req_gap_back_merge(struct request *req, struct bio *bio)
|
|
|
|
{
|
|
|
|
return bio_will_gap(req->q, req, req->biotail, bio);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline bool req_gap_front_merge(struct request *req, struct bio *bio)
|
|
|
|
{
|
|
|
|
return bio_will_gap(req->q, NULL, bio, req->bio);
|
|
|
|
}
|
|
|
|
|
2015-04-24 07:37:18 +02:00
|
|
|
static struct bio *blk_bio_discard_split(struct request_queue *q,
|
|
|
|
struct bio *bio,
|
2015-10-20 17:13:52 +02:00
|
|
|
struct bio_set *bs,
|
|
|
|
unsigned *nsegs)
|
2015-04-24 07:37:18 +02:00
|
|
|
{
|
|
|
|
unsigned int max_discard_sectors, granularity;
|
|
|
|
int alignment;
|
|
|
|
sector_t tmp;
|
|
|
|
unsigned split_sectors;
|
|
|
|
|
2015-10-20 17:13:52 +02:00
|
|
|
*nsegs = 1;
|
|
|
|
|
2015-04-24 07:37:18 +02:00
|
|
|
/* Zero-sector (unknown) and one-sector granularities are the same. */
|
|
|
|
granularity = max(q->limits.discard_granularity >> 9, 1U);
|
|
|
|
|
2018-10-29 13:57:17 +01:00
|
|
|
max_discard_sectors = min(q->limits.max_discard_sectors,
|
|
|
|
bio_allowed_max_sectors(q));
|
2015-04-24 07:37:18 +02:00
|
|
|
max_discard_sectors -= max_discard_sectors % granularity;
|
|
|
|
|
|
|
|
if (unlikely(!max_discard_sectors)) {
|
|
|
|
/* XXX: warn */
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (bio_sectors(bio) <= max_discard_sectors)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
split_sectors = max_discard_sectors;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If the next starting sector would be misaligned, stop the discard at
|
|
|
|
* the previous aligned sector.
|
|
|
|
*/
|
|
|
|
alignment = (q->limits.discard_alignment >> 9) % granularity;
|
|
|
|
|
|
|
|
tmp = bio->bi_iter.bi_sector + split_sectors - alignment;
|
|
|
|
tmp = sector_div(tmp, granularity);
|
|
|
|
|
|
|
|
if (split_sectors > tmp)
|
|
|
|
split_sectors -= tmp;
|
|
|
|
|
|
|
|
return bio_split(bio, split_sectors, GFP_NOIO, bs);
|
|
|
|
}
|
|
|
|
|
2017-04-05 19:21:01 +02:00
|
|
|
static struct bio *blk_bio_write_zeroes_split(struct request_queue *q,
|
|
|
|
struct bio *bio, struct bio_set *bs, unsigned *nsegs)
|
|
|
|
{
|
|
|
|
*nsegs = 1;
|
|
|
|
|
|
|
|
if (!q->limits.max_write_zeroes_sectors)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
if (bio_sectors(bio) <= q->limits.max_write_zeroes_sectors)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
return bio_split(bio, q->limits.max_write_zeroes_sectors, GFP_NOIO, bs);
|
|
|
|
}
|
|
|
|
|
2015-04-24 07:37:18 +02:00
|
|
|
static struct bio *blk_bio_write_same_split(struct request_queue *q,
|
|
|
|
struct bio *bio,
|
2015-10-20 17:13:52 +02:00
|
|
|
struct bio_set *bs,
|
|
|
|
unsigned *nsegs)
|
2015-04-24 07:37:18 +02:00
|
|
|
{
|
2015-10-20 17:13:52 +02:00
|
|
|
*nsegs = 1;
|
|
|
|
|
2015-04-24 07:37:18 +02:00
|
|
|
if (!q->limits.max_write_same_sectors)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
if (bio_sectors(bio) <= q->limits.max_write_same_sectors)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
return bio_split(bio, q->limits.max_write_same_sectors, GFP_NOIO, bs);
|
|
|
|
}
|
|
|
|
|
2016-01-23 01:05:33 +01:00
|
|
|
static inline unsigned get_max_io_size(struct request_queue *q,
|
|
|
|
struct bio *bio)
|
|
|
|
{
|
|
|
|
unsigned sectors = blk_max_size_offset(q, bio->bi_iter.bi_sector);
|
|
|
|
unsigned mask = queue_logical_block_size(q) - 1;
|
|
|
|
|
|
|
|
/* aligned to logical block size */
|
|
|
|
sectors &= ~(mask >> 9);
|
|
|
|
|
|
|
|
return sectors;
|
|
|
|
}
|
|
|
|
|
2019-02-15 12:13:12 +01:00
|
|
|
static unsigned get_max_segment_size(struct request_queue *q,
|
|
|
|
unsigned offset)
|
|
|
|
{
|
|
|
|
unsigned long mask = queue_segment_boundary(q);
|
|
|
|
|
|
|
|
/* default segment boundary mask means no boundary limit */
|
|
|
|
if (mask == BLK_SEG_BOUNDARY_MASK)
|
|
|
|
return queue_max_segment_size(q);
|
|
|
|
|
|
|
|
return min_t(unsigned long, mask - (mask & offset) + 1,
|
|
|
|
queue_max_segment_size(q));
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Split the bvec @bv into segments, and update all kinds of
|
|
|
|
* variables.
|
|
|
|
*/
|
|
|
|
static bool bvec_split_segs(struct request_queue *q, struct bio_vec *bv,
|
2019-05-21 09:01:43 +02:00
|
|
|
unsigned *nsegs, unsigned *sectors, unsigned max_segs)
|
2019-02-15 12:13:12 +01:00
|
|
|
{
|
|
|
|
unsigned len = bv->bv_len;
|
|
|
|
unsigned total_len = 0;
|
|
|
|
unsigned new_nsegs = 0, seg_size = 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Multi-page bvec may be too big to hold in one segment, so the
|
|
|
|
* current bvec has to be splitted as multiple segments.
|
|
|
|
*/
|
2019-03-03 14:17:48 +01:00
|
|
|
while (len && new_nsegs + *nsegs < max_segs) {
|
2019-02-15 12:13:12 +01:00
|
|
|
seg_size = get_max_segment_size(q, bv->bv_offset + total_len);
|
|
|
|
seg_size = min(seg_size, len);
|
|
|
|
|
|
|
|
new_nsegs++;
|
|
|
|
total_len += seg_size;
|
|
|
|
len -= seg_size;
|
|
|
|
|
|
|
|
if ((bv->bv_offset + total_len) & queue_virt_boundary(q))
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2019-05-21 09:01:43 +02:00
|
|
|
if (new_nsegs) {
|
|
|
|
*nsegs += new_nsegs;
|
|
|
|
if (sectors)
|
|
|
|
*sectors += total_len >> 9;
|
2019-02-15 12:13:12 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
/* split in the middle of the bvec if len != 0 */
|
|
|
|
return !!len;
|
|
|
|
}
|
|
|
|
|
2015-04-24 07:37:18 +02:00
|
|
|
static struct bio *blk_bio_segment_split(struct request_queue *q,
|
|
|
|
struct bio *bio,
|
2015-10-20 17:13:52 +02:00
|
|
|
struct bio_set *bs,
|
|
|
|
unsigned *segs)
|
2015-04-24 07:37:18 +02:00
|
|
|
{
|
2015-09-03 00:46:02 +02:00
|
|
|
struct bio_vec bv, bvprv, *bvprvp = NULL;
|
2015-04-24 07:37:18 +02:00
|
|
|
struct bvec_iter iter;
|
2019-05-21 09:01:43 +02:00
|
|
|
unsigned nsegs = 0, sectors = 0;
|
2015-11-24 03:35:30 +01:00
|
|
|
bool do_split = true;
|
|
|
|
struct bio *new = NULL;
|
2016-01-23 01:05:33 +01:00
|
|
|
const unsigned max_sectors = get_max_io_size(q, bio);
|
2019-03-03 14:17:48 +01:00
|
|
|
const unsigned max_segs = queue_max_segments(q);
|
2015-04-24 07:37:18 +02:00
|
|
|
|
2019-02-15 12:13:12 +01:00
|
|
|
bio_for_each_bvec(bv, bio, iter) {
|
2015-04-24 07:37:18 +02:00
|
|
|
/*
|
|
|
|
* If the queue doesn't support SG gaps and adding this
|
|
|
|
* offset would create a gap, disallow it.
|
|
|
|
*/
|
2015-09-03 00:46:02 +02:00
|
|
|
if (bvprvp && bvec_gap_to_prev(q, bvprvp, bv.bv_offset))
|
2015-04-24 07:37:18 +02:00
|
|
|
goto split;
|
|
|
|
|
2016-01-23 01:05:33 +01:00
|
|
|
if (sectors + (bv.bv_len >> 9) > max_sectors) {
|
2016-01-12 23:08:39 +01:00
|
|
|
/*
|
|
|
|
* Consider this a new segment if we're splitting in
|
|
|
|
* the middle of this vector.
|
|
|
|
*/
|
2019-03-03 14:17:48 +01:00
|
|
|
if (nsegs < max_segs &&
|
2016-01-23 01:05:33 +01:00
|
|
|
sectors < max_sectors) {
|
2019-02-15 12:13:12 +01:00
|
|
|
/* split in the middle of bvec */
|
|
|
|
bv.bv_len = (max_sectors - sectors) << 9;
|
|
|
|
bvec_split_segs(q, &bv, &nsegs,
|
2019-03-03 14:17:48 +01:00
|
|
|
§ors, max_segs);
|
2016-01-12 23:08:39 +01:00
|
|
|
}
|
2017-12-18 13:22:16 +01:00
|
|
|
goto split;
|
2016-01-12 23:08:39 +01:00
|
|
|
}
|
|
|
|
|
2019-03-03 14:17:48 +01:00
|
|
|
if (nsegs == max_segs)
|
2015-04-24 07:37:18 +02:00
|
|
|
goto split;
|
|
|
|
|
|
|
|
bvprv = bv;
|
2015-11-24 03:35:29 +01:00
|
|
|
bvprvp = &bvprv;
|
2019-02-15 12:13:12 +01:00
|
|
|
|
2019-02-27 13:40:12 +01:00
|
|
|
if (bv.bv_offset + bv.bv_len <= PAGE_SIZE) {
|
|
|
|
nsegs++;
|
|
|
|
sectors += bv.bv_len >> 9;
|
2019-05-21 09:01:43 +02:00
|
|
|
} else if (bvec_split_segs(q, &bv, &nsegs, §ors,
|
|
|
|
max_segs)) {
|
2019-02-15 12:13:12 +01:00
|
|
|
goto split;
|
2019-02-27 13:40:12 +01:00
|
|
|
}
|
2015-04-24 07:37:18 +02:00
|
|
|
}
|
|
|
|
|
2015-11-24 03:35:30 +01:00
|
|
|
do_split = false;
|
2015-04-24 07:37:18 +02:00
|
|
|
split:
|
2015-10-20 17:13:52 +02:00
|
|
|
*segs = nsegs;
|
2015-11-24 03:35:30 +01:00
|
|
|
|
|
|
|
if (do_split) {
|
|
|
|
new = bio_split(bio, sectors, GFP_NOIO, bs);
|
|
|
|
if (new)
|
|
|
|
bio = new;
|
|
|
|
}
|
|
|
|
|
|
|
|
return do_split ? new : NULL;
|
2015-04-24 07:37:18 +02:00
|
|
|
}
|
|
|
|
|
2017-06-18 06:38:57 +02:00
|
|
|
void blk_queue_split(struct request_queue *q, struct bio **bio)
|
2015-04-24 07:37:18 +02:00
|
|
|
{
|
2015-10-20 17:13:52 +02:00
|
|
|
struct bio *split, *res;
|
|
|
|
unsigned nsegs;
|
2015-04-24 07:37:18 +02:00
|
|
|
|
2016-08-16 09:59:35 +02:00
|
|
|
switch (bio_op(*bio)) {
|
|
|
|
case REQ_OP_DISCARD:
|
|
|
|
case REQ_OP_SECURE_ERASE:
|
2018-05-21 00:25:47 +02:00
|
|
|
split = blk_bio_discard_split(q, *bio, &q->bio_split, &nsegs);
|
2016-08-16 09:59:35 +02:00
|
|
|
break;
|
2016-11-30 21:28:59 +01:00
|
|
|
case REQ_OP_WRITE_ZEROES:
|
2018-05-21 00:25:47 +02:00
|
|
|
split = blk_bio_write_zeroes_split(q, *bio, &q->bio_split, &nsegs);
|
2016-11-30 21:28:59 +01:00
|
|
|
break;
|
2016-08-16 09:59:35 +02:00
|
|
|
case REQ_OP_WRITE_SAME:
|
2018-05-21 00:25:47 +02:00
|
|
|
split = blk_bio_write_same_split(q, *bio, &q->bio_split, &nsegs);
|
2016-08-16 09:59:35 +02:00
|
|
|
break;
|
|
|
|
default:
|
2018-05-21 00:25:47 +02:00
|
|
|
split = blk_bio_segment_split(q, *bio, &q->bio_split, &nsegs);
|
2016-08-16 09:59:35 +02:00
|
|
|
break;
|
|
|
|
}
|
2015-10-20 17:13:52 +02:00
|
|
|
|
|
|
|
/* physical segments can be figured out during splitting */
|
|
|
|
res = split ? split : *bio;
|
|
|
|
res->bi_phys_segments = nsegs;
|
|
|
|
bio_set_flag(res, BIO_SEG_VALID);
|
2015-04-24 07:37:18 +02:00
|
|
|
|
|
|
|
if (split) {
|
2015-10-20 17:13:53 +02:00
|
|
|
/* there isn't chance to merge the splitted bio */
|
2016-08-05 23:35:16 +02:00
|
|
|
split->bi_opf |= REQ_NOMERGE;
|
2015-10-20 17:13:53 +02:00
|
|
|
|
2019-01-27 14:35:28 +01:00
|
|
|
/*
|
|
|
|
* Since we're recursing into make_request here, ensure
|
|
|
|
* that we mark this bio as already having entered the queue.
|
|
|
|
* If not, and the queue is going away, we can get stuck
|
|
|
|
* forever on waiting for the queue reference to drop. But
|
|
|
|
* that will never happen, as we're already holding a
|
|
|
|
* reference to it.
|
|
|
|
*/
|
|
|
|
bio_set_flag(*bio, BIO_QUEUE_ENTERED);
|
|
|
|
|
2015-04-24 07:37:18 +02:00
|
|
|
bio_chain(split, *bio);
|
2015-12-03 15:32:30 +01:00
|
|
|
trace_block_split(q, split, (*bio)->bi_iter.bi_sector);
|
2015-04-24 07:37:18 +02:00
|
|
|
generic_make_request(*bio);
|
|
|
|
*bio = split;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(blk_queue_split);
|
|
|
|
|
2009-02-23 09:03:10 +01:00
|
|
|
static unsigned int __blk_recalc_rq_segments(struct request_queue *q,
|
2019-02-15 12:13:23 +01:00
|
|
|
struct bio *bio)
|
2008-01-29 14:04:06 +01:00
|
|
|
{
|
2019-05-21 09:01:43 +02:00
|
|
|
unsigned int nr_phys_segs = 0;
|
2013-11-24 02:19:00 +01:00
|
|
|
struct bvec_iter iter;
|
2019-05-21 09:01:43 +02:00
|
|
|
struct bio_vec bv;
|
2008-01-29 14:04:06 +01:00
|
|
|
|
2009-02-23 09:03:10 +01:00
|
|
|
if (!bio)
|
|
|
|
return 0;
|
2008-01-29 14:04:06 +01:00
|
|
|
|
2016-11-30 21:28:59 +01:00
|
|
|
switch (bio_op(bio)) {
|
|
|
|
case REQ_OP_DISCARD:
|
|
|
|
case REQ_OP_SECURE_ERASE:
|
|
|
|
case REQ_OP_WRITE_ZEROES:
|
2016-12-08 23:20:32 +01:00
|
|
|
return 0;
|
|
|
|
case REQ_OP_WRITE_SAME:
|
2014-02-07 21:53:46 +01:00
|
|
|
return 1;
|
2016-11-30 21:28:59 +01:00
|
|
|
}
|
2014-02-07 21:53:46 +01:00
|
|
|
|
2009-02-23 09:03:10 +01:00
|
|
|
for_each_bio(bio) {
|
2019-05-21 09:01:43 +02:00
|
|
|
bio_for_each_bvec(bv, bio, iter)
|
|
|
|
bvec_split_segs(q, &bv, &nr_phys_segs, NULL, UINT_MAX);
|
2008-01-29 14:04:06 +01:00
|
|
|
}
|
|
|
|
|
2009-02-23 09:03:10 +01:00
|
|
|
return nr_phys_segs;
|
|
|
|
}
|
|
|
|
|
|
|
|
void blk_recalc_rq_segments(struct request *rq)
|
|
|
|
{
|
2019-02-15 12:13:23 +01:00
|
|
|
rq->nr_phys_segments = __blk_recalc_rq_segments(rq->q, rq->bio);
|
2008-01-29 14:04:06 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
void blk_recount_segments(struct request_queue *q, struct bio *bio)
|
|
|
|
{
|
2019-02-15 12:13:23 +01:00
|
|
|
struct bio *nxt = bio->bi_next;
|
2014-05-29 17:53:32 +02:00
|
|
|
|
2019-02-15 12:13:23 +01:00
|
|
|
bio->bi_next = NULL;
|
|
|
|
bio->bi_phys_segments = __blk_recalc_rq_segments(q, bio);
|
|
|
|
bio->bi_next = nxt;
|
2009-02-23 09:03:10 +01:00
|
|
|
|
2015-07-24 20:37:59 +02:00
|
|
|
bio_set_flag(bio, BIO_SEG_VALID);
|
2008-01-29 14:04:06 +01:00
|
|
|
}
|
|
|
|
|
2019-02-27 13:40:11 +01:00
|
|
|
static inline struct scatterlist *blk_next_sg(struct scatterlist **sg,
|
2019-02-15 12:13:13 +01:00
|
|
|
struct scatterlist *sglist)
|
|
|
|
{
|
|
|
|
if (!*sg)
|
|
|
|
return sglist;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If the driver previously mapped a shorter list, we could see a
|
|
|
|
* termination bit prematurely unless it fully inits the sg table
|
|
|
|
* on each mapping. We KNOW that there must be more entries here
|
|
|
|
* or the driver would be buggy, so force clear the termination bit
|
|
|
|
* to avoid doing a full sg_init_table() in drivers for each command.
|
|
|
|
*/
|
|
|
|
sg_unmark_end(*sg);
|
|
|
|
return sg_next(*sg);
|
|
|
|
}
|
|
|
|
|
|
|
|
static unsigned blk_bvec_map_sg(struct request_queue *q,
|
|
|
|
struct bio_vec *bvec, struct scatterlist *sglist,
|
|
|
|
struct scatterlist **sg)
|
|
|
|
{
|
|
|
|
unsigned nbytes = bvec->bv_len;
|
2019-04-11 08:23:27 +02:00
|
|
|
unsigned nsegs = 0, total = 0;
|
2019-02-15 12:13:13 +01:00
|
|
|
|
|
|
|
while (nbytes > 0) {
|
2019-04-11 08:23:27 +02:00
|
|
|
unsigned offset = bvec->bv_offset + total;
|
|
|
|
unsigned len = min(get_max_segment_size(q, offset), nbytes);
|
2019-04-19 08:56:24 +02:00
|
|
|
struct page *page = bvec->bv_page;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Unfortunately a fair number of drivers barf on scatterlists
|
|
|
|
* that have an offset larger than PAGE_SIZE, despite other
|
|
|
|
* subsystems dealing with that invariant just fine. For now
|
|
|
|
* stick to the legacy format where we never present those from
|
|
|
|
* the block layer, but the code below should be removed once
|
|
|
|
* these offenders (mostly MMC/SD drivers) are fixed.
|
|
|
|
*/
|
|
|
|
page += (offset >> PAGE_SHIFT);
|
|
|
|
offset &= ~PAGE_MASK;
|
2019-02-15 12:13:13 +01:00
|
|
|
|
|
|
|
*sg = blk_next_sg(sg, sglist);
|
2019-04-19 08:56:24 +02:00
|
|
|
sg_set_page(*sg, page, len, offset);
|
2019-02-15 12:13:13 +01:00
|
|
|
|
2019-04-11 08:23:27 +02:00
|
|
|
total += len;
|
|
|
|
nbytes -= len;
|
2019-02-15 12:13:13 +01:00
|
|
|
nsegs++;
|
|
|
|
}
|
|
|
|
|
|
|
|
return nsegs;
|
|
|
|
}
|
|
|
|
|
2019-03-17 11:01:11 +01:00
|
|
|
static inline int __blk_bvec_map_sg(struct bio_vec bv,
|
|
|
|
struct scatterlist *sglist, struct scatterlist **sg)
|
|
|
|
{
|
|
|
|
*sg = blk_next_sg(sg, sglist);
|
|
|
|
sg_set_page(*sg, bv.bv_page, bv.bv_len, bv.bv_offset);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2019-03-17 11:01:12 +01:00
|
|
|
/* only try to merge bvecs into one sg if they are from two bios */
|
|
|
|
static inline bool
|
|
|
|
__blk_segment_map_sg_merge(struct request_queue *q, struct bio_vec *bvec,
|
|
|
|
struct bio_vec *bvprv, struct scatterlist **sg)
|
2012-08-02 23:42:03 +02:00
|
|
|
{
|
|
|
|
|
|
|
|
int nbytes = bvec->bv_len;
|
|
|
|
|
2019-03-17 11:01:12 +01:00
|
|
|
if (!*sg)
|
|
|
|
return false;
|
2012-08-02 23:42:03 +02:00
|
|
|
|
2019-03-17 11:01:12 +01:00
|
|
|
if ((*sg)->length + nbytes > queue_max_segment_size(q))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (!biovec_phys_mergeable(q, bvprv, bvec))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
(*sg)->length += nbytes;
|
|
|
|
|
|
|
|
return true;
|
2012-08-02 23:42:03 +02:00
|
|
|
}
|
|
|
|
|
2014-02-07 21:53:46 +01:00
|
|
|
static int __blk_bios_map_sg(struct request_queue *q, struct bio *bio,
|
|
|
|
struct scatterlist *sglist,
|
|
|
|
struct scatterlist **sg)
|
2008-01-29 14:04:06 +01:00
|
|
|
{
|
2019-04-02 04:26:44 +02:00
|
|
|
struct bio_vec uninitialized_var(bvec), bvprv = { NULL };
|
2014-02-07 21:53:46 +01:00
|
|
|
struct bvec_iter iter;
|
2018-12-13 16:17:10 +01:00
|
|
|
int nsegs = 0;
|
2019-03-17 11:01:12 +01:00
|
|
|
bool new_bio = false;
|
2014-02-07 21:53:46 +01:00
|
|
|
|
2019-03-17 11:01:12 +01:00
|
|
|
for_each_bio(bio) {
|
|
|
|
bio_for_each_bvec(bvec, bio, iter) {
|
|
|
|
/*
|
|
|
|
* Only try to merge bvecs from two bios given we
|
|
|
|
* have done bio internal merge when adding pages
|
|
|
|
* to bio
|
|
|
|
*/
|
|
|
|
if (new_bio &&
|
|
|
|
__blk_segment_map_sg_merge(q, &bvec, &bvprv, sg))
|
|
|
|
goto next_bvec;
|
|
|
|
|
|
|
|
if (bvec.bv_offset + bvec.bv_len <= PAGE_SIZE)
|
|
|
|
nsegs += __blk_bvec_map_sg(bvec, sglist, sg);
|
|
|
|
else
|
|
|
|
nsegs += blk_bvec_map_sg(q, &bvec, sglist, sg);
|
|
|
|
next_bvec:
|
|
|
|
new_bio = false;
|
|
|
|
}
|
2019-04-02 04:26:44 +02:00
|
|
|
if (likely(bio->bi_iter.bi_size)) {
|
|
|
|
bvprv = bvec;
|
|
|
|
new_bio = true;
|
|
|
|
}
|
2019-03-17 11:01:12 +01:00
|
|
|
}
|
2008-01-29 14:04:06 +01:00
|
|
|
|
2014-02-07 21:53:46 +01:00
|
|
|
return nsegs;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* map a request to scatterlist, return number of sg entries setup. Caller
|
|
|
|
* must make sure sg can hold rq->nr_phys_segments entries
|
|
|
|
*/
|
|
|
|
int blk_rq_map_sg(struct request_queue *q, struct request *rq,
|
|
|
|
struct scatterlist *sglist)
|
|
|
|
{
|
|
|
|
struct scatterlist *sg = NULL;
|
|
|
|
int nsegs = 0;
|
|
|
|
|
2016-12-08 23:20:32 +01:00
|
|
|
if (rq->rq_flags & RQF_SPECIAL_PAYLOAD)
|
2019-03-17 11:01:10 +01:00
|
|
|
nsegs = __blk_bvec_map_sg(rq->special_vec, sglist, &sg);
|
2016-12-08 23:20:32 +01:00
|
|
|
else if (rq->bio && bio_op(rq->bio) == REQ_OP_WRITE_SAME)
|
2019-03-17 11:01:10 +01:00
|
|
|
nsegs = __blk_bvec_map_sg(bio_iovec(rq->bio), sglist, &sg);
|
2016-12-08 23:20:32 +01:00
|
|
|
else if (rq->bio)
|
2014-02-07 21:53:46 +01:00
|
|
|
nsegs = __blk_bios_map_sg(q, rq->bio, sglist, &sg);
|
2008-04-11 12:56:52 +02:00
|
|
|
|
2016-10-20 15:12:13 +02:00
|
|
|
if (unlikely(rq->rq_flags & RQF_COPY_USER) &&
|
2009-05-07 15:24:41 +02:00
|
|
|
(blk_rq_bytes(rq) & q->dma_pad_mask)) {
|
|
|
|
unsigned int pad_len =
|
|
|
|
(q->dma_pad_mask & ~blk_rq_bytes(rq)) + 1;
|
2008-04-11 12:56:52 +02:00
|
|
|
|
|
|
|
sg->length += pad_len;
|
|
|
|
rq->extra_len += pad_len;
|
|
|
|
}
|
|
|
|
|
2008-02-19 11:36:53 +01:00
|
|
|
if (q->dma_drain_size && q->dma_drain_needed(rq)) {
|
2016-06-05 21:31:45 +02:00
|
|
|
if (op_is_write(req_op(rq)))
|
2008-02-19 11:36:55 +01:00
|
|
|
memset(q->dma_drain_buffer, 0, q->dma_drain_size);
|
|
|
|
|
2015-08-07 18:15:14 +02:00
|
|
|
sg_unmark_end(sg);
|
2008-01-29 14:04:06 +01:00
|
|
|
sg = sg_next(sg);
|
|
|
|
sg_set_page(sg, virt_to_page(q->dma_drain_buffer),
|
|
|
|
q->dma_drain_size,
|
|
|
|
((unsigned long)q->dma_drain_buffer) &
|
|
|
|
(PAGE_SIZE - 1));
|
|
|
|
nsegs++;
|
2008-03-04 11:17:11 +01:00
|
|
|
rq->extra_len += q->dma_drain_size;
|
2008-01-29 14:04:06 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
if (sg)
|
|
|
|
sg_mark_end(sg);
|
|
|
|
|
2015-11-24 03:35:31 +01:00
|
|
|
/*
|
|
|
|
* Something must have been wrong if the figured number of
|
|
|
|
* segment is bigger than number of req's physical segments
|
|
|
|
*/
|
2016-12-08 23:20:32 +01:00
|
|
|
WARN_ON(nsegs > blk_rq_nr_phys_segments(rq));
|
2015-11-24 03:35:31 +01:00
|
|
|
|
2008-01-29 14:04:06 +01:00
|
|
|
return nsegs;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(blk_rq_map_sg);
|
|
|
|
|
|
|
|
static inline int ll_new_hw_segment(struct request_queue *q,
|
|
|
|
struct request *req,
|
|
|
|
struct bio *bio)
|
|
|
|
{
|
|
|
|
int nr_phys_segs = bio_phys_segments(q, bio);
|
|
|
|
|
2010-09-10 20:50:10 +02:00
|
|
|
if (req->nr_phys_segments + nr_phys_segs > queue_max_segments(q))
|
|
|
|
goto no_merge;
|
|
|
|
|
2014-09-27 01:20:06 +02:00
|
|
|
if (blk_integrity_merge_bio(q, req, bio) == false)
|
2010-09-10 20:50:10 +02:00
|
|
|
goto no_merge;
|
2008-01-29 14:04:06 +01:00
|
|
|
|
|
|
|
/*
|
|
|
|
* This will form the start of a new hw segment. Bump both
|
|
|
|
* counters.
|
|
|
|
*/
|
|
|
|
req->nr_phys_segments += nr_phys_segs;
|
|
|
|
return 1;
|
2010-09-10 20:50:10 +02:00
|
|
|
|
|
|
|
no_merge:
|
2016-12-01 16:36:16 +01:00
|
|
|
req_set_nomerge(q, req);
|
2010-09-10 20:50:10 +02:00
|
|
|
return 0;
|
2008-01-29 14:04:06 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
int ll_back_merge_fn(struct request_queue *q, struct request *req,
|
|
|
|
struct bio *bio)
|
|
|
|
{
|
2015-09-03 18:28:20 +02:00
|
|
|
if (req_gap_back_merge(req, bio))
|
|
|
|
return 0;
|
2015-09-11 17:03:04 +02:00
|
|
|
if (blk_integrity_rq(req) &&
|
|
|
|
integrity_req_gap_back_merge(req, bio))
|
|
|
|
return 0;
|
2012-09-18 18:19:26 +02:00
|
|
|
if (blk_rq_sectors(req) + bio_sectors(bio) >
|
2016-07-21 05:40:47 +02:00
|
|
|
blk_rq_get_max_sectors(req, blk_rq_pos(req))) {
|
2016-12-01 16:36:16 +01:00
|
|
|
req_set_nomerge(q, req);
|
2008-01-29 14:04:06 +01:00
|
|
|
return 0;
|
|
|
|
}
|
2008-05-07 09:33:55 +02:00
|
|
|
if (!bio_flagged(req->biotail, BIO_SEG_VALID))
|
2008-01-29 14:04:06 +01:00
|
|
|
blk_recount_segments(q, req->biotail);
|
2008-05-07 09:33:55 +02:00
|
|
|
if (!bio_flagged(bio, BIO_SEG_VALID))
|
2008-01-29 14:04:06 +01:00
|
|
|
blk_recount_segments(q, bio);
|
|
|
|
|
|
|
|
return ll_new_hw_segment(q, req, bio);
|
|
|
|
}
|
|
|
|
|
2008-01-31 13:03:55 +01:00
|
|
|
int ll_front_merge_fn(struct request_queue *q, struct request *req,
|
2008-01-29 14:04:06 +01:00
|
|
|
struct bio *bio)
|
|
|
|
{
|
2015-09-03 18:28:20 +02:00
|
|
|
|
|
|
|
if (req_gap_front_merge(req, bio))
|
|
|
|
return 0;
|
2015-09-11 17:03:04 +02:00
|
|
|
if (blk_integrity_rq(req) &&
|
|
|
|
integrity_req_gap_front_merge(req, bio))
|
|
|
|
return 0;
|
2012-09-18 18:19:26 +02:00
|
|
|
if (blk_rq_sectors(req) + bio_sectors(bio) >
|
2016-07-21 05:40:47 +02:00
|
|
|
blk_rq_get_max_sectors(req, bio->bi_iter.bi_sector)) {
|
2016-12-01 16:36:16 +01:00
|
|
|
req_set_nomerge(q, req);
|
2008-01-29 14:04:06 +01:00
|
|
|
return 0;
|
|
|
|
}
|
2008-05-07 09:33:55 +02:00
|
|
|
if (!bio_flagged(bio, BIO_SEG_VALID))
|
2008-01-29 14:04:06 +01:00
|
|
|
blk_recount_segments(q, bio);
|
2008-05-07 09:33:55 +02:00
|
|
|
if (!bio_flagged(req->bio, BIO_SEG_VALID))
|
2008-01-29 14:04:06 +01:00
|
|
|
blk_recount_segments(q, req->bio);
|
|
|
|
|
|
|
|
return ll_new_hw_segment(q, req, bio);
|
|
|
|
}
|
|
|
|
|
2018-02-01 22:01:02 +01:00
|
|
|
static bool req_attempt_discard_merge(struct request_queue *q, struct request *req,
|
|
|
|
struct request *next)
|
|
|
|
{
|
|
|
|
unsigned short segments = blk_rq_nr_discard_segments(req);
|
|
|
|
|
|
|
|
if (segments >= queue_max_discard_segments(q))
|
|
|
|
goto no_merge;
|
|
|
|
if (blk_rq_sectors(req) + bio_sectors(next->bio) >
|
|
|
|
blk_rq_get_max_sectors(req, blk_rq_pos(req)))
|
|
|
|
goto no_merge;
|
|
|
|
|
|
|
|
req->nr_phys_segments = segments + blk_rq_nr_discard_segments(next);
|
|
|
|
return true;
|
|
|
|
no_merge:
|
|
|
|
req_set_nomerge(q, req);
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2008-01-29 14:04:06 +01:00
|
|
|
static int ll_merge_requests_fn(struct request_queue *q, struct request *req,
|
|
|
|
struct request *next)
|
|
|
|
{
|
|
|
|
int total_phys_segments;
|
|
|
|
|
2015-09-03 18:28:20 +02:00
|
|
|
if (req_gap_back_merge(req, next->bio))
|
2015-02-11 16:20:13 +01:00
|
|
|
return 0;
|
|
|
|
|
2008-01-29 14:04:06 +01:00
|
|
|
/*
|
|
|
|
* Will it become too large?
|
|
|
|
*/
|
2012-09-18 18:19:26 +02:00
|
|
|
if ((blk_rq_sectors(req) + blk_rq_sectors(next)) >
|
2016-07-21 05:40:47 +02:00
|
|
|
blk_rq_get_max_sectors(req, blk_rq_pos(req)))
|
2008-01-29 14:04:06 +01:00
|
|
|
return 0;
|
|
|
|
|
|
|
|
total_phys_segments = req->nr_phys_segments + next->nr_phys_segments;
|
2010-02-26 06:20:39 +01:00
|
|
|
if (total_phys_segments > queue_max_segments(q))
|
2008-01-29 14:04:06 +01:00
|
|
|
return 0;
|
|
|
|
|
2014-09-27 01:20:06 +02:00
|
|
|
if (blk_integrity_merge_rq(q, req, next) == false)
|
2010-09-10 20:50:10 +02:00
|
|
|
return 0;
|
|
|
|
|
2008-01-29 14:04:06 +01:00
|
|
|
/* Merge is OK... */
|
|
|
|
req->nr_phys_segments = total_phys_segments;
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2009-07-03 10:48:17 +02:00
|
|
|
/**
|
|
|
|
* blk_rq_set_mixed_merge - mark a request as mixed merge
|
|
|
|
* @rq: request to mark as mixed merge
|
|
|
|
*
|
|
|
|
* Description:
|
|
|
|
* @rq is about to be mixed merged. Make sure the attributes
|
|
|
|
* which can be mixed are set in each bio and mark @rq as mixed
|
|
|
|
* merged.
|
|
|
|
*/
|
|
|
|
void blk_rq_set_mixed_merge(struct request *rq)
|
|
|
|
{
|
|
|
|
unsigned int ff = rq->cmd_flags & REQ_FAILFAST_MASK;
|
|
|
|
struct bio *bio;
|
|
|
|
|
2016-10-20 15:12:13 +02:00
|
|
|
if (rq->rq_flags & RQF_MIXED_MERGE)
|
2009-07-03 10:48:17 +02:00
|
|
|
return;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* @rq will no longer represent mixable attributes for all the
|
|
|
|
* contained bios. It will just track those of the first one.
|
|
|
|
* Distributes the attributs to each bio.
|
|
|
|
*/
|
|
|
|
for (bio = rq->bio; bio; bio = bio->bi_next) {
|
2016-08-05 23:35:16 +02:00
|
|
|
WARN_ON_ONCE((bio->bi_opf & REQ_FAILFAST_MASK) &&
|
|
|
|
(bio->bi_opf & REQ_FAILFAST_MASK) != ff);
|
|
|
|
bio->bi_opf |= ff;
|
2009-07-03 10:48:17 +02:00
|
|
|
}
|
2016-10-20 15:12:13 +02:00
|
|
|
rq->rq_flags |= RQF_MIXED_MERGE;
|
2009-07-03 10:48:17 +02:00
|
|
|
}
|
|
|
|
|
2009-03-27 10:31:51 +01:00
|
|
|
static void blk_account_io_merge(struct request *req)
|
|
|
|
{
|
|
|
|
if (blk_do_io_stat(req)) {
|
|
|
|
struct hd_struct *part;
|
|
|
|
|
2018-12-06 17:41:18 +01:00
|
|
|
part_stat_lock();
|
2011-01-05 16:57:38 +01:00
|
|
|
part = req->part;
|
2009-03-27 10:31:51 +01:00
|
|
|
|
2017-07-01 05:55:08 +02:00
|
|
|
part_dec_in_flight(req->q, part, rq_data_dir(req));
|
2009-03-27 10:31:51 +01:00
|
|
|
|
2011-01-07 08:43:37 +01:00
|
|
|
hd_struct_put(part);
|
2009-03-27 10:31:51 +01:00
|
|
|
part_stat_unlock();
|
|
|
|
}
|
|
|
|
}
|
2018-10-27 13:52:14 +02:00
|
|
|
/*
|
|
|
|
* Two cases of handling DISCARD merge:
|
|
|
|
* If max_discard_segments > 1, the driver takes every bio
|
|
|
|
* as a range and send them to controller together. The ranges
|
|
|
|
* needn't to be contiguous.
|
|
|
|
* Otherwise, the bios/requests will be handled as same as
|
|
|
|
* others which should be contiguous.
|
|
|
|
*/
|
|
|
|
static inline bool blk_discard_mergable(struct request *req)
|
|
|
|
{
|
|
|
|
if (req_op(req) == REQ_OP_DISCARD &&
|
|
|
|
queue_max_discard_segments(req->q) > 1)
|
|
|
|
return true;
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2018-11-15 02:19:46 +01:00
|
|
|
static enum elv_merge blk_try_req_merge(struct request *req,
|
|
|
|
struct request *next)
|
2018-10-27 13:52:14 +02:00
|
|
|
{
|
|
|
|
if (blk_discard_mergable(req))
|
|
|
|
return ELEVATOR_DISCARD_MERGE;
|
|
|
|
else if (blk_rq_pos(req) + blk_rq_sectors(req) == blk_rq_pos(next))
|
|
|
|
return ELEVATOR_BACK_MERGE;
|
|
|
|
|
|
|
|
return ELEVATOR_NO_MERGE;
|
|
|
|
}
|
2009-03-27 10:31:51 +01:00
|
|
|
|
2008-01-29 14:04:06 +01:00
|
|
|
/*
|
2017-02-02 16:54:40 +01:00
|
|
|
* For non-mq, this has to be called with the request spinlock acquired.
|
|
|
|
* For mq with scheduling, the appropriate queue wide lock should be held.
|
2008-01-29 14:04:06 +01:00
|
|
|
*/
|
2017-02-02 16:54:40 +01:00
|
|
|
static struct request *attempt_merge(struct request_queue *q,
|
|
|
|
struct request *req, struct request *next)
|
2008-01-29 14:04:06 +01:00
|
|
|
{
|
|
|
|
if (!rq_mergeable(req) || !rq_mergeable(next))
|
2017-02-02 16:54:40 +01:00
|
|
|
return NULL;
|
2008-01-29 14:04:06 +01:00
|
|
|
|
2016-06-09 16:00:36 +02:00
|
|
|
if (req_op(req) != req_op(next))
|
2017-02-02 16:54:40 +01:00
|
|
|
return NULL;
|
2012-09-18 18:19:26 +02:00
|
|
|
|
2008-01-29 14:04:06 +01:00
|
|
|
if (rq_data_dir(req) != rq_data_dir(next)
|
2018-10-12 20:39:10 +02:00
|
|
|
|| req->rq_disk != next->rq_disk)
|
2017-02-02 16:54:40 +01:00
|
|
|
return NULL;
|
2008-01-29 14:04:06 +01:00
|
|
|
|
2016-06-05 21:32:15 +02:00
|
|
|
if (req_op(req) == REQ_OP_WRITE_SAME &&
|
2012-09-18 18:19:27 +02:00
|
|
|
!blk_write_same_mergeable(req->bio, next->bio))
|
2017-02-02 16:54:40 +01:00
|
|
|
return NULL;
|
2012-09-18 18:19:27 +02:00
|
|
|
|
2017-06-27 17:22:02 +02:00
|
|
|
/*
|
|
|
|
* Don't allow merge of different write hints, or for a hint with
|
|
|
|
* non-hint IO.
|
|
|
|
*/
|
|
|
|
if (req->write_hint != next->write_hint)
|
|
|
|
return NULL;
|
|
|
|
|
2018-11-20 02:52:37 +01:00
|
|
|
if (req->ioprio != next->ioprio)
|
|
|
|
return NULL;
|
|
|
|
|
2008-01-29 14:04:06 +01:00
|
|
|
/*
|
|
|
|
* If we are allowed to merge, then append bio list
|
|
|
|
* from next to rq and release next. merge_requests_fn
|
|
|
|
* will have updated segment counts, update sector
|
2018-02-01 22:01:02 +01:00
|
|
|
* counts here. Handle DISCARDs separately, as they
|
|
|
|
* have separate settings.
|
2008-01-29 14:04:06 +01:00
|
|
|
*/
|
2018-10-27 13:52:14 +02:00
|
|
|
|
|
|
|
switch (blk_try_req_merge(req, next)) {
|
|
|
|
case ELEVATOR_DISCARD_MERGE:
|
2018-02-01 22:01:02 +01:00
|
|
|
if (!req_attempt_discard_merge(q, req, next))
|
|
|
|
return NULL;
|
2018-10-27 13:52:14 +02:00
|
|
|
break;
|
|
|
|
case ELEVATOR_BACK_MERGE:
|
|
|
|
if (!ll_merge_requests_fn(q, req, next))
|
|
|
|
return NULL;
|
|
|
|
break;
|
|
|
|
default:
|
2017-02-02 16:54:40 +01:00
|
|
|
return NULL;
|
2018-10-27 13:52:14 +02:00
|
|
|
}
|
2008-01-29 14:04:06 +01:00
|
|
|
|
2009-07-03 10:48:17 +02:00
|
|
|
/*
|
|
|
|
* If failfast settings disagree or any of the two is already
|
|
|
|
* a mixed merge, mark both as mixed before proceeding. This
|
|
|
|
* makes sure that all involved bios have mixable attributes
|
|
|
|
* set properly.
|
|
|
|
*/
|
2016-10-20 15:12:13 +02:00
|
|
|
if (((req->rq_flags | next->rq_flags) & RQF_MIXED_MERGE) ||
|
2009-07-03 10:48:17 +02:00
|
|
|
(req->cmd_flags & REQ_FAILFAST_MASK) !=
|
|
|
|
(next->cmd_flags & REQ_FAILFAST_MASK)) {
|
|
|
|
blk_rq_set_mixed_merge(req);
|
|
|
|
blk_rq_set_mixed_merge(next);
|
|
|
|
}
|
|
|
|
|
2008-01-29 14:04:06 +01:00
|
|
|
/*
|
block: consolidate struct request timestamp fields
Currently, struct request has four timestamp fields:
- A start time, set at get_request time, in jiffies, used for iostats
- An I/O start time, set at start_request time, in ktime nanoseconds,
used for blk-stats (i.e., wbt, kyber, hybrid polling)
- Another start time and another I/O start time, used for cfq and bfq
These can all be consolidated into one start time and one I/O start
time, both in ktime nanoseconds, shaving off up to 16 bytes from struct
request depending on the kernel config.
Signed-off-by: Omar Sandoval <osandov@fb.com>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
2018-05-09 11:08:53 +02:00
|
|
|
* At this point we have either done a back merge or front merge. We
|
|
|
|
* need the smaller start_time_ns of the merged requests to be the
|
|
|
|
* current request for accounting purposes.
|
2008-01-29 14:04:06 +01:00
|
|
|
*/
|
block: consolidate struct request timestamp fields
Currently, struct request has four timestamp fields:
- A start time, set at get_request time, in jiffies, used for iostats
- An I/O start time, set at start_request time, in ktime nanoseconds,
used for blk-stats (i.e., wbt, kyber, hybrid polling)
- Another start time and another I/O start time, used for cfq and bfq
These can all be consolidated into one start time and one I/O start
time, both in ktime nanoseconds, shaving off up to 16 bytes from struct
request depending on the kernel config.
Signed-off-by: Omar Sandoval <osandov@fb.com>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
2018-05-09 11:08:53 +02:00
|
|
|
if (next->start_time_ns < req->start_time_ns)
|
|
|
|
req->start_time_ns = next->start_time_ns;
|
2008-01-29 14:04:06 +01:00
|
|
|
|
|
|
|
req->biotail->bi_next = next->bio;
|
|
|
|
req->biotail = next->biotail;
|
|
|
|
|
2009-05-07 15:24:44 +02:00
|
|
|
req->__data_len += blk_rq_bytes(next);
|
2008-01-29 14:04:06 +01:00
|
|
|
|
2018-11-30 17:38:18 +01:00
|
|
|
if (!blk_discard_mergable(req))
|
2018-02-01 22:01:02 +01:00
|
|
|
elv_merge_requests(q, req, next);
|
2008-01-29 14:04:06 +01:00
|
|
|
|
2009-04-22 14:01:49 +02:00
|
|
|
/*
|
|
|
|
* 'next' is going away, so update stats accordingly
|
|
|
|
*/
|
|
|
|
blk_account_io_merge(next);
|
2008-01-29 14:04:06 +01:00
|
|
|
|
2017-02-03 17:48:28 +01:00
|
|
|
/*
|
|
|
|
* ownership of bio passed from next to req, return 'next' for
|
|
|
|
* the caller to free
|
|
|
|
*/
|
2009-03-24 12:35:07 +01:00
|
|
|
next->bio = NULL;
|
2017-02-02 16:54:40 +01:00
|
|
|
return next;
|
2008-01-29 14:04:06 +01:00
|
|
|
}
|
|
|
|
|
2017-02-02 16:54:40 +01:00
|
|
|
struct request *attempt_back_merge(struct request_queue *q, struct request *rq)
|
2008-01-29 14:04:06 +01:00
|
|
|
{
|
|
|
|
struct request *next = elv_latter_request(q, rq);
|
|
|
|
|
|
|
|
if (next)
|
|
|
|
return attempt_merge(q, rq, next);
|
|
|
|
|
2017-02-02 16:54:40 +01:00
|
|
|
return NULL;
|
2008-01-29 14:04:06 +01:00
|
|
|
}
|
|
|
|
|
2017-02-02 16:54:40 +01:00
|
|
|
struct request *attempt_front_merge(struct request_queue *q, struct request *rq)
|
2008-01-29 14:04:06 +01:00
|
|
|
{
|
|
|
|
struct request *prev = elv_former_request(q, rq);
|
|
|
|
|
|
|
|
if (prev)
|
|
|
|
return attempt_merge(q, prev, rq);
|
|
|
|
|
2017-02-02 16:54:40 +01:00
|
|
|
return NULL;
|
2008-01-29 14:04:06 +01:00
|
|
|
}
|
2011-03-21 10:14:27 +01:00
|
|
|
|
|
|
|
int blk_attempt_req_merge(struct request_queue *q, struct request *rq,
|
|
|
|
struct request *next)
|
|
|
|
{
|
2017-02-03 17:48:28 +01:00
|
|
|
struct request *free;
|
2016-07-07 20:48:22 +02:00
|
|
|
|
2017-02-03 17:48:28 +01:00
|
|
|
free = attempt_merge(q, rq, next);
|
|
|
|
if (free) {
|
2018-10-24 21:52:28 +02:00
|
|
|
blk_put_request(free);
|
2017-02-03 17:48:28 +01:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
2011-03-21 10:14:27 +01:00
|
|
|
}
|
2012-02-08 09:19:38 +01:00
|
|
|
|
|
|
|
bool blk_rq_merge_ok(struct request *rq, struct bio *bio)
|
|
|
|
{
|
2012-09-18 18:19:25 +02:00
|
|
|
if (!rq_mergeable(rq) || !bio_mergeable(bio))
|
2012-02-08 09:19:38 +01:00
|
|
|
return false;
|
|
|
|
|
2016-06-09 16:00:36 +02:00
|
|
|
if (req_op(rq) != bio_op(bio))
|
2012-09-18 18:19:26 +02:00
|
|
|
return false;
|
|
|
|
|
2012-02-08 09:19:38 +01:00
|
|
|
/* different data direction or already started, don't merge */
|
|
|
|
if (bio_data_dir(bio) != rq_data_dir(rq))
|
|
|
|
return false;
|
|
|
|
|
2018-10-12 20:39:10 +02:00
|
|
|
/* must be same device */
|
|
|
|
if (rq->rq_disk != bio->bi_disk)
|
2012-02-08 09:19:38 +01:00
|
|
|
return false;
|
|
|
|
|
|
|
|
/* only merge integrity protected bio into ditto rq */
|
2014-09-27 01:20:06 +02:00
|
|
|
if (blk_integrity_merge_bio(rq->q, rq, bio) == false)
|
2012-02-08 09:19:38 +01:00
|
|
|
return false;
|
|
|
|
|
2012-09-18 18:19:27 +02:00
|
|
|
/* must be using the same buffer */
|
2016-06-05 21:32:15 +02:00
|
|
|
if (req_op(rq) == REQ_OP_WRITE_SAME &&
|
2012-09-18 18:19:27 +02:00
|
|
|
!blk_write_same_mergeable(rq->bio, bio))
|
|
|
|
return false;
|
|
|
|
|
2017-06-27 17:22:02 +02:00
|
|
|
/*
|
|
|
|
* Don't allow merge of different write hints, or for a hint with
|
|
|
|
* non-hint IO.
|
|
|
|
*/
|
|
|
|
if (rq->write_hint != bio->bi_write_hint)
|
|
|
|
return false;
|
|
|
|
|
2018-11-20 02:52:37 +01:00
|
|
|
if (rq->ioprio != bio_prio(bio))
|
|
|
|
return false;
|
|
|
|
|
2012-02-08 09:19:38 +01:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2017-02-08 14:46:48 +01:00
|
|
|
enum elv_merge blk_try_merge(struct request *rq, struct bio *bio)
|
2012-02-08 09:19:38 +01:00
|
|
|
{
|
2018-10-27 13:52:14 +02:00
|
|
|
if (blk_discard_mergable(rq))
|
2017-02-08 14:46:49 +01:00
|
|
|
return ELEVATOR_DISCARD_MERGE;
|
|
|
|
else if (blk_rq_pos(rq) + blk_rq_sectors(rq) == bio->bi_iter.bi_sector)
|
2012-02-08 09:19:38 +01:00
|
|
|
return ELEVATOR_BACK_MERGE;
|
2013-10-12 00:44:27 +02:00
|
|
|
else if (blk_rq_pos(rq) - bio_sectors(bio) == bio->bi_iter.bi_sector)
|
2012-02-08 09:19:38 +01:00
|
|
|
return ELEVATOR_FRONT_MERGE;
|
|
|
|
return ELEVATOR_NO_MERGE;
|
|
|
|
}
|