All of lore.kernel.org
 help / color / mirror / Atom feed
From: Ming Lei <ming.lei@redhat.com>
To: Jens Axboe <axboe@kernel.dk>
Cc: Mike Snitzer <snitzer@redhat.com>,
	linux-mm@kvack.org, dm-devel@redhat.com,
	Christoph Hellwig <hch@lst.de>, Sagi Grimberg <sagi@grimberg.me>,
	"Darrick J . Wong" <darrick.wong@oracle.com>,
	Omar Sandoval <osandov@fb.com>,
	cluster-devel@redhat.com, linux-ext4@vger.kernel.org,
	Shaohua Li <shli@kernel.org>,
	Kent Overstreet <kent.overstreet@gmail.com>,
	Boaz Harrosh <ooo@electrozaur.com>,
	Gao Xiang <gaoxiang25@huawei.com>, Coly Li <colyli@suse.de>,
	linux-raid@vger.kernel.org, Bob Peterson <rpeterso@redhat.com>,
	linux-bcache@vger.kernel.org,
	Alexander Viro <viro@zeniv.linux.org.uk>,
	Dave Chinner <dchinner@redhat.com>,
	David Sterba <dsterba@suse.com>, Ming Lei <ming.lei@redhat.com>,
	linux-block@vger.kernel.org, Theodore Ts'o <tytso@mit.edu>,
	linux-kernel@vger.kernel.org, linux-xfs@vger.kernel.org,
	linux-fsdevel@vger.kernel.org, linux-btrfs@vger.kernel.org
Subject: [PATCH V12 03/20] block: remove the "cluster" flag
Date: Mon, 26 Nov 2018 10:17:03 +0800	[thread overview]
Message-ID: <20181126021720.19471-4-ming.lei@redhat.com> (raw)
In-Reply-To: <20181126021720.19471-1-ming.lei@redhat.com>

From: Christoph Hellwig <hch@lst.de>

The cluster flag implements some very old SCSI behavior.  As far as I
can tell the original intent was to enable or disable any kind of
segment merging.  But the actually visible effect to the LLDD is that
it limits each segments to be inside a single page, which we can
also affect by setting the maximum segment size and the segment
boundary.

Signed-off-by: Christoph Hellwig <hch@lst.de>

Replace virt boundary with segment boundary limit.

Signed-off-by: Ming Lei <ming.lei@redhat.com>
---
 block/blk-merge.c       | 20 ++++++++------------
 block/blk-settings.c    |  3 ---
 block/blk-sysfs.c       |  5 +----
 drivers/scsi/scsi_lib.c | 20 ++++++++++++++++----
 include/linux/blkdev.h  |  6 ------
 5 files changed, 25 insertions(+), 29 deletions(-)

diff --git a/block/blk-merge.c b/block/blk-merge.c
index 6be04ef8da5b..e69d8f8ba819 100644
--- a/block/blk-merge.c
+++ b/block/blk-merge.c
@@ -195,7 +195,7 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
 			goto split;
 		}
 
-		if (bvprvp && blk_queue_cluster(q)) {
+		if (bvprvp) {
 			if (seg_size + bv.bv_len > queue_max_segment_size(q))
 				goto new_segment;
 			if (!biovec_phys_mergeable(q, bvprvp, &bv))
@@ -295,10 +295,10 @@ static unsigned int __blk_recalc_rq_segments(struct request_queue *q,
 					     bool no_sg_merge)
 {
 	struct bio_vec bv, bvprv = { NULL };
-	int cluster, prev = 0;
 	unsigned int seg_size, nr_phys_segs;
 	struct bio *fbio, *bbio;
 	struct bvec_iter iter;
+	bool prev = false;
 
 	if (!bio)
 		return 0;
@@ -313,7 +313,6 @@ static unsigned int __blk_recalc_rq_segments(struct request_queue *q,
 	}
 
 	fbio = bio;
-	cluster = blk_queue_cluster(q);
 	seg_size = 0;
 	nr_phys_segs = 0;
 	for_each_bio(bio) {
@@ -325,7 +324,7 @@ static unsigned int __blk_recalc_rq_segments(struct request_queue *q,
 			if (no_sg_merge)
 				goto new_segment;
 
-			if (prev && cluster) {
+			if (prev) {
 				if (seg_size + bv.bv_len
 				    > queue_max_segment_size(q))
 					goto new_segment;
@@ -343,7 +342,7 @@ static unsigned int __blk_recalc_rq_segments(struct request_queue *q,
 
 			nr_phys_segs++;
 			bvprv = bv;
-			prev = 1;
+			prev = true;
 			seg_size = bv.bv_len;
 		}
 		bbio = bio;
@@ -396,9 +395,6 @@ static int blk_phys_contig_segment(struct request_queue *q, struct bio *bio,
 {
 	struct bio_vec end_bv = { NULL }, nxt_bv;
 
-	if (!blk_queue_cluster(q))
-		return 0;
-
 	if (bio->bi_seg_back_size + nxt->bi_seg_front_size >
 	    queue_max_segment_size(q))
 		return 0;
@@ -415,12 +411,12 @@ static int blk_phys_contig_segment(struct request_queue *q, struct bio *bio,
 static inline void
 __blk_segment_map_sg(struct request_queue *q, struct bio_vec *bvec,
 		     struct scatterlist *sglist, struct bio_vec *bvprv,
-		     struct scatterlist **sg, int *nsegs, int *cluster)
+		     struct scatterlist **sg, int *nsegs)
 {
 
 	int nbytes = bvec->bv_len;
 
-	if (*sg && *cluster) {
+	if (*sg) {
 		if ((*sg)->length + nbytes > queue_max_segment_size(q))
 			goto new_segment;
 		if (!biovec_phys_mergeable(q, bvprv, bvec))
@@ -466,12 +462,12 @@ static int __blk_bios_map_sg(struct request_queue *q, struct bio *bio,
 {
 	struct bio_vec bvec, bvprv = { NULL };
 	struct bvec_iter iter;
-	int cluster = blk_queue_cluster(q), nsegs = 0;
+	int nsegs = 0;
 
 	for_each_bio(bio)
 		bio_for_each_segment(bvec, bio, iter)
 			__blk_segment_map_sg(q, &bvec, sglist, &bvprv, sg,
-					     &nsegs, &cluster);
+					     &nsegs);
 
 	return nsegs;
 }
diff --git a/block/blk-settings.c b/block/blk-settings.c
index 3abe831e92c8..3e7038e475ee 100644
--- a/block/blk-settings.c
+++ b/block/blk-settings.c
@@ -56,7 +56,6 @@ void blk_set_default_limits(struct queue_limits *lim)
 	lim->alignment_offset = 0;
 	lim->io_opt = 0;
 	lim->misaligned = 0;
-	lim->cluster = 1;
 	lim->zoned = BLK_ZONED_NONE;
 }
 EXPORT_SYMBOL(blk_set_default_limits);
@@ -547,8 +546,6 @@ int blk_stack_limits(struct queue_limits *t, struct queue_limits *b,
 	t->io_min = max(t->io_min, b->io_min);
 	t->io_opt = lcm_not_zero(t->io_opt, b->io_opt);
 
-	t->cluster &= b->cluster;
-
 	/* Physical block size a multiple of the logical block size? */
 	if (t->physical_block_size & (t->logical_block_size - 1)) {
 		t->physical_block_size = t->logical_block_size;
diff --git a/block/blk-sysfs.c b/block/blk-sysfs.c
index 80eef48fddc8..ef7b844a3e00 100644
--- a/block/blk-sysfs.c
+++ b/block/blk-sysfs.c
@@ -132,10 +132,7 @@ static ssize_t queue_max_integrity_segments_show(struct request_queue *q, char *
 
 static ssize_t queue_max_segment_size_show(struct request_queue *q, char *page)
 {
-	if (blk_queue_cluster(q))
-		return queue_var_show(queue_max_segment_size(q), (page));
-
-	return queue_var_show(PAGE_SIZE, (page));
+	return queue_var_show(queue_max_segment_size(q), page);
 }
 
 static ssize_t queue_logical_block_size_show(struct request_queue *q, char *page)
diff --git a/drivers/scsi/scsi_lib.c b/drivers/scsi/scsi_lib.c
index 0df15cb738d2..78d6d05992b0 100644
--- a/drivers/scsi/scsi_lib.c
+++ b/drivers/scsi/scsi_lib.c
@@ -1810,6 +1810,8 @@ static int scsi_map_queues(struct blk_mq_tag_set *set)
 void __scsi_init_queue(struct Scsi_Host *shost, struct request_queue *q)
 {
 	struct device *dev = shost->dma_dev;
+	unsigned max_segment_size = dma_get_max_seg_size(dev);
+	unsigned long segment_boundary = shost->dma_boundary;
 
 	/*
 	 * this limit is imposed by hardware restrictions
@@ -1828,13 +1830,23 @@ void __scsi_init_queue(struct Scsi_Host *shost, struct request_queue *q)
 	blk_queue_max_hw_sectors(q, shost->max_sectors);
 	if (shost->unchecked_isa_dma)
 		blk_queue_bounce_limit(q, BLK_BOUNCE_ISA);
-	blk_queue_segment_boundary(q, shost->dma_boundary);
 	dma_set_seg_boundary(dev, shost->dma_boundary);
 
-	blk_queue_max_segment_size(q, dma_get_max_seg_size(dev));
+	/*
+	 * Clustering is a really old concept from the stone age of Linux
+	 * SCSI support.  But the basic idea is that we never give the
+	 * driver a segment that spans multiple pages.  For that we need
+	 * to limit the segment size, and set the segment boundary so that
+	 * we never merge a second segment which is no page aligned.
+	 */
+	if (!shost->use_clustering) {
+		max_segment_size = min_t(unsigned, max_segment_size, PAGE_SIZE);
+		segment_boundary = min_t(unsigned, segment_boundary,
+					 PAGE_SIZE - 1);
+	}
 
-	if (!shost->use_clustering)
-		q->limits.cluster = 0;
+	blk_queue_max_segment_size(q, max_segment_size);
+	blk_queue_segment_boundary(q, segment_boundary);
 
 	/*
 	 * Set a reasonable default alignment:  The larger of 32-byte (dword),
diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h
index 9b53db06ad08..399a7a415609 100644
--- a/include/linux/blkdev.h
+++ b/include/linux/blkdev.h
@@ -341,7 +341,6 @@ struct queue_limits {
 
 	unsigned char		misaligned;
 	unsigned char		discard_misaligned;
-	unsigned char		cluster;
 	unsigned char		raid_partial_stripes_expensive;
 	enum blk_zoned_model	zoned;
 };
@@ -660,11 +659,6 @@ static inline bool queue_is_mq(struct request_queue *q)
 	return q->mq_ops;
 }
 
-static inline unsigned int blk_queue_cluster(struct request_queue *q)
-{
-	return q->limits.cluster;
-}
-
 static inline enum blk_zoned_model
 blk_queue_zoned_model(struct request_queue *q)
 {
-- 
2.9.5

WARNING: multiple messages have this Message-ID (diff)
From: Ming Lei <ming.lei@redhat.com>
To: Jens Axboe <axboe@kernel.dk>
Cc: linux-block@vger.kernel.org, linux-kernel@vger.kernel.org,
	linux-mm@kvack.org, Theodore Ts'o <tytso@mit.edu>,
	Omar Sandoval <osandov@fb.com>, Sagi Grimberg <sagi@grimberg.me>,
	Dave Chinner <dchinner@redhat.com>,
	Kent Overstreet <kent.overstreet@gmail.com>,
	Mike Snitzer <snitzer@redhat.com>,
	dm-devel@redhat.com, Alexander Viro <viro@zeniv.linux.org.uk>,
	linux-fsdevel@vger.kernel.org, Shaohua Li <shli@kernel.org>,
	linux-raid@vger.kernel.org, David Sterba <dsterba@suse.com>,
	linux-btrfs@vger.kernel.org,
	"Darrick J . Wong" <darrick.wong@oracle.com>,
	linux-xfs@vger.kernel.org, Gao Xiang <gaoxiang25@huawei.com>,
	Christoph Hellwig <hch@lst.de>,
	linux-ext4@vger.kernel.org, Coly Li <colyli@suse.de>,
	linux-bcache@vger.kernel.org, Boaz Harrosh <ooo@electrozaur.com>,
	Bob Peterson <rpeterso@redhat.com>,
	cluster-devel@redhat.com, Ming Lei <ming.lei@redhat.com>
Subject: [PATCH V12 03/20] block: remove the "cluster" flag
Date: Mon, 26 Nov 2018 10:17:03 +0800	[thread overview]
Message-ID: <20181126021720.19471-4-ming.lei@redhat.com> (raw)
In-Reply-To: <20181126021720.19471-1-ming.lei@redhat.com>

From: Christoph Hellwig <hch@lst.de>

The cluster flag implements some very old SCSI behavior.  As far as I
can tell the original intent was to enable or disable any kind of
segment merging.  But the actually visible effect to the LLDD is that
it limits each segments to be inside a single page, which we can
also affect by setting the maximum segment size and the segment
boundary.

Signed-off-by: Christoph Hellwig <hch@lst.de>

Replace virt boundary with segment boundary limit.

Signed-off-by: Ming Lei <ming.lei@redhat.com>
---
 block/blk-merge.c       | 20 ++++++++------------
 block/blk-settings.c    |  3 ---
 block/blk-sysfs.c       |  5 +----
 drivers/scsi/scsi_lib.c | 20 ++++++++++++++++----
 include/linux/blkdev.h  |  6 ------
 5 files changed, 25 insertions(+), 29 deletions(-)

diff --git a/block/blk-merge.c b/block/blk-merge.c
index 6be04ef8da5b..e69d8f8ba819 100644
--- a/block/blk-merge.c
+++ b/block/blk-merge.c
@@ -195,7 +195,7 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
 			goto split;
 		}
 
-		if (bvprvp && blk_queue_cluster(q)) {
+		if (bvprvp) {
 			if (seg_size + bv.bv_len > queue_max_segment_size(q))
 				goto new_segment;
 			if (!biovec_phys_mergeable(q, bvprvp, &bv))
@@ -295,10 +295,10 @@ static unsigned int __blk_recalc_rq_segments(struct request_queue *q,
 					     bool no_sg_merge)
 {
 	struct bio_vec bv, bvprv = { NULL };
-	int cluster, prev = 0;
 	unsigned int seg_size, nr_phys_segs;
 	struct bio *fbio, *bbio;
 	struct bvec_iter iter;
+	bool prev = false;
 
 	if (!bio)
 		return 0;
@@ -313,7 +313,6 @@ static unsigned int __blk_recalc_rq_segments(struct request_queue *q,
 	}
 
 	fbio = bio;
-	cluster = blk_queue_cluster(q);
 	seg_size = 0;
 	nr_phys_segs = 0;
 	for_each_bio(bio) {
@@ -325,7 +324,7 @@ static unsigned int __blk_recalc_rq_segments(struct request_queue *q,
 			if (no_sg_merge)
 				goto new_segment;
 
-			if (prev && cluster) {
+			if (prev) {
 				if (seg_size + bv.bv_len
 				    > queue_max_segment_size(q))
 					goto new_segment;
@@ -343,7 +342,7 @@ static unsigned int __blk_recalc_rq_segments(struct request_queue *q,
 
 			nr_phys_segs++;
 			bvprv = bv;
-			prev = 1;
+			prev = true;
 			seg_size = bv.bv_len;
 		}
 		bbio = bio;
@@ -396,9 +395,6 @@ static int blk_phys_contig_segment(struct request_queue *q, struct bio *bio,
 {
 	struct bio_vec end_bv = { NULL }, nxt_bv;
 
-	if (!blk_queue_cluster(q))
-		return 0;
-
 	if (bio->bi_seg_back_size + nxt->bi_seg_front_size >
 	    queue_max_segment_size(q))
 		return 0;
@@ -415,12 +411,12 @@ static int blk_phys_contig_segment(struct request_queue *q, struct bio *bio,
 static inline void
 __blk_segment_map_sg(struct request_queue *q, struct bio_vec *bvec,
 		     struct scatterlist *sglist, struct bio_vec *bvprv,
-		     struct scatterlist **sg, int *nsegs, int *cluster)
+		     struct scatterlist **sg, int *nsegs)
 {
 
 	int nbytes = bvec->bv_len;
 
-	if (*sg && *cluster) {
+	if (*sg) {
 		if ((*sg)->length + nbytes > queue_max_segment_size(q))
 			goto new_segment;
 		if (!biovec_phys_mergeable(q, bvprv, bvec))
@@ -466,12 +462,12 @@ static int __blk_bios_map_sg(struct request_queue *q, struct bio *bio,
 {
 	struct bio_vec bvec, bvprv = { NULL };
 	struct bvec_iter iter;
-	int cluster = blk_queue_cluster(q), nsegs = 0;
+	int nsegs = 0;
 
 	for_each_bio(bio)
 		bio_for_each_segment(bvec, bio, iter)
 			__blk_segment_map_sg(q, &bvec, sglist, &bvprv, sg,
-					     &nsegs, &cluster);
+					     &nsegs);
 
 	return nsegs;
 }
diff --git a/block/blk-settings.c b/block/blk-settings.c
index 3abe831e92c8..3e7038e475ee 100644
--- a/block/blk-settings.c
+++ b/block/blk-settings.c
@@ -56,7 +56,6 @@ void blk_set_default_limits(struct queue_limits *lim)
 	lim->alignment_offset = 0;
 	lim->io_opt = 0;
 	lim->misaligned = 0;
-	lim->cluster = 1;
 	lim->zoned = BLK_ZONED_NONE;
 }
 EXPORT_SYMBOL(blk_set_default_limits);
@@ -547,8 +546,6 @@ int blk_stack_limits(struct queue_limits *t, struct queue_limits *b,
 	t->io_min = max(t->io_min, b->io_min);
 	t->io_opt = lcm_not_zero(t->io_opt, b->io_opt);
 
-	t->cluster &= b->cluster;
-
 	/* Physical block size a multiple of the logical block size? */
 	if (t->physical_block_size & (t->logical_block_size - 1)) {
 		t->physical_block_size = t->logical_block_size;
diff --git a/block/blk-sysfs.c b/block/blk-sysfs.c
index 80eef48fddc8..ef7b844a3e00 100644
--- a/block/blk-sysfs.c
+++ b/block/blk-sysfs.c
@@ -132,10 +132,7 @@ static ssize_t queue_max_integrity_segments_show(struct request_queue *q, char *
 
 static ssize_t queue_max_segment_size_show(struct request_queue *q, char *page)
 {
-	if (blk_queue_cluster(q))
-		return queue_var_show(queue_max_segment_size(q), (page));
-
-	return queue_var_show(PAGE_SIZE, (page));
+	return queue_var_show(queue_max_segment_size(q), page);
 }
 
 static ssize_t queue_logical_block_size_show(struct request_queue *q, char *page)
diff --git a/drivers/scsi/scsi_lib.c b/drivers/scsi/scsi_lib.c
index 0df15cb738d2..78d6d05992b0 100644
--- a/drivers/scsi/scsi_lib.c
+++ b/drivers/scsi/scsi_lib.c
@@ -1810,6 +1810,8 @@ static int scsi_map_queues(struct blk_mq_tag_set *set)
 void __scsi_init_queue(struct Scsi_Host *shost, struct request_queue *q)
 {
 	struct device *dev = shost->dma_dev;
+	unsigned max_segment_size = dma_get_max_seg_size(dev);
+	unsigned long segment_boundary = shost->dma_boundary;
 
 	/*
 	 * this limit is imposed by hardware restrictions
@@ -1828,13 +1830,23 @@ void __scsi_init_queue(struct Scsi_Host *shost, struct request_queue *q)
 	blk_queue_max_hw_sectors(q, shost->max_sectors);
 	if (shost->unchecked_isa_dma)
 		blk_queue_bounce_limit(q, BLK_BOUNCE_ISA);
-	blk_queue_segment_boundary(q, shost->dma_boundary);
 	dma_set_seg_boundary(dev, shost->dma_boundary);
 
-	blk_queue_max_segment_size(q, dma_get_max_seg_size(dev));
+	/*
+	 * Clustering is a really old concept from the stone age of Linux
+	 * SCSI support.  But the basic idea is that we never give the
+	 * driver a segment that spans multiple pages.  For that we need
+	 * to limit the segment size, and set the segment boundary so that
+	 * we never merge a second segment which is no page aligned.
+	 */
+	if (!shost->use_clustering) {
+		max_segment_size = min_t(unsigned, max_segment_size, PAGE_SIZE);
+		segment_boundary = min_t(unsigned, segment_boundary,
+					 PAGE_SIZE - 1);
+	}
 
-	if (!shost->use_clustering)
-		q->limits.cluster = 0;
+	blk_queue_max_segment_size(q, max_segment_size);
+	blk_queue_segment_boundary(q, segment_boundary);
 
 	/*
 	 * Set a reasonable default alignment:  The larger of 32-byte (dword),
diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h
index 9b53db06ad08..399a7a415609 100644
--- a/include/linux/blkdev.h
+++ b/include/linux/blkdev.h
@@ -341,7 +341,6 @@ struct queue_limits {
 
 	unsigned char		misaligned;
 	unsigned char		discard_misaligned;
-	unsigned char		cluster;
 	unsigned char		raid_partial_stripes_expensive;
 	enum blk_zoned_model	zoned;
 };
@@ -660,11 +659,6 @@ static inline bool queue_is_mq(struct request_queue *q)
 	return q->mq_ops;
 }
 
-static inline unsigned int blk_queue_cluster(struct request_queue *q)
-{
-	return q->limits.cluster;
-}
-
 static inline enum blk_zoned_model
 blk_queue_zoned_model(struct request_queue *q)
 {
-- 
2.9.5


WARNING: multiple messages have this Message-ID (diff)
From: Ming Lei <ming.lei@redhat.com>
To: cluster-devel.redhat.com
Subject: [Cluster-devel] [PATCH V12 03/20] block: remove the "cluster" flag
Date: Mon, 26 Nov 2018 10:17:03 +0800	[thread overview]
Message-ID: <20181126021720.19471-4-ming.lei@redhat.com> (raw)
In-Reply-To: <20181126021720.19471-1-ming.lei@redhat.com>

From: Christoph Hellwig <hch@lst.de>

The cluster flag implements some very old SCSI behavior.  As far as I
can tell the original intent was to enable or disable any kind of
segment merging.  But the actually visible effect to the LLDD is that
it limits each segments to be inside a single page, which we can
also affect by setting the maximum segment size and the segment
boundary.

Signed-off-by: Christoph Hellwig <hch@lst.de>

Replace virt boundary with segment boundary limit.

Signed-off-by: Ming Lei <ming.lei@redhat.com>
---
 block/blk-merge.c       | 20 ++++++++------------
 block/blk-settings.c    |  3 ---
 block/blk-sysfs.c       |  5 +----
 drivers/scsi/scsi_lib.c | 20 ++++++++++++++++----
 include/linux/blkdev.h  |  6 ------
 5 files changed, 25 insertions(+), 29 deletions(-)

diff --git a/block/blk-merge.c b/block/blk-merge.c
index 6be04ef8da5b..e69d8f8ba819 100644
--- a/block/blk-merge.c
+++ b/block/blk-merge.c
@@ -195,7 +195,7 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
 			goto split;
 		}
 
-		if (bvprvp && blk_queue_cluster(q)) {
+		if (bvprvp) {
 			if (seg_size + bv.bv_len > queue_max_segment_size(q))
 				goto new_segment;
 			if (!biovec_phys_mergeable(q, bvprvp, &bv))
@@ -295,10 +295,10 @@ static unsigned int __blk_recalc_rq_segments(struct request_queue *q,
 					     bool no_sg_merge)
 {
 	struct bio_vec bv, bvprv = { NULL };
-	int cluster, prev = 0;
 	unsigned int seg_size, nr_phys_segs;
 	struct bio *fbio, *bbio;
 	struct bvec_iter iter;
+	bool prev = false;
 
 	if (!bio)
 		return 0;
@@ -313,7 +313,6 @@ static unsigned int __blk_recalc_rq_segments(struct request_queue *q,
 	}
 
 	fbio = bio;
-	cluster = blk_queue_cluster(q);
 	seg_size = 0;
 	nr_phys_segs = 0;
 	for_each_bio(bio) {
@@ -325,7 +324,7 @@ static unsigned int __blk_recalc_rq_segments(struct request_queue *q,
 			if (no_sg_merge)
 				goto new_segment;
 
-			if (prev && cluster) {
+			if (prev) {
 				if (seg_size + bv.bv_len
 				    > queue_max_segment_size(q))
 					goto new_segment;
@@ -343,7 +342,7 @@ static unsigned int __blk_recalc_rq_segments(struct request_queue *q,
 
 			nr_phys_segs++;
 			bvprv = bv;
-			prev = 1;
+			prev = true;
 			seg_size = bv.bv_len;
 		}
 		bbio = bio;
@@ -396,9 +395,6 @@ static int blk_phys_contig_segment(struct request_queue *q, struct bio *bio,
 {
 	struct bio_vec end_bv = { NULL }, nxt_bv;
 
-	if (!blk_queue_cluster(q))
-		return 0;
-
 	if (bio->bi_seg_back_size + nxt->bi_seg_front_size >
 	    queue_max_segment_size(q))
 		return 0;
@@ -415,12 +411,12 @@ static int blk_phys_contig_segment(struct request_queue *q, struct bio *bio,
 static inline void
 __blk_segment_map_sg(struct request_queue *q, struct bio_vec *bvec,
 		     struct scatterlist *sglist, struct bio_vec *bvprv,
-		     struct scatterlist **sg, int *nsegs, int *cluster)
+		     struct scatterlist **sg, int *nsegs)
 {
 
 	int nbytes = bvec->bv_len;
 
-	if (*sg && *cluster) {
+	if (*sg) {
 		if ((*sg)->length + nbytes > queue_max_segment_size(q))
 			goto new_segment;
 		if (!biovec_phys_mergeable(q, bvprv, bvec))
@@ -466,12 +462,12 @@ static int __blk_bios_map_sg(struct request_queue *q, struct bio *bio,
 {
 	struct bio_vec bvec, bvprv = { NULL };
 	struct bvec_iter iter;
-	int cluster = blk_queue_cluster(q), nsegs = 0;
+	int nsegs = 0;
 
 	for_each_bio(bio)
 		bio_for_each_segment(bvec, bio, iter)
 			__blk_segment_map_sg(q, &bvec, sglist, &bvprv, sg,
-					     &nsegs, &cluster);
+					     &nsegs);
 
 	return nsegs;
 }
diff --git a/block/blk-settings.c b/block/blk-settings.c
index 3abe831e92c8..3e7038e475ee 100644
--- a/block/blk-settings.c
+++ b/block/blk-settings.c
@@ -56,7 +56,6 @@ void blk_set_default_limits(struct queue_limits *lim)
 	lim->alignment_offset = 0;
 	lim->io_opt = 0;
 	lim->misaligned = 0;
-	lim->cluster = 1;
 	lim->zoned = BLK_ZONED_NONE;
 }
 EXPORT_SYMBOL(blk_set_default_limits);
@@ -547,8 +546,6 @@ int blk_stack_limits(struct queue_limits *t, struct queue_limits *b,
 	t->io_min = max(t->io_min, b->io_min);
 	t->io_opt = lcm_not_zero(t->io_opt, b->io_opt);
 
-	t->cluster &= b->cluster;
-
 	/* Physical block size a multiple of the logical block size? */
 	if (t->physical_block_size & (t->logical_block_size - 1)) {
 		t->physical_block_size = t->logical_block_size;
diff --git a/block/blk-sysfs.c b/block/blk-sysfs.c
index 80eef48fddc8..ef7b844a3e00 100644
--- a/block/blk-sysfs.c
+++ b/block/blk-sysfs.c
@@ -132,10 +132,7 @@ static ssize_t queue_max_integrity_segments_show(struct request_queue *q, char *
 
 static ssize_t queue_max_segment_size_show(struct request_queue *q, char *page)
 {
-	if (blk_queue_cluster(q))
-		return queue_var_show(queue_max_segment_size(q), (page));
-
-	return queue_var_show(PAGE_SIZE, (page));
+	return queue_var_show(queue_max_segment_size(q), page);
 }
 
 static ssize_t queue_logical_block_size_show(struct request_queue *q, char *page)
diff --git a/drivers/scsi/scsi_lib.c b/drivers/scsi/scsi_lib.c
index 0df15cb738d2..78d6d05992b0 100644
--- a/drivers/scsi/scsi_lib.c
+++ b/drivers/scsi/scsi_lib.c
@@ -1810,6 +1810,8 @@ static int scsi_map_queues(struct blk_mq_tag_set *set)
 void __scsi_init_queue(struct Scsi_Host *shost, struct request_queue *q)
 {
 	struct device *dev = shost->dma_dev;
+	unsigned max_segment_size = dma_get_max_seg_size(dev);
+	unsigned long segment_boundary = shost->dma_boundary;
 
 	/*
 	 * this limit is imposed by hardware restrictions
@@ -1828,13 +1830,23 @@ void __scsi_init_queue(struct Scsi_Host *shost, struct request_queue *q)
 	blk_queue_max_hw_sectors(q, shost->max_sectors);
 	if (shost->unchecked_isa_dma)
 		blk_queue_bounce_limit(q, BLK_BOUNCE_ISA);
-	blk_queue_segment_boundary(q, shost->dma_boundary);
 	dma_set_seg_boundary(dev, shost->dma_boundary);
 
-	blk_queue_max_segment_size(q, dma_get_max_seg_size(dev));
+	/*
+	 * Clustering is a really old concept from the stone age of Linux
+	 * SCSI support.  But the basic idea is that we never give the
+	 * driver a segment that spans multiple pages.  For that we need
+	 * to limit the segment size, and set the segment boundary so that
+	 * we never merge a second segment which is no page aligned.
+	 */
+	if (!shost->use_clustering) {
+		max_segment_size = min_t(unsigned, max_segment_size, PAGE_SIZE);
+		segment_boundary = min_t(unsigned, segment_boundary,
+					 PAGE_SIZE - 1);
+	}
 
-	if (!shost->use_clustering)
-		q->limits.cluster = 0;
+	blk_queue_max_segment_size(q, max_segment_size);
+	blk_queue_segment_boundary(q, segment_boundary);
 
 	/*
 	 * Set a reasonable default alignment:  The larger of 32-byte (dword),
diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h
index 9b53db06ad08..399a7a415609 100644
--- a/include/linux/blkdev.h
+++ b/include/linux/blkdev.h
@@ -341,7 +341,6 @@ struct queue_limits {
 
 	unsigned char		misaligned;
 	unsigned char		discard_misaligned;
-	unsigned char		cluster;
 	unsigned char		raid_partial_stripes_expensive;
 	enum blk_zoned_model	zoned;
 };
@@ -660,11 +659,6 @@ static inline bool queue_is_mq(struct request_queue *q)
 	return q->mq_ops;
 }
 
-static inline unsigned int blk_queue_cluster(struct request_queue *q)
-{
-	return q->limits.cluster;
-}
-
 static inline enum blk_zoned_model
 blk_queue_zoned_model(struct request_queue *q)
 {
-- 
2.9.5



  parent reply	other threads:[~2018-11-26  2:17 UTC|newest]

Thread overview: 144+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2018-11-26  2:17 [PATCH V12 00/20] block: support multi-page bvec Ming Lei
2018-11-26  2:17 ` [Cluster-devel] " Ming Lei
2018-11-26  2:17 ` Ming Lei
2018-11-26  2:17 ` [PATCH V12 01/20] btrfs: remove various bio_offset arguments Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26 22:28   ` Omar Sandoval
2018-11-26 22:28     ` [Cluster-devel] " Omar Sandoval
2018-11-26 22:28     ` Omar Sandoval
2018-11-26  2:17 ` [PATCH V12 02/20] btrfs: look at bi_size for repair decisions Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26 22:46   ` Omar Sandoval
2018-11-26 22:46     ` [Cluster-devel] " Omar Sandoval
2018-11-26 22:46     ` Omar Sandoval
2018-11-26  2:17 ` Ming Lei [this message]
2018-11-26  2:17   ` [Cluster-devel] [PATCH V12 03/20] block: remove the "cluster" flag Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26 22:12   ` Omar Sandoval
2018-11-26 22:12     ` [Cluster-devel] " Omar Sandoval
2018-11-26 22:12     ` Omar Sandoval
2018-11-26  2:17 ` [PATCH V12 04/20] block: don't use bio->bi_vcnt to figure out segment number Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26 22:14   ` Omar Sandoval
2018-11-26 22:14     ` [Cluster-devel] " Omar Sandoval
2018-11-26 22:14     ` Omar Sandoval
2018-11-26  2:17 ` [PATCH V12 05/20] block: remove bvec_iter_rewind() Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26 12:53   ` Christoph Hellwig
2018-11-26 12:53     ` [Cluster-devel] " Christoph Hellwig
2018-11-26 12:53     ` Christoph Hellwig
2018-11-26 22:16   ` Omar Sandoval
2018-11-26 22:16     ` [Cluster-devel] " Omar Sandoval
2018-11-26 22:16     ` Omar Sandoval
2018-11-26  2:17 ` [PATCH V12 06/20] block: rename bvec helpers Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26  8:17   ` Miguel Ojeda
2018-11-26  8:17     ` [Cluster-devel] " Miguel Ojeda
2018-11-26 12:54   ` Christoph Hellwig
2018-11-26 12:54     ` [Cluster-devel] " Christoph Hellwig
2018-11-26 12:54     ` Christoph Hellwig
2018-11-26 22:18   ` Omar Sandoval
2018-11-26 22:18     ` [Cluster-devel] " Omar Sandoval
2018-11-26 22:18     ` Omar Sandoval
2018-11-26  2:17 ` [PATCH V12 07/20] block: introduce multi-page " Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26 12:55   ` Christoph Hellwig
2018-11-26 12:55     ` [Cluster-devel] " Christoph Hellwig
2018-11-26 12:55     ` Christoph Hellwig
2018-11-26  2:17 ` [PATCH V12 08/20] block: introduce bio_for_each_bvec() and rq_for_each_bvec() Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26 12:55   ` Christoph Hellwig
2018-11-26 12:55     ` [Cluster-devel] " Christoph Hellwig
2018-11-26 12:55     ` Christoph Hellwig
2018-11-26  2:17 ` [PATCH V12 09/20] block: use bio_for_each_bvec() to compute multi-page bvec count Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26 13:11   ` Christoph Hellwig
2018-11-26 13:11     ` [Cluster-devel] " Christoph Hellwig
2018-11-26 13:11     ` Christoph Hellwig
2018-11-26 22:37   ` Omar Sandoval
2018-11-26 22:37     ` [Cluster-devel] " Omar Sandoval
2018-11-26 22:37     ` Omar Sandoval
2018-11-26  2:17 ` [PATCH V12 10/20] block: use bio_for_each_bvec() to map sg Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26 13:11   ` Christoph Hellwig
2018-11-26 13:11     ` [Cluster-devel] " Christoph Hellwig
2018-11-26 13:11     ` Christoph Hellwig
2018-11-26  2:17 ` [PATCH V12 11/20] block: introduce bvec_last_segment() Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26  2:17 ` [PATCH V12 12/20] fs/buffer.c: use bvec iterator to truncate the bio Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26  2:17 ` [PATCH V12 13/20] block: loop: pass multi-page bvec to iov_iter Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26 22:39   ` Omar Sandoval
2018-11-26 22:39     ` [Cluster-devel] " Omar Sandoval
2018-11-26 22:39     ` Omar Sandoval
2018-11-26  2:17 ` [PATCH V12 14/20] bcache: avoid to use bio_for_each_segment_all() in bch_bio_alloc_pages() Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26 22:40   ` Omar Sandoval
2018-11-26 22:40     ` [Cluster-devel] " Omar Sandoval
2018-11-26 22:40     ` Omar Sandoval
2018-11-26  2:17 ` [PATCH V12 15/20] block: allow bio_for_each_segment_all() to iterate over multi-page bvec Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26 22:42   ` Omar Sandoval
2018-11-26 22:42     ` [Cluster-devel] " Omar Sandoval
2018-11-26 22:42     ` Omar Sandoval
2018-11-26  2:17 ` [PATCH V12 16/20] block: enable multipage bvecs Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26 12:58   ` Christoph Hellwig
2018-11-26 12:58     ` [Cluster-devel] " Christoph Hellwig
2018-11-26 12:58     ` Christoph Hellwig
2018-11-27  2:25     ` Ming Lei
2018-11-27  2:25       ` [Cluster-devel] " Ming Lei
2018-11-27  2:25       ` Ming Lei
2018-11-26 22:56   ` Omar Sandoval
2018-11-26 22:56     ` [Cluster-devel] " Omar Sandoval
2018-11-26 22:56     ` Omar Sandoval
2018-11-26  2:17 ` [PATCH V12 17/20] block: always define BIO_MAX_PAGES as 256 Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26 22:44   ` Omar Sandoval
2018-11-26 22:44     ` [Cluster-devel] " Omar Sandoval
2018-11-26 22:44     ` Omar Sandoval
2018-11-26  2:17 ` [PATCH V12 18/20] block: document usage of bio iterator helpers Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26 12:59   ` Christoph Hellwig
2018-11-26 12:59     ` [Cluster-devel] " Christoph Hellwig
2018-11-26 12:59     ` Christoph Hellwig
2018-11-26 22:43   ` Omar Sandoval
2018-11-26 22:43     ` [Cluster-devel] " Omar Sandoval
2018-11-26 22:43     ` Omar Sandoval
2018-11-26  2:17 ` [PATCH V12 19/20] block: kill QUEUE_FLAG_NO_SG_MERGE Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-26  2:17 ` [PATCH V12 20/20] block: kill BLK_MQ_F_SG_MERGE Ming Lei
2018-11-26  2:17   ` [Cluster-devel] " Ming Lei
2018-11-26  2:17   ` Ming Lei
2018-11-28 13:44 ` [PATCH V12 00/20] block: support multi-page bvec Jens Axboe
2018-11-28 13:44   ` [Cluster-devel] " Jens Axboe
2018-11-28 13:44   ` Jens Axboe
2018-11-29  1:30   ` Ming Lei
2018-11-29  1:30     ` [Cluster-devel] " Ming Lei
2018-11-29  1:30     ` Ming Lei
2018-11-29  1:30     ` Ming Lei
2018-11-29  2:20     ` Jens Axboe
2018-11-29  2:20       ` [Cluster-devel] " Jens Axboe
2018-11-29  2:20       ` Jens Axboe
2018-11-29  3:30       ` Ming Lei
2018-11-29  3:30         ` [Cluster-devel] " Ming Lei
2018-11-29  3:30         ` Ming Lei

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20181126021720.19471-4-ming.lei@redhat.com \
    --to=ming.lei@redhat.com \
    --cc=axboe@kernel.dk \
    --cc=cluster-devel@redhat.com \
    --cc=colyli@suse.de \
    --cc=darrick.wong@oracle.com \
    --cc=dchinner@redhat.com \
    --cc=dm-devel@redhat.com \
    --cc=dsterba@suse.com \
    --cc=gaoxiang25@huawei.com \
    --cc=hch@lst.de \
    --cc=kent.overstreet@gmail.com \
    --cc=linux-bcache@vger.kernel.org \
    --cc=linux-block@vger.kernel.org \
    --cc=linux-btrfs@vger.kernel.org \
    --cc=linux-ext4@vger.kernel.org \
    --cc=linux-fsdevel@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=linux-raid@vger.kernel.org \
    --cc=linux-xfs@vger.kernel.org \
    --cc=ooo@electrozaur.com \
    --cc=osandov@fb.com \
    --cc=rpeterso@redhat.com \
    --cc=sagi@grimberg.me \
    --cc=shli@kernel.org \
    --cc=snitzer@redhat.com \
    --cc=tytso@mit.edu \
    --cc=viro@zeniv.linux.org.uk \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.