All of lore.kernel.org
 help / color / mirror / Atom feed
From: Jan Kara <jack@suse.cz>
To: linux-mm@kvack.org
Cc: linux-fsdevel@vger.kernel.org, linux-nvdimm@lists.01.org,
	Andrew Morton <akpm@linux-foundation.org>,
	Ross Zwisler <ross.zwisler@linux.intel.com>,
	Jan Kara <jack@suse.cz>
Subject: [PATCH 19/21] dax: Make cache flushing protected by entry lock
Date: Tue,  1 Nov 2016 23:36:30 +0100	[thread overview]
Message-ID: <1478039794-20253-25-git-send-email-jack@suse.cz> (raw)
In-Reply-To: <1478039794-20253-1-git-send-email-jack@suse.cz>

Currently, flushing of caches for DAX mappings was ignoring entry lock.
So far this was ok (modulo a bug that a difference in entry lock could
cause cache flushing to be mistakenly skipped) but in the following
patches we will write-protect PTEs on cache flushing and clear dirty
tags. For that we will need more exclusion. So do cache flushing under
an entry lock. This allows us to remove one lock-unlock pair of
mapping->tree_lock as a bonus.

Reviewed-by: Ross Zwisler <ross.zwisler@linux.intel.com>
Signed-off-by: Jan Kara <jack@suse.cz>
---
 fs/dax.c | 61 +++++++++++++++++++++++++++++++++++++++----------------------
 1 file changed, 39 insertions(+), 22 deletions(-)

diff --git a/fs/dax.c b/fs/dax.c
index b4ea4bcca9e7..857117b6db6b 100644
--- a/fs/dax.c
+++ b/fs/dax.c
@@ -619,32 +619,50 @@ static int dax_writeback_one(struct block_device *bdev,
 		struct address_space *mapping, pgoff_t index, void *entry)
 {
 	struct radix_tree_root *page_tree = &mapping->page_tree;
-	struct radix_tree_node *node;
 	struct blk_dax_ctl dax;
-	void **slot;
+	void *entry2, **slot;
 	int ret = 0;
 
-	spin_lock_irq(&mapping->tree_lock);
 	/*
-	 * Regular page slots are stabilized by the page lock even
-	 * without the tree itself locked.  These unlocked entries
-	 * need verification under the tree lock.
+	 * A page got tagged dirty in DAX mapping? Something is seriously
+	 * wrong.
 	 */
-	if (!__radix_tree_lookup(page_tree, index, &node, &slot))
-		goto unlock;
-	if (*slot != entry)
-		goto unlock;
-
-	/* another fsync thread may have already written back this entry */
-	if (!radix_tree_tag_get(page_tree, index, PAGECACHE_TAG_TOWRITE))
-		goto unlock;
+	if (WARN_ON(!radix_tree_exceptional_entry(entry)))
+		return -EIO;
 
+	spin_lock_irq(&mapping->tree_lock);
+	entry2 = get_unlocked_mapping_entry(mapping, index, &slot);
+	/* Entry got punched out / reallocated? */
+	if (!entry2 || !radix_tree_exceptional_entry(entry2))
+		goto put_unlocked;
+	/*
+	 * Entry got reallocated elsewhere? No need to writeback. We have to
+	 * compare sectors as we must not bail out due to difference in lockbit
+	 * or entry type.
+	 */
+	if (dax_radix_sector(entry2) != dax_radix_sector(entry))
+		goto put_unlocked;
 	if (WARN_ON_ONCE(dax_is_empty_entry(entry) ||
 				dax_is_zero_entry(entry))) {
 		ret = -EIO;
-		goto unlock;
+		goto put_unlocked;
 	}
 
+	/* Another fsync thread may have already written back this entry */
+	if (!radix_tree_tag_get(page_tree, index, PAGECACHE_TAG_TOWRITE))
+		goto put_unlocked;
+	/* Lock the entry to serialize with page faults */
+	entry = lock_slot(mapping, slot);
+	/*
+	 * We can clear the tag now but we have to be careful so that concurrent
+	 * dax_writeback_one() calls for the same index cannot finish before we
+	 * actually flush the caches. This is achieved as the calls will look
+	 * at the entry only under tree_lock and once they do that they will
+	 * see the entry locked and wait for it to unlock.
+	 */
+	radix_tree_tag_clear(page_tree, index, PAGECACHE_TAG_TOWRITE);
+	spin_unlock_irq(&mapping->tree_lock);
+
 	/*
 	 * Even if dax_writeback_mapping_range() was given a wbc->range_start
 	 * in the middle of a PMD, the 'index' we are given will be aligned to
@@ -654,15 +672,16 @@ static int dax_writeback_one(struct block_device *bdev,
 	 */
 	dax.sector = dax_radix_sector(entry);
 	dax.size = PAGE_SIZE << dax_radix_order(entry);
-	spin_unlock_irq(&mapping->tree_lock);
 
 	/*
 	 * We cannot hold tree_lock while calling dax_map_atomic() because it
 	 * eventually calls cond_resched().
 	 */
 	ret = dax_map_atomic(bdev, &dax);
-	if (ret < 0)
+	if (ret < 0) {
+		put_locked_mapping_entry(mapping, index, entry);
 		return ret;
+	}
 
 	if (WARN_ON_ONCE(ret < dax.size)) {
 		ret = -EIO;
@@ -670,15 +689,13 @@ static int dax_writeback_one(struct block_device *bdev,
 	}
 
 	wb_cache_pmem(dax.addr, dax.size);
-
-	spin_lock_irq(&mapping->tree_lock);
-	radix_tree_tag_clear(page_tree, index, PAGECACHE_TAG_TOWRITE);
-	spin_unlock_irq(&mapping->tree_lock);
  unmap:
 	dax_unmap_atomic(bdev, &dax);
+	put_locked_mapping_entry(mapping, index, entry);
 	return ret;
 
- unlock:
+ put_unlocked:
+	put_unlocked_mapping_entry(mapping, index, entry2);
 	spin_unlock_irq(&mapping->tree_lock);
 	return ret;
 }
-- 
2.6.6

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

WARNING: multiple messages have this Message-ID (diff)
From: Jan Kara <jack@suse.cz>
To: <linux-mm@kvack.org>
Cc: <linux-fsdevel@vger.kernel.org>,
	linux-nvdimm@lists.01.org,
	Andrew Morton <akpm@linux-foundation.org>,
	Ross Zwisler <ross.zwisler@linux.intel.com>,
	Jan Kara <jack@suse.cz>
Subject: [PATCH 19/21] dax: Make cache flushing protected by entry lock
Date: Tue,  1 Nov 2016 23:36:30 +0100	[thread overview]
Message-ID: <1478039794-20253-25-git-send-email-jack@suse.cz> (raw)
In-Reply-To: <1478039794-20253-1-git-send-email-jack@suse.cz>

Currently, flushing of caches for DAX mappings was ignoring entry lock.
So far this was ok (modulo a bug that a difference in entry lock could
cause cache flushing to be mistakenly skipped) but in the following
patches we will write-protect PTEs on cache flushing and clear dirty
tags. For that we will need more exclusion. So do cache flushing under
an entry lock. This allows us to remove one lock-unlock pair of
mapping->tree_lock as a bonus.

Reviewed-by: Ross Zwisler <ross.zwisler@linux.intel.com>
Signed-off-by: Jan Kara <jack@suse.cz>
---
 fs/dax.c | 61 +++++++++++++++++++++++++++++++++++++++----------------------
 1 file changed, 39 insertions(+), 22 deletions(-)

diff --git a/fs/dax.c b/fs/dax.c
index b4ea4bcca9e7..857117b6db6b 100644
--- a/fs/dax.c
+++ b/fs/dax.c
@@ -619,32 +619,50 @@ static int dax_writeback_one(struct block_device *bdev,
 		struct address_space *mapping, pgoff_t index, void *entry)
 {
 	struct radix_tree_root *page_tree = &mapping->page_tree;
-	struct radix_tree_node *node;
 	struct blk_dax_ctl dax;
-	void **slot;
+	void *entry2, **slot;
 	int ret = 0;
 
-	spin_lock_irq(&mapping->tree_lock);
 	/*
-	 * Regular page slots are stabilized by the page lock even
-	 * without the tree itself locked.  These unlocked entries
-	 * need verification under the tree lock.
+	 * A page got tagged dirty in DAX mapping? Something is seriously
+	 * wrong.
 	 */
-	if (!__radix_tree_lookup(page_tree, index, &node, &slot))
-		goto unlock;
-	if (*slot != entry)
-		goto unlock;
-
-	/* another fsync thread may have already written back this entry */
-	if (!radix_tree_tag_get(page_tree, index, PAGECACHE_TAG_TOWRITE))
-		goto unlock;
+	if (WARN_ON(!radix_tree_exceptional_entry(entry)))
+		return -EIO;
 
+	spin_lock_irq(&mapping->tree_lock);
+	entry2 = get_unlocked_mapping_entry(mapping, index, &slot);
+	/* Entry got punched out / reallocated? */
+	if (!entry2 || !radix_tree_exceptional_entry(entry2))
+		goto put_unlocked;
+	/*
+	 * Entry got reallocated elsewhere? No need to writeback. We have to
+	 * compare sectors as we must not bail out due to difference in lockbit
+	 * or entry type.
+	 */
+	if (dax_radix_sector(entry2) != dax_radix_sector(entry))
+		goto put_unlocked;
 	if (WARN_ON_ONCE(dax_is_empty_entry(entry) ||
 				dax_is_zero_entry(entry))) {
 		ret = -EIO;
-		goto unlock;
+		goto put_unlocked;
 	}
 
+	/* Another fsync thread may have already written back this entry */
+	if (!radix_tree_tag_get(page_tree, index, PAGECACHE_TAG_TOWRITE))
+		goto put_unlocked;
+	/* Lock the entry to serialize with page faults */
+	entry = lock_slot(mapping, slot);
+	/*
+	 * We can clear the tag now but we have to be careful so that concurrent
+	 * dax_writeback_one() calls for the same index cannot finish before we
+	 * actually flush the caches. This is achieved as the calls will look
+	 * at the entry only under tree_lock and once they do that they will
+	 * see the entry locked and wait for it to unlock.
+	 */
+	radix_tree_tag_clear(page_tree, index, PAGECACHE_TAG_TOWRITE);
+	spin_unlock_irq(&mapping->tree_lock);
+
 	/*
 	 * Even if dax_writeback_mapping_range() was given a wbc->range_start
 	 * in the middle of a PMD, the 'index' we are given will be aligned to
@@ -654,15 +672,16 @@ static int dax_writeback_one(struct block_device *bdev,
 	 */
 	dax.sector = dax_radix_sector(entry);
 	dax.size = PAGE_SIZE << dax_radix_order(entry);
-	spin_unlock_irq(&mapping->tree_lock);
 
 	/*
 	 * We cannot hold tree_lock while calling dax_map_atomic() because it
 	 * eventually calls cond_resched().
 	 */
 	ret = dax_map_atomic(bdev, &dax);
-	if (ret < 0)
+	if (ret < 0) {
+		put_locked_mapping_entry(mapping, index, entry);
 		return ret;
+	}
 
 	if (WARN_ON_ONCE(ret < dax.size)) {
 		ret = -EIO;
@@ -670,15 +689,13 @@ static int dax_writeback_one(struct block_device *bdev,
 	}
 
 	wb_cache_pmem(dax.addr, dax.size);
-
-	spin_lock_irq(&mapping->tree_lock);
-	radix_tree_tag_clear(page_tree, index, PAGECACHE_TAG_TOWRITE);
-	spin_unlock_irq(&mapping->tree_lock);
  unmap:
 	dax_unmap_atomic(bdev, &dax);
+	put_locked_mapping_entry(mapping, index, entry);
 	return ret;
 
- unlock:
+ put_unlocked:
+	put_unlocked_mapping_entry(mapping, index, entry2);
 	spin_unlock_irq(&mapping->tree_lock);
 	return ret;
 }
-- 
2.6.6

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

  parent reply	other threads:[~2016-11-01 22:36 UTC|newest]

Thread overview: 111+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2016-11-01 22:36 [PATCH 0/21 v4] dax: Clear dirty bits after flushing caches Jan Kara
2016-11-01 22:36 ` Jan Kara
2016-11-01 22:36 ` Jan Kara
2016-11-01 22:36 ` [PATCH 01/20] mm: Change type of vmf->virtual_address Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-02  9:55   ` Kirill A. Shutemov
2016-11-02  9:55     ` Kirill A. Shutemov
2016-11-01 22:36 ` [PATCH 01/21] mm: Join struct fault_env and vm_fault Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-02  9:58   ` Kirill A. Shutemov
2016-11-02  9:58     ` Kirill A. Shutemov
2016-11-04  4:32     ` Jan Kara
2016-11-04  4:32       ` Jan Kara
2016-11-01 22:36 ` [PATCH 02/20] " Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 02/21] mm: Use vmf->address instead of of vmf->virtual_address Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-02  4:18   ` Hillf Danton
2016-11-02  4:18     ` Hillf Danton
2016-11-02  4:18     ` Hillf Danton
2016-11-04  3:46     ` Jan Kara
2016-11-04  3:46       ` Jan Kara
2016-11-01 22:36 ` [PATCH 03/21] mm: Use pgoff in struct vm_fault instead of passing it separately Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 04/21] mm: Use passed vm_fault structure in __do_fault() Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 05/21] mm: Trim __do_fault() arguments Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 06/21] mm: Use passed vm_fault structure for in wp_pfn_shared() Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 06/20] mm: Use pass " Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 07/21] mm: Add orig_pte field into vm_fault Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 08/21] mm: Allow full handling of COW faults in ->fault handlers Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 09/21] mm: Factor out functionality to finish page faults Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 10/21] mm: Move handling of COW faults into DAX code Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 11/21] mm: Remove unnecessary vma->vm_ops check Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 12/21] mm: Factor out common parts of write fault handling Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 13/21] mm: Pass vm_fault structure into do_page_mkwrite() Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 14/21] mm: Use vmf->page during WP faults Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 15/21] mm: Move part of wp_page_reuse() into the single call site Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 16/21] mm: Provide helper for finishing mkwrite faults Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 17/21] mm: Change return values of finish_mkwrite_fault() Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 17/20] mm: Export follow_pte() Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 18/20] dax: Make cache flushing protected by entry lock Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 18/21] mm: Export follow_pte() Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` Jan Kara [this message]
2016-11-01 22:36   ` [PATCH 19/21] dax: Make cache flushing protected by entry lock Jan Kara
2016-11-01 22:36 ` [PATCH 19/20] dax: Protect PTE modification on WP fault by radix tree " Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 20/20] dax: Clear dirty entry tags on cache flush Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 20/21] dax: Protect PTE modification on WP fault by radix tree entry lock Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 22:36 ` [PATCH 21/21] dax: Clear dirty entry tags on cache flush Jan Kara
2016-11-01 22:36   ` Jan Kara
2016-11-01 23:13 ` [PATCH 0/21 v4] dax: Clear dirty bits after flushing caches Jan Kara
2016-11-01 23:13   ` Jan Kara
2016-11-02 10:02   ` Kirill A. Shutemov
2016-11-02 10:02     ` Kirill A. Shutemov
2016-11-03 20:46     ` Jan Kara
2016-11-03 20:46       ` Jan Kara
2016-11-02  5:17 ` Ross Zwisler
2016-11-02  5:17   ` Ross Zwisler
2016-11-04  4:46   ` Jan Kara
2016-11-04  4:46     ` Jan Kara
2016-11-04 18:14     ` Jan Kara
2016-11-04 18:14       ` Jan Kara
2016-11-04  4:24 [PATCH 0/21 v4 RESEND] " Jan Kara
2016-11-04  4:25 ` [PATCH 19/21] dax: Make cache flushing protected by entry lock Jan Kara
2016-11-04  4:25   ` Jan Kara
2016-11-04  4:25   ` Jan Kara

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=1478039794-20253-25-git-send-email-jack@suse.cz \
    --to=jack@suse.cz \
    --cc=akpm@linux-foundation.org \
    --cc=linux-fsdevel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=linux-nvdimm@lists.01.org \
    --cc=ross.zwisler@linux.intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.