All of lore.kernel.org
 help / color / mirror / Atom feed
From: Jeff Layton <jlayton@kernel.org>
To: Dave Chinner <david@fromorbit.com>
Cc: "J. Bruce Fields" <bfields@fieldses.org>,
	linux-fsdevel@vger.kernel.org, linux-kernel@vger.kernel.org,
	hch@lst.de, neilb@suse.de, amir73il@gmail.com, jack@suse.de,
	viro@zeniv.linux.org.uk
Subject: Re: [PATCH 00/19] fs: rework and optimize i_version handling in filesystems
Date: Thu, 14 Dec 2017 09:14:47 -0500	[thread overview]
Message-ID: <1513260887.3504.12.camel@kernel.org> (raw)
In-Reply-To: <1513209746.3498.59.camel@kernel.org>

On Wed, 2017-12-13 at 19:02 -0500, Jeff Layton wrote:
> On Thu, 2017-12-14 at 10:03 +1100, Dave Chinner wrote:
> > On Wed, Dec 13, 2017 at 03:14:28PM -0500, Jeff Layton wrote:
> > > On Wed, 2017-12-13 at 10:05 -0500, J. Bruce Fields wrote:
> > > > This is great, thanks.
> > > > 
> > > > On Wed, Dec 13, 2017 at 09:19:58AM -0500, Jeff Layton wrote:
> > > > > With this, we reduce inode metadata updates across all 3 filesystems
> > > > > down to roughly the frequency of the timestamp granularity, particularly
> > > > > when it's not being queried (the vastly common case).
> > > > > 
> > > > > The pessimal workload here is 1 byte writes, and it helps that
> > > > > significantly. Of course, that's not what we'd consider a real-world
> > > > > workload.
> > > > > 
> > > > > A tiobench-example.fio workload also shows some modest performance
> > > > > gains, and I've gotten mails from the kernel test robot that show some
> > > > > significant performance gains on some microbenchmarks (case-msync-mt in
> > > > > the vm-scalability testsuite to be specific), with an earlier version of
> > > > > this set.
> > > > > 
> > > > > With larger writes, the gains with this patchset mostly vaporize,
> > > > > but it does not seem to cause performance to regress anywhere, AFAICT.
> > > > > 
> > > > > I'm happy to run other workloads if anyone can suggest them.
> > > > > 
> > > > > At this point, the patchset works and does what it's expected to do in
> > > > > my own testing. It seems like it's at least a modest performance win
> > > > > across all 3 major disk-based filesystems. It may also encourage others
> > > > > to implement i_version as well since it reduces the cost.
> > > > 
> > > > Do you have an idea what the remaining cost is?
> > > > 
> > > > Especially in the ext4 case, are you still able to measure any
> > > > difference in performance between the cases where i_version is turned on
> > > > and off, after these patches?
> > > 
> > > Attached is a fio jobfile + the output from 3 different runs using it
> > > with ext4. This one is using 4k writes. There was no querying of
> > > i_version during the runs.  I've done several runs with each and these
> > > are pretty representative of the results:
> > > 
> > > old = 4.15-rc3, i_version enabled
> > > ivers = 4.15-rc3 + these patches, i_version enabled
> > > noivers = 4.15-rc3 + these patches, i_version disabled
> > > 
> > > To snip out the run status lines:
> > > 
> > > old:
> > > WRITE: bw=85.6MiB/s (89.8MB/s), 9994KiB/s-11.1MiB/s (10.2MB/s-11.7MB/s), io=50.2GiB (53.8GB), run=600001-600001msec
> > > 
> > > ivers:
> > > WRITE: bw=110MiB/s (115MB/s), 13.5MiB/s-14.2MiB/s (14.1MB/s-14.9MB/s), io=64.3GiB (69.0GB), run=600001-600001msec
> > > 
> > > noivers:
> > > WRITE: bw=117MiB/s (123MB/s), 14.2MiB/s-15.2MiB/s (14.9MB/s-15.9MB/s), io=68.7GiB (73.8GB), run=600001-600001msec
> > > 
> > > So, I see some performance degradation with -o iversion compared to not
> > > having it enabled (maybe due to the extra atomic fetches?), but this set
> > > erases most of the difference.
> > 
> > So what is the performance difference when something is actively
> > querying the i_version counter as fast as it can (e.g. file being
> > constantly stat()d via NFS whilst being modified)? How does the
> > performance compare to the old code in that case?
> > 
> 
> I haven't benchmarked that with the latest set, but I did with the set
> that I posted around a year ago. Basically I just ran a similar test to
> this, and had another shell open doing statx(..., STATX_VERSION, ...);
> the thing in a tight loop.
> 
> I did see some performance hit vs. the case where no one is viewing it,
> but it was still significantly faster than the unpatched version that
> was incrementing the counter every time.
> 
> That was on a different test rig, and the patchset has some small
> differences now. I'll see if I can get some hard numbers with such a
> testcase soon.

I reran the test on xfs, with another shell running test-statx to query
the i_version value in a tight loop:

    WRITE: bw=110MiB/s (115MB/s), 13.6MiB/s-14.0MiB/s (14.2MB/s-14.7MB/s), io=64.5GiB (69.3GB), run=600001-600001msec

...contrast that with the run where I was not doing any queries:

   WRITE: bw=129MiB/s (136MB/s), 15.8MiB/s-16.6MiB/s (16.6MB/s-17.4MB/s), io=75.9GiB (81.5GB), run=600001-600001msec

...vs the unpatched kernel:

   WRITE: bw=86.7MiB/s (90.0MB/s), 9689KiB/s-11.7MiB/s (9921kB/s-12.2MB/s), io=50.8GiB (54.6GB), run=600001-600002msec

There is some clear peformance impact when you are running frequent
queries of the i_version.

My gut feeling is that you could probably make the new code perform
worse than the old if you were to _really_ hammer the inode with queries
for the i_version (probably from many threads in parallel) while doing a
lot of small writes to it.

That'd be a pretty unusual workload though.
-- 
Jeff Layton <jlayton@kernel.org>

  reply	other threads:[~2017-12-14 14:14 UTC|newest]

Thread overview: 46+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2017-12-13 14:19 [PATCH 00/19] fs: rework and optimize i_version handling in filesystems Jeff Layton
2017-12-13 14:19 ` [PATCH 01/19] fs: new API for handling inode->i_version Jeff Layton
2017-12-13 22:04   ` NeilBrown
2017-12-14  0:27     ` Jeff Layton
2017-12-16  4:17       ` NeilBrown
2017-12-17 13:01         ` Jeff Layton
2017-12-18 14:03         ` Jeff Layton
2017-12-13 14:20 ` [PATCH 02/19] fs: don't take the i_lock in inode_inc_iversion Jeff Layton
2017-12-13 21:52   ` Jeff Layton
2017-12-13 22:07     ` NeilBrown
2017-12-13 14:20 ` [PATCH 03/19] fat: convert to new i_version API Jeff Layton
2017-12-13 14:20 ` [PATCH 04/19] affs: " Jeff Layton
2017-12-13 14:20 ` [PATCH 05/19] afs: " Jeff Layton
2017-12-13 14:20 ` [PATCH 06/19] btrfs: " Jeff Layton
2017-12-13 14:20 ` [PATCH 07/19] exofs: switch " Jeff Layton
2017-12-13 14:20 ` [PATCH 08/19] ext2: convert " Jeff Layton
2017-12-18 12:47   ` Jan Kara
2017-12-13 14:20 ` [PATCH 09/19] ext4: " Jeff Layton
2017-12-14 21:52   ` Theodore Ts'o
2017-12-13 14:20 ` [PATCH 10/19] nfs: " Jeff Layton
2017-12-13 14:20 ` [PATCH 11/19] nfsd: " Jeff Layton
2017-12-13 14:20 ` [PATCH 12/19] ocfs2: " Jeff Layton
2017-12-18 12:49   ` Jan Kara
2017-12-13 14:20 ` [PATCH 13/19] ufs: use " Jeff Layton
2017-12-13 14:20 ` [PATCH 14/19] xfs: convert to " Jeff Layton
2017-12-13 22:48   ` Dave Chinner
2017-12-13 23:25     ` Dave Chinner
2017-12-14  0:10       ` Jeff Layton
2017-12-14  2:17         ` Dave Chinner
2017-12-14 11:16           ` Jeff Layton
2017-12-13 14:20 ` [PATCH 15/19] IMA: switch IMA over " Jeff Layton
2017-12-13 14:20 ` [PATCH 16/19] fs: only set S_VERSION when updating times if necessary Jeff Layton
2017-12-15 12:59   ` Jeff Layton
2017-12-13 14:20 ` [PATCH 17/19] xfs: avoid setting XFS_ILOG_CORE if i_version doesn't need incrementing Jeff Layton
2017-12-13 14:20 ` [PATCH 18/19] btrfs: only dirty the inode in btrfs_update_time if something was changed Jeff Layton
2017-12-15 13:03   ` Jeff Layton
2017-12-13 14:20 ` [PATCH 19/19] fs: handle inode->i_version more efficiently Jeff Layton
2017-12-13 15:05 ` [PATCH 00/19] fs: rework and optimize i_version handling in filesystems J. Bruce Fields
2017-12-13 20:14   ` Jeff Layton
2017-12-13 22:10     ` Jeff Layton
2017-12-13 23:03     ` Dave Chinner
2017-12-14  0:02       ` Jeff Layton
2017-12-14 14:14         ` Jeff Layton [this message]
2017-12-14 15:14           ` J. Bruce Fields
2017-12-15 15:15             ` Jeff Layton
2017-12-15 15:26               ` J. Bruce Fields

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=1513260887.3504.12.camel@kernel.org \
    --to=jlayton@kernel.org \
    --cc=amir73il@gmail.com \
    --cc=bfields@fieldses.org \
    --cc=david@fromorbit.com \
    --cc=hch@lst.de \
    --cc=jack@suse.de \
    --cc=linux-fsdevel@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=neilb@suse.de \
    --cc=viro@zeniv.linux.org.uk \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.