All of lore.kernel.org
 help / color / mirror / Atom feed
From: Waiman Long <waiman.long@hpe.com>
To: Dave Chinner <david@fromorbit.com>
Cc: linux-kernel@vger.kernel.org, peterz@infradead.org,
	Waiman.Long@hp.com, torvalds@linux-foundation.org
Subject: Re: [4.2, Regression] Queued spinlocks cause major XFS performance regression
Date: Wed, 09 Sep 2015 22:01:54 -0400	[thread overview]
Message-ID: <55F0E492.4060607@hpe.com> (raw)
In-Reply-To: <20150904054820.GY3902@dastard>

On 09/04/2015 01:48 AM, Dave Chinner wrote:
> Hi Waiman,
>
> For the first time in months I just turned of spinlock debugging on
> my performance test machine and I just got an unpleasant surprise on
> my standard inode allocation and reclaim test.  I've described this
> test to you before, because it's found regressions in your previous
> lock scaling changes:
>
> http://permalink.gmane.org/gmane.linux.kernel/1768786
>
> This time it is the fsmark run that I use to populate the filesystem
> that is demonstrating a locking regression. I'll asked you before
> if you could add this test to your lock scaling regression test
> suite; please do it this time.
>
> Now, the regression.  With spinlock debugging turned on, the
> performance of my usual XFS inode allocation benchmark using fsmark
> reports performance like this:
>
> FSUse%        Count         Size    Files/sec     App Overhead
>       0      1600000            0     312594.0          9944159
>       0      3200000            0     295668.6         10399679
>       0      4800000            0     279026.1         11397617
> .....
>
> This has been pretty stable for several releases - it varies +/- a
> few percent, but it's pretty much been like this since about 3.2
> when CONFIG_XFS_DEBUG=n, with or without basic spinlock debugging.
>
> When I turned spinlock debugging off on 4.2 to get some perf numbers
> a request from Linus, I got this:
>
> FSUse%        Count         Size    Files/sec     App Overhead
>       0      1600000            0     114143.9          9597599
>       0      3200000            0      95486.9          9460413
>       0      4800000            0      93918.2          9784699
> .

I am sorry that I was on vacation over the past weekend and so was not 
able to respond in a timely manner. As Peter already has a patch to 
address the root cause of this problem. I think this problem is all set.

Cheers,
Longman


      parent reply	other threads:[~2015-09-10  2:11 UTC|newest]

Thread overview: 32+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2015-09-04  5:48 [4.2, Regression] Queued spinlocks cause major XFS performance regression Dave Chinner
2015-09-04  6:39 ` Linus Torvalds
2015-09-04  7:11   ` Dave Chinner
2015-09-04  7:31     ` Juergen Gross
2015-09-04  7:55     ` Peter Zijlstra
2015-09-04  8:29     ` Dave Chinner
2015-09-04 15:05       ` Linus Torvalds
2015-09-04 15:14         ` Peter Zijlstra
2015-09-04 15:21           ` Linus Torvalds
2015-09-04 15:30             ` Peter Zijlstra
2015-09-04 15:54               ` Peter Zijlstra
2015-09-10  2:06                 ` Waiman Long
2015-09-04 15:58               ` Linus Torvalds
2015-09-05 17:45                 ` Peter Zijlstra
2015-09-04 15:25           ` Peter Zijlstra
2015-09-06 23:32             ` Dave Chinner
2015-09-07  0:05             ` Davidlohr Bueso
2015-09-07  6:57               ` Peter Zijlstra
2015-09-07 20:45                 ` Linus Torvalds
2015-09-08  6:37                   ` Davidlohr Bueso
2015-09-08 10:05                   ` Peter Zijlstra
2015-09-08 17:45                     ` Linus Torvalds
2015-09-13 10:55             ` [tip:locking/core] locking/qspinlock/x86: Fix performance regression under unaccelerated VMs tip-bot for Peter Zijlstra
2015-09-04  7:39   ` [4.2, Regression] Queued spinlocks cause major XFS performance regression Peter Zijlstra
2015-09-04  8:12     ` Dave Chinner
2015-09-04 11:32       ` Peter Zijlstra
2015-09-04 22:03         ` Dave Chinner
2015-09-06 23:47         ` Dave Chinner
2015-09-10  2:09           ` Waiman Long
     [not found]         ` <CAC=cRTOraeOeu3Z8C1qx6w=GMSzD_4VevrEzn0mMhrqy=7n3wQ@mail.gmail.com>
     [not found]           ` <56094F05.4090809@hpe.com>
2015-09-29  0:47             ` huang ying
2015-09-29  2:57               ` Waiman Long
2015-09-10  2:01 ` Waiman Long [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=55F0E492.4060607@hpe.com \
    --to=waiman.long@hpe.com \
    --cc=Waiman.Long@hp.com \
    --cc=david@fromorbit.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=peterz@infradead.org \
    --cc=torvalds@linux-foundation.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.