All of lore.kernel.org
 help / color / mirror / Atom feed
From: NeilBrown <neilb@suse.de>
To: "J.Bruce Fields" <bfields@citi.umich.edu>
Cc: Olga Kornievskaia <aglo@citi.umich.edu>, NFS <linux-nfs@vger.kernel.org>
Subject: Re: Is tcp autotuning really what NFS wants?
Date: Mon, 15 Jul 2013 14:32:03 +1000	[thread overview]
Message-ID: <20130715143203.51bc583b@notabene.brown> (raw)
In-Reply-To: <20130710190727.GA22305@fieldses.org>

[-- Attachment #1: Type: text/plain, Size: 6806 bytes --]

On Wed, 10 Jul 2013 15:07:28 -0400 "J.Bruce Fields" <bfields@citi.umich.edu>
wrote:

> On Wed, Jul 10, 2013 at 02:32:33PM +1000, NeilBrown wrote:
> > On Tue, 9 Jul 2013 22:27:35 -0400 "J.Bruce Fields" <bfields@citi.umich.edu>
> > wrote:
> > 
> > > On Wed, Jul 10, 2013 at 09:22:55AM +1000, NeilBrown wrote:
> > > > 
> > > > Hi,
> > > >  I just noticed this commit:
> > > > 
> > > > commit 9660439861aa8dbd5e2b8087f33e20760c2c9afc
> > > > Author: Olga Kornievskaia <aglo@citi.umich.edu>
> > > > Date:   Tue Oct 21 14:13:47 2008 -0400
> > > > 
> > > >     svcrpc: take advantage of tcp autotuning
> > > > 
> > > > 
> > > > which I must confess surprised me.  I wonder if the full implications of
> > > > removing that functionality were understood.
> > > > 
> > > > Previously nfsd would set the transmit buffer space for a connection to
> > > > ensure there is plenty to hold all replies.  Now it doesn't.
> > > > 
> > > > nfsd refuses to accept a request if there isn't enough space in the transmit
> > > > buffer to send a reply.  This is important to ensure that each reply gets
> > > > sent atomically without blocking and there is no risk of replies getting
> > > > interleaved.
> 
> By the way, it's xpt_mutex that really guarantees non-interleaving,
> isn't it?

Yes.  I had the two different things confused.  The issue is only about
blocking on writes and consequently tying up nfsd threads.

> 
> I think of the svc_tcp_has_wspace checks as solving a problem of
> fairness between clients.  If we only had one client, everything would
> work without them.  But when we have multiple clients we don't want a
> single slow client to be able to tie block every thread waiting for that
> client to receive read data.  Is that accurate?

It agrees with my understanding - yes.


> 
> > > > The server starts out with a large estimate of the reply space (1M) and for
> > > > NFSv3 and v2 it quickly adjusts this down to something realistic.  For NFSv4
> > > > it is much harder to estimate the space needed so it just assumes every
> > > > reply will require 1M of space.
> > > > 
> > > > This means that with NFSv4, as soon as you have enough concurrent requests
> > > > such that 1M each reserves all of whatever window size was auto-tuned, new
> > > > requests on that connection will be ignored.
> > > >
> > > > This could significantly limit the amount of parallelism that can be achieved
> > > > for a single TCP connection (and given that the Linux client strongly prefers
> > > > a single connection now, this could become more of an issue).
> > > 
> > > Worse, I believe it can deadlock completely if the transmit buffer
> > > shrinks too far, and people really have run into this:
> > > 
> > > 	http://mid.gmane.org/<20130125185748.GC29596@fieldses.org>
> > > 
> > > Trond's suggestion looked at the time like it might work and be doable:
> > > 
> > > 	http://mid.gmane.org/<4FA345DA4F4AE44899BD2B03EEEC2FA91833C1D8@sacexcmbx05-prd.hq.netapp.com>
> > > 
> > > but I dropped it.
> > 
> > I would probably generalise Trond's suggestion and allow "N" extra requests
> > through that exceed the reservation, when N is related to the number of idle
> > threads.  squareroot might be nice, but half is probably easiest.
> > 
> > If any send takes more than 30 seconds the sk_sndtimeo will kick in and close
> > the connection so a really bad connection won't block threads indefinitely.
> > 
> > 
> > And yes - a nice test case would be good.
> > 
> > What do you think of the following (totally untested - just for comment)?
> 
> In cases where the disk is the bottleneck, there aren't actually going
> to be idle threads, are there?  In which case I don't think this helps
> save a stuck client.

Do we care about the disk being a bottleneck?  We don't currently try to
handle that situation at all - if the disk is slow, everything slows down.
Unless you have to separate clients using two separate devices, then I think
that is the right think to do.  Limiting each filesystem to some number of
threads might be interesting but I don't think there would be much call for
it.

The present issue is when a client or network is slow and could thereby
interfere with the response to other clients.

Certainly the deadlock that can happen if write buffer space drops below 1Meg
would be an issue when very few clients are active.


> 
> Trond's suggestion doesn't have the same limitation, if I understand it
> correctly.

I think I missed the "hand off to a workqueue" bit the first time.

I guess something like changing the "mutex_lock" in svc_send to
"mutex_try_lock" and if that fails, punt to a work-queue. We would need to
split the buffer space off the rqstp and attach it to the work struct.
Each rqstp could have a small array of work_structs and that sets the limit
on the number of requests that can be accepted when there is no wspace.

We probably don't even need a workqueue.  After completing a send, svc_send
can check if another reply has been queued and if so, it starts sending it -
if there is enough wspace.  If no wspace just do nothing.  The wspace
callback can queue the socket, and svc_xprt_enqueue can queue up an nfsd
thread if there is write work to do.  svc_handle_xprt() can do the actual
send.

This would require modifying all the xpo_sendto routines and the routines
they call to take an index into the list of send buffers per socket.  Or
maybe a pointer to a "delayed reply" structure.


For each delayed reply we would need:
 - struct xdr_buf  - to hold the reply
 - struct page *[RPCSVC_MAXPAGES]  - to hold the pages so we can free them
                                     when done
 - struct sockaddr_storage         - to hold source address for UDP reply
 - struct sockaddr_storage         - to hold the dest address for UDP
 - maybe something else for RDMA?  - svc_rdma_sendto uses rqstp->rq_arg - not
                                     sure why.

We could possible allocate these as needed.  If the kmalloc fails, just use
"mutex_lock" instead of "mutex_trylock" and block for a while.

Note that while this structure would be smaller than "struct svc_rqst", it is
not completely trivial.  It might be just as easy to allocate a new "svc_rqst"
structure.  Or fork new nfsd threads.


I think the key issue is really to allow some nfsd threads to block on
writing the reply, but to firmly limit the number which do.  This is what my
patch tried to do.
Punting to a work queue might allow a little less memory to be allocated
to end up with the same number of spare threads, but would require a lot more
code churn.

I wonder if it is time to revisit dynamic sizing of the nfsd thread pool.

NeilBrown


[-- Attachment #2: signature.asc --]
[-- Type: application/pgp-signature, Size: 828 bytes --]

  reply	other threads:[~2013-07-15  4:32 UTC|newest]

Thread overview: 27+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
     [not found] <20130710092255.0240a36d@notabene.brown>
2013-07-10  2:27 ` Is tcp autotuning really what NFS wants? J.Bruce Fields
2013-07-10  4:32   ` NeilBrown
2013-07-10 19:07     ` J.Bruce Fields
2013-07-15  4:32       ` NeilBrown [this message]
2013-07-16  1:58         ` J.Bruce Fields
2013-07-16  4:00           ` NeilBrown
2013-07-16 14:24             ` J.Bruce Fields
2013-07-18  0:03               ` Ben Myers
2013-07-24 21:07                 ` J.Bruce Fields
2013-07-25  1:30                   ` [PATCH] NFSD/sunrpc: avoid deadlock on TCP connection due to memory pressure NeilBrown
2013-07-25 12:35                     ` Jim Rees
2013-07-25 20:18                     ` J.Bruce Fields
2013-07-25 20:33                       ` NeilBrown
2013-07-26 14:19                         ` J.Bruce Fields
2013-07-30  2:48                           ` NeilBrown
2013-08-01  2:49                             ` J.Bruce Fields
2013-07-10 17:33   ` Is tcp autotuning really what NFS wants? Dean
2013-07-10 17:39     ` Ben Greear
2013-07-15  4:35       ` NeilBrown
2013-07-15 23:32         ` Ben Greear
2013-07-16  4:46           ` NeilBrown
2013-07-10 19:59     ` Michael Richardson
2013-07-15  1:26   ` Jim Rees
2013-07-15  5:02     ` NeilBrown
2013-07-15 11:57       ` Jim Rees
2013-07-15 13:42   ` Jim Rees
2013-07-16  1:10     ` NeilBrown

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20130715143203.51bc583b@notabene.brown \
    --to=neilb@suse.de \
    --cc=aglo@citi.umich.edu \
    --cc=bfields@citi.umich.edu \
    --cc=linux-nfs@vger.kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.