All of lore.kernel.org
 help / color / mirror / Atom feed
From: David Wysochanski <dwysocha@redhat.com>
To: Trond Myklebust <trondmy@hammerspace.com>,
	Anna Schumaker <anna.schumaker@netapp.com>
Cc: linux-nfs <linux-nfs@vger.kernel.org>
Subject: Re: [PATCH 00/10] Convert NFS fscache read paths to netfs API
Date: Sun, 31 Jan 2021 21:15:59 -0500	[thread overview]
Message-ID: <CALF+zOkaB8=uedDiSy6YheGjnObGSpUiYmuA13K-TqBgreO1eQ@mail.gmail.com> (raw)
In-Reply-To: <1611845708-6752-1-git-send-email-dwysocha@redhat.com>

On Thu, Jan 28, 2021 at 9:59 AM Dave Wysochanski <dwysocha@redhat.com> wrote:
>
> This minimal set of patches update the NFS client to use the new
> readahead method, and convert the NFS fscache to use the new netfs
> IO API, and are at:
> https://github.com/DaveWysochanskiRH/kernel/releases/tag/fscache-iter-lib-nfs-20210128
> https://github.com/DaveWysochanskiRH/kernel/commit/74357eb291c9c292f3ab3bc9ed1227cb76f52c51
>
> The patches are based on David Howells fscache-netfs-lib tree at
> https://git.kernel.org/pub/scm/linux/kernel/git/dhowells/linux-fs.git/log/?h=fscache-netfs-lib
>
> The first 6 patches refactor some of the NFS read code to facilitate
> re-use, the next 4 patches do the conversion to the new API.  Note
> patch 8 converts nfs_readpages to nfs_readahead.
>
> Changes since my last posting on Jan 27, 2021
> - Fix oops with fscache enabled on parallel read unit test
> - Add patches to handle invalidate and releasepage
> - Use #define FSCACHE_USE_NEW_IO_API to select the new API
> - Minor cleanup in nfs_readahead_from_fscache
>
> Still TODO
> 1. Fix known bugs
> a) nfs_issue_op: takes rcu_read_lock but may calls nfs_page_alloc()
>    with GFP_KERNEL which may sleep (dhowells noted this in a review)
> b) nfs_refresh_inode() takes inode->i_lock but may call
>    __fscache_invalidate() which may sleep (found with lockdep)
> c) WARN with xfstest fscache/netapp/pnfs/nfs41

Turns out this is a bit more involved and I would not consider pNFS +
fscache stable right now.
For now I may have to disable fscache if pNFS is enabled unless I can
quickly come up
with a reasonable fix for the problem.

The problem is as follows. Once netfs calls us in "issue_op" for a
given subrequest, it expects
one call back when the subrequest completes.  Now the "clamp_length"
function was developed
so we tell the netfs caller how big of an IO we can handle.  However,
right now it only implements
an 'rsize' check, and it does not take into account pNFS
characteristics such as segments
which may split up the IO into multiple RPCs. Since each of the RPC
have their own
completion, and so far I've not come up with a way to just call back
into netfs when the
last one is done, I am not sure what the right approach is.  One
obvious approach would be
a more sophisticated "clamp_length" function which adds similar logic
as to the *pg_test()
functions.  But I don't want to duplicate that and so it's not really clear.

> 2. Fixup NFS fscache stats (NFSIOS_FSCACHE_*)
> * Compare with netfs stats and determine if still needed
> 3. Cleanup dfprintks and/or convert to tracepoints
> 4. Further tests (see "Not tested yet")
>
> Tests run
> 1. Custom NFS+fscache unit tests for basic operation: PASS
> * vers=3,4.0,4.1,4.2,sec=sys,server=localhost (same kernel)
> 2. cthon04: PASS
> * test options "-b -g -s -l", fsc,vers=3,4.0,4.1,4.2,sec=sys
> * No failures, oopses or hangs
> 3. iozone tests: PASS
> * nofsc,fsc,vers=3,4.0,4.1,4.2,sec=sys,server=rhel7,rhel8
> * No failures, oopses, or hangs
> 4. xfstests/generic: PASS*
> * no hangs or crashes (one WARN); failures unrelated to these patches
> * Ran following configurations
>   * vers=4.1,fsc,sec=sys,rhel7-server: PASS
>   * vers=4.0,fsc,sec=sys,rhel7-server: PASS
>   * vers=3,fsc,sec=sys,rhel7-server: PASS
>   * vers=4.1,nofsc,sec=sys,netapp-server(pnfs/files): PASS
>   * vers=4.1,fsc,sec=sys,netapp-server(pnfs/files): INCOMPLETE
>     * WARN_ON fs/netfs/read_helper.c:616
>     * ran with kernel.panic_on_oops=1
>   * vers=4.2,fsc,sec=sys,rhel7-server: running at generic/438
>   * vers=4.2,fsc,sec=sys,rhel8-server: running at generic/127
> 5. kernel build: PASS
>   * vers=4.2,fsc,sec=sys,rhel8-server: PASS
>
> Not tested yet:
> * error injections (for example, connection disruptions, server errors during IO, etc)
> * many process mixed read/write on same file
> * performance
>
> Dave Wysochanski (10):
>   NFS: Clean up nfs_readpage() and nfs_readpages()
>   NFS: In nfs_readpage() only increment NFSIOS_READPAGES when read
>     succeeds
>   NFS: Refactor nfs_readpage() and nfs_readpage_async() to use
>     nfs_readdesc
>   NFS: Call readpage_async_filler() from nfs_readpage_async()
>   NFS: Add nfs_pageio_complete_read() and remove nfs_readpage_async()
>   NFS: Allow internal use of read structs and functions
>   NFS: Convert to the netfs API and nfs_readpage to use netfs_readpage
>   NFS: Convert readpages to readahead and use netfs_readahead for
>     fscache
>   NFS: Update releasepage to handle new fscache kiocb IO API
>   NFS: update various invalidation code paths for new IO API
>
>  fs/nfs/file.c              |  22 +++--
>  fs/nfs/fscache.c           | 230 +++++++++++++++++++------------------------
>  fs/nfs/fscache.h           | 105 +++-----------------
>  fs/nfs/internal.h          |   8 ++
>  fs/nfs/pagelist.c          |   2 +
>  fs/nfs/read.c              | 240 ++++++++++++++++++++-------------------------
>  fs/nfs/write.c             |  10 +-
>  include/linux/nfs_fs.h     |   5 +-
>  include/linux/nfs_iostat.h |   2 +-
>  include/linux/nfs_page.h   |   1 +
>  include/linux/nfs_xdr.h    |   1 +
>  11 files changed, 257 insertions(+), 369 deletions(-)
>
> --
> 1.8.3.1
>


  parent reply	other threads:[~2021-02-01  2:18 UTC|newest]

Thread overview: 16+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2021-01-28 14:54 [PATCH 00/10] Convert NFS fscache read paths to netfs API Dave Wysochanski
2021-01-28 14:54 ` [PATCH 01/10] NFS: Clean up nfs_readpage() and nfs_readpages() Dave Wysochanski
2021-01-28 14:55 ` [PATCH 02/10] NFS: In nfs_readpage() only increment NFSIOS_READPAGES when read succeeds Dave Wysochanski
2021-01-28 14:55 ` [PATCH 03/10] NFS: Refactor nfs_readpage() and nfs_readpage_async() to use nfs_readdesc Dave Wysochanski
2021-01-28 14:55 ` [PATCH 04/10] NFS: Call readpage_async_filler() from nfs_readpage_async() Dave Wysochanski
2021-01-28 14:55 ` [PATCH 05/10] NFS: Add nfs_pageio_complete_read() and remove nfs_readpage_async() Dave Wysochanski
2021-01-28 14:55 ` [PATCH 06/10] NFS: Allow internal use of read structs and functions Dave Wysochanski
2021-02-05 13:47   ` David Wysochanski
2021-02-05 16:24     ` Anna Schumaker
2021-01-28 14:55 ` [PATCH 07/10] NFS: Convert to the netfs API and nfs_readpage to use netfs_readpage Dave Wysochanski
2021-01-28 14:55 ` [PATCH 08/10] NFS: Convert readpages to readahead and use netfs_readahead for fscache Dave Wysochanski
2021-01-28 14:55 ` [PATCH 09/10] NFS: Update releasepage to handle new fscache kiocb IO API Dave Wysochanski
2021-01-28 14:55 ` [PATCH 10/10] NFS: update various invalidation code paths for new " Dave Wysochanski
2021-02-01  2:15 ` David Wysochanski [this message]
2021-02-01 14:30   ` [PATCH 00/10] Convert NFS fscache read paths to netfs API Anna Schumaker
2021-02-02 12:19     ` David Wysochanski

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to='CALF+zOkaB8=uedDiSy6YheGjnObGSpUiYmuA13K-TqBgreO1eQ@mail.gmail.com' \
    --to=dwysocha@redhat.com \
    --cc=anna.schumaker@netapp.com \
    --cc=linux-nfs@vger.kernel.org \
    --cc=trondmy@hammerspace.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.