From: Ralph Campbell <rcampbell@nvidia.com>
To: "Christoph Hellwig" <hch@lst.de>,
"Dan Williams" <dan.j.williams@intel.com>,
"Jérôme Glisse" <jglisse@redhat.com>,
"Jason Gunthorpe" <jgg@mellanox.com>,
"Ben Skeggs" <bskeggs@redhat.com>
Cc: Ira Weiny <ira.weiny@intel.com>, <linux-mm@kvack.org>,
<nouveau@lists.freedesktop.org>,
<dri-devel@lists.freedesktop.org>, <linux-nvdimm@lists.01.org>,
<linux-pci@vger.kernel.org>, <linux-kernel@vger.kernel.org>
Subject: Re: [PATCH 20/22] mm: move hmm_vma_fault to nouveau
Date: Wed, 3 Jul 2019 10:48:56 -0700 [thread overview]
Message-ID: <a3108540-e431-2513-650e-3bb143f7f161@nvidia.com> (raw)
In-Reply-To: <20190701062020.19239-21-hch@lst.de>
On 6/30/19 11:20 PM, Christoph Hellwig wrote:
> hmm_vma_fault is marked as a legacy API to get rid of, but quite suites
> the current nouvea flow. Move it to the only user in preparation for
I didn't quite parse the phrase "quite suites the current nouvea flow."
s/nouvea/nouveau/
> fixing a locking bug involving caller and callee.
>
> Signed-off-by: Christoph Hellwig <hch@lst.de>
I see where you are going with this and it
looks like straightforward code movement so,
Reviewed-by: Ralph Campbell <rcampbell@nvidia.com>
> ---
> drivers/gpu/drm/nouveau/nouveau_svm.c | 54 ++++++++++++++++++++++++++-
> include/linux/hmm.h | 54 ---------------------------
> 2 files changed, 53 insertions(+), 55 deletions(-)
>
> diff --git a/drivers/gpu/drm/nouveau/nouveau_svm.c b/drivers/gpu/drm/nouveau/nouveau_svm.c
> index 9d40114d7949..e831f4184a17 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_svm.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_svm.c
> @@ -36,6 +36,13 @@
> #include <linux/sort.h>
> #include <linux/hmm.h>
>
> +/*
> + * When waiting for mmu notifiers we need some kind of time out otherwise we
> + * could potentialy wait for ever, 1000ms ie 1s sounds like a long time to
> + * wait already.
> + */
> +#define NOUVEAU_RANGE_FAULT_TIMEOUT 1000
> +
> struct nouveau_svm {
> struct nouveau_drm *drm;
> struct mutex mutex;
> @@ -475,6 +482,51 @@ nouveau_svm_fault_cache(struct nouveau_svm *svm,
> fault->inst, fault->addr, fault->access);
> }
>
> +static int
> +nouveau_range_fault(struct hmm_mirror *mirror, struct hmm_range *range,
> + bool block)
> +{
> + long ret;
> +
> + /*
> + * With the old API the driver must set each individual entries with
> + * the requested flags (valid, write, ...). So here we set the mask to
> + * keep intact the entries provided by the driver and zero out the
> + * default_flags.
> + */
> + range->default_flags = 0;
> + range->pfn_flags_mask = -1UL;
> +
> + ret = hmm_range_register(range, mirror,
> + range->start, range->end,
> + PAGE_SHIFT);
> + if (ret)
> + return (int)ret;
> +
> + if (!hmm_range_wait_until_valid(range, NOUVEAU_RANGE_FAULT_TIMEOUT)) {
> + /*
> + * The mmap_sem was taken by driver we release it here and
> + * returns -EAGAIN which correspond to mmap_sem have been
> + * drop in the old API.
> + */
> + up_read(&range->vma->vm_mm->mmap_sem);
> + return -EAGAIN;
> + }
> +
> + ret = hmm_range_fault(range, block);
> + if (ret <= 0) {
> + if (ret == -EBUSY || !ret) {
> + /* Same as above, drop mmap_sem to match old API. */
> + up_read(&range->vma->vm_mm->mmap_sem);
> + ret = -EBUSY;
> + } else if (ret == -EAGAIN)
> + ret = -EBUSY;
> + hmm_range_unregister(range);
> + return ret;
> + }
> + return 0;
> +}
> +
> static int
> nouveau_svm_fault(struct nvif_notify *notify)
> {
> @@ -649,7 +701,7 @@ nouveau_svm_fault(struct nvif_notify *notify)
> range.values = nouveau_svm_pfn_values;
> range.pfn_shift = NVIF_VMM_PFNMAP_V0_ADDR_SHIFT;
> again:
> - ret = hmm_vma_fault(&svmm->mirror, &range, true);
> + ret = nouveau_range_fault(&svmm->mirror, &range, true);
> if (ret == 0) {
> mutex_lock(&svmm->mutex);
> if (!hmm_range_unregister(&range)) {
> diff --git a/include/linux/hmm.h b/include/linux/hmm.h
> index 4b185d286c3b..3457cf9182e5 100644
> --- a/include/linux/hmm.h
> +++ b/include/linux/hmm.h
> @@ -478,60 +478,6 @@ long hmm_range_dma_unmap(struct hmm_range *range,
> dma_addr_t *daddrs,
> bool dirty);
>
> -/*
> - * HMM_RANGE_DEFAULT_TIMEOUT - default timeout (ms) when waiting for a range
> - *
> - * When waiting for mmu notifiers we need some kind of time out otherwise we
> - * could potentialy wait for ever, 1000ms ie 1s sounds like a long time to
> - * wait already.
> - */
> -#define HMM_RANGE_DEFAULT_TIMEOUT 1000
> -
> -/* This is a temporary helper to avoid merge conflict between trees. */
> -static inline int hmm_vma_fault(struct hmm_mirror *mirror,
> - struct hmm_range *range, bool block)
> -{
> - long ret;
> -
> - /*
> - * With the old API the driver must set each individual entries with
> - * the requested flags (valid, write, ...). So here we set the mask to
> - * keep intact the entries provided by the driver and zero out the
> - * default_flags.
> - */
> - range->default_flags = 0;
> - range->pfn_flags_mask = -1UL;
> -
> - ret = hmm_range_register(range, mirror,
> - range->start, range->end,
> - PAGE_SHIFT);
> - if (ret)
> - return (int)ret;
> -
> - if (!hmm_range_wait_until_valid(range, HMM_RANGE_DEFAULT_TIMEOUT)) {
> - /*
> - * The mmap_sem was taken by driver we release it here and
> - * returns -EAGAIN which correspond to mmap_sem have been
> - * drop in the old API.
> - */
> - up_read(&range->vma->vm_mm->mmap_sem);
> - return -EAGAIN;
> - }
> -
> - ret = hmm_range_fault(range, block);
> - if (ret <= 0) {
> - if (ret == -EBUSY || !ret) {
> - /* Same as above, drop mmap_sem to match old API. */
> - up_read(&range->vma->vm_mm->mmap_sem);
> - ret = -EBUSY;
> - } else if (ret == -EAGAIN)
> - ret = -EBUSY;
> - hmm_range_unregister(range);
> - return ret;
> - }
> - return 0;
> -}
> -
> /* Below are for HMM internal use only! Not to be used by device driver! */
> static inline void hmm_mm_init(struct mm_struct *mm)
> {
>
next prev parent reply other threads:[~2019-07-03 17:49 UTC|newest]
Thread overview: 40+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-07-01 6:19 dev_pagemap related cleanups v4 Christoph Hellwig
2019-07-01 6:19 ` [PATCH 01/22] mm/hmm.c: suppress compilation warnings when CONFIG_HUGETLB_PAGE is not set Christoph Hellwig
2019-07-01 6:20 ` [PATCH 02/22] mm/hmm: update HMM documentation Christoph Hellwig
2019-07-01 6:20 ` [PATCH 03/22] mm/hmm: clean up some coding style and comments Christoph Hellwig
2019-07-01 6:20 ` [PATCH 04/22] mm/hmm: support automatic NUMA balancing Christoph Hellwig
2019-07-01 6:20 ` [PATCH 05/22] mm/hmm: Only set FAULT_FLAG_ALLOW_RETRY for non-blocking Christoph Hellwig
2019-07-01 6:20 ` [PATCH 06/22] mm/hmm: fix use after free with struct hmm in the mmu notifiers Christoph Hellwig
2019-07-01 6:20 ` [PATCH 07/22] mm/hmm: Use hmm_mirror not mm as an argument for hmm_range_register Christoph Hellwig
2019-07-01 6:20 ` [PATCH 08/22] mm/hmm: Hold a mmgrab from hmm to mm Christoph Hellwig
2019-07-01 6:20 ` [PATCH 09/22] mm/hmm: Simplify hmm_get_or_create and make it reliable Christoph Hellwig
2019-07-01 6:20 ` [PATCH 10/22] mm/hmm: Remove duplicate condition test before wait_event_timeout Christoph Hellwig
2019-07-01 6:20 ` [PATCH 11/22] mm/hmm: Do not use list*_rcu() for hmm->ranges Christoph Hellwig
2019-07-01 6:20 ` [PATCH 12/22] mm/hmm: Hold on to the mmget for the lifetime of the range Christoph Hellwig
2019-07-01 6:20 ` [PATCH 13/22] mm/hmm: Use lockdep instead of comments Christoph Hellwig
2019-07-01 6:20 ` [PATCH 14/22] mm/hmm: Remove racy protection against double-unregistration Christoph Hellwig
2019-07-01 6:20 ` [PATCH 15/22] mm/hmm: Poison hmm_range during unregister Christoph Hellwig
2019-07-01 6:20 ` [PATCH 16/22] mm/hmm: Remove confusing comment and logic from hmm_release Christoph Hellwig
2019-07-01 6:20 ` [PATCH 17/22] mm/hmm: Fix error flows in hmm_invalidate_range_start Christoph Hellwig
2019-07-01 6:20 ` [PATCH 18/22] mm: return valid info from hmm_range_unregister Christoph Hellwig
2019-07-03 17:22 ` Ralph Campbell
2019-07-01 6:20 ` [PATCH 19/22] mm: always return EBUSY for invalid ranges in hmm_range_{fault,snapshot} Christoph Hellwig
2019-07-02 21:43 ` Kuehling, Felix
2019-07-03 17:32 ` Ralph Campbell
2019-07-01 6:20 ` [PATCH 20/22] mm: move hmm_vma_fault to nouveau Christoph Hellwig
2019-07-03 17:48 ` Ralph Campbell [this message]
2019-07-03 17:50 ` [Nouveau] " Ilia Mirkin
2019-07-03 18:03 ` Jason Gunthorpe
2019-07-03 18:05 ` Christoph Hellwig
2019-07-03 18:13 ` Jason Gunthorpe
2019-07-01 6:20 ` [PATCH 21/22] nouveau: unlock mmap_sem on all errors from nouveau_range_fault Christoph Hellwig
2019-07-01 6:20 ` [PATCH 22/22] mm: remove the legacy hmm_pfn_* APIs Christoph Hellwig
2019-07-03 18:01 ` Jason Gunthorpe
2019-07-03 18:03 ` Christoph Hellwig
2019-07-03 18:15 ` Jason Gunthorpe
2019-07-01 8:25 ` dev_pagemap related cleanups v4 Christoph Hellwig
2019-07-02 18:42 ` Jason Gunthorpe
2019-07-02 22:45 ` Weiny, Ira
2019-07-02 22:47 ` Christoph Hellwig
2019-07-02 23:17 ` Dan Williams
2019-07-03 1:08 ` Jason Gunthorpe
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=a3108540-e431-2513-650e-3bb143f7f161@nvidia.com \
--to=rcampbell@nvidia.com \
--cc=bskeggs@redhat.com \
--cc=dan.j.williams@intel.com \
--cc=dri-devel@lists.freedesktop.org \
--cc=hch@lst.de \
--cc=ira.weiny@intel.com \
--cc=jgg@mellanox.com \
--cc=jglisse@redhat.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=linux-nvdimm@lists.01.org \
--cc=linux-pci@vger.kernel.org \
--cc=nouveau@lists.freedesktop.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).