From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-8.3 required=3.0 tests=HEADER_FROM_DIFFERENT_DOMAINS, INCLUDES_PATCH,MAILING_LIST_MULTI,SIGNED_OFF_BY,SPF_HELO_NONE,SPF_PASS, USER_AGENT_SANE_1 autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 1E649C43331 for ; Mon, 11 Nov 2019 03:21:18 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id C5D162080F for ; Mon, 11 Nov 2019 03:21:17 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org C5D162080F Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=intel.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=owner-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix) id 31C516B0005; Sun, 10 Nov 2019 22:21:17 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id 2CCD56B0006; Sun, 10 Nov 2019 22:21:17 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 1E3A86B0007; Sun, 10 Nov 2019 22:21:17 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0192.hostedemail.com [216.40.44.192]) by kanga.kvack.org (Postfix) with ESMTP id 06A8C6B0005 for ; Sun, 10 Nov 2019 22:21:17 -0500 (EST) Received: from smtpin18.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay04.hostedemail.com (Postfix) with SMTP id A98AE3A97 for ; Mon, 11 Nov 2019 03:21:16 +0000 (UTC) X-FDA: 76142545752.18.roof73_4537e80163926 X-HE-Tag: roof73_4537e80163926 X-Filterd-Recvd-Size: 3192 Received: from mga01.intel.com (mga01.intel.com [192.55.52.88]) by imf22.hostedemail.com (Postfix) with ESMTP for ; Mon, 11 Nov 2019 03:21:15 +0000 (UTC) X-Amp-Result: UNKNOWN X-Amp-Original-Verdict: FILE UNKNOWN X-Amp-File-Uploaded: False Received: from orsmga006.jf.intel.com ([10.7.209.51]) by fmsmga101.fm.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 10 Nov 2019 19:21:12 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.68,291,1569308400"; d="scan'208";a="207006324" Received: from iweiny-desk2.sc.intel.com ([10.3.52.157]) by orsmga006.jf.intel.com with ESMTP; 10 Nov 2019 19:21:11 -0800 Date: Sun, 10 Nov 2019 19:21:11 -0800 From: Ira Weiny To: John Hubbard Cc: Andrew Morton , Jason Gunthorpe , linux-rdma@vger.kernel.org, linux-mm@kvack.org, LKML Subject: Re: [PATCH 1/1] IB/umem: use get_user_pages_fast() to pin DMA pages Message-ID: <20191111032111.GA30123@iweiny-DESK2.sc.intel.com> References: <20191109020434.389855-1-jhubbard@nvidia.com> <20191109020434.389855-2-jhubbard@nvidia.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20191109020434.389855-2-jhubbard@nvidia.com> User-Agent: Mutt/1.11.1 (2018-12-01) X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: On Fri, Nov 08, 2019 at 06:04:34PM -0800, John Hubbard wrote: > And get rid of the mmap_sem calls, as part of that. Note > that get_user_pages_fast() will, if necessary, fall back to > __gup_longterm_unlocked(), which takes the mmap_sem as needed. > > Cc: Jason Gunthorpe > Cc: Ira Weiny Reviewed-by: Ira Weiny > Signed-off-by: John Hubbard > --- > drivers/infiniband/core/umem.c | 17 ++++++----------- > 1 file changed, 6 insertions(+), 11 deletions(-) > > diff --git a/drivers/infiniband/core/umem.c b/drivers/infiniband/core/umem.c > index 24244a2f68cc..3d664a2539eb 100644 > --- a/drivers/infiniband/core/umem.c > +++ b/drivers/infiniband/core/umem.c > @@ -271,16 +271,13 @@ struct ib_umem *ib_umem_get(struct ib_udata *udata, unsigned long addr, > sg = umem->sg_head.sgl; > > while (npages) { > - down_read(&mm->mmap_sem); > - ret = get_user_pages(cur_base, > - min_t(unsigned long, npages, > - PAGE_SIZE / sizeof (struct page *)), > - gup_flags | FOLL_LONGTERM, > - page_list, NULL); > - if (ret < 0) { > - up_read(&mm->mmap_sem); > + ret = get_user_pages_fast(cur_base, > + min_t(unsigned long, npages, > + PAGE_SIZE / > + sizeof(struct page *)), > + gup_flags | FOLL_LONGTERM, page_list); > + if (ret < 0) > goto umem_release; > - } > > cur_base += ret * PAGE_SIZE; > npages -= ret; > @@ -288,8 +285,6 @@ struct ib_umem *ib_umem_get(struct ib_udata *udata, unsigned long addr, > sg = ib_umem_add_sg_table(sg, page_list, ret, > dma_get_max_seg_size(context->device->dma_device), > &umem->sg_nents); > - > - up_read(&mm->mmap_sem); > } > > sg_mark_end(sg); > -- > 2.24.0 >