From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-16.2 required=3.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,HK_RANDOM_FROM,INCLUDES_PATCH, MAILING_LIST_MULTI,MENTIONS_GIT_HOSTING,NICE_REPLY_A,SIGNED_OFF_BY, SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED,USER_AGENT_SANE_1 autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 1593DC2D0A8 for ; Wed, 23 Sep 2020 09:52:44 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id C9FE62076A for ; Wed, 23 Sep 2020 09:52:43 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726415AbgIWJwm (ORCPT ); Wed, 23 Sep 2020 05:52:42 -0400 Received: from mga03.intel.com ([134.134.136.65]:63175 "EHLO mga03.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726130AbgIWJwm (ORCPT ); Wed, 23 Sep 2020 05:52:42 -0400 IronPort-SDR: 3ckRgmtJUqY6hrIGd+VsfBsGXTiohl7lHZ596fEpa1coVCKOiaX05rolkUSMN5kP3NW8e/u1wE inbZECR1cXDw== X-IronPort-AV: E=McAfee;i="6000,8403,9752"; a="160921537" X-IronPort-AV: E=Sophos;i="5.77,293,1596524400"; d="scan'208";a="160921537" X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from orsmga005.jf.intel.com ([10.7.209.41]) by orsmga103.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 23 Sep 2020 02:52:41 -0700 IronPort-SDR: seIxyGE1Vw8wLigC1YT863o99ElqSWQAiuSZRkgjZoMQsFRRWMV/uMP79du30Xlum1gHB9JcsL DrdLmgIHayQA== X-IronPort-AV: E=Sophos;i="5.77,293,1596524400"; d="scan'208";a="486376229" Received: from yymichae-mobl.ger.corp.intel.com (HELO [10.214.208.219]) ([10.214.208.219]) by orsmga005-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 23 Sep 2020 02:52:36 -0700 Subject: Re: [Intel-gfx] [PATCH 4/6] drm/i915: use vmap in i915_gem_object_map To: Christoph Hellwig , Andrew Morton Cc: Juergen Gross , Stefano Stabellini , linux-mm@kvack.org, Peter Zijlstra , intel-gfx@lists.freedesktop.org, x86@kernel.org, linux-kernel@vger.kernel.org, Minchan Kim , dri-devel@lists.freedesktop.org, xen-devel@lists.xenproject.org, Boris Ostrovsky , Nitin Gupta References: <20200918163724.2511-1-hch@lst.de> <20200918163724.2511-5-hch@lst.de> From: Tvrtko Ursulin Organization: Intel Corporation UK Plc Message-ID: <9b5d40af-7378-9e68-ca51-73b2148287f3@linux.intel.com> Date: Wed, 23 Sep 2020 10:52:33 +0100 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.10.0 MIME-Version: 1.0 In-Reply-To: <20200918163724.2511-5-hch@lst.de> Content-Type: text/plain; charset=utf-8; format=flowed Content-Language: en-US Content-Transfer-Encoding: 7bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 18/09/2020 17:37, Christoph Hellwig wrote: > i915_gem_object_map implements fairly low-level vmap functionality in > a driver. Split it into two helpers, one for remapping kernel memory > which can use vmap, and one for I/O memory that uses vmap_pfn. > > The only practical difference is that alloc_vm_area prefeaults the > vmalloc area PTEs, which doesn't seem to be required here for the > kernel memory case (and could be added to vmap using a flag if actually > required). Patch looks good to me. Series did not get a CI run from our side because of a different base so I don't know if you would like to have a run there? If so you would need to rebase against git://anongit.freedesktop.org/drm-tip drm-tip and you could even send a series to intel-gfx-trybot@lists.freedesktop.org, suppressing cc, to check it out without sending a copy to the real mailing list. Regards, Tvrtko > Signed-off-by: Christoph Hellwig > --- > drivers/gpu/drm/i915/Kconfig | 1 + > drivers/gpu/drm/i915/gem/i915_gem_pages.c | 101 ++++++++++------------ > 2 files changed, 47 insertions(+), 55 deletions(-) > > diff --git a/drivers/gpu/drm/i915/Kconfig b/drivers/gpu/drm/i915/Kconfig > index 9afa5c4a6bf006..1e1cb245fca778 100644 > --- a/drivers/gpu/drm/i915/Kconfig > +++ b/drivers/gpu/drm/i915/Kconfig > @@ -25,6 +25,7 @@ config DRM_I915 > select CRC32 > select SND_HDA_I915 if SND_HDA_CORE > select CEC_CORE if CEC_NOTIFIER > + select VMAP_PFN > help > Choose this option if you have a system that has "Intel Graphics > Media Accelerator" or "HD Graphics" integrated graphics, > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_pages.c b/drivers/gpu/drm/i915/gem/i915_gem_pages.c > index e8a083743e0927..90029ea83aede9 100644 > --- a/drivers/gpu/drm/i915/gem/i915_gem_pages.c > +++ b/drivers/gpu/drm/i915/gem/i915_gem_pages.c > @@ -234,50 +234,24 @@ int __i915_gem_object_put_pages(struct drm_i915_gem_object *obj) > return err; > } > > -static inline pte_t iomap_pte(resource_size_t base, > - dma_addr_t offset, > - pgprot_t prot) > -{ > - return pte_mkspecial(pfn_pte((base + offset) >> PAGE_SHIFT, prot)); > -} > - > /* The 'mapping' part of i915_gem_object_pin_map() below */ > -static void *i915_gem_object_map(struct drm_i915_gem_object *obj, > +static void *i915_gem_object_map_page(struct drm_i915_gem_object *obj, > enum i915_map_type type) > { > - unsigned long n_pte = obj->base.size >> PAGE_SHIFT; > - struct sg_table *sgt = obj->mm.pages; > - pte_t *stack[32], **mem; > - struct vm_struct *area; > + unsigned long n_pages = obj->base.size >> PAGE_SHIFT, i; > + struct page *stack[32], **pages = stack, *page; > + struct sgt_iter iter; > pgprot_t pgprot; > - > - if (!i915_gem_object_has_struct_page(obj) && type != I915_MAP_WC) > - return NULL; > - > - /* A single page can always be kmapped */ > - if (n_pte == 1 && type == I915_MAP_WB) > - return kmap(sg_page(sgt->sgl)); > - > - mem = stack; > - if (n_pte > ARRAY_SIZE(stack)) { > - /* Too big for stack -- allocate temporary array instead */ > - mem = kvmalloc_array(n_pte, sizeof(*mem), GFP_KERNEL); > - if (!mem) > - return NULL; > - } > - > - area = alloc_vm_area(obj->base.size, mem); > - if (!area) { > - if (mem != stack) > - kvfree(mem); > - return NULL; > - } > + void *vaddr; > > switch (type) { > default: > MISSING_CASE(type); > fallthrough; /* to use PAGE_KERNEL anyway */ > case I915_MAP_WB: > + /* A single page can always be kmapped */ > + if (n_pages == 1) > + return kmap(sg_page(obj->mm.pages->sgl)); > pgprot = PAGE_KERNEL; > break; > case I915_MAP_WC: > @@ -285,30 +259,44 @@ static void *i915_gem_object_map(struct drm_i915_gem_object *obj, > break; > } > > - if (i915_gem_object_has_struct_page(obj)) { > - struct sgt_iter iter; > - struct page *page; > - pte_t **ptes = mem; > - > - for_each_sgt_page(page, iter, sgt) > - **ptes++ = mk_pte(page, pgprot); > - } else { > - resource_size_t iomap; > - struct sgt_iter iter; > - pte_t **ptes = mem; > - dma_addr_t addr; > + if (n_pages > ARRAY_SIZE(stack)) { > + /* Too big for stack -- allocate temporary array instead */ > + pages = kvmalloc_array(n_pages, sizeof(*pages), GFP_KERNEL); > + if (!pages) > + return NULL; > + } > > - iomap = obj->mm.region->iomap.base; > - iomap -= obj->mm.region->region.start; > + for_each_sgt_page(page, iter, obj->mm.pages) > + pages[i++] = page; > + vaddr = vmap(pages, n_pages, 0, pgprot); > + if (pages != stack) > + kvfree(pages); > + return vaddr; > +} > > - for_each_sgt_daddr(addr, iter, sgt) > - **ptes++ = iomap_pte(iomap, addr, pgprot); > +static void *i915_gem_object_map_pfn(struct drm_i915_gem_object *obj) > +{ > + resource_size_t iomap = obj->mm.region->iomap.base - > + obj->mm.region->region.start; > + unsigned long n_pfn = obj->base.size >> PAGE_SHIFT; > + unsigned long stack[32], *pfns = stack, i; > + struct sgt_iter iter; > + dma_addr_t addr; > + void *vaddr; > + > + if (n_pfn > ARRAY_SIZE(stack)) { > + /* Too big for stack -- allocate temporary array instead */ > + pfns = kvmalloc_array(n_pfn, sizeof(*pfns), GFP_KERNEL); > + if (!pfns) > + return NULL; > } > > - if (mem != stack) > - kvfree(mem); > - > - return area->addr; > + for_each_sgt_daddr(addr, iter, obj->mm.pages) > + pfns[i++] = (iomap + addr) >> PAGE_SHIFT; > + vaddr = vmap_pfn(pfns, n_pfn, pgprot_writecombine(PAGE_KERNEL_IO)); > + if (pfns != stack) > + kvfree(pfns); > + return vaddr; > } > > /* get, pin, and map the pages of the object into kernel space */ > @@ -360,7 +348,10 @@ void *i915_gem_object_pin_map(struct drm_i915_gem_object *obj, > } > > if (!ptr) { > - ptr = i915_gem_object_map(obj, type); > + if (i915_gem_object_has_struct_page(obj)) > + ptr = i915_gem_object_map_page(obj, type); > + else if (type == I915_MAP_WC) > + ptr = i915_gem_object_map_pfn(obj); > if (!ptr) { > err = -ENOMEM; > goto err_unpin; > From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-16.2 required=3.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,HK_RANDOM_FROM,INCLUDES_PATCH, MAILING_LIST_MULTI,MENTIONS_GIT_HOSTING,NICE_REPLY_A,SIGNED_OFF_BY, SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED,USER_AGENT_SANE_1 autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id E7D4CC2D0A8 for ; Wed, 23 Sep 2020 09:52:46 +0000 (UTC) Received: from gabe.freedesktop.org (gabe.freedesktop.org [131.252.210.177]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 7C10D214D8 for ; Wed, 23 Sep 2020 09:52:46 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 7C10D214D8 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=linux.intel.com Authentication-Results: mail.kernel.org; spf=none smtp.mailfrom=dri-devel-bounces@lists.freedesktop.org Received: from gabe.freedesktop.org (localhost [127.0.0.1]) by gabe.freedesktop.org (Postfix) with ESMTP id 187EA6E093; Wed, 23 Sep 2020 09:52:44 +0000 (UTC) Received: from mga02.intel.com (mga02.intel.com [134.134.136.20]) by gabe.freedesktop.org (Postfix) with ESMTPS id 65CB66E093; Wed, 23 Sep 2020 09:52:43 +0000 (UTC) IronPort-SDR: UQo/hK2K3gff/26nznxVtZmu69MENX3p8U4srk5hVmrvjCKPV6lpEWa9ETox+E2d7bjjRxvGAT wIE/V+vdUupQ== X-IronPort-AV: E=McAfee;i="6000,8403,9752"; a="148500533" X-IronPort-AV: E=Sophos;i="5.77,293,1596524400"; d="scan'208";a="148500533" X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from orsmga005.jf.intel.com ([10.7.209.41]) by orsmga101.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 23 Sep 2020 02:52:41 -0700 IronPort-SDR: seIxyGE1Vw8wLigC1YT863o99ElqSWQAiuSZRkgjZoMQsFRRWMV/uMP79du30Xlum1gHB9JcsL DrdLmgIHayQA== X-IronPort-AV: E=Sophos;i="5.77,293,1596524400"; d="scan'208";a="486376229" Received: from yymichae-mobl.ger.corp.intel.com (HELO [10.214.208.219]) ([10.214.208.219]) by orsmga005-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 23 Sep 2020 02:52:36 -0700 Subject: Re: [Intel-gfx] [PATCH 4/6] drm/i915: use vmap in i915_gem_object_map To: Christoph Hellwig , Andrew Morton References: <20200918163724.2511-1-hch@lst.de> <20200918163724.2511-5-hch@lst.de> From: Tvrtko Ursulin Organization: Intel Corporation UK Plc Message-ID: <9b5d40af-7378-9e68-ca51-73b2148287f3@linux.intel.com> Date: Wed, 23 Sep 2020 10:52:33 +0100 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.10.0 MIME-Version: 1.0 In-Reply-To: <20200918163724.2511-5-hch@lst.de> Content-Language: en-US X-BeenThere: dri-devel@lists.freedesktop.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: Direct Rendering Infrastructure - Development List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Juergen Gross , Stefano Stabellini , Minchan Kim , Peter Zijlstra , intel-gfx@lists.freedesktop.org, x86@kernel.org, linux-kernel@vger.kernel.org, dri-devel@lists.freedesktop.org, linux-mm@kvack.org, xen-devel@lists.xenproject.org, Boris Ostrovsky , Nitin Gupta Content-Transfer-Encoding: 7bit Content-Type: text/plain; charset="us-ascii"; Format="flowed" Errors-To: dri-devel-bounces@lists.freedesktop.org Sender: "dri-devel" On 18/09/2020 17:37, Christoph Hellwig wrote: > i915_gem_object_map implements fairly low-level vmap functionality in > a driver. Split it into two helpers, one for remapping kernel memory > which can use vmap, and one for I/O memory that uses vmap_pfn. > > The only practical difference is that alloc_vm_area prefeaults the > vmalloc area PTEs, which doesn't seem to be required here for the > kernel memory case (and could be added to vmap using a flag if actually > required). Patch looks good to me. Series did not get a CI run from our side because of a different base so I don't know if you would like to have a run there? If so you would need to rebase against git://anongit.freedesktop.org/drm-tip drm-tip and you could even send a series to intel-gfx-trybot@lists.freedesktop.org, suppressing cc, to check it out without sending a copy to the real mailing list. Regards, Tvrtko > Signed-off-by: Christoph Hellwig > --- > drivers/gpu/drm/i915/Kconfig | 1 + > drivers/gpu/drm/i915/gem/i915_gem_pages.c | 101 ++++++++++------------ > 2 files changed, 47 insertions(+), 55 deletions(-) > > diff --git a/drivers/gpu/drm/i915/Kconfig b/drivers/gpu/drm/i915/Kconfig > index 9afa5c4a6bf006..1e1cb245fca778 100644 > --- a/drivers/gpu/drm/i915/Kconfig > +++ b/drivers/gpu/drm/i915/Kconfig > @@ -25,6 +25,7 @@ config DRM_I915 > select CRC32 > select SND_HDA_I915 if SND_HDA_CORE > select CEC_CORE if CEC_NOTIFIER > + select VMAP_PFN > help > Choose this option if you have a system that has "Intel Graphics > Media Accelerator" or "HD Graphics" integrated graphics, > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_pages.c b/drivers/gpu/drm/i915/gem/i915_gem_pages.c > index e8a083743e0927..90029ea83aede9 100644 > --- a/drivers/gpu/drm/i915/gem/i915_gem_pages.c > +++ b/drivers/gpu/drm/i915/gem/i915_gem_pages.c > @@ -234,50 +234,24 @@ int __i915_gem_object_put_pages(struct drm_i915_gem_object *obj) > return err; > } > > -static inline pte_t iomap_pte(resource_size_t base, > - dma_addr_t offset, > - pgprot_t prot) > -{ > - return pte_mkspecial(pfn_pte((base + offset) >> PAGE_SHIFT, prot)); > -} > - > /* The 'mapping' part of i915_gem_object_pin_map() below */ > -static void *i915_gem_object_map(struct drm_i915_gem_object *obj, > +static void *i915_gem_object_map_page(struct drm_i915_gem_object *obj, > enum i915_map_type type) > { > - unsigned long n_pte = obj->base.size >> PAGE_SHIFT; > - struct sg_table *sgt = obj->mm.pages; > - pte_t *stack[32], **mem; > - struct vm_struct *area; > + unsigned long n_pages = obj->base.size >> PAGE_SHIFT, i; > + struct page *stack[32], **pages = stack, *page; > + struct sgt_iter iter; > pgprot_t pgprot; > - > - if (!i915_gem_object_has_struct_page(obj) && type != I915_MAP_WC) > - return NULL; > - > - /* A single page can always be kmapped */ > - if (n_pte == 1 && type == I915_MAP_WB) > - return kmap(sg_page(sgt->sgl)); > - > - mem = stack; > - if (n_pte > ARRAY_SIZE(stack)) { > - /* Too big for stack -- allocate temporary array instead */ > - mem = kvmalloc_array(n_pte, sizeof(*mem), GFP_KERNEL); > - if (!mem) > - return NULL; > - } > - > - area = alloc_vm_area(obj->base.size, mem); > - if (!area) { > - if (mem != stack) > - kvfree(mem); > - return NULL; > - } > + void *vaddr; > > switch (type) { > default: > MISSING_CASE(type); > fallthrough; /* to use PAGE_KERNEL anyway */ > case I915_MAP_WB: > + /* A single page can always be kmapped */ > + if (n_pages == 1) > + return kmap(sg_page(obj->mm.pages->sgl)); > pgprot = PAGE_KERNEL; > break; > case I915_MAP_WC: > @@ -285,30 +259,44 @@ static void *i915_gem_object_map(struct drm_i915_gem_object *obj, > break; > } > > - if (i915_gem_object_has_struct_page(obj)) { > - struct sgt_iter iter; > - struct page *page; > - pte_t **ptes = mem; > - > - for_each_sgt_page(page, iter, sgt) > - **ptes++ = mk_pte(page, pgprot); > - } else { > - resource_size_t iomap; > - struct sgt_iter iter; > - pte_t **ptes = mem; > - dma_addr_t addr; > + if (n_pages > ARRAY_SIZE(stack)) { > + /* Too big for stack -- allocate temporary array instead */ > + pages = kvmalloc_array(n_pages, sizeof(*pages), GFP_KERNEL); > + if (!pages) > + return NULL; > + } > > - iomap = obj->mm.region->iomap.base; > - iomap -= obj->mm.region->region.start; > + for_each_sgt_page(page, iter, obj->mm.pages) > + pages[i++] = page; > + vaddr = vmap(pages, n_pages, 0, pgprot); > + if (pages != stack) > + kvfree(pages); > + return vaddr; > +} > > - for_each_sgt_daddr(addr, iter, sgt) > - **ptes++ = iomap_pte(iomap, addr, pgprot); > +static void *i915_gem_object_map_pfn(struct drm_i915_gem_object *obj) > +{ > + resource_size_t iomap = obj->mm.region->iomap.base - > + obj->mm.region->region.start; > + unsigned long n_pfn = obj->base.size >> PAGE_SHIFT; > + unsigned long stack[32], *pfns = stack, i; > + struct sgt_iter iter; > + dma_addr_t addr; > + void *vaddr; > + > + if (n_pfn > ARRAY_SIZE(stack)) { > + /* Too big for stack -- allocate temporary array instead */ > + pfns = kvmalloc_array(n_pfn, sizeof(*pfns), GFP_KERNEL); > + if (!pfns) > + return NULL; > } > > - if (mem != stack) > - kvfree(mem); > - > - return area->addr; > + for_each_sgt_daddr(addr, iter, obj->mm.pages) > + pfns[i++] = (iomap + addr) >> PAGE_SHIFT; > + vaddr = vmap_pfn(pfns, n_pfn, pgprot_writecombine(PAGE_KERNEL_IO)); > + if (pfns != stack) > + kvfree(pfns); > + return vaddr; > } > > /* get, pin, and map the pages of the object into kernel space */ > @@ -360,7 +348,10 @@ void *i915_gem_object_pin_map(struct drm_i915_gem_object *obj, > } > > if (!ptr) { > - ptr = i915_gem_object_map(obj, type); > + if (i915_gem_object_has_struct_page(obj)) > + ptr = i915_gem_object_map_page(obj, type); > + else if (type == I915_MAP_WC) > + ptr = i915_gem_object_map_pfn(obj); > if (!ptr) { > err = -ENOMEM; > goto err_unpin; > _______________________________________________ dri-devel mailing list dri-devel@lists.freedesktop.org https://lists.freedesktop.org/mailman/listinfo/dri-devel From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-16.2 required=3.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,HK_RANDOM_FROM,INCLUDES_PATCH, MAILING_LIST_MULTI,MENTIONS_GIT_HOSTING,NICE_REPLY_A,SIGNED_OFF_BY, SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED,USER_AGENT_SANE_1 autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 0CBC3C4727E for ; Wed, 23 Sep 2020 09:52:45 +0000 (UTC) Received: from gabe.freedesktop.org (gabe.freedesktop.org [131.252.210.177]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id B56F82076A for ; Wed, 23 Sep 2020 09:52:44 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org B56F82076A Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=linux.intel.com Authentication-Results: mail.kernel.org; spf=none smtp.mailfrom=intel-gfx-bounces@lists.freedesktop.org Received: from gabe.freedesktop.org (localhost [127.0.0.1]) by gabe.freedesktop.org (Postfix) with ESMTP id 1FBEA6E16D; Wed, 23 Sep 2020 09:52:44 +0000 (UTC) Received: from mga02.intel.com (mga02.intel.com [134.134.136.20]) by gabe.freedesktop.org (Postfix) with ESMTPS id 65CB66E093; Wed, 23 Sep 2020 09:52:43 +0000 (UTC) IronPort-SDR: UQo/hK2K3gff/26nznxVtZmu69MENX3p8U4srk5hVmrvjCKPV6lpEWa9ETox+E2d7bjjRxvGAT wIE/V+vdUupQ== X-IronPort-AV: E=McAfee;i="6000,8403,9752"; a="148500533" X-IronPort-AV: E=Sophos;i="5.77,293,1596524400"; d="scan'208";a="148500533" X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from orsmga005.jf.intel.com ([10.7.209.41]) by orsmga101.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 23 Sep 2020 02:52:41 -0700 IronPort-SDR: seIxyGE1Vw8wLigC1YT863o99ElqSWQAiuSZRkgjZoMQsFRRWMV/uMP79du30Xlum1gHB9JcsL DrdLmgIHayQA== X-IronPort-AV: E=Sophos;i="5.77,293,1596524400"; d="scan'208";a="486376229" Received: from yymichae-mobl.ger.corp.intel.com (HELO [10.214.208.219]) ([10.214.208.219]) by orsmga005-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 23 Sep 2020 02:52:36 -0700 To: Christoph Hellwig , Andrew Morton References: <20200918163724.2511-1-hch@lst.de> <20200918163724.2511-5-hch@lst.de> From: Tvrtko Ursulin Organization: Intel Corporation UK Plc Message-ID: <9b5d40af-7378-9e68-ca51-73b2148287f3@linux.intel.com> Date: Wed, 23 Sep 2020 10:52:33 +0100 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.10.0 MIME-Version: 1.0 In-Reply-To: <20200918163724.2511-5-hch@lst.de> Content-Language: en-US Subject: Re: [Intel-gfx] [PATCH 4/6] drm/i915: use vmap in i915_gem_object_map X-BeenThere: intel-gfx@lists.freedesktop.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: Intel graphics driver community testing & development List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Juergen Gross , Stefano Stabellini , Minchan Kim , Peter Zijlstra , intel-gfx@lists.freedesktop.org, x86@kernel.org, linux-kernel@vger.kernel.org, dri-devel@lists.freedesktop.org, linux-mm@kvack.org, xen-devel@lists.xenproject.org, Boris Ostrovsky , Nitin Gupta Content-Transfer-Encoding: 7bit Content-Type: text/plain; charset="us-ascii"; Format="flowed" Errors-To: intel-gfx-bounces@lists.freedesktop.org Sender: "Intel-gfx" On 18/09/2020 17:37, Christoph Hellwig wrote: > i915_gem_object_map implements fairly low-level vmap functionality in > a driver. Split it into two helpers, one for remapping kernel memory > which can use vmap, and one for I/O memory that uses vmap_pfn. > > The only practical difference is that alloc_vm_area prefeaults the > vmalloc area PTEs, which doesn't seem to be required here for the > kernel memory case (and could be added to vmap using a flag if actually > required). Patch looks good to me. Series did not get a CI run from our side because of a different base so I don't know if you would like to have a run there? If so you would need to rebase against git://anongit.freedesktop.org/drm-tip drm-tip and you could even send a series to intel-gfx-trybot@lists.freedesktop.org, suppressing cc, to check it out without sending a copy to the real mailing list. Regards, Tvrtko > Signed-off-by: Christoph Hellwig > --- > drivers/gpu/drm/i915/Kconfig | 1 + > drivers/gpu/drm/i915/gem/i915_gem_pages.c | 101 ++++++++++------------ > 2 files changed, 47 insertions(+), 55 deletions(-) > > diff --git a/drivers/gpu/drm/i915/Kconfig b/drivers/gpu/drm/i915/Kconfig > index 9afa5c4a6bf006..1e1cb245fca778 100644 > --- a/drivers/gpu/drm/i915/Kconfig > +++ b/drivers/gpu/drm/i915/Kconfig > @@ -25,6 +25,7 @@ config DRM_I915 > select CRC32 > select SND_HDA_I915 if SND_HDA_CORE > select CEC_CORE if CEC_NOTIFIER > + select VMAP_PFN > help > Choose this option if you have a system that has "Intel Graphics > Media Accelerator" or "HD Graphics" integrated graphics, > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_pages.c b/drivers/gpu/drm/i915/gem/i915_gem_pages.c > index e8a083743e0927..90029ea83aede9 100644 > --- a/drivers/gpu/drm/i915/gem/i915_gem_pages.c > +++ b/drivers/gpu/drm/i915/gem/i915_gem_pages.c > @@ -234,50 +234,24 @@ int __i915_gem_object_put_pages(struct drm_i915_gem_object *obj) > return err; > } > > -static inline pte_t iomap_pte(resource_size_t base, > - dma_addr_t offset, > - pgprot_t prot) > -{ > - return pte_mkspecial(pfn_pte((base + offset) >> PAGE_SHIFT, prot)); > -} > - > /* The 'mapping' part of i915_gem_object_pin_map() below */ > -static void *i915_gem_object_map(struct drm_i915_gem_object *obj, > +static void *i915_gem_object_map_page(struct drm_i915_gem_object *obj, > enum i915_map_type type) > { > - unsigned long n_pte = obj->base.size >> PAGE_SHIFT; > - struct sg_table *sgt = obj->mm.pages; > - pte_t *stack[32], **mem; > - struct vm_struct *area; > + unsigned long n_pages = obj->base.size >> PAGE_SHIFT, i; > + struct page *stack[32], **pages = stack, *page; > + struct sgt_iter iter; > pgprot_t pgprot; > - > - if (!i915_gem_object_has_struct_page(obj) && type != I915_MAP_WC) > - return NULL; > - > - /* A single page can always be kmapped */ > - if (n_pte == 1 && type == I915_MAP_WB) > - return kmap(sg_page(sgt->sgl)); > - > - mem = stack; > - if (n_pte > ARRAY_SIZE(stack)) { > - /* Too big for stack -- allocate temporary array instead */ > - mem = kvmalloc_array(n_pte, sizeof(*mem), GFP_KERNEL); > - if (!mem) > - return NULL; > - } > - > - area = alloc_vm_area(obj->base.size, mem); > - if (!area) { > - if (mem != stack) > - kvfree(mem); > - return NULL; > - } > + void *vaddr; > > switch (type) { > default: > MISSING_CASE(type); > fallthrough; /* to use PAGE_KERNEL anyway */ > case I915_MAP_WB: > + /* A single page can always be kmapped */ > + if (n_pages == 1) > + return kmap(sg_page(obj->mm.pages->sgl)); > pgprot = PAGE_KERNEL; > break; > case I915_MAP_WC: > @@ -285,30 +259,44 @@ static void *i915_gem_object_map(struct drm_i915_gem_object *obj, > break; > } > > - if (i915_gem_object_has_struct_page(obj)) { > - struct sgt_iter iter; > - struct page *page; > - pte_t **ptes = mem; > - > - for_each_sgt_page(page, iter, sgt) > - **ptes++ = mk_pte(page, pgprot); > - } else { > - resource_size_t iomap; > - struct sgt_iter iter; > - pte_t **ptes = mem; > - dma_addr_t addr; > + if (n_pages > ARRAY_SIZE(stack)) { > + /* Too big for stack -- allocate temporary array instead */ > + pages = kvmalloc_array(n_pages, sizeof(*pages), GFP_KERNEL); > + if (!pages) > + return NULL; > + } > > - iomap = obj->mm.region->iomap.base; > - iomap -= obj->mm.region->region.start; > + for_each_sgt_page(page, iter, obj->mm.pages) > + pages[i++] = page; > + vaddr = vmap(pages, n_pages, 0, pgprot); > + if (pages != stack) > + kvfree(pages); > + return vaddr; > +} > > - for_each_sgt_daddr(addr, iter, sgt) > - **ptes++ = iomap_pte(iomap, addr, pgprot); > +static void *i915_gem_object_map_pfn(struct drm_i915_gem_object *obj) > +{ > + resource_size_t iomap = obj->mm.region->iomap.base - > + obj->mm.region->region.start; > + unsigned long n_pfn = obj->base.size >> PAGE_SHIFT; > + unsigned long stack[32], *pfns = stack, i; > + struct sgt_iter iter; > + dma_addr_t addr; > + void *vaddr; > + > + if (n_pfn > ARRAY_SIZE(stack)) { > + /* Too big for stack -- allocate temporary array instead */ > + pfns = kvmalloc_array(n_pfn, sizeof(*pfns), GFP_KERNEL); > + if (!pfns) > + return NULL; > } > > - if (mem != stack) > - kvfree(mem); > - > - return area->addr; > + for_each_sgt_daddr(addr, iter, obj->mm.pages) > + pfns[i++] = (iomap + addr) >> PAGE_SHIFT; > + vaddr = vmap_pfn(pfns, n_pfn, pgprot_writecombine(PAGE_KERNEL_IO)); > + if (pfns != stack) > + kvfree(pfns); > + return vaddr; > } > > /* get, pin, and map the pages of the object into kernel space */ > @@ -360,7 +348,10 @@ void *i915_gem_object_pin_map(struct drm_i915_gem_object *obj, > } > > if (!ptr) { > - ptr = i915_gem_object_map(obj, type); > + if (i915_gem_object_has_struct_page(obj)) > + ptr = i915_gem_object_map_page(obj, type); > + else if (type == I915_MAP_WC) > + ptr = i915_gem_object_map_pfn(obj); > if (!ptr) { > err = -ENOMEM; > goto err_unpin; > _______________________________________________ Intel-gfx mailing list Intel-gfx@lists.freedesktop.org https://lists.freedesktop.org/mailman/listinfo/intel-gfx