From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-4.1 required=3.0 tests=DKIM_SIGNED,DKIM_VALID, DKIM_VALID_AU,FREEMAIL_FORGED_FROMDOMAIN,FREEMAIL_FROM, HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_HELO_NONE, SPF_PASS,URIBL_BLOCKED autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id DB9EFC49ED7 for ; Mon, 16 Sep 2019 04:45:20 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id 8D52420890 for ; Mon, 16 Sep 2019 04:45:20 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (2048-bit key) header.d=gmail.com header.i=@gmail.com header.b="DXztuFoV" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727937AbfIPEpT (ORCPT ); Mon, 16 Sep 2019 00:45:19 -0400 Received: from mail-vs1-f65.google.com ([209.85.217.65]:40093 "EHLO mail-vs1-f65.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726128AbfIPEpT (ORCPT ); Mon, 16 Sep 2019 00:45:19 -0400 Received: by mail-vs1-f65.google.com with SMTP id v10so19148114vsc.7 for ; Sun, 15 Sep 2019 21:45:18 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=ckAPzt2hjbX190/rHtJWex8pJTA//TPllI5bTrQOOAc=; b=DXztuFoVa7VP57BPbHaSDCPaj5/A8YmfItUOjZ7epZ3+UOst4zFEAk93OshNZybXmW VgYvgr5wuVOKDjaqYKoHyMwKcs0V3MlsDBUcMIbiUhd8DUyBfahWlN/itYoJKPyciQJR ku7PXuqJ/puh2/2DyvBNAPDtnU++IaWqONzFwNV9u8it/UtAmXIUrrIcBifTGkEXYBL0 DLugW+D6l3Lc+Si+tT0VrW1EbQQUlRYaBAuK5vEKAlUk/rjbJyuyziWu1gSnmMUV9NvW W0Y2UT4WvDl6j5qSz/LWlzrV7/k7Wad40jMBOSXv/ms0XdhaTPuvXUC8/4CCRwlAw853 ODbA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=ckAPzt2hjbX190/rHtJWex8pJTA//TPllI5bTrQOOAc=; b=bWWjnd1YBQHs+zFZGJkIYU8v2dXA1njbJbz2FhZdnyWvJXjePnOqOaQlnWNSi3Uh7f GVrVVShimUMkrRwbkprfXyKxrOEUrtxznHJDpux7m4ev6CYGruGNcD/iH56no6U2abVW nMNhMBT/AJFVDAW9eFKAxhhS0pljZaZkZpTBV7ymHbU4a+nfFMJ1R0kSYEqdpJFYH8OL DIm75WioCm3PiwaUuQ1fH7rPbgAsRa3FufriTwqrIgyHD/m9cfCkTEn32iSLBUN412d7 AME5D6PkneD1ehQfvrUsAE4Bof13CIsYiuqPtRSJknjES7JFQHVHI2S8dxaCMJQKOa2w /jkg== X-Gm-Message-State: APjAAAXsK9Bgx9Km/NLk3arYsi2Dej3ALjVMRjegwksQMU6ZqH5VcoZZ o9auWASEFnpMiwMHTkEnI7wNzJzRzbL6xY9Zf5w= X-Google-Smtp-Source: APXvYqzWHds35vLSToJKhbTRshf6wIelAL7txb2fwkaJSCNZt8c4jkUDu6Coe2OAQgvSC0fvK/daaeiUUAEfqQkG/BI= X-Received: by 2002:a67:5f45:: with SMTP id t66mr21145461vsb.204.1568609117965; Sun, 15 Sep 2019 21:45:17 -0700 (PDT) MIME-Version: 1.0 References: <20190805140119.7337-1-kraxel@redhat.com> <20190805140119.7337-9-kraxel@redhat.com> <20190813151115.GA29955@ulmo> <20190814055827.6hrxj6daovxxnnvw@sirius.home.kraxel.org> <20190814093524.GA31345@ulmo> <20190814101411.lj3p6zjzbjvnnjf4@sirius.home.kraxel.org> <20190821115523.GA21839@ulmo> <20190910215259.GA7525@ulmo> In-Reply-To: <20190910215259.GA7525@ulmo> From: Ben Skeggs Date: Mon, 16 Sep 2019 14:45:06 +1000 Message-ID: Subject: Re: [Nouveau] [Intel-gfx] [PATCH v6 08/17] drm/ttm: use gem vma_node To: Thierry Reding Cc: Ilia Mirkin , ML dri-devel , ML nouveau , intel-gfx , LKML , amd-gfx mailing list , linux-graphics-maintainer@vmware.com, Gerd Hoffmann , spice-devel@lists.freedesktop.org, Ben Skeggs Content-Type: text/plain; charset="UTF-8" Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Wed, 11 Sep 2019 at 07:53, Thierry Reding wrote: > > On Sat, Sep 07, 2019 at 09:58:46PM -0400, Ilia Mirkin wrote: > > On Wed, Aug 21, 2019 at 7:55 AM Thierry Reding wrote: > > > > > > On Wed, Aug 21, 2019 at 04:33:58PM +1000, Ben Skeggs wrote: > > > > On Wed, 14 Aug 2019 at 20:14, Gerd Hoffmann wrote: > > > > > > > > > > Hi, > > > > > > > > > > > > Changing the order doesn't look hard. Patch attached (untested, have no > > > > > > > test hardware). But maybe I missed some detail ... > > > > > > > > > > > > I came up with something very similar by splitting up nouveau_bo_new() > > > > > > into allocation and initialization steps, so that when necessary the GEM > > > > > > object can be initialized in between. I think that's slightly more > > > > > > flexible and easier to understand than a boolean flag. > > > > > > > > > > Yes, that should work too. > > > > > > > > > > Acked-by: Gerd Hoffmann > > > > Acked-by: Ben Skeggs > > > > > > Thanks guys, applied to drm-misc-next. > > > > Hi Thierry, > > > > Initial investigations suggest that this commit currently in drm-next > > > > commit 019cbd4a4feb3aa3a917d78e7110e3011bbff6d5 > > Author: Thierry Reding > > Date: Wed Aug 14 11:00:48 2019 +0200 > > > > drm/nouveau: Initialize GEM object before TTM object > > > > breaks nouveau userspace which tries to allocate GEM objects with a > > non-page-aligned size. Previously nouveau_gem_new would just call > > nouveau_bo_init which would call nouveau_bo_fixup_align before > > initializing the GEM object. With this change, it is done after. What > > do you think -- OK to just move that bit of logic into the new > > nouveau_bo_alloc() (and make size/align be pointers so that they can > > be fixed up?) > > Hi Ilia, > > sorry, got side-tracked earlier and forgot to send this out. I'll turn > this into a proper patch, but if you manage to find the time to test > this while I work out the userspace issues that are preventing me from > testing this more thoroughly, that'd be great. I can confirm both I can reproduce the bug, and that the fix here appears to do the trick nicely. Ben. > > Thierry > > --- >8 --- > diff --git a/drivers/gpu/drm/nouveau/nouveau_bo.c b/drivers/gpu/drm/nouveau/nouveau_bo.c > index e918b437af17..7d5ede756711 100644 > --- a/drivers/gpu/drm/nouveau/nouveau_bo.c > +++ b/drivers/gpu/drm/nouveau/nouveau_bo.c > @@ -186,8 +186,8 @@ nouveau_bo_fixup_align(struct nouveau_bo *nvbo, u32 flags, > } > > struct nouveau_bo * > -nouveau_bo_alloc(struct nouveau_cli *cli, u64 size, u32 flags, u32 tile_mode, > - u32 tile_flags) > +nouveau_bo_alloc(struct nouveau_cli *cli, u64 *size, int *align, u32 flags, > + u32 tile_mode, u32 tile_flags) > { > struct nouveau_drm *drm = cli->drm; > struct nouveau_bo *nvbo; > @@ -195,8 +195,8 @@ nouveau_bo_alloc(struct nouveau_cli *cli, u64 size, u32 flags, u32 tile_mode, > struct nvif_vmm *vmm = cli->svm.cli ? &cli->svm.vmm : &cli->vmm.vmm; > int i, pi = -1; > > - if (!size) { > - NV_WARN(drm, "skipped size %016llx\n", size); > + if (!*size) { > + NV_WARN(drm, "skipped size %016llx\n", *size); > return ERR_PTR(-EINVAL); > } > > @@ -266,7 +266,7 @@ nouveau_bo_alloc(struct nouveau_cli *cli, u64 size, u32 flags, u32 tile_mode, > pi = i; > > /* Stop once the buffer is larger than the current page size. */ > - if (size >= 1ULL << vmm->page[i].shift) > + if (*size >= 1ULL << vmm->page[i].shift) > break; > } > > @@ -281,6 +281,8 @@ nouveau_bo_alloc(struct nouveau_cli *cli, u64 size, u32 flags, u32 tile_mode, > } > nvbo->page = vmm->page[pi].shift; > > + nouveau_bo_fixup_align(nvbo, flags, align, size); > + > return nvbo; > } > > @@ -292,12 +294,11 @@ nouveau_bo_init(struct nouveau_bo *nvbo, u64 size, int align, u32 flags, > size_t acc_size; > int ret; > > - acc_size = ttm_bo_dma_acc_size(nvbo->bo.bdev, size, sizeof(*nvbo)); > - > - nouveau_bo_fixup_align(nvbo, flags, &align, &size); > nvbo->bo.mem.num_pages = size >> PAGE_SHIFT; > nouveau_bo_placement_set(nvbo, flags, 0); > > + acc_size = ttm_bo_dma_acc_size(nvbo->bo.bdev, size, sizeof(*nvbo)); > + > ret = ttm_bo_init(nvbo->bo.bdev, &nvbo->bo, size, type, > &nvbo->placement, align >> PAGE_SHIFT, false, > acc_size, sg, robj, nouveau_bo_del_ttm); > @@ -318,7 +319,8 @@ nouveau_bo_new(struct nouveau_cli *cli, u64 size, int align, > struct nouveau_bo *nvbo; > int ret; > > - nvbo = nouveau_bo_alloc(cli, size, flags, tile_mode, tile_flags); > + nvbo = nouveau_bo_alloc(cli, &size, &align, flags, tile_mode, > + tile_flags); > if (IS_ERR(nvbo)) > return PTR_ERR(nvbo); > > diff --git a/drivers/gpu/drm/nouveau/nouveau_bo.h b/drivers/gpu/drm/nouveau/nouveau_bo.h > index 62930d834fba..38f9d8350963 100644 > --- a/drivers/gpu/drm/nouveau/nouveau_bo.h > +++ b/drivers/gpu/drm/nouveau/nouveau_bo.h > @@ -71,8 +71,8 @@ nouveau_bo_ref(struct nouveau_bo *ref, struct nouveau_bo **pnvbo) > extern struct ttm_bo_driver nouveau_bo_driver; > > void nouveau_bo_move_init(struct nouveau_drm *); > -struct nouveau_bo *nouveau_bo_alloc(struct nouveau_cli *, u64 size, u32 flags, > - u32 tile_mode, u32 tile_flags); > +struct nouveau_bo *nouveau_bo_alloc(struct nouveau_cli *, u64 *size, int *align, > + u32 flags, u32 tile_mode, u32 tile_flags); > int nouveau_bo_init(struct nouveau_bo *, u64 size, int align, u32 flags, > struct sg_table *sg, struct dma_resv *robj); > int nouveau_bo_new(struct nouveau_cli *, u64 size, int align, u32 flags, > diff --git a/drivers/gpu/drm/nouveau/nouveau_gem.c b/drivers/gpu/drm/nouveau/nouveau_gem.c > index c2bfc0591909..1bdffd714456 100644 > --- a/drivers/gpu/drm/nouveau/nouveau_gem.c > +++ b/drivers/gpu/drm/nouveau/nouveau_gem.c > @@ -188,7 +188,8 @@ nouveau_gem_new(struct nouveau_cli *cli, u64 size, int align, uint32_t domain, > if (domain & NOUVEAU_GEM_DOMAIN_COHERENT) > flags |= TTM_PL_FLAG_UNCACHED; > > - nvbo = nouveau_bo_alloc(cli, size, flags, tile_mode, tile_flags); > + nvbo = nouveau_bo_alloc(cli, &size, &align, flags, tile_mode, > + tile_flags); > if (IS_ERR(nvbo)) > return PTR_ERR(nvbo); > > diff --git a/drivers/gpu/drm/nouveau/nouveau_prime.c b/drivers/gpu/drm/nouveau/nouveau_prime.c > index 84658d434225..656c334ee7d9 100644 > --- a/drivers/gpu/drm/nouveau/nouveau_prime.c > +++ b/drivers/gpu/drm/nouveau/nouveau_prime.c > @@ -62,14 +62,15 @@ struct drm_gem_object *nouveau_gem_prime_import_sg_table(struct drm_device *dev, > struct nouveau_drm *drm = nouveau_drm(dev); > struct nouveau_bo *nvbo; > struct dma_resv *robj = attach->dmabuf->resv; > - size_t size = attach->dmabuf->size; > + u64 size = attach->dmabuf->size; > u32 flags = 0; > + int align = 0; > int ret; > > flags = TTM_PL_FLAG_TT; > > dma_resv_lock(robj, NULL); > - nvbo = nouveau_bo_alloc(&drm->client, size, flags, 0, 0); > + nvbo = nouveau_bo_alloc(&drm->client, &size, &align, flags, 0, 0); > dma_resv_unlock(robj); > if (IS_ERR(nvbo)) > return ERR_CAST(nvbo); > @@ -84,7 +85,7 @@ struct drm_gem_object *nouveau_gem_prime_import_sg_table(struct drm_device *dev, > return ERR_PTR(-ENOMEM); > } > > - ret = nouveau_bo_init(nvbo, size, 0, flags, sg, robj); > + ret = nouveau_bo_init(nvbo, size, align, flags, sg, robj); > if (ret) { > nouveau_bo_ref(NULL, &nvbo); > return ERR_PTR(ret);