From: "Christian König" <christian.koenig@amd.com>
To: Andrey Grodzovsky <Andrey.Grodzovsky@amd.com>,
Daniel Vetter <daniel.vetter@ffwll.ch>
Cc: "Michel Dänzer" <michel@daenzer.net>,
"amd-gfx list" <amd-gfx@lists.freedesktop.org>,
dri-devel <dri-devel@lists.freedesktop.org>
Subject: Re: [PATCH v2 1/8] drm: Add dummy page per device or GEM object
Date: Mon, 16 Nov 2020 21:36:54 +0100 [thread overview]
Message-ID: <42c5cd8d-f534-0b10-bb27-92f767a7d382@amd.com> (raw)
In-Reply-To: <f231ffef-683a-7f34-2ae6-346da6e1cbcb@amd.com>
Am 16.11.20 um 20:00 schrieb Andrey Grodzovsky:
>
> On 11/16/20 4:48 AM, Christian König wrote:
>> Am 15.11.20 um 07:34 schrieb Andrey Grodzovsky:
>>>
>>> On 11/14/20 4:51 AM, Daniel Vetter wrote:
>>>> On Sat, Nov 14, 2020 at 9:41 AM Christian König
>>>> <ckoenig.leichtzumerken@gmail.com> wrote:
>>>>> Am 13.11.20 um 21:52 schrieb Andrey Grodzovsky:
>>>>>> On 6/22/20 1:50 PM, Daniel Vetter wrote:
>>>>>>> On Mon, Jun 22, 2020 at 7:45 PM Christian König
>>>>>>> <christian.koenig@amd.com> wrote:
>>>>>>>> Am 22.06.20 um 16:32 schrieb Andrey Grodzovsky:
>>>>>>>>> On 6/22/20 9:18 AM, Christian König wrote:
>>>>>>>>>> Am 21.06.20 um 08:03 schrieb Andrey Grodzovsky:
>>>>>>>>>>> Will be used to reroute CPU mapped BO's page faults once
>>>>>>>>>>> device is removed.
>>>>>>>>>>>
>>>>>>>>>>> Signed-off-by: Andrey Grodzovsky <andrey.grodzovsky@amd.com>
>>>>>>>>>>> ---
>>>>>>>>>>> drivers/gpu/drm/drm_file.c | 8 ++++++++
>>>>>>>>>>> drivers/gpu/drm/drm_prime.c | 10 ++++++++++
>>>>>>>>>>> include/drm/drm_file.h | 2 ++
>>>>>>>>>>> include/drm/drm_gem.h | 2 ++
>>>>>>>>>>> 4 files changed, 22 insertions(+)
>>>>>>>>>>>
>>>>>>>>>>> diff --git a/drivers/gpu/drm/drm_file.c
>>>>>>>>>>> b/drivers/gpu/drm/drm_file.c
>>>>>>>>>>> index c4c704e..67c0770 100644
>>>>>>>>>>> --- a/drivers/gpu/drm/drm_file.c
>>>>>>>>>>> +++ b/drivers/gpu/drm/drm_file.c
>>>>>>>>>>> @@ -188,6 +188,12 @@ struct drm_file *drm_file_alloc(struct
>>>>>>>>>>> drm_minor *minor)
>>>>>>>>>>> goto out_prime_destroy;
>>>>>>>>>>> }
>>>>>>>>>>> + file->dummy_page = alloc_page(GFP_KERNEL |
>>>>>>>>>>> __GFP_ZERO);
>>>>>>>>>>> + if (!file->dummy_page) {
>>>>>>>>>>> + ret = -ENOMEM;
>>>>>>>>>>> + goto out_prime_destroy;
>>>>>>>>>>> + }
>>>>>>>>>>> +
>>>>>>>>>>> return file;
>>>>>>>>>>> out_prime_destroy:
>>>>>>>>>>> @@ -284,6 +290,8 @@ void drm_file_free(struct drm_file *file)
>>>>>>>>>>> if (dev->driver->postclose)
>>>>>>>>>>> dev->driver->postclose(dev, file);
>>>>>>>>>>> + __free_page(file->dummy_page);
>>>>>>>>>>> +
>>>>>>>>>>> drm_prime_destroy_file_private(&file->prime);
>>>>>>>>>>> WARN_ON(!list_empty(&file->event_list));
>>>>>>>>>>> diff --git a/drivers/gpu/drm/drm_prime.c
>>>>>>>>>>> b/drivers/gpu/drm/drm_prime.c
>>>>>>>>>>> index 1de2cde..c482e9c 100644
>>>>>>>>>>> --- a/drivers/gpu/drm/drm_prime.c
>>>>>>>>>>> +++ b/drivers/gpu/drm/drm_prime.c
>>>>>>>>>>> @@ -335,6 +335,13 @@ int drm_gem_prime_fd_to_handle(struct
>>>>>>>>>>> drm_device *dev,
>>>>>>>>>>> ret = drm_prime_add_buf_handle(&file_priv->prime,
>>>>>>>>>>> dma_buf, *handle);
>>>>>>>>>>> +
>>>>>>>>>>> + if (!ret) {
>>>>>>>>>>> + obj->dummy_page = alloc_page(GFP_KERNEL | __GFP_ZERO);
>>>>>>>>>>> + if (!obj->dummy_page)
>>>>>>>>>>> + ret = -ENOMEM;
>>>>>>>>>>> + }
>>>>>>>>>>> +
>>>>>>>>>> While the per file case still looks acceptable this is a
>>>>>>>>>> clear NAK
>>>>>>>>>> since it will massively increase the memory needed for a prime
>>>>>>>>>> exported object.
>>>>>>>>>>
>>>>>>>>>> I think that this is quite overkill in the first place and
>>>>>>>>>> for the
>>>>>>>>>> hot unplug case we can just use the global dummy page as well.
>>>>>>>>>>
>>>>>>>>>> Christian.
>>>>>>>>> Global dummy page is good for read access, what do you do on
>>>>>>>>> write
>>>>>>>>> access ? My first approach was indeed to map at first global
>>>>>>>>> dummy
>>>>>>>>> page as read only and mark the vma->vm_flags as !VM_SHARED
>>>>>>>>> assuming
>>>>>>>>> that this would trigger Copy On Write flow in core mm
>>>>>>>>> (https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Felixir.bootlin.com%2Flinux%2Fv5.7-rc7%2Fsource%2Fmm%2Fmemory.c%23L3977&data=04%7C01%7CAndrey.Grodzovsky%40amd.com%7C00053e9d983041ed63ae08d88882ed87%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C637409443224016377%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=kghiG3VpCJod6YefExoDVPl9X03zNhw3SN5GAxgbnmU%3D&reserved=0)
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> on the next page fault to same address triggered by a write
>>>>>>>>> access but
>>>>>>>>> then i realized a new COW page will be allocated for each such
>>>>>>>>> mapping
>>>>>>>>> and this is much more wasteful then having a dedicated page
>>>>>>>>> per GEM
>>>>>>>>> object.
>>>>>>>> Yeah, but this is only for a very very small corner cases. What
>>>>>>>> we need
>>>>>>>> to prevent is increasing the memory usage during normal
>>>>>>>> operation to
>>>>>>>> much.
>>>>>>>>
>>>>>>>> Using memory during the unplug is completely unproblematic
>>>>>>>> because we
>>>>>>>> just released quite a bunch of it by releasing all those system
>>>>>>>> memory
>>>>>>>> buffers.
>>>>>>>>
>>>>>>>> And I'm pretty sure that COWed pages are correctly accounted
>>>>>>>> towards
>>>>>>>> the
>>>>>>>> used memory of a process.
>>>>>>>>
>>>>>>>> So I think if that approach works as intended and the COW pages
>>>>>>>> are
>>>>>>>> released again on unmapping it would be the perfect solution to
>>>>>>>> the
>>>>>>>> problem.
>>>>>>>>
>>>>>>>> Daniel what do you think?
>>>>>>> If COW works, sure sounds reasonable. And if we can make sure we
>>>>>>> managed to drop all the system allocations (otherwise suddenly 2x
>>>>>>> memory usage, worst case). But I have no idea whether we can
>>>>>>> retroshoehorn that into an established vma, you might have fun
>>>>>>> stuff
>>>>>>> like a mkwrite handler there (which I thought is the COW handler
>>>>>>> thing, but really no idea).
>>>>>>>
>>>>>>> If we need to massively change stuff then I think rw dummy page,
>>>>>>> allocated on first fault after hotunplug (maybe just make it one
>>>>>>> per
>>>>>>> object, that's simplest) seems like the much safer option. Much
>>>>>>> less
>>>>>>> code that can go wrong.
>>>>>>> -Daniel
>>>>>>
>>>>>> Regarding COW, i was looking into how to properly implement it from
>>>>>> within the fault handler (i.e. ttm_bo_vm_fault)
>>>>>> and the main obstacle I hit is that of exclusive access to the
>>>>>> vm_area_struct, i need to be able to modify
>>>>>> vma->vm_flags (and vm_page_prot) to remove VM_SHARED bit so COW can
>>>>>> be triggered on subsequent write access
>>>>>> fault (here
>>>>>> https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Felixir.bootlin.com%2Flinux%2Flatest%2Fsource%2Fmm%2Fmemory.c%23L4128&data=04%7C01%7CAndrey.Grodzovsky%40amd.com%7C00053e9d983041ed63ae08d88882ed87%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C637409443224016377%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=ziHJtqyHuLrlb0uYKhoWCWhUAZnX0JquE%2BkBJ5Fx%2BNo%3D&reserved=0)
>>>>>>
>>>>>> but core mm takes only read side mm_sem (here for example
>>>>>> https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Felixir.bootlin.com%2Flinux%2Flatest%2Fsource%2Fdrivers%2Fiommu%2Famd%2Fiommu_v2.c%23L488&data=04%7C01%7CAndrey.Grodzovsky%40amd.com%7C00053e9d983041ed63ae08d88882ed87%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C637409443224016377%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=h360c75Upl3%2FW7im7M1%2BxY%2FXy4gxin%2BkCF1Ui2zFXMs%3D&reserved=0)
>>>>>>
>>>>>> and so I am not supposed to modify vm_area_struct in this case. I am
>>>>>> not sure if it's legit to write lock tthe mm_sem from this point.
>>>>>> I found some discussions about this here
>>>>>> https://nam11.safelinks.protection.outlook.com/?url=http%3A%2F%2Flkml.iu.edu%2Fhypermail%2Flinux%2Fkernel%2F1909.1%2F02754.html&data=04%7C01%7CAndrey.Grodzovsky%40amd.com%7C00053e9d983041ed63ae08d88882ed87%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C637409443224021379%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=sx6s1lH%2FvxbIZajc4Yr49vFhxvPEnBHZlTt52D8qvZA%3D&reserved=0
>>>>>> but it
>>>>>> wasn't really clear to me
>>>>>> what's the solution.
>>>>>>
>>>>>> In any case, seems to me that easier and more memory saving solution
>>>>>> would be to just switch to per ttm bo dumy rw page that
>>>>>> would be allocated on demand as you suggested here. This should also
>>>>>> take care of imported BOs and flink cases.
>>>>>> Then i can drop the per device FD and per GEM object FD dummy BO and
>>>>>> the ugly loop i am using in patch 2 to match faulting BO to the
>>>>>> right
>>>>>> dummy page.
>>>>>>
>>>>>> Does this makes sense ?
>>>>> I still don't see the information leak as much of a problem, but if
>>>>> Daniel insists we should probably do this.
>>>> Well amdgpu doesn't clear buffers by default, so indeed you guys are a
>>>> lot more laissez-faire here. But in general we really don't do that
>>>> kind of leaking. Iirc there's even radeonsi bugs because else clears,
>>>> and radeonsi happily displays gunk :-)
>>>>
>>>>> But could we at least have only one page per client instead of per
>>>>> BO?
>>>> I think you can do one page per file descriptor or something like
>>>> that. But gets annoying with shared bo, especially with dma_buf_mmap
>>>> forwarding.
>>>> -Daniel
>>>
>>>
>>> Christian - is your concern more with too much page allocations or
>>> with extra pointer member
>>> cluttering TTM BO struct ?
>>
>> Yes, that is one problem.
>>
>>> Because we can allocate the dummy page on demand only when
>>> needed. It's just seems to me that keeping it per BO streamlines the
>>> code as I don't need to
>>> have different handling for local vs imported BOs.
>>
>> Why should you have a difference between local vs imported BOs?
>
>
> For local BO seems like Daniel's suggestion to use
> vm_area_struct->vm_file->private_data
> should work as this points to drm_file. For imported BOs private_data
> will point to dma_buf structure
> since each imported BO is backed by a pseudo file (created in
> dma_buf_getfile).
Oh, good point. But we could easily fix that now. That should make the
mapping code less complex as well.
Regards,
Christian.
> If so,where should we store the dummy RW BO in this case ? In current
> implementation it's stored in drm_gem_object.
>
> P.S For FLINK case it seems to me the handling should be no different
> then with local BO as the
> FD used for mmap in this case is still the same one associated with
> the DRM file.
>
> Andrey
>
>
>>
>> Christian.
>>
>>>
>>> Andrey
>>
_______________________________________________
dri-devel mailing list
dri-devel@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/dri-devel
next prev parent reply other threads:[~2020-11-16 20:37 UTC|newest]
Thread overview: 97+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-06-21 6:03 [PATCH v2 0/8] RFC Support hot device unplug in amdgpu Andrey Grodzovsky
2020-06-21 6:03 ` [PATCH v2 1/8] drm: Add dummy page per device or GEM object Andrey Grodzovsky
2020-06-22 9:35 ` Daniel Vetter
2020-06-22 14:21 ` Pekka Paalanen
2020-06-22 14:24 ` Daniel Vetter
2020-06-22 14:28 ` Pekka Paalanen
2020-11-09 20:34 ` Andrey Grodzovsky
2020-11-15 6:39 ` Andrey Grodzovsky
2020-06-22 13:18 ` Christian König
2020-06-22 14:23 ` Daniel Vetter
2020-06-22 14:32 ` Andrey Grodzovsky
2020-06-22 17:45 ` Christian König
2020-06-22 17:50 ` Daniel Vetter
2020-11-09 20:53 ` Andrey Grodzovsky
2020-11-13 20:52 ` Andrey Grodzovsky
2020-11-14 8:41 ` Christian König
2020-11-14 9:51 ` Daniel Vetter
2020-11-14 9:57 ` Daniel Vetter
2020-11-16 9:42 ` Michel Dänzer
2020-11-15 6:34 ` Andrey Grodzovsky
2020-11-16 9:48 ` Christian König
2020-11-16 19:00 ` Andrey Grodzovsky
2020-11-16 20:36 ` Christian König [this message]
2020-11-16 20:42 ` Andrey Grodzovsky
2020-11-19 10:01 ` Christian König
2020-06-21 6:03 ` [PATCH v2 2/8] drm/ttm: Remap all page faults to per process dummy page Andrey Grodzovsky
2020-06-22 9:41 ` Daniel Vetter
2020-06-24 3:31 ` Andrey Grodzovsky
2020-06-24 7:19 ` Daniel Vetter
2020-11-10 17:41 ` Andrey Grodzovsky
2020-06-22 19:30 ` Christian König
2020-06-21 6:03 ` [PATCH v2 3/8] drm/ttm: Add unampping of the entire device address space Andrey Grodzovsky
2020-06-22 9:45 ` Daniel Vetter
2020-06-23 5:00 ` Andrey Grodzovsky
2020-06-23 10:25 ` Daniel Vetter
2020-06-23 12:55 ` Christian König
2020-06-22 19:37 ` Christian König
2020-06-22 19:47 ` Alex Deucher
2020-06-21 6:03 ` [PATCH v2 4/8] drm/amdgpu: Split amdgpu_device_fini into early and late Andrey Grodzovsky
2020-06-22 9:48 ` Daniel Vetter
2020-11-12 4:19 ` Andrey Grodzovsky
2020-11-12 9:29 ` Daniel Vetter
2020-06-21 6:03 ` [PATCH v2 5/8] drm/amdgpu: Refactor sysfs removal Andrey Grodzovsky
2020-06-22 9:51 ` Daniel Vetter
2020-06-22 11:21 ` Greg KH
2020-06-22 16:07 ` Andrey Grodzovsky
2020-06-22 16:45 ` Greg KH
2020-06-23 4:51 ` Andrey Grodzovsky
2020-06-23 6:05 ` Greg KH
2020-06-24 3:04 ` Andrey Grodzovsky
2020-06-24 6:11 ` Greg KH
2020-06-25 1:52 ` Andrey Grodzovsky
2020-11-10 17:54 ` Andrey Grodzovsky
2020-11-10 17:59 ` Greg KH
2020-11-11 15:13 ` Andrey Grodzovsky
2020-11-11 15:34 ` Greg KH
2020-11-11 15:45 ` Andrey Grodzovsky
2020-11-11 16:06 ` Greg KH
2020-11-11 16:34 ` Andrey Grodzovsky
2020-12-02 15:48 ` Andrey Grodzovsky
2020-12-02 17:34 ` Greg KH
2020-12-02 18:02 ` Andrey Grodzovsky
2020-12-02 18:20 ` Greg KH
2020-12-02 18:40 ` Andrey Grodzovsky
2020-06-22 13:19 ` Christian König
2020-06-21 6:03 ` [PATCH v2 6/8] drm/amdgpu: Unmap entire device address space on device remove Andrey Grodzovsky
2020-06-22 9:56 ` Daniel Vetter
2020-06-22 19:38 ` Christian König
2020-06-22 19:48 ` Alex Deucher
2020-06-23 10:22 ` Daniel Vetter
2020-06-23 13:16 ` Christian König
2020-06-24 3:12 ` Andrey Grodzovsky
2020-06-21 6:03 ` [PATCH v2 7/8] drm/amdgpu: Fix sdma code crash post device unplug Andrey Grodzovsky
2020-06-22 9:55 ` Daniel Vetter
2020-06-22 19:40 ` Christian König
2020-06-23 5:11 ` Andrey Grodzovsky
2020-06-23 7:14 ` Christian König
2020-06-21 6:03 ` [PATCH v2 8/8] drm/amdgpu: Prevent any job recoveries after device is unplugged Andrey Grodzovsky
2020-06-22 9:53 ` Daniel Vetter
2020-11-17 18:38 ` Andrey Grodzovsky
2020-11-17 18:52 ` Daniel Vetter
2020-11-17 19:18 ` Andrey Grodzovsky
2020-11-17 19:49 ` Daniel Vetter
2020-11-17 20:07 ` Andrey Grodzovsky
2020-11-18 7:39 ` Daniel Vetter
2020-11-18 12:01 ` Christian König
2020-11-18 15:43 ` Luben Tuikov
2020-11-18 16:20 ` Andrey Grodzovsky
2020-11-19 7:55 ` Christian König
2020-11-19 15:02 ` Andrey Grodzovsky
2020-11-19 15:29 ` Daniel Vetter
2020-11-19 21:24 ` Andrey Grodzovsky
2020-11-18 0:46 ` Luben Tuikov
2020-06-22 9:46 ` [PATCH v2 0/8] RFC Support hot device unplug in amdgpu Daniel Vetter
2020-06-23 5:14 ` Andrey Grodzovsky
2020-06-23 9:04 ` Michel Dänzer
2020-06-24 3:21 ` Andrey Grodzovsky
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=42c5cd8d-f534-0b10-bb27-92f767a7d382@amd.com \
--to=christian.koenig@amd.com \
--cc=Andrey.Grodzovsky@amd.com \
--cc=amd-gfx@lists.freedesktop.org \
--cc=daniel.vetter@ffwll.ch \
--cc=dri-devel@lists.freedesktop.org \
--cc=michel@daenzer.net \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).