On Mon, May 10, 2021 at 11:23:24AM -0400, Vivek Goyal wrote: > On Mon, May 10, 2021 at 10:05:09AM +0100, Stefan Hajnoczi wrote: > > On Thu, May 06, 2021 at 12:02:23PM -0400, Vivek Goyal wrote: > > > On Thu, May 06, 2021 at 04:37:04PM +0100, Stefan Hajnoczi wrote: > > > > On Wed, Apr 28, 2021 at 12:01:00PM +0100, Dr. David Alan Gilbert (git) wrote: > > > > > From: Vivek Goyal > > > > > > > > > > If qemu guest asked to drop CAP_FSETID upon write, send that info > > > > > to qemu in SLAVE_FS_IO message so that qemu can drop capability > > > > > before WRITE. This is to make sure that any setuid bit is killed > > > > > on fd (if there is one set). > > > > > > > > > > Signed-off-by: Vivek Goyal > > > > > > > > I'm not sure if the QEMU FSETID patches make sense. QEMU shouldn't be > > > > running with FSETID because QEMU is untrusted. FSETGID would allow QEMU > > > > to create setgid files, thereby potentially allowing an attacker to gain > > > > any GID. > > > > > > Sure, its not recommended to run QEMU as root, but we don't block that > > > either and I do regularly test with qemu running as root. > > > > > > > > > > > I think it's better not to implement QEMU FSETID functionality at all > > > > and to handle it another way. > > > > > > One way could be that virtiofsd tries to clear setuid bit after I/O > > > has finished. But that will be non-atomic operation and it is filled > > > with perils as it requires virtiofsd to know what all kernel will > > > do if this write has been done with CAP_FSETID dropped. > > > > > > > In the worst case I/O requests should just > > > > fail, it seems like a rare case anyway: > > > > > > Is there a way for virtiofsd to know if qemu is running with CAP_FSETID > > > or not. If there is one, it might be reasonable to error out. If we > > > don't know, then we can't fail all the operations. > > > > > > > I/O to a setuid/setgid file with > > > > a memory buffer that is not mapped in virtiofsd. > > > > > > With DAX it is easily triggerable. User has to append to a setuid file > > > in virtiofs and this path will trigger. > > > > > > I am fine with not supporting this patch but will also need a reaosonable > > > alternative solution. > > > > One way to avoid this problem is by introducing DMA read/write functions > > into the vhost-user protocol that can be used by all device types, not > > just virtio-fs. > > > > Today virtio-fs uses the IO slave request when it cannot access a region > > of guest memory. It sends the file descriptor to QEMU and QEMU performs > > the pread(2)/pwrite(2) on behalf of virtiofsd. > > > > I mentioned in the past that this solution is over-specialized. It > > doesn't solve the larger problem that vhost-user processes do not have > > full access to the guest memory space (e.g. DAX window). > > > > Instead of sending file I/O requests over to QEMU, the vhost-user > > protocol should offer DMA read/write requests so any vhost-user process > > can access the guest memory space where vhost's shared memory mechanism > > is insufficient. > > > > Here is how it would work: > > > > 1. Drop the IO slave request, replace it with DMA read/write slave > > requests. > > > > Note that these new requests can also be used in environments where > > maximum vIOMMU isolation is needed for security reasons and sharing > > all of guest RAM with the vhost-user process is considered > > unacceptable. > > > > 2. When virtqueue buffer mapping fails, send DMA read/write slave > > requests to transfer the data from/to QEMU. virtiofsd calls > > pread(2)/pwrite(2) itself with virtiofsd's Linux capabilities. > > Can you elaborate a bit more how will this new DMA read/write vhost-user > commands can be implemented. I am assuming its not a real DMA and just > sort of emulation of DMA. Effectively we have two processes and one > process needs to read/write to/from address space of other process. > > We were also wondering if we can make use of process_vm_readv() > and process_vm_write() syscalls to achieve this. But this atleast > requires virtiofsd to be more priviliged than qemu and also virtiofsd > needs to know where DAX mapping window is. We briefly discussed this here. > > https://lore.kernel.org/qemu-devel/20210421200746.GH1579961@redhat.com/ I wasn't thinking of directly allowing QEMU virtual memory access via process_vm_readv/writev(). That would be more efficient but requires privileges and also exposes internals of QEMU's virtual memory layout and vIOMMU translation to the vhost-user process. Instead I was thinking about VHOST_USER_DMA_READ/WRITE messages containing the address (a device IOVA, it could just be a guest physical memory address in most cases) and the length. The WRITE message would also contain the data that the vhost-user device wishes to write. The READ message reply would contain the data that the device read from QEMU. QEMU would implement this using QEMU's address_space_read/write() APIs. So basically just a new vhost-user protocol message to do a memcpy(), but with guest addresses and vIOMMU support :). The vhost-user device will need to do bounce buffering so using these new messages is slower than zero-copy I/O to shared guest RAM. Stefan