From mboxrd@z Thu Jan 1 00:00:00 1970 From: "Michael S. Tsirkin" Subject: Re: [PATCH 1/2] Revert "vhost: access vq metadata through kernel virtual address" Date: Fri, 6 Sep 2019 09:46:44 -0400 Message-ID: <20190906094332-mutt-send-email-mst__39893.9769993135$1567777628$gmane$org@kernel.org> References: <20190905122736.19768-1-jasowang@redhat.com> <20190905122736.19768-2-jasowang@redhat.com> Mime-Version: 1.0 Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit Return-path: Content-Disposition: inline In-Reply-To: <20190905122736.19768-2-jasowang@redhat.com> List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Sender: virtualization-bounces@lists.linux-foundation.org Errors-To: virtualization-bounces@lists.linux-foundation.org To: Jason Wang Cc: aarcange@redhat.com, kvm@vger.kernel.org, netdev@vger.kernel.org, linux-kernel@vger.kernel.org, virtualization@lists.linux-foundation.org, linux-mm@kvack.org, jglisse@redhat.com, jgg@mellanox.com List-Id: virtualization@lists.linuxfoundation.org On Thu, Sep 05, 2019 at 08:27:35PM +0800, Jason Wang wrote: > It was reported that metadata acceleration introduces several issues, > so this patch reverts commit ff466032dc9e5a61217f22ea34b2df932786bbfc, > 73f628ec9e6bcc45b77c53fe6d0c0ec55eaf82af and > 0b4a7092ffe568a55bf8f3cefdf79ff666586d91. > > We will rework it on the next version. > > Cc: Jason Gunthorpe > Signed-off-by: Jason Wang I am confused by the above. What I see upstream is 7f466032dc. commit 7f466032dc9e5a61217f22ea34b2df932786bbfc Author: Jason Wang Date: Fri May 24 04:12:18 2019 -0400 vhost: access vq metadata through kernel virtual address so this is what I reverted. Pls take a look, and let me know if you see issues. Thanks! > --- > drivers/vhost/vhost.c | 515 +----------------------------------------- > drivers/vhost/vhost.h | 41 ---- > 2 files changed, 3 insertions(+), 553 deletions(-) > > diff --git a/drivers/vhost/vhost.c b/drivers/vhost/vhost.c > index 0536f8526359..791562e03fe0 100644 > --- a/drivers/vhost/vhost.c > +++ b/drivers/vhost/vhost.c > @@ -298,160 +298,6 @@ static void vhost_vq_meta_reset(struct vhost_dev *d) > __vhost_vq_meta_reset(d->vqs[i]); > } > > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > -static void vhost_map_unprefetch(struct vhost_map *map) > -{ > - kfree(map->pages); > - map->pages = NULL; > - map->npages = 0; > - map->addr = NULL; > -} > - > -static void vhost_uninit_vq_maps(struct vhost_virtqueue *vq) > -{ > - struct vhost_map *map[VHOST_NUM_ADDRS]; > - int i; > - > - spin_lock(&vq->mmu_lock); > - for (i = 0; i < VHOST_NUM_ADDRS; i++) { > - map[i] = rcu_dereference_protected(vq->maps[i], > - lockdep_is_held(&vq->mmu_lock)); > - if (map[i]) > - rcu_assign_pointer(vq->maps[i], NULL); > - } > - spin_unlock(&vq->mmu_lock); > - > - synchronize_rcu(); > - > - for (i = 0; i < VHOST_NUM_ADDRS; i++) > - if (map[i]) > - vhost_map_unprefetch(map[i]); > - > -} > - > -static void vhost_reset_vq_maps(struct vhost_virtqueue *vq) > -{ > - int i; > - > - vhost_uninit_vq_maps(vq); > - for (i = 0; i < VHOST_NUM_ADDRS; i++) > - vq->uaddrs[i].size = 0; > -} > - > -static bool vhost_map_range_overlap(struct vhost_uaddr *uaddr, > - unsigned long start, > - unsigned long end) > -{ > - if (unlikely(!uaddr->size)) > - return false; > - > - return !(end < uaddr->uaddr || start > uaddr->uaddr - 1 + uaddr->size); > -} > - > -static void vhost_invalidate_vq_start(struct vhost_virtqueue *vq, > - int index, > - unsigned long start, > - unsigned long end) > -{ > - struct vhost_uaddr *uaddr = &vq->uaddrs[index]; > - struct vhost_map *map; > - int i; > - > - if (!vhost_map_range_overlap(uaddr, start, end)) > - return; > - > - spin_lock(&vq->mmu_lock); > - ++vq->invalidate_count; > - > - map = rcu_dereference_protected(vq->maps[index], > - lockdep_is_held(&vq->mmu_lock)); > - if (map) { > - if (uaddr->write) { > - for (i = 0; i < map->npages; i++) > - set_page_dirty(map->pages[i]); > - } > - rcu_assign_pointer(vq->maps[index], NULL); > - } > - spin_unlock(&vq->mmu_lock); > - > - if (map) { > - synchronize_rcu(); > - vhost_map_unprefetch(map); > - } > -} > - > -static void vhost_invalidate_vq_end(struct vhost_virtqueue *vq, > - int index, > - unsigned long start, > - unsigned long end) > -{ > - if (!vhost_map_range_overlap(&vq->uaddrs[index], start, end)) > - return; > - > - spin_lock(&vq->mmu_lock); > - --vq->invalidate_count; > - spin_unlock(&vq->mmu_lock); > -} > - > -static int vhost_invalidate_range_start(struct mmu_notifier *mn, > - const struct mmu_notifier_range *range) > -{ > - struct vhost_dev *dev = container_of(mn, struct vhost_dev, > - mmu_notifier); > - int i, j; > - > - if (!mmu_notifier_range_blockable(range)) > - return -EAGAIN; > - > - for (i = 0; i < dev->nvqs; i++) { > - struct vhost_virtqueue *vq = dev->vqs[i]; > - > - for (j = 0; j < VHOST_NUM_ADDRS; j++) > - vhost_invalidate_vq_start(vq, j, > - range->start, > - range->end); > - } > - > - return 0; > -} > - > -static void vhost_invalidate_range_end(struct mmu_notifier *mn, > - const struct mmu_notifier_range *range) > -{ > - struct vhost_dev *dev = container_of(mn, struct vhost_dev, > - mmu_notifier); > - int i, j; > - > - for (i = 0; i < dev->nvqs; i++) { > - struct vhost_virtqueue *vq = dev->vqs[i]; > - > - for (j = 0; j < VHOST_NUM_ADDRS; j++) > - vhost_invalidate_vq_end(vq, j, > - range->start, > - range->end); > - } > -} > - > -static const struct mmu_notifier_ops vhost_mmu_notifier_ops = { > - .invalidate_range_start = vhost_invalidate_range_start, > - .invalidate_range_end = vhost_invalidate_range_end, > -}; > - > -static void vhost_init_maps(struct vhost_dev *dev) > -{ > - struct vhost_virtqueue *vq; > - int i, j; > - > - dev->mmu_notifier.ops = &vhost_mmu_notifier_ops; > - > - for (i = 0; i < dev->nvqs; ++i) { > - vq = dev->vqs[i]; > - for (j = 0; j < VHOST_NUM_ADDRS; j++) > - RCU_INIT_POINTER(vq->maps[j], NULL); > - } > -} > -#endif > - > static void vhost_vq_reset(struct vhost_dev *dev, > struct vhost_virtqueue *vq) > { > @@ -480,11 +326,7 @@ static void vhost_vq_reset(struct vhost_dev *dev, > vq->busyloop_timeout = 0; > vq->umem = NULL; > vq->iotlb = NULL; > - vq->invalidate_count = 0; > __vhost_vq_meta_reset(vq); > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - vhost_reset_vq_maps(vq); > -#endif > } > > static int vhost_worker(void *data) > @@ -634,9 +476,7 @@ void vhost_dev_init(struct vhost_dev *dev, > INIT_LIST_HEAD(&dev->read_list); > INIT_LIST_HEAD(&dev->pending_list); > spin_lock_init(&dev->iotlb_lock); > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - vhost_init_maps(dev); > -#endif > + > > for (i = 0; i < dev->nvqs; ++i) { > vq = dev->vqs[i]; > @@ -645,7 +485,6 @@ void vhost_dev_init(struct vhost_dev *dev, > vq->heads = NULL; > vq->dev = dev; > mutex_init(&vq->mutex); > - spin_lock_init(&vq->mmu_lock); > vhost_vq_reset(dev, vq); > if (vq->handle_kick) > vhost_poll_init(&vq->poll, vq->handle_kick, > @@ -725,18 +564,7 @@ long vhost_dev_set_owner(struct vhost_dev *dev) > if (err) > goto err_cgroup; > > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - err = mmu_notifier_register(&dev->mmu_notifier, dev->mm); > - if (err) > - goto err_mmu_notifier; > -#endif > - > return 0; > - > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > -err_mmu_notifier: > - vhost_dev_free_iovecs(dev); > -#endif > err_cgroup: > kthread_stop(worker); > dev->worker = NULL; > @@ -827,107 +655,6 @@ static void vhost_clear_msg(struct vhost_dev *dev) > spin_unlock(&dev->iotlb_lock); > } > > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > -static void vhost_setup_uaddr(struct vhost_virtqueue *vq, > - int index, unsigned long uaddr, > - size_t size, bool write) > -{ > - struct vhost_uaddr *addr = &vq->uaddrs[index]; > - > - addr->uaddr = uaddr; > - addr->size = size; > - addr->write = write; > -} > - > -static void vhost_setup_vq_uaddr(struct vhost_virtqueue *vq) > -{ > - vhost_setup_uaddr(vq, VHOST_ADDR_DESC, > - (unsigned long)vq->desc, > - vhost_get_desc_size(vq, vq->num), > - false); > - vhost_setup_uaddr(vq, VHOST_ADDR_AVAIL, > - (unsigned long)vq->avail, > - vhost_get_avail_size(vq, vq->num), > - false); > - vhost_setup_uaddr(vq, VHOST_ADDR_USED, > - (unsigned long)vq->used, > - vhost_get_used_size(vq, vq->num), > - true); > -} > - > -static int vhost_map_prefetch(struct vhost_virtqueue *vq, > - int index) > -{ > - struct vhost_map *map; > - struct vhost_uaddr *uaddr = &vq->uaddrs[index]; > - struct page **pages; > - int npages = DIV_ROUND_UP(uaddr->size, PAGE_SIZE); > - int npinned; > - void *vaddr, *v; > - int err; > - int i; > - > - spin_lock(&vq->mmu_lock); > - > - err = -EFAULT; > - if (vq->invalidate_count) > - goto err; > - > - err = -ENOMEM; > - map = kmalloc(sizeof(*map), GFP_ATOMIC); > - if (!map) > - goto err; > - > - pages = kmalloc_array(npages, sizeof(struct page *), GFP_ATOMIC); > - if (!pages) > - goto err_pages; > - > - err = EFAULT; > - npinned = __get_user_pages_fast(uaddr->uaddr, npages, > - uaddr->write, pages); > - if (npinned > 0) > - release_pages(pages, npinned); > - if (npinned != npages) > - goto err_gup; > - > - for (i = 0; i < npinned; i++) > - if (PageHighMem(pages[i])) > - goto err_gup; > - > - vaddr = v = page_address(pages[0]); > - > - /* For simplicity, fallback to userspace address if VA is not > - * contigious. > - */ > - for (i = 1; i < npinned; i++) { > - v += PAGE_SIZE; > - if (v != page_address(pages[i])) > - goto err_gup; > - } > - > - map->addr = vaddr + (uaddr->uaddr & (PAGE_SIZE - 1)); > - map->npages = npages; > - map->pages = pages; > - > - rcu_assign_pointer(vq->maps[index], map); > - /* No need for a synchronize_rcu(). This function should be > - * called by dev->worker so we are serialized with all > - * readers. > - */ > - spin_unlock(&vq->mmu_lock); > - > - return 0; > - > -err_gup: > - kfree(pages); > -err_pages: > - kfree(map); > -err: > - spin_unlock(&vq->mmu_lock); > - return err; > -} > -#endif > - > void vhost_dev_cleanup(struct vhost_dev *dev) > { > int i; > @@ -957,16 +684,8 @@ void vhost_dev_cleanup(struct vhost_dev *dev) > kthread_stop(dev->worker); > dev->worker = NULL; > } > - if (dev->mm) { > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - mmu_notifier_unregister(&dev->mmu_notifier, dev->mm); > -#endif > + if (dev->mm) > mmput(dev->mm); > - } > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - for (i = 0; i < dev->nvqs; i++) > - vhost_uninit_vq_maps(dev->vqs[i]); > -#endif > dev->mm = NULL; > } > EXPORT_SYMBOL_GPL(vhost_dev_cleanup); > @@ -1195,26 +914,6 @@ static inline void __user *__vhost_get_user(struct vhost_virtqueue *vq, > > static inline int vhost_put_avail_event(struct vhost_virtqueue *vq) > { > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - struct vhost_map *map; > - struct vring_used *used; > - > - if (!vq->iotlb) { > - rcu_read_lock(); > - > - map = rcu_dereference(vq->maps[VHOST_ADDR_USED]); > - if (likely(map)) { > - used = map->addr; > - *((__virtio16 *)&used->ring[vq->num]) = > - cpu_to_vhost16(vq, vq->avail_idx); > - rcu_read_unlock(); > - return 0; > - } > - > - rcu_read_unlock(); > - } > -#endif > - > return vhost_put_user(vq, cpu_to_vhost16(vq, vq->avail_idx), > vhost_avail_event(vq)); > } > @@ -1223,27 +922,6 @@ static inline int vhost_put_used(struct vhost_virtqueue *vq, > struct vring_used_elem *head, int idx, > int count) > { > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - struct vhost_map *map; > - struct vring_used *used; > - size_t size; > - > - if (!vq->iotlb) { > - rcu_read_lock(); > - > - map = rcu_dereference(vq->maps[VHOST_ADDR_USED]); > - if (likely(map)) { > - used = map->addr; > - size = count * sizeof(*head); > - memcpy(used->ring + idx, head, size); > - rcu_read_unlock(); > - return 0; > - } > - > - rcu_read_unlock(); > - } > -#endif > - > return vhost_copy_to_user(vq, vq->used->ring + idx, head, > count * sizeof(*head)); > } > @@ -1251,25 +929,6 @@ static inline int vhost_put_used(struct vhost_virtqueue *vq, > static inline int vhost_put_used_flags(struct vhost_virtqueue *vq) > > { > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - struct vhost_map *map; > - struct vring_used *used; > - > - if (!vq->iotlb) { > - rcu_read_lock(); > - > - map = rcu_dereference(vq->maps[VHOST_ADDR_USED]); > - if (likely(map)) { > - used = map->addr; > - used->flags = cpu_to_vhost16(vq, vq->used_flags); > - rcu_read_unlock(); > - return 0; > - } > - > - rcu_read_unlock(); > - } > -#endif > - > return vhost_put_user(vq, cpu_to_vhost16(vq, vq->used_flags), > &vq->used->flags); > } > @@ -1277,25 +936,6 @@ static inline int vhost_put_used_flags(struct vhost_virtqueue *vq) > static inline int vhost_put_used_idx(struct vhost_virtqueue *vq) > > { > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - struct vhost_map *map; > - struct vring_used *used; > - > - if (!vq->iotlb) { > - rcu_read_lock(); > - > - map = rcu_dereference(vq->maps[VHOST_ADDR_USED]); > - if (likely(map)) { > - used = map->addr; > - used->idx = cpu_to_vhost16(vq, vq->last_used_idx); > - rcu_read_unlock(); > - return 0; > - } > - > - rcu_read_unlock(); > - } > -#endif > - > return vhost_put_user(vq, cpu_to_vhost16(vq, vq->last_used_idx), > &vq->used->idx); > } > @@ -1341,50 +981,12 @@ static void vhost_dev_unlock_vqs(struct vhost_dev *d) > static inline int vhost_get_avail_idx(struct vhost_virtqueue *vq, > __virtio16 *idx) > { > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - struct vhost_map *map; > - struct vring_avail *avail; > - > - if (!vq->iotlb) { > - rcu_read_lock(); > - > - map = rcu_dereference(vq->maps[VHOST_ADDR_AVAIL]); > - if (likely(map)) { > - avail = map->addr; > - *idx = avail->idx; > - rcu_read_unlock(); > - return 0; > - } > - > - rcu_read_unlock(); > - } > -#endif > - > return vhost_get_avail(vq, *idx, &vq->avail->idx); > } > > static inline int vhost_get_avail_head(struct vhost_virtqueue *vq, > __virtio16 *head, int idx) > { > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - struct vhost_map *map; > - struct vring_avail *avail; > - > - if (!vq->iotlb) { > - rcu_read_lock(); > - > - map = rcu_dereference(vq->maps[VHOST_ADDR_AVAIL]); > - if (likely(map)) { > - avail = map->addr; > - *head = avail->ring[idx & (vq->num - 1)]; > - rcu_read_unlock(); > - return 0; > - } > - > - rcu_read_unlock(); > - } > -#endif > - > return vhost_get_avail(vq, *head, > &vq->avail->ring[idx & (vq->num - 1)]); > } > @@ -1392,98 +994,24 @@ static inline int vhost_get_avail_head(struct vhost_virtqueue *vq, > static inline int vhost_get_avail_flags(struct vhost_virtqueue *vq, > __virtio16 *flags) > { > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - struct vhost_map *map; > - struct vring_avail *avail; > - > - if (!vq->iotlb) { > - rcu_read_lock(); > - > - map = rcu_dereference(vq->maps[VHOST_ADDR_AVAIL]); > - if (likely(map)) { > - avail = map->addr; > - *flags = avail->flags; > - rcu_read_unlock(); > - return 0; > - } > - > - rcu_read_unlock(); > - } > -#endif > - > return vhost_get_avail(vq, *flags, &vq->avail->flags); > } > > static inline int vhost_get_used_event(struct vhost_virtqueue *vq, > __virtio16 *event) > { > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - struct vhost_map *map; > - struct vring_avail *avail; > - > - if (!vq->iotlb) { > - rcu_read_lock(); > - map = rcu_dereference(vq->maps[VHOST_ADDR_AVAIL]); > - if (likely(map)) { > - avail = map->addr; > - *event = (__virtio16)avail->ring[vq->num]; > - rcu_read_unlock(); > - return 0; > - } > - rcu_read_unlock(); > - } > -#endif > - > return vhost_get_avail(vq, *event, vhost_used_event(vq)); > } > > static inline int vhost_get_used_idx(struct vhost_virtqueue *vq, > __virtio16 *idx) > { > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - struct vhost_map *map; > - struct vring_used *used; > - > - if (!vq->iotlb) { > - rcu_read_lock(); > - > - map = rcu_dereference(vq->maps[VHOST_ADDR_USED]); > - if (likely(map)) { > - used = map->addr; > - *idx = used->idx; > - rcu_read_unlock(); > - return 0; > - } > - > - rcu_read_unlock(); > - } > -#endif > - > return vhost_get_used(vq, *idx, &vq->used->idx); > } > > static inline int vhost_get_desc(struct vhost_virtqueue *vq, > struct vring_desc *desc, int idx) > { > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - struct vhost_map *map; > - struct vring_desc *d; > - > - if (!vq->iotlb) { > - rcu_read_lock(); > - > - map = rcu_dereference(vq->maps[VHOST_ADDR_DESC]); > - if (likely(map)) { > - d = map->addr; > - *desc = *(d + idx); > - rcu_read_unlock(); > - return 0; > - } > - > - rcu_read_unlock(); > - } > -#endif > - > return vhost_copy_from_user(vq, desc, vq->desc + idx, sizeof(*desc)); > } > > @@ -1824,32 +1352,12 @@ static bool iotlb_access_ok(struct vhost_virtqueue *vq, > return true; > } > > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > -static void vhost_vq_map_prefetch(struct vhost_virtqueue *vq) > -{ > - struct vhost_map __rcu *map; > - int i; > - > - for (i = 0; i < VHOST_NUM_ADDRS; i++) { > - rcu_read_lock(); > - map = rcu_dereference(vq->maps[i]); > - rcu_read_unlock(); > - if (unlikely(!map)) > - vhost_map_prefetch(vq, i); > - } > -} > -#endif > - > int vq_meta_prefetch(struct vhost_virtqueue *vq) > { > unsigned int num = vq->num; > > - if (!vq->iotlb) { > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - vhost_vq_map_prefetch(vq); > -#endif > + if (!vq->iotlb) > return 1; > - } > > return iotlb_access_ok(vq, VHOST_ACCESS_RO, (u64)(uintptr_t)vq->desc, > vhost_get_desc_size(vq, num), VHOST_ADDR_DESC) && > @@ -2060,16 +1568,6 @@ static long vhost_vring_set_num_addr(struct vhost_dev *d, > > mutex_lock(&vq->mutex); > > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - /* Unregister MMU notifer to allow invalidation callback > - * can access vq->uaddrs[] without holding a lock. > - */ > - if (d->mm) > - mmu_notifier_unregister(&d->mmu_notifier, d->mm); > - > - vhost_uninit_vq_maps(vq); > -#endif > - > switch (ioctl) { > case VHOST_SET_VRING_NUM: > r = vhost_vring_set_num(d, vq, argp); > @@ -2081,13 +1579,6 @@ static long vhost_vring_set_num_addr(struct vhost_dev *d, > BUG(); > } > > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - vhost_setup_vq_uaddr(vq); > - > - if (d->mm) > - mmu_notifier_register(&d->mmu_notifier, d->mm); > -#endif > - > mutex_unlock(&vq->mutex); > > return r; > diff --git a/drivers/vhost/vhost.h b/drivers/vhost/vhost.h > index 42a8c2a13ab1..e9ed2722b633 100644 > --- a/drivers/vhost/vhost.h > +++ b/drivers/vhost/vhost.h > @@ -12,9 +12,6 @@ > #include > #include > #include > -#include > -#include > -#include > > struct vhost_work; > typedef void (*vhost_work_fn_t)(struct vhost_work *work); > @@ -83,24 +80,6 @@ enum vhost_uaddr_type { > VHOST_NUM_ADDRS = 3, > }; > > -struct vhost_map { > - int npages; > - void *addr; > - struct page **pages; > -}; > - > -struct vhost_uaddr { > - unsigned long uaddr; > - size_t size; > - bool write; > -}; > - > -#if defined(CONFIG_MMU_NOTIFIER) && ARCH_IMPLEMENTS_FLUSH_DCACHE_PAGE == 0 > -#define VHOST_ARCH_CAN_ACCEL_UACCESS 0 > -#else > -#define VHOST_ARCH_CAN_ACCEL_UACCESS 0 > -#endif > - > /* The virtqueue structure describes a queue attached to a device. */ > struct vhost_virtqueue { > struct vhost_dev *dev; > @@ -111,22 +90,7 @@ struct vhost_virtqueue { > struct vring_desc __user *desc; > struct vring_avail __user *avail; > struct vring_used __user *used; > - > -#if VHOST_ARCH_CAN_ACCEL_UACCESS > - /* Read by memory accessors, modified by meta data > - * prefetching, MMU notifier and vring ioctl(). > - * Synchonrized through mmu_lock (writers) and RCU (writers > - * and readers). > - */ > - struct vhost_map __rcu *maps[VHOST_NUM_ADDRS]; > - /* Read by MMU notifier, modified by vring ioctl(), > - * synchronized through MMU notifier > - * registering/unregistering. > - */ > - struct vhost_uaddr uaddrs[VHOST_NUM_ADDRS]; > -#endif > const struct vhost_umem_node *meta_iotlb[VHOST_NUM_ADDRS]; > - > struct file *kick; > struct eventfd_ctx *call_ctx; > struct eventfd_ctx *error_ctx; > @@ -181,8 +145,6 @@ struct vhost_virtqueue { > bool user_be; > #endif > u32 busyloop_timeout; > - spinlock_t mmu_lock; > - int invalidate_count; > }; > > struct vhost_msg_node { > @@ -196,9 +158,6 @@ struct vhost_msg_node { > > struct vhost_dev { > struct mm_struct *mm; > -#ifdef CONFIG_MMU_NOTIFIER > - struct mmu_notifier mmu_notifier; > -#endif > struct mutex mutex; > struct vhost_virtqueue **vqs; > int nvqs; > -- > 2.19.1