All of lore.kernel.org
 help / color / mirror / Atom feed
From: Jason Wang <jasowang@redhat.com>
To: Xuan Zhuo <xuanzhuo@linux.alibaba.com>
Cc: Jeff Dike <jdike@addtoit.com>,
	Richard Weinberger <richard@nod.at>,
	Anton Ivanov <anton.ivanov@cambridgegreys.com>,
	"Michael S. Tsirkin" <mst@redhat.com>,
	"David S. Miller" <davem@davemloft.net>,
	Jakub Kicinski <kuba@kernel.org>,
	Hans de Goede <hdegoede@redhat.com>,
	Mark Gross <markgross@kernel.org>,
	Vadim Pasternak <vadimp@nvidia.com>,
	Bjorn Andersson <bjorn.andersson@linaro.org>,
	Mathieu Poirier <mathieu.poirier@linaro.org>,
	Cornelia Huck <cohuck@redhat.com>,
	Halil Pasic <pasic@linux.ibm.com>,
	Heiko Carstens <hca@linux.ibm.com>,
	Vasily Gorbik <gor@linux.ibm.com>,
	Christian Borntraeger <borntraeger@linux.ibm.com>,
	Alexander Gordeev <agordeev@linux.ibm.com>,
	Sven Schnelle <svens@linux.ibm.com>,
	Alexei Starovoitov <ast@kernel.org>,
	Daniel Borkmann <daniel@iogearbox.net>,
	Jesper Dangaard Brouer <hawk@kernel.org>,
	John Fastabend <john.fastabend@gmail.com>,
	Johannes Berg <johannes.berg@intel.com>,
	Vincent Whitchurch <vincent.whitchurch@axis.com>,
	linux-um@lists.infradead.org, netdev <netdev@vger.kernel.org>,
	platform-driver-x86@vger.kernel.org,
	linux-remoteproc@vger.kernel.org, linux-s390@vger.kernel.org,
	kvm <kvm@vger.kernel.org>,
	"open list:XDP (eXpress Data Path)" <bpf@vger.kernel.org>,
	virtualization <virtualization@lists.linux-foundation.org>
Subject: Re: [PATCH v9 31/32] virtio_net: support rx/tx queue resize
Date: Thu, 14 Apr 2022 17:30:02 +0800	[thread overview]
Message-ID: <CACGkMEvPH1k76xB_cHq_S9hvMXgGruoXpKLfoMZvJZ-L7wM9iw@mail.gmail.com> (raw)
In-Reply-To: <1649838917.6726515-10-xuanzhuo@linux.alibaba.com>

On Wed, Apr 13, 2022 at 4:47 PM Xuan Zhuo <xuanzhuo@linux.alibaba.com> wrote:
>
> On Wed, 13 Apr 2022 16:00:18 +0800, Jason Wang <jasowang@redhat.com> wrote:
> >
> > 在 2022/4/6 上午11:43, Xuan Zhuo 写道:
> > > This patch implements the resize function of the rx, tx queues.
> > > Based on this function, it is possible to modify the ring num of the
> > > queue.
> > >
> > > There may be an exception during the resize process, the resize may
> > > fail, or the vq can no longer be used. Either way, we must execute
> > > napi_enable(). Because napi_disable is similar to a lock, napi_enable
> > > must be called after calling napi_disable.
> > >
> > > Signed-off-by: Xuan Zhuo <xuanzhuo@linux.alibaba.com>
> > > ---
> > >   drivers/net/virtio_net.c | 81 ++++++++++++++++++++++++++++++++++++++++
> > >   1 file changed, 81 insertions(+)
> > >
> > > diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c
> > > index b8bf00525177..ba6859f305f7 100644
> > > --- a/drivers/net/virtio_net.c
> > > +++ b/drivers/net/virtio_net.c
> > > @@ -251,6 +251,9 @@ struct padded_vnet_hdr {
> > >     char padding[4];
> > >   };
> > >
> > > +static void virtnet_sq_free_unused_buf(struct virtqueue *vq, void *buf);
> > > +static void virtnet_rq_free_unused_buf(struct virtqueue *vq, void *buf);
> > > +
> > >   static bool is_xdp_frame(void *ptr)
> > >   {
> > >     return (unsigned long)ptr & VIRTIO_XDP_FLAG;
> > > @@ -1369,6 +1372,15 @@ static void virtnet_napi_enable(struct virtqueue *vq, struct napi_struct *napi)
> > >   {
> > >     napi_enable(napi);
> > >
> > > +   /* Check if vq is in reset state. The normal reset/resize process will
> > > +    * be protected by napi. However, the protection of napi is only enabled
> > > +    * during the operation, and the protection of napi will end after the
> > > +    * operation is completed. If re-enable fails during the process, vq
> > > +    * will remain unavailable with reset state.
> > > +    */
> > > +   if (vq->reset)
> > > +           return;
> >
> >
> > I don't get when could we hit this condition.
>
>
> In patch 23, the code to implement re-enable vq is as follows:
>
> +static int vp_modern_enable_reset_vq(struct virtqueue *vq)
> +{
> +       struct virtio_pci_device *vp_dev = to_vp_device(vq->vdev);
> +       struct virtio_pci_modern_device *mdev = &vp_dev->mdev;
> +       struct virtio_pci_vq_info *info;
> +       unsigned long flags, index;
> +       int err;
> +
> +       if (!vq->reset)
> +               return -EBUSY;
> +
> +       index = vq->index;
> +       info = vp_dev->vqs[index];
> +
> +       /* check queue reset status */
> +       if (vp_modern_get_queue_reset(mdev, index) != 1)
> +               return -EBUSY;
> +
> +       err = vp_active_vq(vq, info->msix_vector);
> +       if (err)
> +               return err;
> +
> +       if (vq->callback) {
> +               spin_lock_irqsave(&vp_dev->lock, flags);
> +               list_add(&info->node, &vp_dev->virtqueues);
> +               spin_unlock_irqrestore(&vp_dev->lock, flags);
> +       } else {
> +               INIT_LIST_HEAD(&info->node);
> +       }
> +
> +       vp_modern_set_queue_enable(&vp_dev->mdev, index, true);
> +
> +       if (vp_dev->per_vq_vectors && info->msix_vector != VIRTIO_MSI_NO_VECTOR)
> +               enable_irq(pci_irq_vector(vp_dev->pci_dev, info->msix_vector));
> +
> +       vq->reset = false;
> +
> +       return 0;
> +}
>
>
> There are three situations where an error will be returned. These are the
> situations I want to handle.

Right, but it looks harmless if we just schedule the NAPI without the check.

>
> But I'm rethinking the question, and I feel like you're right, although the
> hardware setup may fail. We can no longer sync with the hardware. But using it
> as a normal vq doesn't have any problems.

Note that we should make sure the buggy(malicous) device won't crash
the codes by changing the queue_reset value at its will.

>
> >
> >
> > > +
> > >     /* If all buffers were filled by other side before we napi_enabled, we
> > >      * won't get another interrupt, so process any outstanding packets now.
> > >      * Call local_bh_enable after to trigger softIRQ processing.
> > > @@ -1413,6 +1425,15 @@ static void refill_work(struct work_struct *work)
> > >             struct receive_queue *rq = &vi->rq[i];
> > >
> > >             napi_disable(&rq->napi);
> > > +
> > > +           /* Check if vq is in reset state. See more in
> > > +            * virtnet_napi_enable()
> > > +            */
> > > +           if (rq->vq->reset) {
> > > +                   virtnet_napi_enable(rq->vq, &rq->napi);
> > > +                   continue;
> > > +           }
> >
> >
> > Can we do something similar in virtnet_close() by canceling the work?
>
> I think there is no need to cancel the work here, because napi_disable will wait
> for the napi_enable of the resize. So if the re-enable failed vq is used as a normal
> vq, this logic can be removed.

Actually I meant the part of virtnet_rx_resize().

If we don't synchronize with the refill work, it might enable NAPI unexpectedly?

Thanks

>
>
> >
> >
> > > +
> > >             still_empty = !try_fill_recv(vi, rq, GFP_KERNEL);
> > >             virtnet_napi_enable(rq->vq, &rq->napi);
> > >
> > > @@ -1523,6 +1544,10 @@ static void virtnet_poll_cleantx(struct receive_queue *rq)
> > >     if (!sq->napi.weight || is_xdp_raw_buffer_queue(vi, index))
> > >             return;
> > >
> > > +   /* Check if vq is in reset state. See more in virtnet_napi_enable() */
> > > +   if (sq->vq->reset)
> > > +           return;
> >
> >
> > We've disabled TX napi, any chance we can still hit this?
>
> Same as above.
>
> >
> >
> > > +
> > >     if (__netif_tx_trylock(txq)) {
> > >             do {
> > >                     virtqueue_disable_cb(sq->vq);
> > > @@ -1769,6 +1794,62 @@ static netdev_tx_t start_xmit(struct sk_buff *skb, struct net_device *dev)
> > >     return NETDEV_TX_OK;
> > >   }
> > >
> > > +static int virtnet_rx_resize(struct virtnet_info *vi,
> > > +                        struct receive_queue *rq, u32 ring_num)
> > > +{
> > > +   int err;
> > > +
> > > +   napi_disable(&rq->napi);
> > > +
> > > +   err = virtqueue_resize(rq->vq, ring_num, virtnet_rq_free_unused_buf);
> > > +   if (err)
> > > +           goto err;
> > > +
> > > +   if (!try_fill_recv(vi, rq, GFP_KERNEL))
> > > +           schedule_delayed_work(&vi->refill, 0);
> > > +
> > > +   virtnet_napi_enable(rq->vq, &rq->napi);
> > > +   return 0;
> > > +
> > > +err:
> > > +   netdev_err(vi->dev,
> > > +              "reset rx reset vq fail: rx queue index: %td err: %d\n",
> > > +              rq - vi->rq, err);
> > > +   virtnet_napi_enable(rq->vq, &rq->napi);
> > > +   return err;
> > > +}
> > > +
> > > +static int virtnet_tx_resize(struct virtnet_info *vi,
> > > +                        struct send_queue *sq, u32 ring_num)
> > > +{
> > > +   struct netdev_queue *txq;
> > > +   int err, qindex;
> > > +
> > > +   qindex = sq - vi->sq;
> > > +
> > > +   virtnet_napi_tx_disable(&sq->napi);
> > > +
> > > +   txq = netdev_get_tx_queue(vi->dev, qindex);
> > > +   __netif_tx_lock_bh(txq);
> > > +   netif_stop_subqueue(vi->dev, qindex);
> > > +   __netif_tx_unlock_bh(txq);
> > > +
> > > +   err = virtqueue_resize(sq->vq, ring_num, virtnet_sq_free_unused_buf);
> > > +   if (err)
> > > +           goto err;
> > > +
> > > +   netif_start_subqueue(vi->dev, qindex);
> > > +   virtnet_napi_tx_enable(vi, sq->vq, &sq->napi);
> > > +   return 0;
> > > +
> > > +err:
> >
> >
> > I guess we can still start the queue in this case? (Since we don't
> > change the queue if resize fails).
>
> Yes, you are right.
>
> Thanks.
>
> >
> >
> > > +   netdev_err(vi->dev,
> > > +              "reset tx reset vq fail: tx queue index: %td err: %d\n",
> > > +              sq - vi->sq, err);
> > > +   virtnet_napi_tx_enable(vi, sq->vq, &sq->napi);
> > > +   return err;
> > > +}
> > > +
> > >   /*
> > >    * Send command via the control virtqueue and check status.  Commands
> > >    * supported by the hypervisor, as indicated by feature bits, should
> >
>


WARNING: multiple messages have this Message-ID (diff)
From: Jason Wang <jasowang@redhat.com>
To: Xuan Zhuo <xuanzhuo@linux.alibaba.com>
Cc: Vadim Pasternak <vadimp@nvidia.com>,
	"Michael S. Tsirkin" <mst@redhat.com>,
	linux-remoteproc@vger.kernel.org,
	Alexei Starovoitov <ast@kernel.org>,
	Bjorn Andersson <bjorn.andersson@linaro.org>,
	Alexander Gordeev <agordeev@linux.ibm.com>,
	Anton Ivanov <anton.ivanov@cambridgegreys.com>,
	linux-s390@vger.kernel.org,
	Johannes Berg <johannes.berg@intel.com>,
	Daniel Borkmann <daniel@iogearbox.net>,
	Richard Weinberger <richard@nod.at>,
	Vincent Whitchurch <vincent.whitchurch@axis.com>,
	John Fastabend <john.fastabend@gmail.com>,
	Halil Pasic <pasic@linux.ibm.com>,
	Jakub Kicinski <kuba@kernel.org>,
	virtualization <virtualization@lists.linux-foundation.org>,
	Heiko Carstens <hca@linux.ibm.com>,
	Jesper Dangaard Brouer <hawk@kernel.org>,
	Vasily Gorbik <gor@linux.ibm.com>, Jeff Dike <jdike@addtoit.com>,
	linux-um@lists.infradead.org, Mark Gross <markgross@kernel.org>,
	Hans de Goede <hdegoede@redhat.com>, kvm <kvm@vger.kernel.org>,
	platform-driver-x86@vger.kernel.org,
	Mathieu Poirier <mathieu.poirier@linaro.org>,
	netdev <netdev@vger.kernel.org>,
	Cornelia Huck <cohuck@redhat.com>,
	Sven Schnelle <svens@linux.ibm.com>,
	"open list:XDP \(eXpress Data Path\)" <bpf@vger.kernel.org>,
	"David S. Miller" <davem@davemloft.net>
Subject: Re: [PATCH v9 31/32] virtio_net: support rx/tx queue resize
Date: Thu, 14 Apr 2022 17:30:02 +0800	[thread overview]
Message-ID: <CACGkMEvPH1k76xB_cHq_S9hvMXgGruoXpKLfoMZvJZ-L7wM9iw@mail.gmail.com> (raw)
In-Reply-To: <1649838917.6726515-10-xuanzhuo@linux.alibaba.com>

On Wed, Apr 13, 2022 at 4:47 PM Xuan Zhuo <xuanzhuo@linux.alibaba.com> wrote:
>
> On Wed, 13 Apr 2022 16:00:18 +0800, Jason Wang <jasowang@redhat.com> wrote:
> >
> > 在 2022/4/6 上午11:43, Xuan Zhuo 写道:
> > > This patch implements the resize function of the rx, tx queues.
> > > Based on this function, it is possible to modify the ring num of the
> > > queue.
> > >
> > > There may be an exception during the resize process, the resize may
> > > fail, or the vq can no longer be used. Either way, we must execute
> > > napi_enable(). Because napi_disable is similar to a lock, napi_enable
> > > must be called after calling napi_disable.
> > >
> > > Signed-off-by: Xuan Zhuo <xuanzhuo@linux.alibaba.com>
> > > ---
> > >   drivers/net/virtio_net.c | 81 ++++++++++++++++++++++++++++++++++++++++
> > >   1 file changed, 81 insertions(+)
> > >
> > > diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c
> > > index b8bf00525177..ba6859f305f7 100644
> > > --- a/drivers/net/virtio_net.c
> > > +++ b/drivers/net/virtio_net.c
> > > @@ -251,6 +251,9 @@ struct padded_vnet_hdr {
> > >     char padding[4];
> > >   };
> > >
> > > +static void virtnet_sq_free_unused_buf(struct virtqueue *vq, void *buf);
> > > +static void virtnet_rq_free_unused_buf(struct virtqueue *vq, void *buf);
> > > +
> > >   static bool is_xdp_frame(void *ptr)
> > >   {
> > >     return (unsigned long)ptr & VIRTIO_XDP_FLAG;
> > > @@ -1369,6 +1372,15 @@ static void virtnet_napi_enable(struct virtqueue *vq, struct napi_struct *napi)
> > >   {
> > >     napi_enable(napi);
> > >
> > > +   /* Check if vq is in reset state. The normal reset/resize process will
> > > +    * be protected by napi. However, the protection of napi is only enabled
> > > +    * during the operation, and the protection of napi will end after the
> > > +    * operation is completed. If re-enable fails during the process, vq
> > > +    * will remain unavailable with reset state.
> > > +    */
> > > +   if (vq->reset)
> > > +           return;
> >
> >
> > I don't get when could we hit this condition.
>
>
> In patch 23, the code to implement re-enable vq is as follows:
>
> +static int vp_modern_enable_reset_vq(struct virtqueue *vq)
> +{
> +       struct virtio_pci_device *vp_dev = to_vp_device(vq->vdev);
> +       struct virtio_pci_modern_device *mdev = &vp_dev->mdev;
> +       struct virtio_pci_vq_info *info;
> +       unsigned long flags, index;
> +       int err;
> +
> +       if (!vq->reset)
> +               return -EBUSY;
> +
> +       index = vq->index;
> +       info = vp_dev->vqs[index];
> +
> +       /* check queue reset status */
> +       if (vp_modern_get_queue_reset(mdev, index) != 1)
> +               return -EBUSY;
> +
> +       err = vp_active_vq(vq, info->msix_vector);
> +       if (err)
> +               return err;
> +
> +       if (vq->callback) {
> +               spin_lock_irqsave(&vp_dev->lock, flags);
> +               list_add(&info->node, &vp_dev->virtqueues);
> +               spin_unlock_irqrestore(&vp_dev->lock, flags);
> +       } else {
> +               INIT_LIST_HEAD(&info->node);
> +       }
> +
> +       vp_modern_set_queue_enable(&vp_dev->mdev, index, true);
> +
> +       if (vp_dev->per_vq_vectors && info->msix_vector != VIRTIO_MSI_NO_VECTOR)
> +               enable_irq(pci_irq_vector(vp_dev->pci_dev, info->msix_vector));
> +
> +       vq->reset = false;
> +
> +       return 0;
> +}
>
>
> There are three situations where an error will be returned. These are the
> situations I want to handle.

Right, but it looks harmless if we just schedule the NAPI without the check.

>
> But I'm rethinking the question, and I feel like you're right, although the
> hardware setup may fail. We can no longer sync with the hardware. But using it
> as a normal vq doesn't have any problems.

Note that we should make sure the buggy(malicous) device won't crash
the codes by changing the queue_reset value at its will.

>
> >
> >
> > > +
> > >     /* If all buffers were filled by other side before we napi_enabled, we
> > >      * won't get another interrupt, so process any outstanding packets now.
> > >      * Call local_bh_enable after to trigger softIRQ processing.
> > > @@ -1413,6 +1425,15 @@ static void refill_work(struct work_struct *work)
> > >             struct receive_queue *rq = &vi->rq[i];
> > >
> > >             napi_disable(&rq->napi);
> > > +
> > > +           /* Check if vq is in reset state. See more in
> > > +            * virtnet_napi_enable()
> > > +            */
> > > +           if (rq->vq->reset) {
> > > +                   virtnet_napi_enable(rq->vq, &rq->napi);
> > > +                   continue;
> > > +           }
> >
> >
> > Can we do something similar in virtnet_close() by canceling the work?
>
> I think there is no need to cancel the work here, because napi_disable will wait
> for the napi_enable of the resize. So if the re-enable failed vq is used as a normal
> vq, this logic can be removed.

Actually I meant the part of virtnet_rx_resize().

If we don't synchronize with the refill work, it might enable NAPI unexpectedly?

Thanks

>
>
> >
> >
> > > +
> > >             still_empty = !try_fill_recv(vi, rq, GFP_KERNEL);
> > >             virtnet_napi_enable(rq->vq, &rq->napi);
> > >
> > > @@ -1523,6 +1544,10 @@ static void virtnet_poll_cleantx(struct receive_queue *rq)
> > >     if (!sq->napi.weight || is_xdp_raw_buffer_queue(vi, index))
> > >             return;
> > >
> > > +   /* Check if vq is in reset state. See more in virtnet_napi_enable() */
> > > +   if (sq->vq->reset)
> > > +           return;
> >
> >
> > We've disabled TX napi, any chance we can still hit this?
>
> Same as above.
>
> >
> >
> > > +
> > >     if (__netif_tx_trylock(txq)) {
> > >             do {
> > >                     virtqueue_disable_cb(sq->vq);
> > > @@ -1769,6 +1794,62 @@ static netdev_tx_t start_xmit(struct sk_buff *skb, struct net_device *dev)
> > >     return NETDEV_TX_OK;
> > >   }
> > >
> > > +static int virtnet_rx_resize(struct virtnet_info *vi,
> > > +                        struct receive_queue *rq, u32 ring_num)
> > > +{
> > > +   int err;
> > > +
> > > +   napi_disable(&rq->napi);
> > > +
> > > +   err = virtqueue_resize(rq->vq, ring_num, virtnet_rq_free_unused_buf);
> > > +   if (err)
> > > +           goto err;
> > > +
> > > +   if (!try_fill_recv(vi, rq, GFP_KERNEL))
> > > +           schedule_delayed_work(&vi->refill, 0);
> > > +
> > > +   virtnet_napi_enable(rq->vq, &rq->napi);
> > > +   return 0;
> > > +
> > > +err:
> > > +   netdev_err(vi->dev,
> > > +              "reset rx reset vq fail: rx queue index: %td err: %d\n",
> > > +              rq - vi->rq, err);
> > > +   virtnet_napi_enable(rq->vq, &rq->napi);
> > > +   return err;
> > > +}
> > > +
> > > +static int virtnet_tx_resize(struct virtnet_info *vi,
> > > +                        struct send_queue *sq, u32 ring_num)
> > > +{
> > > +   struct netdev_queue *txq;
> > > +   int err, qindex;
> > > +
> > > +   qindex = sq - vi->sq;
> > > +
> > > +   virtnet_napi_tx_disable(&sq->napi);
> > > +
> > > +   txq = netdev_get_tx_queue(vi->dev, qindex);
> > > +   __netif_tx_lock_bh(txq);
> > > +   netif_stop_subqueue(vi->dev, qindex);
> > > +   __netif_tx_unlock_bh(txq);
> > > +
> > > +   err = virtqueue_resize(sq->vq, ring_num, virtnet_sq_free_unused_buf);
> > > +   if (err)
> > > +           goto err;
> > > +
> > > +   netif_start_subqueue(vi->dev, qindex);
> > > +   virtnet_napi_tx_enable(vi, sq->vq, &sq->napi);
> > > +   return 0;
> > > +
> > > +err:
> >
> >
> > I guess we can still start the queue in this case? (Since we don't
> > change the queue if resize fails).
>
> Yes, you are right.
>
> Thanks.
>
> >
> >
> > > +   netdev_err(vi->dev,
> > > +              "reset tx reset vq fail: tx queue index: %td err: %d\n",
> > > +              sq - vi->sq, err);
> > > +   virtnet_napi_tx_enable(vi, sq->vq, &sq->napi);
> > > +   return err;
> > > +}
> > > +
> > >   /*
> > >    * Send command via the control virtqueue and check status.  Commands
> > >    * supported by the hypervisor, as indicated by feature bits, should
> >
>

_______________________________________________
Virtualization mailing list
Virtualization@lists.linux-foundation.org
https://lists.linuxfoundation.org/mailman/listinfo/virtualization

WARNING: multiple messages have this Message-ID (diff)
From: Jason Wang <jasowang@redhat.com>
To: Xuan Zhuo <xuanzhuo@linux.alibaba.com>
Cc: Jeff Dike <jdike@addtoit.com>,
	Richard Weinberger <richard@nod.at>,
	Anton Ivanov <anton.ivanov@cambridgegreys.com>,
	"Michael S. Tsirkin" <mst@redhat.com>,
	"David S. Miller" <davem@davemloft.net>,
	Jakub Kicinski <kuba@kernel.org>,
	Hans de Goede <hdegoede@redhat.com>,
	Mark Gross <markgross@kernel.org>,
	Vadim Pasternak <vadimp@nvidia.com>,
	Bjorn Andersson <bjorn.andersson@linaro.org>,
	Mathieu Poirier <mathieu.poirier@linaro.org>,
	Cornelia Huck <cohuck@redhat.com>,
	Halil Pasic <pasic@linux.ibm.com>,
	Heiko Carstens <hca@linux.ibm.com>,
	Vasily Gorbik <gor@linux.ibm.com>,
	Christian Borntraeger <borntraeger@linux.ibm.com>,
	Alexander Gordeev <agordeev@linux.ibm.com>,
	Sven Schnelle <svens@linux.ibm.com>,
	Alexei Starovoitov <ast@kernel.org>,
	Daniel Borkmann <daniel@iogearbox.net>,
	Jesper Dangaard Brouer <hawk@kernel.org>,
	John Fastabend <john.fastabend@gmail.com>,
	Johannes Berg <johannes.berg@intel.com>,
	Vincent Whitchurch <vincent.whitchurch@axis.com>,
	linux-um@lists.infradead.org, netdev <netdev@vger.kernel.org>,
	platform-driver-x86@vger.kernel.org,
	linux-remoteproc@vger.kernel.org, linux-s390@vger.kernel.org,
	kvm <kvm@vger.kernel.org>,
	"open list:XDP (eXpress Data Path)" <bpf@vger.kernel.org>,
	virtualization <virtualization@lists.linux-foundation.org>
Subject: Re: [PATCH v9 31/32] virtio_net: support rx/tx queue resize
Date: Thu, 14 Apr 2022 17:30:02 +0800	[thread overview]
Message-ID: <CACGkMEvPH1k76xB_cHq_S9hvMXgGruoXpKLfoMZvJZ-L7wM9iw@mail.gmail.com> (raw)
In-Reply-To: <1649838917.6726515-10-xuanzhuo@linux.alibaba.com>

On Wed, Apr 13, 2022 at 4:47 PM Xuan Zhuo <xuanzhuo@linux.alibaba.com> wrote:
>
> On Wed, 13 Apr 2022 16:00:18 +0800, Jason Wang <jasowang@redhat.com> wrote:
> >
> > 在 2022/4/6 上午11:43, Xuan Zhuo 写道:
> > > This patch implements the resize function of the rx, tx queues.
> > > Based on this function, it is possible to modify the ring num of the
> > > queue.
> > >
> > > There may be an exception during the resize process, the resize may
> > > fail, or the vq can no longer be used. Either way, we must execute
> > > napi_enable(). Because napi_disable is similar to a lock, napi_enable
> > > must be called after calling napi_disable.
> > >
> > > Signed-off-by: Xuan Zhuo <xuanzhuo@linux.alibaba.com>
> > > ---
> > >   drivers/net/virtio_net.c | 81 ++++++++++++++++++++++++++++++++++++++++
> > >   1 file changed, 81 insertions(+)
> > >
> > > diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c
> > > index b8bf00525177..ba6859f305f7 100644
> > > --- a/drivers/net/virtio_net.c
> > > +++ b/drivers/net/virtio_net.c
> > > @@ -251,6 +251,9 @@ struct padded_vnet_hdr {
> > >     char padding[4];
> > >   };
> > >
> > > +static void virtnet_sq_free_unused_buf(struct virtqueue *vq, void *buf);
> > > +static void virtnet_rq_free_unused_buf(struct virtqueue *vq, void *buf);
> > > +
> > >   static bool is_xdp_frame(void *ptr)
> > >   {
> > >     return (unsigned long)ptr & VIRTIO_XDP_FLAG;
> > > @@ -1369,6 +1372,15 @@ static void virtnet_napi_enable(struct virtqueue *vq, struct napi_struct *napi)
> > >   {
> > >     napi_enable(napi);
> > >
> > > +   /* Check if vq is in reset state. The normal reset/resize process will
> > > +    * be protected by napi. However, the protection of napi is only enabled
> > > +    * during the operation, and the protection of napi will end after the
> > > +    * operation is completed. If re-enable fails during the process, vq
> > > +    * will remain unavailable with reset state.
> > > +    */
> > > +   if (vq->reset)
> > > +           return;
> >
> >
> > I don't get when could we hit this condition.
>
>
> In patch 23, the code to implement re-enable vq is as follows:
>
> +static int vp_modern_enable_reset_vq(struct virtqueue *vq)
> +{
> +       struct virtio_pci_device *vp_dev = to_vp_device(vq->vdev);
> +       struct virtio_pci_modern_device *mdev = &vp_dev->mdev;
> +       struct virtio_pci_vq_info *info;
> +       unsigned long flags, index;
> +       int err;
> +
> +       if (!vq->reset)
> +               return -EBUSY;
> +
> +       index = vq->index;
> +       info = vp_dev->vqs[index];
> +
> +       /* check queue reset status */
> +       if (vp_modern_get_queue_reset(mdev, index) != 1)
> +               return -EBUSY;
> +
> +       err = vp_active_vq(vq, info->msix_vector);
> +       if (err)
> +               return err;
> +
> +       if (vq->callback) {
> +               spin_lock_irqsave(&vp_dev->lock, flags);
> +               list_add(&info->node, &vp_dev->virtqueues);
> +               spin_unlock_irqrestore(&vp_dev->lock, flags);
> +       } else {
> +               INIT_LIST_HEAD(&info->node);
> +       }
> +
> +       vp_modern_set_queue_enable(&vp_dev->mdev, index, true);
> +
> +       if (vp_dev->per_vq_vectors && info->msix_vector != VIRTIO_MSI_NO_VECTOR)
> +               enable_irq(pci_irq_vector(vp_dev->pci_dev, info->msix_vector));
> +
> +       vq->reset = false;
> +
> +       return 0;
> +}
>
>
> There are three situations where an error will be returned. These are the
> situations I want to handle.

Right, but it looks harmless if we just schedule the NAPI without the check.

>
> But I'm rethinking the question, and I feel like you're right, although the
> hardware setup may fail. We can no longer sync with the hardware. But using it
> as a normal vq doesn't have any problems.

Note that we should make sure the buggy(malicous) device won't crash
the codes by changing the queue_reset value at its will.

>
> >
> >
> > > +
> > >     /* If all buffers were filled by other side before we napi_enabled, we
> > >      * won't get another interrupt, so process any outstanding packets now.
> > >      * Call local_bh_enable after to trigger softIRQ processing.
> > > @@ -1413,6 +1425,15 @@ static void refill_work(struct work_struct *work)
> > >             struct receive_queue *rq = &vi->rq[i];
> > >
> > >             napi_disable(&rq->napi);
> > > +
> > > +           /* Check if vq is in reset state. See more in
> > > +            * virtnet_napi_enable()
> > > +            */
> > > +           if (rq->vq->reset) {
> > > +                   virtnet_napi_enable(rq->vq, &rq->napi);
> > > +                   continue;
> > > +           }
> >
> >
> > Can we do something similar in virtnet_close() by canceling the work?
>
> I think there is no need to cancel the work here, because napi_disable will wait
> for the napi_enable of the resize. So if the re-enable failed vq is used as a normal
> vq, this logic can be removed.

Actually I meant the part of virtnet_rx_resize().

If we don't synchronize with the refill work, it might enable NAPI unexpectedly?

Thanks

>
>
> >
> >
> > > +
> > >             still_empty = !try_fill_recv(vi, rq, GFP_KERNEL);
> > >             virtnet_napi_enable(rq->vq, &rq->napi);
> > >
> > > @@ -1523,6 +1544,10 @@ static void virtnet_poll_cleantx(struct receive_queue *rq)
> > >     if (!sq->napi.weight || is_xdp_raw_buffer_queue(vi, index))
> > >             return;
> > >
> > > +   /* Check if vq is in reset state. See more in virtnet_napi_enable() */
> > > +   if (sq->vq->reset)
> > > +           return;
> >
> >
> > We've disabled TX napi, any chance we can still hit this?
>
> Same as above.
>
> >
> >
> > > +
> > >     if (__netif_tx_trylock(txq)) {
> > >             do {
> > >                     virtqueue_disable_cb(sq->vq);
> > > @@ -1769,6 +1794,62 @@ static netdev_tx_t start_xmit(struct sk_buff *skb, struct net_device *dev)
> > >     return NETDEV_TX_OK;
> > >   }
> > >
> > > +static int virtnet_rx_resize(struct virtnet_info *vi,
> > > +                        struct receive_queue *rq, u32 ring_num)
> > > +{
> > > +   int err;
> > > +
> > > +   napi_disable(&rq->napi);
> > > +
> > > +   err = virtqueue_resize(rq->vq, ring_num, virtnet_rq_free_unused_buf);
> > > +   if (err)
> > > +           goto err;
> > > +
> > > +   if (!try_fill_recv(vi, rq, GFP_KERNEL))
> > > +           schedule_delayed_work(&vi->refill, 0);
> > > +
> > > +   virtnet_napi_enable(rq->vq, &rq->napi);
> > > +   return 0;
> > > +
> > > +err:
> > > +   netdev_err(vi->dev,
> > > +              "reset rx reset vq fail: rx queue index: %td err: %d\n",
> > > +              rq - vi->rq, err);
> > > +   virtnet_napi_enable(rq->vq, &rq->napi);
> > > +   return err;
> > > +}
> > > +
> > > +static int virtnet_tx_resize(struct virtnet_info *vi,
> > > +                        struct send_queue *sq, u32 ring_num)
> > > +{
> > > +   struct netdev_queue *txq;
> > > +   int err, qindex;
> > > +
> > > +   qindex = sq - vi->sq;
> > > +
> > > +   virtnet_napi_tx_disable(&sq->napi);
> > > +
> > > +   txq = netdev_get_tx_queue(vi->dev, qindex);
> > > +   __netif_tx_lock_bh(txq);
> > > +   netif_stop_subqueue(vi->dev, qindex);
> > > +   __netif_tx_unlock_bh(txq);
> > > +
> > > +   err = virtqueue_resize(sq->vq, ring_num, virtnet_sq_free_unused_buf);
> > > +   if (err)
> > > +           goto err;
> > > +
> > > +   netif_start_subqueue(vi->dev, qindex);
> > > +   virtnet_napi_tx_enable(vi, sq->vq, &sq->napi);
> > > +   return 0;
> > > +
> > > +err:
> >
> >
> > I guess we can still start the queue in this case? (Since we don't
> > change the queue if resize fails).
>
> Yes, you are right.
>
> Thanks.
>
> >
> >
> > > +   netdev_err(vi->dev,
> > > +              "reset tx reset vq fail: tx queue index: %td err: %d\n",
> > > +              sq - vi->sq, err);
> > > +   virtnet_napi_tx_enable(vi, sq->vq, &sq->napi);
> > > +   return err;
> > > +}
> > > +
> > >   /*
> > >    * Send command via the control virtqueue and check status.  Commands
> > >    * supported by the hypervisor, as indicated by feature bits, should
> >
>


_______________________________________________
linux-um mailing list
linux-um@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/linux-um

  reply	other threads:[~2022-04-14  9:30 UTC|newest]

Thread overview: 255+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-04-06  3:43 [PATCH v9 00/32] virtio pci support VIRTIO_F_RING_RESET (refactor vring) Xuan Zhuo
2022-04-06  3:43 ` Xuan Zhuo
2022-04-06  3:43 ` Xuan Zhuo
2022-04-06  3:43 ` [PATCH v9 01/32] virtio: add helper virtqueue_get_vring_max_size() Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  2:41   ` Jason Wang
2022-04-12  2:41     ` Jason Wang
2022-04-12  2:41     ` Jason Wang
2022-04-13  2:24     ` Xuan Zhuo
2022-04-13  2:24       ` Xuan Zhuo
2022-04-13  2:24       ` Xuan Zhuo
2022-04-14  9:16       ` Jason Wang
2022-04-14  9:16         ` Jason Wang
2022-04-14  9:16         ` Jason Wang
2022-04-06  3:43 ` [PATCH v9 02/32] virtio: struct virtio_config_ops add callbacks for queue_reset Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  2:47   ` Jason Wang
2022-04-12  2:47     ` Jason Wang
2022-04-12  2:47     ` Jason Wang
2022-04-12  2:49     ` Jason Wang
2022-04-12  2:49       ` Jason Wang
2022-04-12  2:49       ` Jason Wang
2022-04-06  3:43 ` [PATCH v9 03/32] virtio_ring: update the document of the virtqueue_detach_unused_buf for queue reset Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  2:49   ` Jason Wang
2022-04-12  2:49     ` Jason Wang
2022-04-12  2:49     ` Jason Wang
2022-04-06  3:43 ` [PATCH v9 04/32] virtio_ring: remove the arg vq of vring_alloc_desc_extra() Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  2:53   ` Jason Wang
2022-04-12  2:53     ` Jason Wang
2022-04-12  2:53     ` Jason Wang
2022-04-06  3:43 ` [PATCH v9 05/32] virtio_ring: extract the logic of freeing vring Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  2:58   ` Jason Wang
2022-04-12  2:58     ` Jason Wang
2022-04-12  2:58     ` Jason Wang
2022-04-06  3:43 ` [PATCH v9 06/32] virtio_ring: split: extract the logic of alloc queue Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  3:22   ` Jason Wang
2022-04-12  3:22     ` Jason Wang
2022-04-12  3:22     ` Jason Wang
2022-04-13  6:52     ` Xuan Zhuo
2022-04-13  6:52       ` Xuan Zhuo
2022-04-13  6:52       ` Xuan Zhuo
2022-04-06  3:43 ` [PATCH v9 07/32] virtio_ring: split: extract the logic of alloc state and extra Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  3:26   ` Jason Wang
2022-04-12  3:26     ` Jason Wang
2022-04-12  3:26     ` Jason Wang
2022-04-13  6:48     ` Xuan Zhuo
2022-04-13  6:48       ` Xuan Zhuo
2022-04-13  6:48       ` Xuan Zhuo
2022-04-06  3:43 ` [PATCH v9 08/32] virtio_ring: split: extract the logic of attach vring Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  3:31   ` Jason Wang
2022-04-12  3:31     ` Jason Wang
2022-04-12  3:31     ` Jason Wang
2022-04-13  6:44     ` Xuan Zhuo
2022-04-13  6:44       ` Xuan Zhuo
2022-04-13  6:44       ` Xuan Zhuo
2022-04-06  3:43 ` [PATCH v9 09/32] virtio_ring: split: extract the logic of vq init Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  3:42   ` Jason Wang
2022-04-12  3:42     ` Jason Wang
2022-04-12  3:42     ` Jason Wang
2022-04-13  7:04     ` Xuan Zhuo
2022-04-13  7:04       ` Xuan Zhuo
2022-04-13  7:04       ` Xuan Zhuo
2022-04-06  3:43 ` [PATCH v9 10/32] virtio_ring: split: introduce virtqueue_reinit_split() Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  5:49   ` Jason Wang
2022-04-12  5:49     ` Jason Wang
2022-04-12  5:49     ` Jason Wang
2022-04-06  3:43 ` [PATCH v9 11/32] virtio_ring: split: introduce virtqueue_resize_split() Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  5:53   ` Jason Wang
2022-04-12  5:53     ` Jason Wang
2022-04-12  5:53     ` Jason Wang
2022-04-13  6:32     ` Xuan Zhuo
2022-04-13  6:32       ` Xuan Zhuo
2022-04-13  6:32       ` Xuan Zhuo
2022-04-06  3:43 ` [PATCH v9 12/32] virtio_ring: packed: extract the logic of alloc queue Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  6:28   ` Jason Wang
2022-04-12  6:28     ` Jason Wang
2022-04-12  6:28     ` Jason Wang
2022-04-13  3:23     ` Xuan Zhuo
2022-04-13  3:23       ` Xuan Zhuo
2022-04-13  3:23       ` Xuan Zhuo
2022-04-14  9:18       ` Jason Wang
2022-04-14  9:18         ` Jason Wang
2022-04-14  9:18         ` Jason Wang
2022-04-06  3:43 ` [PATCH v9 13/32] virtio_ring: packed: extract the logic of alloc state and extra Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43 ` [PATCH v9 14/32] virtio_ring: packed: extract the logic of attach vring Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43 ` [PATCH v9 15/32] virtio_ring: packed: extract the logic of vq init Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43 ` [PATCH v9 16/32] virtio_ring: packed: introduce virtqueue_reinit_packed() Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43 ` [PATCH v9 17/32] virtio_ring: packed: introduce virtqueue_resize_packed() Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43 ` [PATCH v9 18/32] virtio_ring: introduce virtqueue_resize() Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  6:41   ` Jason Wang
2022-04-12  6:41     ` Jason Wang
2022-04-12  6:41     ` Jason Wang
2022-04-13 12:14     ` Xuan Zhuo
2022-04-13 12:14       ` Xuan Zhuo
2022-04-13 12:14       ` Xuan Zhuo
2022-04-13 12:21     ` Xuan Zhuo
2022-04-13 12:21       ` Xuan Zhuo
2022-04-13 12:21       ` Xuan Zhuo
2022-04-06  3:43 ` [PATCH v9 19/32] virtio_pci: struct virtio_pci_common_cfg add queue_notify_data Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43 ` [PATCH v9 20/32] virtio: queue_reset: add VIRTIO_F_RING_RESET Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43 ` [PATCH v9 21/32] virtio_pci: queue_reset: update struct virtio_pci_common_cfg and option functions Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  6:43   ` Jason Wang
2022-04-12  6:43     ` Jason Wang
2022-04-12  6:43     ` Jason Wang
2022-04-06  3:43 ` [PATCH v9 22/32] virtio_pci: queue_reset: extract the logic of active vq for modern pci Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  6:58   ` Jason Wang
2022-04-12  6:58     ` Jason Wang
2022-04-12  6:58     ` Jason Wang
2022-04-14  6:22     ` Xuan Zhuo
2022-04-14  6:22       ` Xuan Zhuo
2022-04-14  6:22       ` Xuan Zhuo
2022-04-14  9:37       ` Jason Wang
2022-04-14  9:37         ` Jason Wang
2022-04-14  9:37         ` Jason Wang
2022-04-06  3:43 ` [PATCH v9 23/32] virtio_pci: queue_reset: support VIRTIO_F_RING_RESET Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  7:07   ` Jason Wang
2022-04-12  7:07     ` Jason Wang
2022-04-12  7:07     ` Jason Wang
2022-04-13  3:21     ` Xuan Zhuo
2022-04-13  3:21       ` Xuan Zhuo
2022-04-13  3:21       ` Xuan Zhuo
2022-04-14  9:17       ` Jason Wang
2022-04-14  9:17         ` Jason Wang
2022-04-14  9:17         ` Jason Wang
2022-04-13  8:48     ` Xuan Zhuo
2022-04-13  8:48       ` Xuan Zhuo
2022-04-13  8:48       ` Xuan Zhuo
2022-04-06  3:43 ` [PATCH v9 24/32] virtio: find_vqs() add arg sizes Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  7:10   ` Jason Wang
2022-04-12  7:10     ` Jason Wang
2022-04-12  7:10     ` Jason Wang
2022-04-06  3:43 ` [PATCH v9 25/32] virtio_pci: support the arg sizes of find_vqs() Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  7:12   ` Jason Wang
2022-04-12  7:12     ` Jason Wang
2022-04-12  7:12     ` Jason Wang
2022-04-06  3:43 ` [PATCH v9 26/32] virtio_mmio: " Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  7:13   ` Jason Wang
2022-04-12  7:13     ` Jason Wang
2022-04-12  7:13     ` Jason Wang
2022-04-06  3:43 ` [PATCH v9 27/32] virtio: add helper virtio_find_vqs_ctx_size() Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  7:15   ` Jason Wang
2022-04-12  7:15     ` Jason Wang
2022-04-12  7:15     ` Jason Wang
2022-04-06  3:43 ` [PATCH v9 28/32] virtio_net: set the default max ring size by find_vqs() Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  7:18   ` Jason Wang
2022-04-12  7:18     ` Jason Wang
2022-04-12  7:18     ` Jason Wang
2022-04-06  3:43 ` [PATCH v9 29/32] virtio_net: get ringparam by virtqueue_get_vring_max_size() Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-12  7:19   ` Jason Wang
2022-04-12  7:19     ` Jason Wang
2022-04-12  7:19     ` Jason Wang
2022-04-06  3:43 ` [PATCH v9 30/32] virtio_net: split free_unused_bufs() Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43 ` [PATCH v9 31/32] virtio_net: support rx/tx queue resize Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-13  8:00   ` Jason Wang
2022-04-13  8:00     ` Jason Wang
2022-04-13  8:00     ` Jason Wang
2022-04-13  8:35     ` Xuan Zhuo
2022-04-13  8:35       ` Xuan Zhuo
2022-04-13  8:35       ` Xuan Zhuo
2022-04-14  9:30       ` Jason Wang [this message]
2022-04-14  9:30         ` Jason Wang
2022-04-14  9:30         ` Jason Wang
2022-04-15  2:18         ` Xuan Zhuo
2022-04-15  2:18           ` Xuan Zhuo
2022-04-15  2:18           ` Xuan Zhuo
2022-04-15  5:53           ` Jason Wang
2022-04-15  5:53             ` Jason Wang
2022-04-15  5:53             ` Jason Wang
2022-04-15  9:17             ` Xuan Zhuo
2022-04-15  9:17               ` Xuan Zhuo
2022-04-15  9:17               ` Xuan Zhuo
2022-04-18  7:57               ` Jason Wang
2022-04-18  7:57                 ` Jason Wang
2022-04-18  7:57                 ` Jason Wang
2022-04-18  3:21     ` Xuan Zhuo
2022-04-18  3:21       ` Xuan Zhuo
2022-04-18  3:21       ` Xuan Zhuo
2022-04-18  7:49       ` Jason Wang
2022-04-18  7:49         ` Jason Wang
2022-04-18  7:49         ` Jason Wang
2022-04-18  8:48         ` Xuan Zhuo
2022-04-18  8:48           ` Xuan Zhuo
2022-04-18  8:48           ` Xuan Zhuo
2022-04-06  3:43 ` [PATCH v9 32/32] virtio_net: support set_ringparam Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-06  3:43   ` Xuan Zhuo
2022-04-13  8:06   ` Jason Wang
2022-04-13  8:06     ` Jason Wang
2022-04-13  8:06     ` Jason Wang
2022-04-26  9:55 ` [PATCH v9 00/32] virtio pci support VIRTIO_F_RING_RESET (refactor vring) Michael S. Tsirkin
2022-04-26  9:55   ` Michael S. Tsirkin
2022-04-26  9:55   ` Michael S. Tsirkin
2022-04-26  9:59   ` Xuan Zhuo
2022-04-26  9:59     ` Xuan Zhuo
2022-04-26  9:59     ` Xuan Zhuo

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=CACGkMEvPH1k76xB_cHq_S9hvMXgGruoXpKLfoMZvJZ-L7wM9iw@mail.gmail.com \
    --to=jasowang@redhat.com \
    --cc=agordeev@linux.ibm.com \
    --cc=anton.ivanov@cambridgegreys.com \
    --cc=ast@kernel.org \
    --cc=bjorn.andersson@linaro.org \
    --cc=borntraeger@linux.ibm.com \
    --cc=bpf@vger.kernel.org \
    --cc=cohuck@redhat.com \
    --cc=daniel@iogearbox.net \
    --cc=davem@davemloft.net \
    --cc=gor@linux.ibm.com \
    --cc=hawk@kernel.org \
    --cc=hca@linux.ibm.com \
    --cc=hdegoede@redhat.com \
    --cc=jdike@addtoit.com \
    --cc=johannes.berg@intel.com \
    --cc=john.fastabend@gmail.com \
    --cc=kuba@kernel.org \
    --cc=kvm@vger.kernel.org \
    --cc=linux-remoteproc@vger.kernel.org \
    --cc=linux-s390@vger.kernel.org \
    --cc=linux-um@lists.infradead.org \
    --cc=markgross@kernel.org \
    --cc=mathieu.poirier@linaro.org \
    --cc=mst@redhat.com \
    --cc=netdev@vger.kernel.org \
    --cc=pasic@linux.ibm.com \
    --cc=platform-driver-x86@vger.kernel.org \
    --cc=richard@nod.at \
    --cc=svens@linux.ibm.com \
    --cc=vadimp@nvidia.com \
    --cc=vincent.whitchurch@axis.com \
    --cc=virtualization@lists.linux-foundation.org \
    --cc=xuanzhuo@linux.alibaba.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.