All of lore.kernel.org
 help / color / mirror / Atom feed
From: Jason Wang <jasowang@redhat.com>
To: Xuan Zhuo <xuanzhuo@linux.alibaba.com>,
	virtualization@lists.linux-foundation.org,
	netdev@vger.kernel.org
Cc: Jeff Dike <jdike@addtoit.com>,
	Richard Weinberger <richard@nod.at>,
	Anton Ivanov <anton.ivanov@cambridgegreys.com>,
	"Michael S. Tsirkin" <mst@redhat.com>,
	"David S. Miller" <davem@davemloft.net>,
	Jakub Kicinski <kuba@kernel.org>,
	Hans de Goede <hdegoede@redhat.com>,
	Mark Gross <markgross@kernel.org>,
	Vadim Pasternak <vadimp@nvidia.com>,
	Bjorn Andersson <bjorn.andersson@linaro.org>,
	Mathieu Poirier <mathieu.poirier@linaro.org>,
	Cornelia Huck <cohuck@redhat.com>,
	Halil Pasic <pasic@linux.ibm.com>,
	Heiko Carstens <hca@linux.ibm.com>,
	Vasily Gorbik <gor@linux.ibm.com>,
	Christian Borntraeger <borntraeger@linux.ibm.com>,
	Alexander Gordeev <agordeev@linux.ibm.com>,
	Sven Schnelle <svens@linux.ibm.com>,
	Alexei Starovoitov <ast@kernel.org>,
	Daniel Borkmann <daniel@iogearbox.net>,
	Jesper Dangaard Brouer <hawk@kernel.org>,
	John Fastabend <john.fastabend@gmail.com>,
	Johannes Berg <johannes.berg@intel.com>,
	Vincent Whitchurch <vincent.whitchurch@axis.com>,
	linux-um@lists.infradead.org,
	platform-driver-x86@vger.kernel.org,
	linux-remoteproc@vger.kernel.org, linux-s390@vger.kernel.org,
	kvm@vger.kernel.org, bpf@vger.kernel.org
Subject: Re: [PATCH v7 24/26] virtio_net: support rx/tx queue reset
Date: Wed, 9 Mar 2022 17:14:34 +0800	[thread overview]
Message-ID: <7ff78ff8-bdd0-bb5e-1cea-cf1126226feb@redhat.com> (raw)
In-Reply-To: <20220308123518.33800-25-xuanzhuo@linux.alibaba.com>


在 2022/3/8 下午8:35, Xuan Zhuo 写道:
> This patch implements the reset function of the rx, tx queues.
>
> Based on this function, it is possible to modify the ring num of the
> queue. And quickly recycle the buffer in the queue.
>
> In the process of the queue disable, in theory, as long as virtio
> supports queue reset, there will be no exceptions.
>
> However, in the process of the queue enable, there may be exceptions due to
> memory allocation.  In this case, vq is not available, but we still have
> to execute napi_enable(). Because napi_disable is similar to a lock,
> napi_enable must be called after calling napi_disable.
>
> Signed-off-by: Xuan Zhuo <xuanzhuo@linux.alibaba.com>
> ---
>   drivers/net/virtio_net.c | 107 +++++++++++++++++++++++++++++++++++++++
>   1 file changed, 107 insertions(+)
>
> diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c
> index 409a8e180918..ffff323dcef0 100644
> --- a/drivers/net/virtio_net.c
> +++ b/drivers/net/virtio_net.c
> @@ -251,6 +251,11 @@ struct padded_vnet_hdr {
>   	char padding[4];
>   };
>   
> +static void virtnet_sq_free_unused_bufs(struct virtnet_info *vi,
> +					struct send_queue *sq);
> +static void virtnet_rq_free_unused_bufs(struct virtnet_info *vi,
> +					struct receive_queue *rq);
> +
>   static bool is_xdp_frame(void *ptr)
>   {
>   	return (unsigned long)ptr & VIRTIO_XDP_FLAG;
> @@ -1369,6 +1374,9 @@ static void virtnet_napi_enable(struct virtqueue *vq, struct napi_struct *napi)
>   {
>   	napi_enable(napi);
>   
> +	if (vq->reset)
> +		return;
> +


Let's WARN_ONCE() here?


>   	/* If all buffers were filled by other side before we napi_enabled, we
>   	 * won't get another interrupt, so process any outstanding packets now.
>   	 * Call local_bh_enable after to trigger softIRQ processing.
> @@ -1413,6 +1421,10 @@ static void refill_work(struct work_struct *work)
>   		struct receive_queue *rq = &vi->rq[i];
>   
>   		napi_disable(&rq->napi);
> +		if (rq->vq->reset) {
> +			virtnet_napi_enable(rq->vq, &rq->napi);
> +			continue;
> +		}


This seems racy and it's a hint that we need sync with the refill work 
during reset like what we did in virtnet_close():

         /* Make sure refill_work doesn't re-enable napi! */
         cancel_delayed_work_sync(&vi->refill);


>   		still_empty = !try_fill_recv(vi, rq, GFP_KERNEL);
>   		virtnet_napi_enable(rq->vq, &rq->napi);
>   
> @@ -1523,6 +1535,9 @@ static void virtnet_poll_cleantx(struct receive_queue *rq)
>   	if (!sq->napi.weight || is_xdp_raw_buffer_queue(vi, index))
>   		return;
>   
> +	if (sq->vq->reset)
> +		return;


It looks to me we'd better either WARN or just remove this. Since it 
looks like a workaround for the un-synchronized NAPI somehow.


> +
>   	if (__netif_tx_trylock(txq)) {
>   		do {
>   			virtqueue_disable_cb(sq->vq);
> @@ -1769,6 +1784,98 @@ static netdev_tx_t start_xmit(struct sk_buff *skb, struct net_device *dev)
>   	return NETDEV_TX_OK;
>   }
>   
> +static int virtnet_rx_vq_reset(struct virtnet_info *vi,
> +			       struct receive_queue *rq, u32 ring_num)


It's better to rename this as virtnet_rx_resize().


> +{
> +	int err;
> +
> +	/* stop napi */
> +	napi_disable(&rq->napi);
> +


Here, as discussed above, we need synchronize with the refill work.


> +	/* reset the queue */
> +	err = virtio_reset_vq(rq->vq);
> +	if (err)
> +		goto err;


Btw, most comment of this function seems useless since code already 
explain themselves.


> +
> +	/* free bufs */
> +	virtnet_rq_free_unused_bufs(vi, rq);
> +
> +	/* reset vring. */
> +	err = virtqueue_reset_vring(rq->vq, ring_num);
> +	if (err)
> +		goto err;
> +
> +	/* enable reset queue */
> +	err = virtio_enable_resetq(rq->vq);
> +	if (err)
> +		goto err;
> +
> +	/* fill recv */
> +	if (!try_fill_recv(vi, rq, GFP_KERNEL))
> +		schedule_delayed_work(&vi->refill, 0);
> +
> +	/* enable napi */
> +	virtnet_napi_enable(rq->vq, &rq->napi);
> +	return 0;
> +
> +err:
> +	netdev_err(vi->dev,
> +		   "reset rx reset vq fail: rx queue index: %ld err: %d\n",
> +		   rq - vi->rq, err);
> +	virtnet_napi_enable(rq->vq, &rq->napi);
> +	return err;
> +}
> +
> +static int virtnet_tx_vq_reset(struct virtnet_info *vi,
> +			       struct send_queue *sq, u32 ring_num)
> +{


It looks to me it's better to rename this as "virtnet_rx_resize()"


> +	struct netdev_queue *txq;
> +	int err, qindex;
> +
> +	qindex = sq - vi->sq;
> +
> +	txq = netdev_get_tx_queue(vi->dev, qindex);
> +	__netif_tx_lock_bh(txq);
> +
> +	/* stop tx queue and napi */
> +	netif_stop_subqueue(vi->dev, qindex);
> +	virtnet_napi_tx_disable(&sq->napi);


There's no need to hold tx lock for napi disable.

Thanks


> +
> +	__netif_tx_unlock_bh(txq);
> +
> +	/* reset the queue */
> +	err = virtio_reset_vq(sq->vq);
> +	if (err) {
> +		netif_start_subqueue(vi->dev, qindex);
> +		goto err;
> +	}
> +
> +	/* free bufs */
> +	virtnet_sq_free_unused_bufs(vi, sq);
> +
> +	/* reset vring. */
> +	err = virtqueue_reset_vring(sq->vq, ring_num);
> +	if (err)
> +		goto err;
> +
> +	/* enable reset queue */
> +	err = virtio_enable_resetq(sq->vq);
> +	if (err)
> +		goto err;
> +
> +	/* start tx queue and napi */
> +	netif_start_subqueue(vi->dev, qindex);
> +	virtnet_napi_tx_enable(vi, sq->vq, &sq->napi);
> +	return 0;
> +
> +err:
> +	netdev_err(vi->dev,
> +		   "reset tx reset vq fail: tx queue index: %ld err: %d\n",
> +		   sq - vi->sq, err);
> +	virtnet_napi_tx_enable(vi, sq->vq, &sq->napi);
> +	return err;
> +}
> +
>   /*
>    * Send command via the control virtqueue and check status.  Commands
>    * supported by the hypervisor, as indicated by feature bits, should


WARNING: multiple messages have this Message-ID (diff)
From: Jason Wang <jasowang@redhat.com>
To: Xuan Zhuo <xuanzhuo@linux.alibaba.com>,
	virtualization@lists.linux-foundation.org,
	netdev@vger.kernel.org
Cc: Vadim Pasternak <vadimp@nvidia.com>,
	"Michael S. Tsirkin" <mst@redhat.com>,
	linux-remoteproc@vger.kernel.org,
	Alexei Starovoitov <ast@kernel.org>,
	Bjorn Andersson <bjorn.andersson@linaro.org>,
	Alexander Gordeev <agordeev@linux.ibm.com>,
	Anton Ivanov <anton.ivanov@cambridgegreys.com>,
	linux-s390@vger.kernel.org,
	Johannes Berg <johannes.berg@intel.com>,
	Daniel Borkmann <daniel@iogearbox.net>,
	Richard Weinberger <richard@nod.at>,
	Vincent Whitchurch <vincent.whitchurch@axis.com>,
	John Fastabend <john.fastabend@gmail.com>,
	Halil Pasic <pasic@linux.ibm.com>,
	Jakub Kicinski <kuba@kernel.org>,
	Heiko Carstens <hca@linux.ibm.com>,
	Jesper Dangaard Brouer <hawk@kernel.org>,
	Vasily Gorbik <gor@linux.ibm.com>, Jeff Dike <jdike@addtoit.com>,
	linux-um@lists.infradead.org, Mark Gross <markgross@kernel.org>,
	Hans de Goede <hdegoede@redhat.com>,
	kvm@vger.kernel.org, platform-driver-x86@vger.kernel.org,
	Mathieu Poirier <mathieu.poirier@linaro.org>,
	Cornelia Huck <cohuck@redhat.com>,
	Sven Schnelle <svens@linux.ibm.com>,
	bpf@vger.kernel.org, "David S. Miller" <davem@davemloft.net>
Subject: Re: [PATCH v7 24/26] virtio_net: support rx/tx queue reset
Date: Wed, 9 Mar 2022 17:14:34 +0800	[thread overview]
Message-ID: <7ff78ff8-bdd0-bb5e-1cea-cf1126226feb@redhat.com> (raw)
In-Reply-To: <20220308123518.33800-25-xuanzhuo@linux.alibaba.com>


在 2022/3/8 下午8:35, Xuan Zhuo 写道:
> This patch implements the reset function of the rx, tx queues.
>
> Based on this function, it is possible to modify the ring num of the
> queue. And quickly recycle the buffer in the queue.
>
> In the process of the queue disable, in theory, as long as virtio
> supports queue reset, there will be no exceptions.
>
> However, in the process of the queue enable, there may be exceptions due to
> memory allocation.  In this case, vq is not available, but we still have
> to execute napi_enable(). Because napi_disable is similar to a lock,
> napi_enable must be called after calling napi_disable.
>
> Signed-off-by: Xuan Zhuo <xuanzhuo@linux.alibaba.com>
> ---
>   drivers/net/virtio_net.c | 107 +++++++++++++++++++++++++++++++++++++++
>   1 file changed, 107 insertions(+)
>
> diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c
> index 409a8e180918..ffff323dcef0 100644
> --- a/drivers/net/virtio_net.c
> +++ b/drivers/net/virtio_net.c
> @@ -251,6 +251,11 @@ struct padded_vnet_hdr {
>   	char padding[4];
>   };
>   
> +static void virtnet_sq_free_unused_bufs(struct virtnet_info *vi,
> +					struct send_queue *sq);
> +static void virtnet_rq_free_unused_bufs(struct virtnet_info *vi,
> +					struct receive_queue *rq);
> +
>   static bool is_xdp_frame(void *ptr)
>   {
>   	return (unsigned long)ptr & VIRTIO_XDP_FLAG;
> @@ -1369,6 +1374,9 @@ static void virtnet_napi_enable(struct virtqueue *vq, struct napi_struct *napi)
>   {
>   	napi_enable(napi);
>   
> +	if (vq->reset)
> +		return;
> +


Let's WARN_ONCE() here?


>   	/* If all buffers were filled by other side before we napi_enabled, we
>   	 * won't get another interrupt, so process any outstanding packets now.
>   	 * Call local_bh_enable after to trigger softIRQ processing.
> @@ -1413,6 +1421,10 @@ static void refill_work(struct work_struct *work)
>   		struct receive_queue *rq = &vi->rq[i];
>   
>   		napi_disable(&rq->napi);
> +		if (rq->vq->reset) {
> +			virtnet_napi_enable(rq->vq, &rq->napi);
> +			continue;
> +		}


This seems racy and it's a hint that we need sync with the refill work 
during reset like what we did in virtnet_close():

         /* Make sure refill_work doesn't re-enable napi! */
         cancel_delayed_work_sync(&vi->refill);


>   		still_empty = !try_fill_recv(vi, rq, GFP_KERNEL);
>   		virtnet_napi_enable(rq->vq, &rq->napi);
>   
> @@ -1523,6 +1535,9 @@ static void virtnet_poll_cleantx(struct receive_queue *rq)
>   	if (!sq->napi.weight || is_xdp_raw_buffer_queue(vi, index))
>   		return;
>   
> +	if (sq->vq->reset)
> +		return;


It looks to me we'd better either WARN or just remove this. Since it 
looks like a workaround for the un-synchronized NAPI somehow.


> +
>   	if (__netif_tx_trylock(txq)) {
>   		do {
>   			virtqueue_disable_cb(sq->vq);
> @@ -1769,6 +1784,98 @@ static netdev_tx_t start_xmit(struct sk_buff *skb, struct net_device *dev)
>   	return NETDEV_TX_OK;
>   }
>   
> +static int virtnet_rx_vq_reset(struct virtnet_info *vi,
> +			       struct receive_queue *rq, u32 ring_num)


It's better to rename this as virtnet_rx_resize().


> +{
> +	int err;
> +
> +	/* stop napi */
> +	napi_disable(&rq->napi);
> +


Here, as discussed above, we need synchronize with the refill work.


> +	/* reset the queue */
> +	err = virtio_reset_vq(rq->vq);
> +	if (err)
> +		goto err;


Btw, most comment of this function seems useless since code already 
explain themselves.


> +
> +	/* free bufs */
> +	virtnet_rq_free_unused_bufs(vi, rq);
> +
> +	/* reset vring. */
> +	err = virtqueue_reset_vring(rq->vq, ring_num);
> +	if (err)
> +		goto err;
> +
> +	/* enable reset queue */
> +	err = virtio_enable_resetq(rq->vq);
> +	if (err)
> +		goto err;
> +
> +	/* fill recv */
> +	if (!try_fill_recv(vi, rq, GFP_KERNEL))
> +		schedule_delayed_work(&vi->refill, 0);
> +
> +	/* enable napi */
> +	virtnet_napi_enable(rq->vq, &rq->napi);
> +	return 0;
> +
> +err:
> +	netdev_err(vi->dev,
> +		   "reset rx reset vq fail: rx queue index: %ld err: %d\n",
> +		   rq - vi->rq, err);
> +	virtnet_napi_enable(rq->vq, &rq->napi);
> +	return err;
> +}
> +
> +static int virtnet_tx_vq_reset(struct virtnet_info *vi,
> +			       struct send_queue *sq, u32 ring_num)
> +{


It looks to me it's better to rename this as "virtnet_rx_resize()"


> +	struct netdev_queue *txq;
> +	int err, qindex;
> +
> +	qindex = sq - vi->sq;
> +
> +	txq = netdev_get_tx_queue(vi->dev, qindex);
> +	__netif_tx_lock_bh(txq);
> +
> +	/* stop tx queue and napi */
> +	netif_stop_subqueue(vi->dev, qindex);
> +	virtnet_napi_tx_disable(&sq->napi);


There's no need to hold tx lock for napi disable.

Thanks


> +
> +	__netif_tx_unlock_bh(txq);
> +
> +	/* reset the queue */
> +	err = virtio_reset_vq(sq->vq);
> +	if (err) {
> +		netif_start_subqueue(vi->dev, qindex);
> +		goto err;
> +	}
> +
> +	/* free bufs */
> +	virtnet_sq_free_unused_bufs(vi, sq);
> +
> +	/* reset vring. */
> +	err = virtqueue_reset_vring(sq->vq, ring_num);
> +	if (err)
> +		goto err;
> +
> +	/* enable reset queue */
> +	err = virtio_enable_resetq(sq->vq);
> +	if (err)
> +		goto err;
> +
> +	/* start tx queue and napi */
> +	netif_start_subqueue(vi->dev, qindex);
> +	virtnet_napi_tx_enable(vi, sq->vq, &sq->napi);
> +	return 0;
> +
> +err:
> +	netdev_err(vi->dev,
> +		   "reset tx reset vq fail: tx queue index: %ld err: %d\n",
> +		   sq - vi->sq, err);
> +	virtnet_napi_tx_enable(vi, sq->vq, &sq->napi);
> +	return err;
> +}
> +
>   /*
>    * Send command via the control virtqueue and check status.  Commands
>    * supported by the hypervisor, as indicated by feature bits, should

_______________________________________________
Virtualization mailing list
Virtualization@lists.linux-foundation.org
https://lists.linuxfoundation.org/mailman/listinfo/virtualization

WARNING: multiple messages have this Message-ID (diff)
From: Jason Wang <jasowang@redhat.com>
To: Xuan Zhuo <xuanzhuo@linux.alibaba.com>,
	virtualization@lists.linux-foundation.org,
	netdev@vger.kernel.org
Cc: Jeff Dike <jdike@addtoit.com>,
	Richard Weinberger <richard@nod.at>,
	Anton Ivanov <anton.ivanov@cambridgegreys.com>,
	"Michael S. Tsirkin" <mst@redhat.com>,
	"David S. Miller" <davem@davemloft.net>,
	Jakub Kicinski <kuba@kernel.org>,
	Hans de Goede <hdegoede@redhat.com>,
	Mark Gross <markgross@kernel.org>,
	Vadim Pasternak <vadimp@nvidia.com>,
	Bjorn Andersson <bjorn.andersson@linaro.org>,
	Mathieu Poirier <mathieu.poirier@linaro.org>,
	Cornelia Huck <cohuck@redhat.com>,
	Halil Pasic <pasic@linux.ibm.com>,
	Heiko Carstens <hca@linux.ibm.com>,
	Vasily Gorbik <gor@linux.ibm.com>,
	Christian Borntraeger <borntraeger@linux.ibm.com>,
	Alexander Gordeev <agordeev@linux.ibm.com>,
	Sven Schnelle <svens@linux.ibm.com>,
	Alexei Starovoitov <ast@kernel.org>,
	Daniel Borkmann <daniel@iogearbox.net>,
	Jesper Dangaard Brouer <hawk@kernel.org>,
	John Fastabend <john.fastabend@gmail.com>,
	Johannes Berg <johannes.berg@intel.com>,
	Vincent Whitchurch <vincent.whitchurch@axis.com>,
	linux-um@lists.infradead.org,
	platform-driver-x86@vger.kernel.org,
	linux-remoteproc@vger.kernel.org, linux-s390@vger.kernel.org,
	kvm@vger.kernel.org, bpf@vger.kernel.org
Subject: Re: [PATCH v7 24/26] virtio_net: support rx/tx queue reset
Date: Wed, 9 Mar 2022 17:14:34 +0800	[thread overview]
Message-ID: <7ff78ff8-bdd0-bb5e-1cea-cf1126226feb@redhat.com> (raw)
In-Reply-To: <20220308123518.33800-25-xuanzhuo@linux.alibaba.com>


在 2022/3/8 下午8:35, Xuan Zhuo 写道:
> This patch implements the reset function of the rx, tx queues.
>
> Based on this function, it is possible to modify the ring num of the
> queue. And quickly recycle the buffer in the queue.
>
> In the process of the queue disable, in theory, as long as virtio
> supports queue reset, there will be no exceptions.
>
> However, in the process of the queue enable, there may be exceptions due to
> memory allocation.  In this case, vq is not available, but we still have
> to execute napi_enable(). Because napi_disable is similar to a lock,
> napi_enable must be called after calling napi_disable.
>
> Signed-off-by: Xuan Zhuo <xuanzhuo@linux.alibaba.com>
> ---
>   drivers/net/virtio_net.c | 107 +++++++++++++++++++++++++++++++++++++++
>   1 file changed, 107 insertions(+)
>
> diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c
> index 409a8e180918..ffff323dcef0 100644
> --- a/drivers/net/virtio_net.c
> +++ b/drivers/net/virtio_net.c
> @@ -251,6 +251,11 @@ struct padded_vnet_hdr {
>   	char padding[4];
>   };
>   
> +static void virtnet_sq_free_unused_bufs(struct virtnet_info *vi,
> +					struct send_queue *sq);
> +static void virtnet_rq_free_unused_bufs(struct virtnet_info *vi,
> +					struct receive_queue *rq);
> +
>   static bool is_xdp_frame(void *ptr)
>   {
>   	return (unsigned long)ptr & VIRTIO_XDP_FLAG;
> @@ -1369,6 +1374,9 @@ static void virtnet_napi_enable(struct virtqueue *vq, struct napi_struct *napi)
>   {
>   	napi_enable(napi);
>   
> +	if (vq->reset)
> +		return;
> +


Let's WARN_ONCE() here?


>   	/* If all buffers were filled by other side before we napi_enabled, we
>   	 * won't get another interrupt, so process any outstanding packets now.
>   	 * Call local_bh_enable after to trigger softIRQ processing.
> @@ -1413,6 +1421,10 @@ static void refill_work(struct work_struct *work)
>   		struct receive_queue *rq = &vi->rq[i];
>   
>   		napi_disable(&rq->napi);
> +		if (rq->vq->reset) {
> +			virtnet_napi_enable(rq->vq, &rq->napi);
> +			continue;
> +		}


This seems racy and it's a hint that we need sync with the refill work 
during reset like what we did in virtnet_close():

         /* Make sure refill_work doesn't re-enable napi! */
         cancel_delayed_work_sync(&vi->refill);


>   		still_empty = !try_fill_recv(vi, rq, GFP_KERNEL);
>   		virtnet_napi_enable(rq->vq, &rq->napi);
>   
> @@ -1523,6 +1535,9 @@ static void virtnet_poll_cleantx(struct receive_queue *rq)
>   	if (!sq->napi.weight || is_xdp_raw_buffer_queue(vi, index))
>   		return;
>   
> +	if (sq->vq->reset)
> +		return;


It looks to me we'd better either WARN or just remove this. Since it 
looks like a workaround for the un-synchronized NAPI somehow.


> +
>   	if (__netif_tx_trylock(txq)) {
>   		do {
>   			virtqueue_disable_cb(sq->vq);
> @@ -1769,6 +1784,98 @@ static netdev_tx_t start_xmit(struct sk_buff *skb, struct net_device *dev)
>   	return NETDEV_TX_OK;
>   }
>   
> +static int virtnet_rx_vq_reset(struct virtnet_info *vi,
> +			       struct receive_queue *rq, u32 ring_num)


It's better to rename this as virtnet_rx_resize().


> +{
> +	int err;
> +
> +	/* stop napi */
> +	napi_disable(&rq->napi);
> +


Here, as discussed above, we need synchronize with the refill work.


> +	/* reset the queue */
> +	err = virtio_reset_vq(rq->vq);
> +	if (err)
> +		goto err;


Btw, most comment of this function seems useless since code already 
explain themselves.


> +
> +	/* free bufs */
> +	virtnet_rq_free_unused_bufs(vi, rq);
> +
> +	/* reset vring. */
> +	err = virtqueue_reset_vring(rq->vq, ring_num);
> +	if (err)
> +		goto err;
> +
> +	/* enable reset queue */
> +	err = virtio_enable_resetq(rq->vq);
> +	if (err)
> +		goto err;
> +
> +	/* fill recv */
> +	if (!try_fill_recv(vi, rq, GFP_KERNEL))
> +		schedule_delayed_work(&vi->refill, 0);
> +
> +	/* enable napi */
> +	virtnet_napi_enable(rq->vq, &rq->napi);
> +	return 0;
> +
> +err:
> +	netdev_err(vi->dev,
> +		   "reset rx reset vq fail: rx queue index: %ld err: %d\n",
> +		   rq - vi->rq, err);
> +	virtnet_napi_enable(rq->vq, &rq->napi);
> +	return err;
> +}
> +
> +static int virtnet_tx_vq_reset(struct virtnet_info *vi,
> +			       struct send_queue *sq, u32 ring_num)
> +{


It looks to me it's better to rename this as "virtnet_rx_resize()"


> +	struct netdev_queue *txq;
> +	int err, qindex;
> +
> +	qindex = sq - vi->sq;
> +
> +	txq = netdev_get_tx_queue(vi->dev, qindex);
> +	__netif_tx_lock_bh(txq);
> +
> +	/* stop tx queue and napi */
> +	netif_stop_subqueue(vi->dev, qindex);
> +	virtnet_napi_tx_disable(&sq->napi);


There's no need to hold tx lock for napi disable.

Thanks


> +
> +	__netif_tx_unlock_bh(txq);
> +
> +	/* reset the queue */
> +	err = virtio_reset_vq(sq->vq);
> +	if (err) {
> +		netif_start_subqueue(vi->dev, qindex);
> +		goto err;
> +	}
> +
> +	/* free bufs */
> +	virtnet_sq_free_unused_bufs(vi, sq);
> +
> +	/* reset vring. */
> +	err = virtqueue_reset_vring(sq->vq, ring_num);
> +	if (err)
> +		goto err;
> +
> +	/* enable reset queue */
> +	err = virtio_enable_resetq(sq->vq);
> +	if (err)
> +		goto err;
> +
> +	/* start tx queue and napi */
> +	netif_start_subqueue(vi->dev, qindex);
> +	virtnet_napi_tx_enable(vi, sq->vq, &sq->napi);
> +	return 0;
> +
> +err:
> +	netdev_err(vi->dev,
> +		   "reset tx reset vq fail: tx queue index: %ld err: %d\n",
> +		   sq - vi->sq, err);
> +	virtnet_napi_tx_enable(vi, sq->vq, &sq->napi);
> +	return err;
> +}
> +
>   /*
>    * Send command via the control virtqueue and check status.  Commands
>    * supported by the hypervisor, as indicated by feature bits, should


_______________________________________________
linux-um mailing list
linux-um@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/linux-um

  reply	other threads:[~2022-03-09  9:14 UTC|newest]

Thread overview: 218+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-03-08 12:34 [PATCH v7 00/26] virtio pci support VIRTIO_F_RING_RESET Xuan Zhuo
2022-03-08 12:34 ` Xuan Zhuo
2022-03-08 12:34 ` Xuan Zhuo
2022-03-08 12:34 ` [PATCH v7 01/26] virtio_pci: struct virtio_pci_common_cfg add queue_notify_data Xuan Zhuo
2022-03-08 12:34   ` Xuan Zhuo
2022-03-08 12:34   ` Xuan Zhuo
2022-03-09  4:48   ` Jason Wang
2022-03-09  4:48     ` Jason Wang
2022-03-09  4:48     ` Jason Wang
2022-03-08 12:34 ` [PATCH v7 02/26] virtio: queue_reset: add VIRTIO_F_RING_RESET Xuan Zhuo
2022-03-08 12:34   ` Xuan Zhuo
2022-03-08 12:34   ` Xuan Zhuo
2022-03-09  6:14   ` Jason Wang
2022-03-09  6:14     ` Jason Wang
2022-03-09  6:14     ` Jason Wang
2022-03-08 12:34 ` [PATCH v7 03/26] virtio: add helper virtqueue_get_vring_max_size() Xuan Zhuo
2022-03-08 12:34   ` Xuan Zhuo
2022-03-08 12:34   ` Xuan Zhuo
2022-03-09  6:16   ` Jason Wang
2022-03-09  6:16     ` Jason Wang
2022-03-09  6:16     ` Jason Wang
2022-03-08 12:34 ` [PATCH v7 04/26] virtio_ring: split: extract the logic of creating vring Xuan Zhuo
2022-03-08 12:34   ` Xuan Zhuo
2022-03-08 12:34   ` Xuan Zhuo
2022-03-09  6:46   ` Jason Wang
2022-03-09  6:46     ` Jason Wang
2022-03-09  6:46     ` Jason Wang
2022-03-09  9:20     ` Xuan Zhuo
2022-03-09  9:20       ` Xuan Zhuo
2022-03-09  9:20       ` Xuan Zhuo
2022-03-08 12:34 ` [PATCH v7 05/26] virtio_ring: split: extract the logic of init vq and attach vring Xuan Zhuo
2022-03-08 12:34   ` Xuan Zhuo
2022-03-08 12:34   ` Xuan Zhuo
2022-03-09  7:36   ` Jason Wang
2022-03-09  7:36     ` Jason Wang
2022-03-09  7:36     ` Jason Wang
2022-03-09  9:21     ` Xuan Zhuo
2022-03-09  9:21       ` Xuan Zhuo
2022-03-09  9:21       ` Xuan Zhuo
2022-03-08 12:34 ` [PATCH v7 06/26] virtio_ring: packed: extract the logic of creating vring Xuan Zhuo
2022-03-08 12:34   ` Xuan Zhuo
2022-03-08 12:34   ` Xuan Zhuo
2022-03-08 12:34 ` [PATCH v7 07/26] virtio_ring: packed: extract the logic of init vq and attach vring Xuan Zhuo
2022-03-08 12:34   ` Xuan Zhuo
2022-03-08 12:34   ` Xuan Zhuo
2022-03-08 12:35 ` [PATCH v7 08/26] virtio_ring: extract the logic of freeing vring Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-09  7:51   ` Jason Wang
2022-03-09  7:51     ` Jason Wang
2022-03-09  7:51     ` Jason Wang
2022-03-09  9:22     ` Xuan Zhuo
2022-03-09  9:22       ` Xuan Zhuo
2022-03-09  9:22       ` Xuan Zhuo
2022-03-08 12:35 ` [PATCH v7 09/26] virtio_ring: split: implement virtqueue_reset_vring_split() Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-09  7:55   ` Jason Wang
2022-03-09  7:55     ` Jason Wang
2022-03-09  7:55     ` Jason Wang
2022-03-09  9:24     ` Xuan Zhuo
2022-03-09  9:24       ` Xuan Zhuo
2022-03-09  9:24       ` Xuan Zhuo
2022-03-10  4:46     ` Xuan Zhuo
2022-03-10  4:46       ` Xuan Zhuo
2022-03-10  4:46       ` Xuan Zhuo
2022-03-11  5:01       ` Jason Wang
2022-03-11  5:01         ` Jason Wang
2022-03-11  5:01         ` Jason Wang
2022-03-10  7:00   ` Michael S. Tsirkin
2022-03-10  7:00     ` Michael S. Tsirkin
2022-03-10  7:00     ` Michael S. Tsirkin
2022-03-10  7:17     ` Xuan Zhuo
2022-03-10  7:17       ` Xuan Zhuo
2022-03-10  7:17       ` Xuan Zhuo
2022-03-10  8:07       ` Michael S. Tsirkin
2022-03-10  8:07         ` Michael S. Tsirkin
2022-03-10  8:07         ` Michael S. Tsirkin
2022-03-10  8:14         ` Xuan Zhuo
2022-03-10  8:14           ` Xuan Zhuo
2022-03-10  8:14           ` Xuan Zhuo
2022-03-10 12:17           ` Michael S. Tsirkin
2022-03-10 12:17             ` Michael S. Tsirkin
2022-03-10 12:17             ` Michael S. Tsirkin
2022-03-10 12:33             ` Xuan Zhuo
2022-03-10 12:33               ` Xuan Zhuo
2022-03-10 12:33               ` Xuan Zhuo
2022-03-10 13:04               ` Michael S. Tsirkin
2022-03-10 13:04                 ` Michael S. Tsirkin
2022-03-10 13:04                 ` Michael S. Tsirkin
2022-03-10 14:09                 ` Xuan Zhuo
2022-03-10 14:09                   ` Xuan Zhuo
2022-03-10 14:09                   ` Xuan Zhuo
2022-03-08 12:35 ` [PATCH v7 10/26] virtio_ring: packed: implement virtqueue_reset_vring_packed() Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35 ` [PATCH v7 11/26] virtio_ring: introduce virtqueue_reset_vring() Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35 ` [PATCH v7 12/26] virtio_ring: update the document of the virtqueue_detach_unused_buf for queue reset Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35 ` [PATCH v7 13/26] virtio: queue_reset: struct virtio_config_ops add callbacks for queue_reset Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-09  8:47   ` Jason Wang
2022-03-09  8:47     ` Jason Wang
2022-03-09  8:47     ` Jason Wang
2022-03-09  9:25     ` Xuan Zhuo
2022-03-09  9:25       ` Xuan Zhuo
2022-03-09  9:25       ` Xuan Zhuo
2022-03-08 12:35 ` [PATCH v7 14/26] virtio: add helper for queue reset Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-09  8:48   ` Jason Wang
2022-03-09  8:48     ` Jason Wang
2022-03-09  8:48     ` Jason Wang
2022-03-09  9:27     ` Xuan Zhuo
2022-03-09  9:27       ` Xuan Zhuo
2022-03-09  9:27       ` Xuan Zhuo
2022-03-08 12:35 ` [PATCH v7 15/26] virtio_pci: queue_reset: update struct virtio_pci_common_cfg and option functions Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35 ` [PATCH v7 16/26] virtio_pci: queue_reset: extract the logic of active vq for modern pci Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35 ` [PATCH v7 17/26] virtio_pci: queue_reset: support VIRTIO_F_RING_RESET Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-09  8:54   ` Jason Wang
2022-03-09  8:54     ` Jason Wang
2022-03-09  8:54     ` Jason Wang
2022-03-09  9:32     ` Xuan Zhuo
2022-03-09  9:32       ` Xuan Zhuo
2022-03-09  9:32       ` Xuan Zhuo
2022-03-11  5:05       ` Jason Wang
2022-03-11  5:05         ` Jason Wang
2022-03-11  5:05         ` Jason Wang
2022-03-10  8:20     ` Xuan Zhuo
2022-03-10  8:20       ` Xuan Zhuo
2022-03-10  8:20       ` Xuan Zhuo
2022-03-11  5:09       ` Jason Wang
2022-03-11  5:09         ` Jason Wang
2022-03-11  5:09         ` Jason Wang
2022-03-08 12:35 ` [PATCH v7 18/26] virtio: find_vqs() add arg sizes Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 13:13   ` Hans de Goede
2022-03-08 13:13     ` Hans de Goede
2022-03-08 13:13     ` Hans de Goede
2022-03-09  8:59   ` Jason Wang
2022-03-09  8:59     ` Jason Wang
2022-03-09  8:59     ` Jason Wang
2022-03-09  9:34     ` Xuan Zhuo
2022-03-09  9:34       ` Xuan Zhuo
2022-03-09  9:34       ` Xuan Zhuo
2022-03-09 18:18   ` Mathieu Poirier
2022-03-09 18:18     ` Mathieu Poirier
2022-03-08 12:35 ` [PATCH v7 19/26] virtio_pci: support the arg sizes of find_vqs() Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35 ` [PATCH v7 20/26] virtio_mmio: " Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35 ` [PATCH v7 21/26] virtio: add helper virtio_find_vqs_ctx_size() Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-09  9:04   ` Jason Wang
2022-03-09  9:04     ` Jason Wang
2022-03-09  9:04     ` Jason Wang
2022-03-09  9:58     ` Xuan Zhuo
2022-03-09  9:58       ` Xuan Zhuo
2022-03-09  9:58       ` Xuan Zhuo
2022-03-08 12:35 ` [PATCH v7 22/26] virtio_net: get ringparam by virtqueue_get_vring_max_size() Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35 ` [PATCH v7 23/26] virtio_net: split free_unused_bufs() Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-09  9:06   ` Jason Wang
2022-03-09  9:06     ` Jason Wang
2022-03-09  9:06     ` Jason Wang
2022-03-08 12:35 ` [PATCH v7 24/26] virtio_net: support rx/tx queue reset Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-09  9:14   ` Jason Wang [this message]
2022-03-09  9:14     ` Jason Wang
2022-03-09  9:14     ` Jason Wang
2022-03-09 10:05     ` Xuan Zhuo
2022-03-09 10:05       ` Xuan Zhuo
2022-03-09 10:05       ` Xuan Zhuo
2022-03-09 10:16     ` Xuan Zhuo
2022-03-09 10:16       ` Xuan Zhuo
2022-03-09 10:16       ` Xuan Zhuo
2022-03-08 12:35 ` [PATCH v7 25/26] virtio_net: set the default max ring size by find_vqs() Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-09  9:28   ` Jason Wang
2022-03-09  9:28     ` Jason Wang
2022-03-09  9:28     ` Jason Wang
2022-03-09  9:35     ` Xuan Zhuo
2022-03-09  9:35       ` Xuan Zhuo
2022-03-09  9:35       ` Xuan Zhuo
2022-03-08 12:35 ` [PATCH v7 26/26] virtio_net: support set_ringparam Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-08 12:35   ` Xuan Zhuo
2022-03-09  9:29   ` Jason Wang
2022-03-09  9:29     ` Jason Wang
2022-03-09  9:29     ` Jason Wang
2022-03-09  9:48     ` Xuan Zhuo
2022-03-09  9:48       ` Xuan Zhuo
2022-03-09  9:48       ` Xuan Zhuo
2022-03-09  4:45 ` [PATCH v7 00/26] virtio pci support VIRTIO_F_RING_RESET Jason Wang
2022-03-09  4:45   ` Jason Wang
2022-03-09  4:45   ` Jason Wang
2022-03-09  9:02   ` Michael S. Tsirkin
2022-03-09  9:02     ` Michael S. Tsirkin
2022-03-09  9:02     ` Michael S. Tsirkin

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=7ff78ff8-bdd0-bb5e-1cea-cf1126226feb@redhat.com \
    --to=jasowang@redhat.com \
    --cc=agordeev@linux.ibm.com \
    --cc=anton.ivanov@cambridgegreys.com \
    --cc=ast@kernel.org \
    --cc=bjorn.andersson@linaro.org \
    --cc=borntraeger@linux.ibm.com \
    --cc=bpf@vger.kernel.org \
    --cc=cohuck@redhat.com \
    --cc=daniel@iogearbox.net \
    --cc=davem@davemloft.net \
    --cc=gor@linux.ibm.com \
    --cc=hawk@kernel.org \
    --cc=hca@linux.ibm.com \
    --cc=hdegoede@redhat.com \
    --cc=jdike@addtoit.com \
    --cc=johannes.berg@intel.com \
    --cc=john.fastabend@gmail.com \
    --cc=kuba@kernel.org \
    --cc=kvm@vger.kernel.org \
    --cc=linux-remoteproc@vger.kernel.org \
    --cc=linux-s390@vger.kernel.org \
    --cc=linux-um@lists.infradead.org \
    --cc=markgross@kernel.org \
    --cc=mathieu.poirier@linaro.org \
    --cc=mst@redhat.com \
    --cc=netdev@vger.kernel.org \
    --cc=pasic@linux.ibm.com \
    --cc=platform-driver-x86@vger.kernel.org \
    --cc=richard@nod.at \
    --cc=svens@linux.ibm.com \
    --cc=vadimp@nvidia.com \
    --cc=vincent.whitchurch@axis.com \
    --cc=virtualization@lists.linux-foundation.org \
    --cc=xuanzhuo@linux.alibaba.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.