linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Jason Wang <jasowang@redhat.com>
To: mst@redhat.com, kvm@vger.kernel.org,
	virtualization@lists.linux-foundation.org,
	netdev@vger.kernel.org, linux-kernel@vger.kernel.org
Cc: Jason Wang <jasowang@redhat.com>
Subject: [PATCH net-next] vhost_net: stop rx net polling when possible
Date: Fri, 15 Aug 2014 11:40:08 +0800	[thread overview]
Message-ID: <1408074008-4357-1-git-send-email-jasowang@redhat.com> (raw)

After rx vq was enabled, we never stop polling its socket. This is sub optimal
when may lead unnecessary wake-ups after the rx net work has already been
queued. This could be optimized by stopping polling the rx net sock when
processing both rx and tx and restart it afterward. This could save unnecessary
wake-ups and even unnecessary spin locks acquiring with the help of commit
9e641bdcfa4ef4d6e2fbaa59c1be0ad5d1551fd5 "net-tun: restructure tun_do_read for
better sleep/wakeup efficiency".

Test shows significant CPU% savings during almost all the cases:

Guest rx stream:
size(B)/sessions/throughput/cpu/normalized thru/
64/1/+0.7773%   -8.6224%     +10.2866%
64/2/+0.6335%   -13.9109%    +16.8946%
64/4/-0.8182%   -14.8336%    +16.4565%
64/8/+0.4830%   -13.7675%    +16.5256%
256/1/-7.0963%  -12.6880%    +6.4043%
256/2/-1.3982%  -11.5424%    +11.4678%
256/4/-0.0350%  -11.8323%    +13.3806%
256/8/-1.5830%  -12.7693%    +12.8238%
1024/1/-7.4895% -19.1449%       +14.4152%
1024/2/-7.4575% -19.4018%       +14.8195%
1024/4/-0.3881% -9.1183%        +9.6061%
1024/8/+0.4713% -11.0155%       +12.9087%
4096/1/+0.8786%  -8.4050%        +10.1355%
4096/2/+0.0098%  -15.3094%       +18.0885%
4096/4/+0.0445%  -10.8247%       +12.1886%
4096/8/-2.1317%  -12.5111%       +11.8637%
16384/1/-0.0008% -6.1891%        +6.5966%
16384/2/-0.0117% -16.2716%       +19.4198%
16384/4/+0.0001% -5.9197%        +6.2923%
16384/8/+0.0173% -7.6681%        +8.3236%
65535/1/+0.0011% -10.3594%       +11.5578%
65535/2/-0.4108%  -14.4304%       +16.3838%
65535/4/+0.0011%  -10.3594%       +11.5578%
65535/8/-0.4108%  -14.4304%       +16.3838%

Guest tx stream:
size(B)/sessions/throughput/cpu/normalized thru/
64/1/-0.6228%     -2.1936%     +1.6060%
64/2/+0.8646%     -3.5063%     +4.5297%
64/4/+0.8733%     -3.2495%     +4.2613%
64/8/+1.4290%     -3.5593%     +5.1724%
256/1/+7.2098%    -3.1122%     +10.6535%
256/2/-10.1408%   -6.8230%     -3.5607%
256/4/-11.3531%   -6.7085%     -4.9785%
256/8/-10.2723%   -6.5628%     -3.9701%
1024/1/-18.9329%  -13.6162%    -6.1547%
1024/2/-0.3728%   -1.3181%     +0.9580%
1024/4/+0.0125%   -3.6338%     +3.7838%
1024/8/-0.0030%   -2.7282%     +2.8017%
4096/1/+16.9367%  -1.9435%     +19.2543%
4096/2/+0.0121%   -6.1682%     +6.5866%
4096/4/+0.0019%   -3.8510%     +4.0072%
4096/8/-0.0222%   -4.1368%     +4.2922%
16384/1/-0.0026%  -8.6892%     +9.5132%
16384/2/-0.0012%  -10.1676%    +11.3171%
16384/4/+0.0196%  -1.2551%     +1.2908%
16384/8/+0.1303%  -3.2634%     +3.5082%
65535/1/+0.0019%  -3.4694%     +3.5961%
65535/2/-0.0003%  -0.7635%     +0.7690%
65535/4/-0.0219%  -2.7875%     +2.8448%
65535/8/+0.1137%  -2.7922%     +2.9894%

TCP_RR:
size(B)/sessions/throughput/cpu/normalized thru/
256/1/+1.9004%    -4.7985%     +7.0366%
256/25/-4.7366%   -11.0809%    +7.1349%
256/50/+3.9808%   -5.2037%     +9.6887%
4096/1/+2.1619%   -0.7303%     +2.9134%
4096/25/-13.1836% -14.7298%    +1.8134%
4096/50/-11.1990% -15.4763%    +5.0605%

Signed-off-by: Jason Wang <jasowang@redhat.com>
---
 drivers/vhost/net.c | 26 +++++++++++++++++++++-----
 1 file changed, 21 insertions(+), 5 deletions(-)

diff --git a/drivers/vhost/net.c b/drivers/vhost/net.c
index 8dae2f7..d4a9742 100644
--- a/drivers/vhost/net.c
+++ b/drivers/vhost/net.c
@@ -334,6 +334,8 @@ static void vhost_zerocopy_callback(struct ubuf_info *ubuf, bool success)
 static void handle_tx(struct vhost_net *net)
 {
 	struct vhost_net_virtqueue *nvq = &net->vqs[VHOST_NET_VQ_TX];
+	struct vhost_virtqueue *rx_vq = &net->vqs[VHOST_NET_VQ_RX].vq;
+	struct vhost_poll *rx_poll = &net->poll[VHOST_NET_VQ_RX];
 	struct vhost_virtqueue *vq = &nvq->vq;
 	unsigned out, in, s;
 	int head;
@@ -348,15 +350,18 @@ static void handle_tx(struct vhost_net *net)
 	size_t len, total_len = 0;
 	int err;
 	size_t hdr_size;
-	struct socket *sock;
+	struct socket *sock, *rxsock;
 	struct vhost_net_ubuf_ref *uninitialized_var(ubufs);
-	bool zcopy, zcopy_used;
+	bool zcopy, zcopy_used, poll = false;
 
 	mutex_lock(&vq->mutex);
+	mutex_lock(&rx_vq->mutex);
 	sock = vq->private_data;
+	rxsock = rx_vq->private_data;
 	if (!sock)
 		goto out;
 
+	vhost_poll_stop(rx_poll);
 	vhost_disable_notify(&net->dev, vq);
 
 	hdr_size = nvq->vhost_hlen;
@@ -451,11 +456,17 @@ static void handle_tx(struct vhost_net *net)
 		total_len += len;
 		vhost_net_tx_packet(net);
 		if (unlikely(total_len >= VHOST_NET_WEIGHT)) {
-			vhost_poll_queue(&vq->poll);
+			poll = true;
 			break;
 		}
 	}
+
+	if (rxsock)
+		vhost_poll_start(rx_poll, rxsock->file);
+	if (poll)
+		vhost_poll_queue(&vq->poll);
 out:
+	mutex_unlock(&rx_vq->mutex);
 	mutex_unlock(&vq->mutex);
 }
 
@@ -554,6 +565,7 @@ err:
 static void handle_rx(struct vhost_net *net)
 {
 	struct vhost_net_virtqueue *nvq = &net->vqs[VHOST_NET_VQ_RX];
+	struct vhost_poll *poll = &net->poll[VHOST_NET_VQ_RX];
 	struct vhost_virtqueue *vq = &nvq->vq;
 	unsigned uninitialized_var(in), log;
 	struct vhost_log *vq_log;
@@ -580,6 +592,8 @@ static void handle_rx(struct vhost_net *net)
 	sock = vq->private_data;
 	if (!sock)
 		goto out;
+
+	vhost_poll_stop(poll);
 	vhost_disable_notify(&net->dev, vq);
 
 	vhost_hlen = nvq->vhost_hlen;
@@ -660,10 +674,12 @@ static void handle_rx(struct vhost_net *net)
 			vhost_log_write(vq, vq_log, log, vhost_len);
 		total_len += vhost_len;
 		if (unlikely(total_len >= VHOST_NET_WEIGHT)) {
-			vhost_poll_queue(&vq->poll);
-			break;
+			vhost_poll_queue(poll);
+			goto out;
 		}
 	}
+
+	vhost_poll_start(poll, sock->file);
 out:
 	mutex_unlock(&vq->mutex);
 }
-- 
1.9.1


             reply	other threads:[~2014-08-15  3:40 UTC|newest]

Thread overview: 3+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2014-08-15  3:40 Jason Wang [this message]
2014-08-17 10:20 ` [PATCH net-next] vhost_net: stop rx net polling when possible Michael S. Tsirkin
2014-08-18  3:11   ` Jason Wang

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=1408074008-4357-1-git-send-email-jasowang@redhat.com \
    --to=jasowang@redhat.com \
    --cc=kvm@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=mst@redhat.com \
    --cc=netdev@vger.kernel.org \
    --cc=virtualization@lists.linux-foundation.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).