From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1757329Ab2IDOkQ (ORCPT ); Tue, 4 Sep 2012 10:40:16 -0400 Received: from mx1.redhat.com ([209.132.183.28]:56434 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1757152Ab2IDOkN (ORCPT ); Tue, 4 Sep 2012 10:40:13 -0400 Date: Tue, 4 Sep 2012 17:41:31 +0300 From: "Michael S. Tsirkin" To: Paolo Bonzini Cc: linux-kernel@vger.kernel.org, linux-scsi@vger.kernel.org, kvm@vger.kernel.org, rusty@rustcorp.com.au, jasowang@redhat.com, virtualization@lists.linux-foundation.org Subject: Re: [PATCH 5/5] virtio-scsi: introduce multiqueue support Message-ID: <20120904144131.GN9805@redhat.com> References: <1346154857-12487-1-git-send-email-pbonzini@redhat.com> <1346154857-12487-6-git-send-email-pbonzini@redhat.com> <20120904124800.GE9805@redhat.com> <504606F6.4080603@redhat.com> <20120904142154.GL9805@redhat.com> <5046108B.8030609@redhat.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <5046108B.8030609@redhat.com> Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, Sep 04, 2012 at 04:30:35PM +0200, Paolo Bonzini wrote: > Il 04/09/2012 16:21, Michael S. Tsirkin ha scritto: > > > One reason is that, even though in practice I expect roughly the same > > > number of targets and VCPUs, hotplug means the number of targets is > > > difficult to predict and is usually fixed to 256. > > > > > > The other reason is that per-target vq didn't give any performance > > > advantage. The bonus comes from cache locality and less process > > > migrations, more than from the independent virtqueues. > > > > Okay, and why is per-target worse for cache locality? > > Because per-target doesn't have IRQ affinity for a particular CPU. > > Assuming that the thread that is sending requests to the device is > I/O-bound, it is likely to be sleeping at the time the ISR is executed, > and thus executing the ISR on the same processor that sent the requests > is cheap. > > But if you have many such I/O-bound processes, the kernel will execute > the ISR on a random processor, rather than the one that is sending > requests to the device. > > Paolo I see, another case where our irq balancing makes bad decisions. You could do it differently - pin irq to the cpu of the last task that executed, tweak irq affinity when that changes. Still if you want to support 256 targets vector per target is not going to work. Would be nice to add this motivation to commit log I think. -- MST