From: Dario Faggioli <dario.faggioli@citrix.com>
To: Stefano Stabellini <sstabellini@kernel.org>, xen-devel@lists.xen.org
Cc: george.dunlap@eu.citrix.com, edgar.iglesias@xilinx.com,
julien.grall@arm.com
Subject: Re: Xen on ARM IRQ latency and scheduler overhead
Date: Fri, 17 Feb 2017 19:40:45 +0100 [thread overview]
Message-ID: <1487356845.6732.100.camel@citrix.com> (raw)
In-Reply-To: <alpine.DEB.2.10.1702091603240.20549@sstabellini-ThinkPad-X260>
[-- Attachment #1.1: Type: text/plain, Size: 4169 bytes --]
On Thu, 2017-02-09 at 16:54 -0800, Stefano Stabellini wrote:
> These are the results, in nanosec:
>
> AVG MIN MAX WARM MAX
>
> NODEBUG no WFI 1890 1800 3170 2070
> NODEBUG WFI 4850 4810 7030 4980
> NODEBUG no WFI credit2 2217 2090 3420 2650
> NODEBUG WFI credit2 8080 7890 10320 8300
>
> DEBUG no WFI 2252 2080 3320 2650
> DEBUG WFI 6500 6140 8520 8130
> DEBUG WFI, credit2 8050 7870 10680 8450
>
> As you can see, depending on whether the guest issues a WFI or not
> while
> waiting for interrupts, the results change significantly.
> Interestingly,
> credit2 does worse than credit1 in this area.
>
I did some measuring myself, on x86, with different tools. So,
cyclictest is basically something very very similar to the app
Stefano's app.
I've run it both within Dom0, and inside a guest. I also run a Xen
build (in this case, only inside of the guest).
> We are down to 2000-3000ns. Then, I started investigating the
> scheduler.
> I measured how long it takes to run "vcpu_unblock": 1050ns, which is
> significant. I don't know what is causing the remaining 1000-2000ns,
> but
> I bet on another scheduler function. Do you have any suggestions on
> which one?
>
So, vcpu_unblock() calls vcpu_wake(), which then invokes the
scheduler's wakeup related functions.
If you time vcpu_unblock(), from beginning to end of the function, you
actually capture quite a few things. E.g., the scheduler lock is taken
inside vcpu_wake(), so you're basically including time spent waited on
the lock in the estimation.
That is probably ok (as in, lock contention definitely is something
relevant to latency), but it is expected for things to be rather
different between Credit1 and Credit2.
I've, OTOH, tried to time, SCHED_OP(wake) and SCHED_OP(do_schedule),
and here's the result. Numbers are in cycles (I've used RDTSC) and, for
making sure to obtain consistent and comparable numbers, I've set the
frequency scaling governor to performance.
Dom0, [performance]
cyclictest 1us cyclictest 1ms cyclictest 100ms
(cycles) Credit1 Credit2 Credit1 Credit2 Credit1 Credit2
wakeup-avg 2429 2035 1980 1633 2535 1979
wakeup-max 14577 113682 15153 203136 12285 115164
sched-avg 1716 1860 2527 1651 2286 1670
sched-max 16059 15000 12297 101760 15831 13122
VM, [performance]
cyclictest 1us cyclictest 1ms cyclictest 100ms make -j xen
(cycles) Credit1 Credit2 Credit1 Credit2 Credit1 Credit2 Credit1 Credit2
wakeup-avg 2213 2128 1944 2342 2374 2213 2429 1618
wakeup-max 9990 10104 11262 9927 10290 10218 14430 15108
sched-avg 2437 2472 1620 1594 2498 1759 2449 1809
sched-max 14100 14634 10071 9984 10878 8748 16476 14220
Actually, TSC on this box should be stable and invariant, so I guess I
can try with the default governor. Will do that on Monday. Does ARM
have frequency scaling (I did remember something on xen-devel, but I am
not sure whether it landed upstream)?
But anyway. You're seeing big differences between Credit1 and Credit2,
while I, at least as far as the actual schedulers' code is concerned,
don't.
Credit2 shows higher wakeup-max values, but only in cases where the
workload runs in dom0. But it also shows better (lower) averages, in
both the two kind of workload considered and both in the dom0 and VM
case.
I therefore wonder what is actually responsible for the huge
differences between the two scheduler that you are seeing... could be
lock contention, but with only 4 pCPUs and 2 active vCPUs, I honestly
doubt it...
Regards,
Dario
--
<<This happens because I choose it to happen!>> (Raistlin Majere)
-----------------------------------------------------------------
Dario Faggioli, Ph.D, http://about.me/dario.faggioli
Senior Software Engineer, Citrix Systems R&D Ltd., Cambridge (UK)
[-- Attachment #1.2: This is a digitally signed message part --]
[-- Type: application/pgp-signature, Size: 819 bytes --]
[-- Attachment #2: Type: text/plain, Size: 127 bytes --]
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xen.org
https://lists.xen.org/xen-devel
next prev parent reply other threads:[~2017-02-17 18:40 UTC|newest]
Thread overview: 19+ messages / expand[flat|nested] mbox.gz Atom feed top
2017-02-10 0:54 Xen on ARM IRQ latency and scheduler overhead Stefano Stabellini
2017-02-10 8:40 ` Dario Faggioli
2017-02-10 18:32 ` Stefano Stabellini
2017-02-16 12:20 ` Dario Faggioli
2017-02-16 19:52 ` Stefano Stabellini
2017-02-16 23:07 ` Stefano Stabellini
2017-02-17 11:02 ` Dario Faggioli
2017-02-17 19:34 ` Julien Grall
2017-02-17 23:14 ` Stefano Stabellini
2017-02-18 0:02 ` Stefano Stabellini
2017-02-18 0:47 ` Dario Faggioli
2017-02-17 18:40 ` Dario Faggioli [this message]
2017-02-17 19:44 ` Julien Grall
2017-02-17 22:55 ` Stefano Stabellini
2017-02-18 0:59 ` Dario Faggioli
2017-02-20 12:18 ` Punit Agrawal
2017-02-18 0:41 ` Stefano Stabellini
2017-02-20 11:04 ` George Dunlap
2017-02-20 11:40 ` Dario Faggioli
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1487356845.6732.100.camel@citrix.com \
--to=dario.faggioli@citrix.com \
--cc=edgar.iglesias@xilinx.com \
--cc=george.dunlap@eu.citrix.com \
--cc=julien.grall@arm.com \
--cc=sstabellini@kernel.org \
--cc=xen-devel@lists.xen.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.