All of lore.kernel.org
 help / color / mirror / Atom feed
From: John Weekes <lists.xen@nuclearfallout.net>
To: Daniel Stodden <daniel.stodden@citrix.com>
Cc: Ian Pratt <Ian.Pratt@eu.citrix.com>,
	"xen-devel@lists.xensource.com" <xen-devel@lists.xensource.com>,
	Jan Beulich <JBeulich@novell.com>
Subject: Re: OOM problems
Date: Tue, 16 Nov 2010 11:54:11 -0800	[thread overview]
Message-ID: <4CE2E163.2090809@nuclearfallout.net> (raw)
In-Reply-To: <4CE1751F.9020202@nuclearfallout.net>

Performance is noticeably lower with aio on these bursty write 
workloads; I've been getting a number of complaints.

I see that 2.6.36 has some page_writeback changes:
http://www.kernel.org/diff/diffview.cgi?file=%2Fpub%2Flinux%2Fkernel%2Fv2.6%2Fpatch-2.6.36.bz2;z=8379 
. Any thoughts on whether these would make a difference for the problems 
with "file:"? I'm still trying to find a way to reproduce the issue in 
the lab, so I'd have to test the patch in production -- that's not a 
tantalizing prospect, unless there is a real chance that it will affect it.

-John

On 11/15/2010 9:59 AM, John Weekes wrote:
>
>> They are throttled, but the single control I'm aware of
>> is /proc/sys/vm/dirty_ratio (or dirty_bytes, nowadays). Which is only
>> per process, not a global limit. Could well be that's part of the
>> problem -- outwitting mm with just too many writers on too many cores?
>>
>> We had a bit of trouble when switching dom0 to 2.6.32, buffered writes
>> made it much easier than with e.g. 2.6.27 to drive everybody else into
>> costly reclaims.
>>
>> The Oom shown here reports about ~650M in dirty pages. The fact alone
>> that this counts as on oom condition doesn't sound quite right in
>> itself. That qemu might just have dared to ask at the wrong point in
>> time.
>>
>> Just to get an idea -- how many guests did this box carry?
>
> It carries about two dozen guests, with a mix of mostly HVMs (all 
> stubdom-based, some with PV-on-HVM drivers) and some PV.
>
> This problem occurred more often for me under 2.6.32 than 2.6.31, I 
> noticed. Since I made the switch to aio, I haven't seen a crash, but 
> it hasn't been long enough for that to mean much.
>
> Having extra caching in the dom0 is nice because it allows for domUs 
> to get away with having small amounts of free memory, while still 
> having very good (much faster than hardware) write performance. If you 
> have a large number of domUs that are all memory-constrained and use 
> the disk in infrequent, large bursts, this can work out pretty well, 
> since the big communal pool provides a better value proposition than 
> giving each domU a few more megabytes of RAM.
>
> If the OOM problem isn't something that can be fixed, it might be a 
> good idea to print out a warning to the user when a domain using 
> "file:" is started. Or, to go a step further and automatically run 
> "file" based domains as though "aio" was specified, possibly with a 
> warning and a way to override that behavior. It's not really intuitive 
> that "file" would cause crashes.
>
> -John
>
> _______________________________________________
> Xen-devel mailing list
> Xen-devel@lists.xensource.com
> http://lists.xensource.com/xen-devel

  reply	other threads:[~2010-11-16 19:54 UTC|newest]

Thread overview: 23+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2010-11-13  7:57 OOM problems John Weekes
2010-11-13  8:14 ` Ian Pratt
2010-11-13  8:27   ` John Weekes
2010-11-13  9:13     ` Ian Pratt
2010-11-13  9:43       ` John Weekes
2010-11-13 10:19       ` John Weekes
2010-11-14  9:53         ` Daniel Stodden
2010-11-15  8:55       ` Jan Beulich
2010-11-15  9:40         ` Daniel Stodden
2010-11-15  9:57           ` Jan Beulich
2010-11-15 17:59           ` John Weekes
2010-11-16 19:54             ` John Weekes [this message]
2010-11-17 20:10               ` Ian Pratt
2010-11-17 22:02                 ` John Weekes
2010-11-18  0:56                   ` Ian Pratt
2010-11-18  1:23                   ` Daniel Stodden
2010-11-18  3:29                     ` John Weekes
2010-11-18  4:08                       ` Daniel Stodden
2010-11-18  7:15                         ` John Weekes
2010-11-18 10:41                           ` Daniel Stodden
2010-11-19  7:27                             ` John Weekes
2010-11-15 14:17         ` Stefano Stabellini
2010-11-13 18:15 ` George Shuklin

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=4CE2E163.2090809@nuclearfallout.net \
    --to=lists.xen@nuclearfallout.net \
    --cc=Ian.Pratt@eu.citrix.com \
    --cc=JBeulich@novell.com \
    --cc=daniel.stodden@citrix.com \
    --cc=xen-devel@lists.xensource.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.