From: Vlastimil Babka <vbabka@suse.cz>
To: Thomas Jarosch <thomas.jarosch@intra2net.com>,
Johannes Weiner <hannes@cmpxchg.org>
Cc: Andrew Morton <akpm@linux-foundation.org>,
Linus Torvalds <torvalds@linux-foundation.org>,
bugzilla-daemon@bugzilla.kernel.org, linux-mm@kvack.org
Subject: Re: [Bug 64121] New: [BISECTED] "mm" performance regression updating from 3.2 to 3.3
Date: Thu, 21 Jul 2016 16:02:06 +0200 [thread overview]
Message-ID: <b3219832-110d-2b74-5ba9-694ab30589f0@suse.cz> (raw)
In-Reply-To: <b4aff3a2-cc22-c68c-cafc-96db332f86c3@intra2net.com>
On 07/19/2016 12:23 AM, Thomas Jarosch wrote:
> Hi Johannes,
>
> referring to an old kernel bugzilla issue:
> https://bugzilla.kernel.org/show_bug.cgi?id=64121
>
> Am 01.11.2013 um 19:43 wrote Johannes Weiner:
>> It is a combination of two separate things on these setups.
>>
>> Traditionally, only lowmem is considered dirtyable so that dirty pages
>> don't scale with highmem and the kernel doesn't overburden itself with
>> lowmem pressure from buffers etc. This is purely about accounting.
>>
>> My patches on the other hand were about dirty page placement and
>> avoiding writeback from page reclaim: by subtracting the watermark and
>> the lowmem reserve (memory not available for user memory / cache) from
>> each zone's dirtyable memory, we make sure that the zone can always be
>> rebalanced without writeback.
>>
>> The problem now is that the lowmem reserves scale with highmem and
>> there is a point where they entirely overshadow the Normal zone. This
>> means that no page cache at all is allowed in lowmem. Combine this
>> with how dirtyable memory excludes highmem, and the sum of all
>> dirtyable memory is nil. This effectively disables the writeback
>> cache.
>>
>> I figure if anything should be fixed it should be the full exclusion
>> of highmem from dirtyable memory and find a better way to calculate a
>> minimum.
>
> recently we've updated our production mail server from 3.14.69
> to 3.14.73 and it worked fine for a few days. When the box is really
> busy (=incoming malware via email), the I/O speed drops to crawl,
> write speed is about 5 MB/s on Intel SSDs. Yikes.
>
> The box has 16GB RAM, so it should be a safe HIGHMEM configuration.
>
> Downgrading to 3.14.69 or booting with "mem=15000M" works. I've tested
> both approaches and the box was stable. Booting 3.14.73 again triggered
> the problem within minutes.
>
> Clearly something with the automatic calculation of the lowmem reserve
> crossed a tipping point again, even with the previously considered safe
> amount of 16GB RAM for HIGHMEM configs. I don't see anything obvious in
> the changelogs from 3.14.69 to 3.14.73, but I might have missed it.
I don't see anything either, might be some change e.g. under fs/ though.
How about git bisect?
>> HOWEVER,
>>
>> the lowmem reserve is highmem/32 per default. With a Normal zone of
>> around 900M, this requires 28G+ worth of HighMem to eclipse lowmem
>> entirely. This is almost double of what you consider still okay...
>
> is there a way to read out the calculated lowmem reserve via /proc?
Probably not, but might be possible with live crash session.
> It might be interesting the see the lowmem reserve
> when booted with mem=15000M or kernel 3.14.69 for comparison.
>
> Do you think it might be worth tinkering with "lowmem_reserve_ratio"?
>
>
> /proc/meminfo from the box using "mem=15000M" + kernel 3.14.73:
>
> MemTotal: 15001512 kB
> HighTotal: 14219160 kB
> HighFree: 9468936 kB
> LowTotal: 782352 kB
> LowFree: 117696 kB
> Slab: 430612 kB
> SReclaimable: 416752 kB
> SUnreclaim: 13860 kB
>
>
> /proc/meminfo from a similar machine with 16GB RAM + kernel 3.14.73:
> (though that machine is just a firewall, so no real disk I/O)
>
> MemTotal: 16407652 kB
> HighTotal: 15636376 kB
> HighFree: 14415472 kB
> LowTotal: 771276 kB
> LowFree: 562852 kB
> Slab: 34712 kB
> SReclaimable: 20888 kB
> SUnreclaim: 13824 kB
>
>
> Any help is appreciated,
> Thomas
>
> --
> To unsubscribe, send a message with 'unsubscribe linux-mm' in
> the body to majordomo@kvack.org. For more info on Linux MM,
> see: http://www.linux-mm.org/ .
> Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
>
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
next prev parent reply other threads:[~2016-07-21 14:02 UTC|newest]
Thread overview: 9+ messages / expand[flat|nested] mbox.gz Atom feed top
[not found] <bug-64121-27@https.bugzilla.kernel.org/>
2013-10-31 20:46 ` [Bug 64121] New: [BISECTED] "mm" performance regression updating from 3.2 to 3.3 Andrew Morton
2013-11-01 18:43 ` Johannes Weiner
2013-11-04 11:32 ` Thomas Jarosch
2016-07-18 22:23 ` Thomas Jarosch
2016-07-21 14:02 ` Vlastimil Babka [this message]
2016-07-27 9:18 ` Thomas Jarosch
2016-07-27 9:21 ` Thomas Jarosch
2016-07-27 16:44 ` Linus Torvalds
2016-07-29 17:00 ` Thomas Jarosch
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=b3219832-110d-2b74-5ba9-694ab30589f0@suse.cz \
--to=vbabka@suse.cz \
--cc=akpm@linux-foundation.org \
--cc=bugzilla-daemon@bugzilla.kernel.org \
--cc=hannes@cmpxchg.org \
--cc=linux-mm@kvack.org \
--cc=thomas.jarosch@intra2net.com \
--cc=torvalds@linux-foundation.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).