From: Jingbai Ma <jingbai.ma@hp.com> To: HATAYAMA Daisuke <d.hatayama@jp.fujitsu.com> Cc: jingbai.ma@hp.com, vgoyal@redhat.com, ebiederm@xmission.com, cpw@sgi.com, kumagai-atsushi@mxc.nes.nec.co.jp, lisa.mitchell@hp.com, akpm@linux-foundation.org, kexec@lists.infradead.org, linux-kernel@vger.kernel.org Subject: Re: makedumpfile mmap() benchmark Date: Wed, 27 Mar 2013 14:35:17 +0800 [thread overview] Message-ID: <51529325.9070207@hp.com> (raw) In-Reply-To: <20130327.152344.338429402.d.hatayama@jp.fujitsu.com> On 03/27/2013 02:23 PM, HATAYAMA Daisuke wrote: > From: Jingbai Ma<jingbai.ma@hp.com> > Subject: makedumpfile mmap() benchmark > Date: Wed, 27 Mar 2013 13:51:37 +0800 > >> Hi, >> >> I have tested the makedumpfile mmap patch on a machine with 2TB >> memory, here is testing results: > > Thanks for your benchmark. It's very helpful to see the benchmark on > different environments. Thanks for your patch, there is a great performance improvement, very impressive! > >> Test environment: >> Machine: HP ProLiant DL980 G7 with 2TB RAM. >> CPU: Intel(R) Xeon(R) CPU E7- 2860 @ 2.27GHz (8 sockets, 10 cores) >> (Only 1 cpu was enabled the 2nd kernel) >> Kernel: 3.9.0-rc3+ with mmap kernel patch v3 >> vmcore size: 2.0TB >> Dump file size: 3.6GB >> makedumpfile mmap branch with parameters: -c --message-level 23 -d 31 >> --map-size<map-size> > > To reduce the benchmark time, I recommend LZO or snappy compressions > rather than zlib. zlib is used when -c option is specified, and it's > too slow for use of crash dump. That's a very helpful suggestion, I will try it again with LZO/snappy lib again. > > To build makedumpfile with each compression format supports, do > USELZO=on or USESNAPPY=on after installing necessary libraries. > >> All measured time from debug message of makedumpfile. >> >> As a comparison, I also have tested with original kernel and original >> makedumpfile 1.5.1 and 1.5.3. >> I added all [Excluding unnecessary pages] and [Excluding free pages] >> time together as "Filter Pages", and [Copyying Data] as "Copy data" >> here. >> >> makedumjpfile Kernel map-size (KB) Filter pages (s) Copy data (s) >> Total (s) >> 1.5.1 3.7.0-0.36.el7.x86_64 N/A 940.28 1269.25 2209.53 >> 1.5.3 3.7.0-0.36.el7.x86_64 N/A 380.09 992.77 1372.86 >> 1.5.3 v3.9-rc3 N/A 197.77 892.27 1090.04 >> 1.5.3+mmap v3.9-rc3+mmap 0 164.87 606.06 770.93 >> 1.5.3+mmap v3.9-rc3+mmap 4 88.62 576.07 664.69 >> 1.5.3+mmap v3.9-rc3+mmap 1024 83.66 477.23 560.89 >> 1.5.3+mmap v3.9-rc3+mmap 2048 83.44 477.21 560.65 >> 1.5.3+mmap v3.9-rc3+mmap 10240 83.84 476.56 560.4 > > Did you calculate "Filter pages" by adding two [Excluding unnecessary > pages] lines? The first one of the two line is displayed by > get_num_dumpable_cyclic() during the calculation of the total number > of dumpable pages, which is later used to print progress of writing > pages in percentage. > > For example, here is the log, where the number of cycles is 3, and > > mem_map (16399) > mem_map : ffffea0801e00000 > pfn_start : 20078000 > pfn_end : 20080000 > read /proc/vmcore with mmap() > STEP [Excluding unnecessary pages] : 13.703842 seconds<-- this part is by get_num_dumpable_cyclic() > STEP [Excluding unnecessary pages] : 13.842656 seconds > STEP [Excluding unnecessary pages] : 6.857910 seconds > STEP [Excluding unnecessary pages] : 13.554281 seconds<-- this part is by the main filtering processing. > STEP [Excluding unnecessary pages] : 14.103593 seconds > STEP [Excluding unnecessary pages] : 7.114239 seconds > STEP [Copying data ] : 138.442116 seconds > Writing erase info... > offset_eraseinfo: 1f4680e40, size_eraseinfo: 0 > > Original pages : 0x000000001ffc28a4 > <cut> > > So, get_num_dumpable_cyclic() actually does filtering operation but it > should not be included here. > > If so, I guess each measured time would be about 42 seconds, right? > Then, it's almost same as the result I posted today: 35 seconds. Yes, I added them together, the following is one dump message log: <Log> makedumpfile -c --message-level 23 -d 31 --map-size 10240 /proc/vmcore /sysroot/var/crash/vmcore_10240 cyclic buffer size has been changed: 77661798 => 77661184 Excluding unnecessary pages : [100 %] STEP [Excluding unnecessary pages] : 24.777717 seconds Excluding unnecessary pages : [100 %] STEP [Excluding unnecessary pages] : 17.291935 seconds Excluding unnecessary pages : [100 %] STEP [Excluding unnecessary pages] : 24.498559 seconds Excluding unnecessary pages : [100 %] STEP [Excluding unnecessary pages] : 17.278414 seconds Copying data : [100 %] STEP [Copying data ] : 476.563428 seconds Original pages : 0x000000001ffe874d Excluded pages : 0x000000001f79429e Pages filled with zero : 0x00000000002b4c9c Cache pages : 0x00000000000493bc Cache pages + private : 0x00000000000011f3 User process data pages : 0x0000000000005c55 Free pages : 0x000000001f48f3fe Hwpoison pages : 0x0000000000000000 Remaining pages : 0x00000000008544af (The number of pages is reduced to 1%.) Memory Hole : 0x000000001c0178b3 -------------------------------------------------- Total pages : 0x000000003c000000 </Log> > > Thanks. > HATAYAMA, Daisuke > -- Thanks, Jingbai Ma
WARNING: multiple messages have this Message-ID (diff)
From: Jingbai Ma <jingbai.ma@hp.com> To: HATAYAMA Daisuke <d.hatayama@jp.fujitsu.com> Cc: kexec@lists.infradead.org, linux-kernel@vger.kernel.org, lisa.mitchell@hp.com, kumagai-atsushi@mxc.nes.nec.co.jp, jingbai.ma@hp.com, akpm@linux-foundation.org, cpw@sgi.com, vgoyal@redhat.com, ebiederm@xmission.com Subject: Re: makedumpfile mmap() benchmark Date: Wed, 27 Mar 2013 14:35:17 +0800 [thread overview] Message-ID: <51529325.9070207@hp.com> (raw) In-Reply-To: <20130327.152344.338429402.d.hatayama@jp.fujitsu.com> On 03/27/2013 02:23 PM, HATAYAMA Daisuke wrote: > From: Jingbai Ma<jingbai.ma@hp.com> > Subject: makedumpfile mmap() benchmark > Date: Wed, 27 Mar 2013 13:51:37 +0800 > >> Hi, >> >> I have tested the makedumpfile mmap patch on a machine with 2TB >> memory, here is testing results: > > Thanks for your benchmark. It's very helpful to see the benchmark on > different environments. Thanks for your patch, there is a great performance improvement, very impressive! > >> Test environment: >> Machine: HP ProLiant DL980 G7 with 2TB RAM. >> CPU: Intel(R) Xeon(R) CPU E7- 2860 @ 2.27GHz (8 sockets, 10 cores) >> (Only 1 cpu was enabled the 2nd kernel) >> Kernel: 3.9.0-rc3+ with mmap kernel patch v3 >> vmcore size: 2.0TB >> Dump file size: 3.6GB >> makedumpfile mmap branch with parameters: -c --message-level 23 -d 31 >> --map-size<map-size> > > To reduce the benchmark time, I recommend LZO or snappy compressions > rather than zlib. zlib is used when -c option is specified, and it's > too slow for use of crash dump. That's a very helpful suggestion, I will try it again with LZO/snappy lib again. > > To build makedumpfile with each compression format supports, do > USELZO=on or USESNAPPY=on after installing necessary libraries. > >> All measured time from debug message of makedumpfile. >> >> As a comparison, I also have tested with original kernel and original >> makedumpfile 1.5.1 and 1.5.3. >> I added all [Excluding unnecessary pages] and [Excluding free pages] >> time together as "Filter Pages", and [Copyying Data] as "Copy data" >> here. >> >> makedumjpfile Kernel map-size (KB) Filter pages (s) Copy data (s) >> Total (s) >> 1.5.1 3.7.0-0.36.el7.x86_64 N/A 940.28 1269.25 2209.53 >> 1.5.3 3.7.0-0.36.el7.x86_64 N/A 380.09 992.77 1372.86 >> 1.5.3 v3.9-rc3 N/A 197.77 892.27 1090.04 >> 1.5.3+mmap v3.9-rc3+mmap 0 164.87 606.06 770.93 >> 1.5.3+mmap v3.9-rc3+mmap 4 88.62 576.07 664.69 >> 1.5.3+mmap v3.9-rc3+mmap 1024 83.66 477.23 560.89 >> 1.5.3+mmap v3.9-rc3+mmap 2048 83.44 477.21 560.65 >> 1.5.3+mmap v3.9-rc3+mmap 10240 83.84 476.56 560.4 > > Did you calculate "Filter pages" by adding two [Excluding unnecessary > pages] lines? The first one of the two line is displayed by > get_num_dumpable_cyclic() during the calculation of the total number > of dumpable pages, which is later used to print progress of writing > pages in percentage. > > For example, here is the log, where the number of cycles is 3, and > > mem_map (16399) > mem_map : ffffea0801e00000 > pfn_start : 20078000 > pfn_end : 20080000 > read /proc/vmcore with mmap() > STEP [Excluding unnecessary pages] : 13.703842 seconds<-- this part is by get_num_dumpable_cyclic() > STEP [Excluding unnecessary pages] : 13.842656 seconds > STEP [Excluding unnecessary pages] : 6.857910 seconds > STEP [Excluding unnecessary pages] : 13.554281 seconds<-- this part is by the main filtering processing. > STEP [Excluding unnecessary pages] : 14.103593 seconds > STEP [Excluding unnecessary pages] : 7.114239 seconds > STEP [Copying data ] : 138.442116 seconds > Writing erase info... > offset_eraseinfo: 1f4680e40, size_eraseinfo: 0 > > Original pages : 0x000000001ffc28a4 > <cut> > > So, get_num_dumpable_cyclic() actually does filtering operation but it > should not be included here. > > If so, I guess each measured time would be about 42 seconds, right? > Then, it's almost same as the result I posted today: 35 seconds. Yes, I added them together, the following is one dump message log: <Log> makedumpfile -c --message-level 23 -d 31 --map-size 10240 /proc/vmcore /sysroot/var/crash/vmcore_10240 cyclic buffer size has been changed: 77661798 => 77661184 Excluding unnecessary pages : [100 %] STEP [Excluding unnecessary pages] : 24.777717 seconds Excluding unnecessary pages : [100 %] STEP [Excluding unnecessary pages] : 17.291935 seconds Excluding unnecessary pages : [100 %] STEP [Excluding unnecessary pages] : 24.498559 seconds Excluding unnecessary pages : [100 %] STEP [Excluding unnecessary pages] : 17.278414 seconds Copying data : [100 %] STEP [Copying data ] : 476.563428 seconds Original pages : 0x000000001ffe874d Excluded pages : 0x000000001f79429e Pages filled with zero : 0x00000000002b4c9c Cache pages : 0x00000000000493bc Cache pages + private : 0x00000000000011f3 User process data pages : 0x0000000000005c55 Free pages : 0x000000001f48f3fe Hwpoison pages : 0x0000000000000000 Remaining pages : 0x00000000008544af (The number of pages is reduced to 1%.) Memory Hole : 0x000000001c0178b3 -------------------------------------------------- Total pages : 0x000000003c000000 </Log> > > Thanks. > HATAYAMA, Daisuke > -- Thanks, Jingbai Ma _______________________________________________ kexec mailing list kexec@lists.infradead.org http://lists.infradead.org/mailman/listinfo/kexec
next prev parent reply other threads:[~2013-03-27 6:35 UTC|newest] Thread overview: 70+ messages / expand[flat|nested] mbox.gz Atom feed top 2013-02-14 10:11 [PATCH 00/13] kdump, vmcore: support mmap() on /proc/vmcore HATAYAMA Daisuke 2013-02-14 10:11 ` HATAYAMA Daisuke 2013-02-14 10:11 ` [PATCH 01/13] vmcore: allocate buffer for ELF headers on page-size alignment HATAYAMA Daisuke 2013-02-14 10:11 ` HATAYAMA Daisuke 2013-02-15 15:01 ` Vivek Goyal 2013-02-15 15:01 ` Vivek Goyal 2013-02-14 10:11 ` [PATCH 02/13] vmcore: round up buffer size of ELF headers by PAGE_SIZE HATAYAMA Daisuke 2013-02-14 10:11 ` HATAYAMA Daisuke 2013-02-15 15:18 ` Vivek Goyal 2013-02-15 15:18 ` Vivek Goyal 2013-02-18 15:58 ` HATAYAMA Daisuke 2013-02-18 15:58 ` HATAYAMA Daisuke 2013-02-14 10:11 ` [PATCH 03/13] vmcore: fill unused part of buffer for ELF headers with 0 HATAYAMA Daisuke 2013-02-14 10:11 ` HATAYAMA Daisuke 2013-02-14 10:12 ` [PATCH 04/13] vmcore: introduce types for objects copied in 2nd kernel HATAYAMA Daisuke 2013-02-14 10:12 ` HATAYAMA Daisuke 2013-02-15 15:28 ` Vivek Goyal 2013-02-15 15:28 ` Vivek Goyal 2013-02-18 16:06 ` HATAYAMA Daisuke 2013-02-18 16:06 ` HATAYAMA Daisuke 2013-02-19 23:07 ` Vivek Goyal 2013-02-19 23:07 ` Vivek Goyal 2013-02-14 10:12 ` [PATCH 05/13] vmcore: modify ELF32 code according to new type HATAYAMA Daisuke 2013-02-14 10:12 ` HATAYAMA Daisuke 2013-02-15 15:30 ` Vivek Goyal 2013-02-15 15:30 ` Vivek Goyal 2013-02-18 16:11 ` HATAYAMA Daisuke 2013-02-18 16:11 ` HATAYAMA Daisuke 2013-02-14 10:12 ` [PATCH 06/13] vmcore: modify vmcore clean-up function to free buffer on 2nd kernel HATAYAMA Daisuke 2013-02-14 10:12 ` HATAYAMA Daisuke 2013-02-15 15:32 ` Vivek Goyal 2013-02-15 15:32 ` Vivek Goyal 2013-02-14 10:12 ` [PATCH 07/13] vmcore: modify read_vmcore() to read " HATAYAMA Daisuke 2013-02-14 10:12 ` HATAYAMA Daisuke 2013-02-15 15:51 ` Vivek Goyal 2013-02-15 15:51 ` Vivek Goyal 2013-02-14 10:12 ` [PATCH 08/13] vmcore: remove unused helper function HATAYAMA Daisuke 2013-02-14 10:12 ` HATAYAMA Daisuke 2013-02-15 15:52 ` Vivek Goyal 2013-02-15 15:52 ` Vivek Goyal 2013-02-14 10:12 ` [PATCH 09/13] vmcore: copy ELF note segments in buffer on 2nd kernel HATAYAMA Daisuke 2013-02-14 10:12 ` HATAYAMA Daisuke 2013-02-15 16:53 ` Vivek Goyal 2013-02-15 16:53 ` Vivek Goyal 2013-02-18 17:02 ` HATAYAMA Daisuke 2013-02-18 17:02 ` HATAYAMA Daisuke 2013-02-19 23:05 ` Vivek Goyal 2013-02-19 23:05 ` Vivek Goyal 2013-02-14 10:12 ` [PATCH 10/13] vmcore: round-up offset of vmcore object in page-size boundary HATAYAMA Daisuke 2013-02-14 10:12 ` HATAYAMA Daisuke 2013-02-14 10:12 ` [PATCH 11/13] vmcore: count holes generated by round-up operation for vmcore size HATAYAMA Daisuke 2013-02-14 10:12 ` HATAYAMA Daisuke 2013-02-14 10:12 ` [PATCH 12/13] vmcore: copy non page-size aligned head and tail pages in 2nd kernel HATAYAMA Daisuke 2013-02-14 10:12 ` HATAYAMA Daisuke 2013-02-14 10:12 ` [PATCH 13/13] vmcore: introduce mmap_vmcore() HATAYAMA Daisuke 2013-02-14 10:12 ` HATAYAMA Daisuke 2013-02-15 3:57 ` [PATCH 00/13] kdump, vmcore: support mmap() on /proc/vmcore Atsushi Kumagai 2013-02-15 3:57 ` Atsushi Kumagai 2013-02-18 0:16 ` Hatayama, Daisuke 2013-02-18 0:16 ` Hatayama, Daisuke 2013-03-27 5:51 ` makedumpfile mmap() benchmark Jingbai Ma 2013-03-27 5:51 ` Jingbai Ma 2013-03-27 6:23 ` HATAYAMA Daisuke 2013-03-27 6:23 ` HATAYAMA Daisuke 2013-03-27 6:35 ` Jingbai Ma [this message] 2013-03-27 6:35 ` Jingbai Ma 2013-05-03 19:10 Cliff Wickman 2013-05-03 19:10 ` Cliff Wickman 2013-05-07 8:47 ` HATAYAMA Daisuke 2013-05-07 8:47 ` HATAYAMA Daisuke
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=51529325.9070207@hp.com \ --to=jingbai.ma@hp.com \ --cc=akpm@linux-foundation.org \ --cc=cpw@sgi.com \ --cc=d.hatayama@jp.fujitsu.com \ --cc=ebiederm@xmission.com \ --cc=kexec@lists.infradead.org \ --cc=kumagai-atsushi@mxc.nes.nec.co.jp \ --cc=linux-kernel@vger.kernel.org \ --cc=lisa.mitchell@hp.com \ --cc=vgoyal@redhat.com \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: linkBe sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.