xen-devel.lists.xenproject.org archive mirror
 help / color / mirror / Atom feed
* [linux-3.18 test] 97278: regressions - FAIL
@ 2016-07-14  5:03 osstest service owner
  2016-07-14 12:56 ` Linux 3.18 and Linux 4.1 reproducible OOM crashes Ian Jackson
  0 siblings, 1 reply; 7+ messages in thread
From: osstest service owner @ 2016-07-14  5:03 UTC (permalink / raw)
  To: xen-devel, osstest-admin

[-- Attachment #1: Type: text/plain, Size: 23785 bytes --]

flight 97278 linux-3.18 real [real]
http://logs.test-lab.xenproject.org/osstest/logs/97278/

Regressions :-(

Tests which did not succeed and are blocking,
including tests which could not be run:
 test-amd64-amd64-xl-qemut-debianhvm-amd64  6 xen-boot     fail REGR. vs. 96188
 test-amd64-amd64-xl-qemut-winxpsp3  6 xen-boot            fail REGR. vs. 96188
 test-amd64-i386-xl-qemuu-debianhvm-amd64 9 debian-hvm-install fail REGR. vs. 96188
 test-amd64-amd64-i386-pvgrub  6 xen-boot                  fail REGR. vs. 96188
 test-amd64-i386-freebsd10-i386  9 freebsd-install         fail REGR. vs. 96188
 test-amd64-amd64-xl-pvh-intel  6 xen-boot                 fail REGR. vs. 96188
 test-amd64-i386-xl-qemut-winxpsp3  9 windows-install      fail REGR. vs. 96188
 test-amd64-amd64-libvirt-qemuu-debianhvm-amd64-xsm 6 xen-boot fail REGR. vs. 96188
 test-amd64-amd64-xl-credit2   6 xen-boot                  fail REGR. vs. 96188
 test-armhf-armhf-xl           6 xen-boot                  fail REGR. vs. 96188
 test-armhf-armhf-xl-xsm       9 debian-install            fail REGR. vs. 96188
 test-amd64-amd64-xl-qemuu-debianhvm-amd64-xsm  6 xen-boot fail REGR. vs. 96188
 test-amd64-i386-xl-qemut-debianhvm-amd64 9 debian-hvm-install fail REGR. vs. 96188
 test-amd64-i386-qemuu-rhel6hvm-amd  9 redhat-install      fail REGR. vs. 96188
 test-amd64-amd64-xl-qcow2     6 xen-boot                  fail REGR. vs. 96188
 test-amd64-i386-freebsd10-amd64  9 freebsd-install        fail REGR. vs. 96188
 test-amd64-i386-xl            9 debian-install            fail REGR. vs. 96188
 test-amd64-amd64-qemuu-nested-intel  6 xen-boot           fail REGR. vs. 96188
 test-amd64-i386-xl-xsm        9 debian-install            fail REGR. vs. 96188
 test-amd64-amd64-xl-qemuu-winxpsp3  6 xen-boot            fail REGR. vs. 96188
 test-amd64-i386-xl-qemut-debianhvm-amd64-xsm 9 debian-hvm-install fail REGR. vs. 96188
 test-amd64-amd64-pygrub       6 xen-boot                  fail REGR. vs. 96188
 test-amd64-i386-libvirt       9 debian-install            fail REGR. vs. 96188
 test-amd64-amd64-xl-xsm       6 xen-boot                  fail REGR. vs. 96188
 test-amd64-amd64-libvirt-xsm  6 xen-boot                  fail REGR. vs. 96188
 test-amd64-i386-xl-qemuu-winxpsp3  9 windows-install      fail REGR. vs. 96188
 test-amd64-i386-xl-qemuu-debianhvm-amd64-xsm 9 debian-hvm-install fail REGR. vs. 96188
 test-armhf-armhf-xl-credit2   9 debian-install            fail REGR. vs. 96188
 test-armhf-armhf-xl-arndale   9 debian-install            fail REGR. vs. 96188
 test-amd64-amd64-amd64-pvgrub  6 xen-boot                 fail REGR. vs. 96188
 test-amd64-amd64-libvirt      6 xen-boot                  fail REGR. vs. 96188
 test-armhf-armhf-xl-vhd       9 debian-di-install         fail REGR. vs. 96188
 test-armhf-armhf-libvirt      9 debian-install            fail REGR. vs. 96188
 test-armhf-armhf-libvirt-qcow2  9 debian-di-install       fail REGR. vs. 96188
 test-amd64-i386-qemut-rhel6hvm-amd  9 redhat-install      fail REGR. vs. 96188
 test-armhf-armhf-libvirt-xsm  9 debian-install            fail REGR. vs. 96188
 test-armhf-armhf-xl-multivcpu  9 debian-install           fail REGR. vs. 96188
 test-amd64-i386-xl-qemut-win7-amd64  9 windows-install    fail REGR. vs. 96188
 test-amd64-i386-xl-qemut-winxpsp3-vcpus1 9 windows-install fail REGR. vs. 96188
 test-amd64-i386-xl-qemut-stubdom-debianhvm-amd64-xsm 9 debian-hvm-install fail REGR. vs. 96188
 test-amd64-amd64-xl-qemuu-debianhvm-amd64  6 xen-boot     fail REGR. vs. 96188
 test-amd64-amd64-xl-multivcpu  6 xen-boot                 fail REGR. vs. 96188
 test-armhf-armhf-xl-cubietruck  9 debian-install          fail REGR. vs. 96188
 test-amd64-amd64-xl           6 xen-boot                  fail REGR. vs. 96188
 test-amd64-i386-xl-qemuu-winxpsp3-vcpus1 9 windows-install fail REGR. vs. 96188
 test-amd64-i386-xl-raw        9 debian-di-install         fail REGR. vs. 96188
 test-amd64-amd64-xl-qemut-debianhvm-amd64-xsm  6 xen-boot fail REGR. vs. 96188
 test-amd64-i386-qemuu-rhel6hvm-intel  9 redhat-install    fail REGR. vs. 96188
 test-amd64-i386-qemut-rhel6hvm-intel  9 redhat-install    fail REGR. vs. 96188
 test-amd64-i386-xl-qemuu-ovmf-amd64  9 debian-hvm-install fail REGR. vs. 96188
 test-amd64-amd64-xl-qemut-stubdom-debianhvm-amd64-xsm 6 xen-boot fail REGR. vs. 96188
 test-amd64-amd64-xl-pvh-amd   6 xen-boot                  fail REGR. vs. 96188
 test-amd64-amd64-xl-qemut-win7-amd64  6 xen-boot          fail REGR. vs. 96188
 test-amd64-amd64-libvirt-vhd  6 xen-boot                  fail REGR. vs. 96188
 test-amd64-i386-xl-qemuu-win7-amd64  9 windows-install    fail REGR. vs. 96188
 test-amd64-i386-libvirt-xsm   9 debian-install            fail REGR. vs. 96188
 test-amd64-amd64-xl-qemuu-win7-amd64  6 xen-boot          fail REGR. vs. 96188
 test-amd64-amd64-xl-qemuu-ovmf-amd64  6 xen-boot          fail REGR. vs. 96188
 test-amd64-i386-libvirt-pair 15 debian-install/dst_host   fail REGR. vs. 96188
 test-amd64-amd64-qemuu-nested-amd  6 xen-boot             fail REGR. vs. 96188
 test-amd64-i386-pair         15 debian-install/dst_host   fail REGR. vs. 96188
 test-amd64-amd64-pair        10 xen-boot/dst_host         fail REGR. vs. 96188
 test-amd64-amd64-pair         9 xen-boot/src_host         fail REGR. vs. 96188
 test-amd64-amd64-libvirt-pair 10 xen-boot/dst_host        fail REGR. vs. 96188
 test-amd64-amd64-libvirt-pair  9 xen-boot/src_host        fail REGR. vs. 96188

Regressions which are regarded as allowable (not blocking):
 test-amd64-amd64-xl-rtds      6 xen-boot                  fail REGR. vs. 96188
 test-armhf-armhf-xl-rtds      9 debian-install            fail REGR. vs. 96188
 test-amd64-i386-libvirt-qemuu-debianhvm-amd64-xsm 9 debian-hvm-install fail like 96161
 test-armhf-armhf-libvirt-raw  9 debian-di-install            fail   like 96161
 build-amd64-rumpuserxen       6 xen-build                    fail   like 96188
 build-i386-rumpuserxen        6 xen-build                    fail   like 96188

Tests which did not succeed, but are not blocking:
 test-amd64-i386-rumpuserxen-i386  1 build-check(1)               blocked  n/a
 test-amd64-amd64-rumpuserxen-amd64  1 build-check(1)               blocked n/a

version targeted for testing:
 linux                0ac0a856d986c1ab240753479f5e50fdfab82b14
baseline version:
 linux                d420f00c7bfb405884dd71fb7f87974f0d1be455

Last test of basis    96188  2016-06-23 18:03:43 Z   20 days
Testing same since    97278  2016-07-13 23:15:41 Z    0 days    1 attempts

------------------------------------------------------------
People who touched revisions under test:
  Al Viro <viro@zeniv.linux.org.uk>
  Alan Stern <stern@rowland.harvard.edu>
  Alex Deucher <alexander.deucher@amd.com>
  Alexander Kochetkov <al.kochet@gmail.com>
  Alexander Shiyan <shc_work@mail.ru>
  Alexandre Belloni <alexandre.belloni@free-electrons.com>
  Alexei Starovoitov <ast@kernel.org>
  Alexey Brodkin <abrodkin@synopsys.com>
  Alexey Brodkin <Alexey.Brodkin@synopsys.com>
  Allain Legacy <allain.legacy@windriver.com>
  Amr Bekhit <amrbekhit@gmail.com>
  Andrew Goodbody <andrew.goodbody@cambrionix.com>
  Andrew Morton <akpm@linux-foundation.org>
  Andrey Grodzovsky <andrey2805@gmail.com>
  Andrey Ryabinin <aryabinin@virtuozzo.com>
  Andrey Ulanov <andreyu@google.com>
  Anna Schumaker <Anna.Schumaker@Netapp.com>
  Anthony Romano <anthony.romano@coreos.com>
  Anton Blanchard <anton@samba.org>
  Antonio Quartulli <a@unstable.cc>
  Ard Biesheuvel <ard.biesheuvel@linaro.org>
  Arnaldo Carvalho de Melo <acme@redhat.com>
  Arnd Bergmann <arnd@arndb.de>
  Asai Thambi S P <asamymuthupa@micron.com>
  Asai Thambi SP <asamymuthupa@micron.com>
  Babu Moger <babu.moger@oracle.com>
  Ben Hutchings <ben.hutchings@codethink.co.uk>
  Ben Hutchings <ben@decadent.org.uk>
  Ben Skeggs <bskeggs@redhat.com>
  Bill Sommerfeld <wsommerfeld@google.com>
  Bin Liu <b-liu@ti.com>
  Bjorn Andersson <bjorn.andersson@linaro.org>
  Bjorn Helgaas <bhelgaas@google.com>
  Bjørn Mork <bjorn@mork.no>
  Borislav Petkov <bp@suse.de>
  Catalin Marinas <catalin.marinas@arm.com>
  Chanwoo Choi <cw00.choi@samsung.com>
  Chris Friesen <chris.friesen@windriver.com>
  Chris Mason <clm@fb.com>
  Christoffer Dall <christoffer.dall@linaro.org>
  Christoph Hellwig <hch@lst.de>
  Chung-Geol Kim <chunggeol.kim@samsung.com>
  Cong Wang <xiyou.wangcong@gmail.com>
  Cyril Bur <cyrilbur@gmail.com>
  Cyrill Gorcunov <gorcunov@openvz.org>
  Dan Carpenter <dan.carpenter@oracle.com>
  Dan Streetman <dan.streetman@canonical.com>
  Daniel Borkmann <daniel@iogearbox.net>
  Daniel Lezcano <daniel.lezcano@linaro.org>
  Daniel Vetter <daniel.vetter@ffwll.ch>
  Daniel Vetter <daniel.vetter@intel.com>
  Dave Airlie <airlied@redhat.com>
  Dave Chinner <david@fromorbit.com>
  David Howells <dhowells@redhat.com>
  David Rientjes <rientjes@google.com>
  David S. Miller <davem@davemloft.net>
  David Sterba <dsterba@suse.com>
  David Vrabel <david.vrabel@citrix.com>
  Davidlohr Bueso <dave@stgolabs.net>
  Davidlohr Bueso <dbueso@suse.de>
  Diego Viola <diego.viola@gmail.com>
  Dirk Behme <dirk.behme@de.bosch.com>
  Dmitrii Tcvetkov <demfloro@demfloro.ru>
  Dmitry Torokhov <dmitry.torokhov@gmail.com>
  Dmitry Vyukov <dvyukov@google.com>
  Doug Ledford <dledford@redhat.com>
  Edward Cree <ecree@solarflare.com>
  Eric Dumazet <edumazet@google.com>
  Eric Sandeen <sandeen@redhat.com>
  Eryu Guan <guaneryu@gmail.com>
  Fang, Yang A <yang.a.fang@intel.com>
  Felipe Balbi <balbi@ti.com>
  Felipe Balbi <felipe.balbi@linux.intel.com>
  Filipe Manana <fdmanana@suse.com>
  Florian Fainelli <f.fainelli@gmail.com>
  Florian Westphal <fw@strlen.de>
  Fugang Duan <fugang.duan@nxp.com>
  Geert Uytterhoeven <geert@linux-m68k.org>
  Grazvydas Ignotas <notasas@gmail.com>
  Greg Kroah-Hartman <gregkh@linuxfoundation.org>
  Guenter Roeck <linux@roeck-us.net>
  Guillaume Nault <g.nault@alphalink.fr>
  Guo-Fu Tseng <cooldavid@cooldavid.org>
  Haishuang Yan <yanhaishuang@cmss.chinamobile.com>
  Hannes Frederic Sowa <hannes@stressinduktion.org>
  Hans de Goede <hdegoede@redhat.com>
  Heiko Stuebner <heiko@sntech.de>
  Herbert Xu <herbert@gondor.apana.org.au>
  Huacai Chen <chenhc@lemote.com>
  Hugh Dickins <hughd@google.com>
  Hui Wang <hui.wang@canonical.com>
  Huibin Hong <huibin.hong@rock-chips.com>
  Ian Campbell <ian.campbell@docker.com>
  Ignat Korchagin <ignat.korchagin@gmail.com>
  Ingo Molnar <mingo@kernel.org>
  J. Bruce Fields <bfields@redhat.com>
  Jamal Hadi Salim <jhs@mojatatu.com>
  James Hogan <james.hogan@imgtec.com>
  James Morse <james.morse@arm.com>
  Jann Horn <jannh@google.com>
  Jason Gunthorpe <jgunthorpe@obsidianresearch.com>
  Jason Wang <jasowang@redhat.com>
  Javier Martinez Canillas <javier@osg.samsung.com>
  Jeff Mahoney <jeffm@suse.com>
  Jens Axboe <axboe@fb.com>
  Jerome Marchand <jmarchan@redhat.com>
  Jerome Stanislaus <jeromes@codeaurora.org>
  Jim McDonough <jmcd@samba.org>
  Jiri Kosina <jkosina@suse.cz>
  Jiri Slaby <jslaby@suse.cz>
  Joe Lawrence <joe.lawrence@stratus.com>
  Johannes Berg <johannes.berg@intel.com>
  John Stultz <john.stultz@linaro.org>
  Jon Hunter <jonathanh@nvidia.com>
  Joonsoo Kim <iamjoonsoo.kim@lge.com>
  Josef Bacik <jbacik@fb.com>
  Joshua Henderson <joshua.henderson@microchip.com>
  Julian Anastasov <ja@ssi.bg>
  Kangjie Lu <kangjielu@gmail.com>
  Kangjie Lu <kjlu@gatech.edu>
  Karol Herbst <nouveau@karolherbst.de>
  Keerthy <j-keerthy@ti.com>
  Kirill A. Shutemov <kirill.shutemov@linux.intel.com>
  Knut Wohlrab <Knut.Wohlrab@de.bosch.com>
  Konstantin Khlebnikov <khlebnikov@yandex-team.ru>
  Krzysztof Hałasa <khalasa@piap.pl>
  Krzysztof Kozlowski <k.kozlowski@samsung.com>
  Lars Persson <larper@axis.com>
  Lars Persson <lars.persson@axis.com>
  Lars-Peter Clausen <lars@metafoo.de>
  Laurent Pinchart <laurent.pinchart+renesas@ideasonboard.com>
  Linus Lüssing <linus.luessing@c0d3.blue>
  Linus Torvalds <torvalds@linux-foundation.org>
  Linus Walleij <linus.walleij@linaro.org>
  Liviu Dudau <Liviu.Dudau@arm.com>
  Lorenzo Pieralisi <lorenzo.pieralisi@arm.com>
  Lucas Stach <dev@lynxeye.de>
  Lukasz Odzioba <lukasz.odzioba@intel.com>
  Lyude <cpaul@redhat.com>
  Manish Chopra <manish.chopra@qlogic.com>
  Marc Kleine-Budde <mkl@pengutronix.de>
  Marc Zyngier <marc.zyngier@arm.com>
  Marcin Åšlusarz <marcin.slusarz@gmail.com>
  Marco Angaroni <marcoangaroni@gmail.com>
  Marcus Weseloh <mweseloh42@gmail.com>
  Marek Lindner <mareklindner@neomailbox.ch>
  Mark Brown <broonie@kernel.org>
  Markus Pargmann <mpa@pengutronix.de>
  Martin K. Petersen <martin.petersen@oracle.com>
  Martin Schwidefsky <schwidefsky@de.ibm.com>
  Masami Hiramatsu <mhiramat@kernel.org>
  Mathias Krause <minipli@googlemail.com>
  Mathias Nyman <mathias.nyman@linux.intel.com>
  Matt Fleming <matt@codeblueprint.co.uk>
  Mauro Carvalho Chehab <mchehab@osg.samsung.com>
  Maxime Ripard <maxime.ripard@free-electrons.com>
  Meelis Roos <mroos@linux.ee>
  Michael Ellerman <mpe@ellerman.id.au>
  Michael Hennerich <michael.hennerich@analog.com>
  Michael Holzheu <holzheu@linux.vnet.ibm.com>
  Michael S. Tsirkin <mst@redhat.com>
  Michal Hocko <mhocko@suse.com>
  Michal Kazior <michal.kazior@tieto.com>
  Michal Nazarewicz <mina86@mina86.com>
  Michal Suchanek <hramrach@gmail.com>
  Mikael Pettersson <mikpelinux@gmail.com>
  Mike Frysinger <vapier@gentoo.org>
  Miklos Szeredi <mszeredi@redhat.com>
  Minchan Kim <minchan@kernel.org>
  Mitsuhiro Kimura <mitsuhiro.kimura.kc@renesas.com>
  Namhyung Kim <namhyung@kernel.org>
  Naoya Horiguchi <n-horiguchi@ah.jp.nec.com>
  Neal Cardwell <ncardwell@google.com>
  Neil Horman <nhorman@tuxdriver.com>
  NeilBrown <neilb@suse.com>
  Nicolai Stange <nicstange@gmail.com>
  Nicolas Dichtel <nicolas.dichtel@6wind.com>
  Nikolay Aleksandrov <nikolay@cumulusnetworks.com>
  Nitin Gupta <nitin.m.gupta@oracle.com>
  Noel Power <noel.power@suse.com>
  Oleg Drokin <green@linuxhacker.ru>
  Oleksij Rempel <linux@rempel-privat.de>
  Oliver Hartkopp <socketcan@hartkopp.net>
  Oliver Neukum <ONeukum@suse.com>
  Oscar <oscar@naiandei.net>
  Pablo Neira Ayuso <pablo@netfilter.org>
  Pali Rohár <pali.rohar@gmail.com>
  Paolo Abeni <pabeni@redhat.com>
  Paolo Bonzini <pbonzini@redhat.com>
  Patrick Lessard <Patrick.Lessard@cogeco.com>
  Pavel Emelyanov <xemul@virtuozzo.com>
  Pavel Shilovsky <pshilovsky@samba.org>
  Peter Chen <peter.chen@nxp.com>
  Pierre Moreau <pierre.morrow@free.fr>
  Ping Cheng <pingc@wacom.com>
  Ping Cheng <pinglinux@gmail.com>
  Rainer Weikusat <rweikusat@mobileactivedefense.com>
  Rajesh Borundia <rajesh.borundia@qlogic.com>
  Ralf Baechle <ralf@linux-mips.org>
  Richard Weinberger <richard@nod.at>
  Russell King <rmk+kernel@arm.linux.org.uk>
  Ryusuke Konishi <konishi.ryusuke@lab.ntt.co.jp>
  Sam Hung <sam.hung@emc.com.tw>
  Sasha Levin <sasha.levin@oracle.com>
  Scott Bauer <sbauer@plzdonthack.me>
  Sebastian Frias <sf84@laposte.net>
  Selvan Mani <smani@micron.com>
  Sergei Shtylyov <sergei.shtylyov@cogentembedded.com>
  Shaokun Zhang <zhangshaokun@hisilicon.com>
  Shawn Lin <shawn.lin@rock-chips.com>
  Simon Horman <horms@verge.net.au>
  Sony Chacko <sony.chacko@qlogic.com>
  Sowmini Varadhan <sowmini.varadhan@oracle.com>
  Srinivas Kandagatla <srinivas.kandagatla@linaro.org>
  Stefan Wahren <stefan.wahren@i2se.com>
  Stefano Stabellini <stefano.stabellini@eu.citrix.com>
  Steinar H. Gunderson <sesse@google.com>
  Stephen Boyd <sboyd@codeaurora.org>
  Steve Capper <steve.capper@arm.com>
  Steve French <smfrench@gmail.com>
  Steve French <steve.french@primarydata.com>
  Steven Rostedt (Red Hat) <rostedt@goodmis.org>
  Steven Rostedt <rostedt@goodmis.org>
  Subash Abhinov Kasiviswanathan <subashab@codeaurora.org>
  subashab@codeaurora.org <subashab@codeaurora.org>
  Sumit Saxena <sumit.saxena@broadcom.com>
  Sushaanth Srirangapathi <sushaanth.s@ti.com>
  Sven Eckelmann <sven@narfation.org>
  Takashi Iwai <tiwai@suse.de>
  Tejun Heo <tj@kernel.org>
  Thadeu Lima de Souza Cascardo <cascardo@redhat.com>
  Theodore Ts'o <tytso@mit.edu>
  Thierry Reding <treding@nvidia.com>
  Thomas Gleixner <tglx@linutronix.de>
  Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
  Thor Thayer <tthayer@opensource.altera.com>
  Tomi Valkeinen <tomi.valkeinen@ti.com>
  Tony Lindgren <tony@atomide.com>
  Torsten Hilbrich <torsten.hilbrich@secunet.com>
  Trond Myklebust <trond.myklebust@primarydata.com>
  Trond Myklebust <trondmy@primarydata.com>
  Uwe Kleine-König <u.kleine-koenig@pengutronix.de>
  Vineet Gupta <vgupta@synopsys.com>
  Vivek Gautam <gautam.vivek@samsung.com>
  Vladis Dronov <vdronov@redhat.com>
  Vlastimil Babka <vbabka@suse.cz>
  WANG Cong <xiyou.wangcong@gmail.com>
  Wei Fang <fangwei1@huawei.com>
  Wei Tang <tangwei@cmss.chinamobile.com>
  Wenwei Tao <ww.tao0320@gmail.com>
  Wilfried Klaebe <linux-kernel@lebenslange-mailadresse.de>
  Will Deacon <will.deacon@arm.com>
  Willy Tarreau <w@1wt.eu>
  Wolfgang Grandegger <wg@grandegger.com>
  Wolfram Sang <wsa@the-dreams.de>
  Xin Long <lucien.xin@gmail.com>
  Xishi Qiu <qiuxishi@huawei.com>
  Xiubo Li <lixiubo@cmss.chinamobile.com>
  Yoshihiro Kaneko <ykaneko0929@gmail.com>
  Yuchung Cheng <ycheng@google.com>
  Zhang Zhuoyu <zhangzhuoyu@cmss.chinamobile.com>
  洪一竹 <sam.hung@emc.com.tw>

jobs:
 build-amd64-xsm                                              pass    
 build-armhf-xsm                                              pass    
 build-i386-xsm                                               pass    
 build-amd64                                                  pass    
 build-armhf                                                  pass    
 build-i386                                                   pass    
 build-amd64-libvirt                                          pass    
 build-armhf-libvirt                                          pass    
 build-i386-libvirt                                           pass    
 build-amd64-pvops                                            pass    
 build-armhf-pvops                                            pass    
 build-i386-pvops                                             pass    
 build-amd64-rumpuserxen                                      fail    
 build-i386-rumpuserxen                                       fail    
 test-amd64-amd64-xl                                          fail    
 test-armhf-armhf-xl                                          fail    
 test-amd64-i386-xl                                           fail    
 test-amd64-amd64-xl-qemut-debianhvm-amd64-xsm                fail    
 test-amd64-i386-xl-qemut-debianhvm-amd64-xsm                 fail    
 test-amd64-amd64-libvirt-qemuu-debianhvm-amd64-xsm           fail    
 test-amd64-i386-libvirt-qemuu-debianhvm-amd64-xsm            fail    
 test-amd64-amd64-xl-qemuu-debianhvm-amd64-xsm                fail    
 test-amd64-i386-xl-qemuu-debianhvm-amd64-xsm                 fail    
 test-amd64-amd64-xl-qemut-stubdom-debianhvm-amd64-xsm        fail    
 test-amd64-i386-xl-qemut-stubdom-debianhvm-amd64-xsm         fail    
 test-amd64-amd64-libvirt-xsm                                 fail    
 test-armhf-armhf-libvirt-xsm                                 fail    
 test-amd64-i386-libvirt-xsm                                  fail    
 test-amd64-amd64-xl-xsm                                      fail    
 test-armhf-armhf-xl-xsm                                      fail    
 test-amd64-i386-xl-xsm                                       fail    
 test-amd64-amd64-qemuu-nested-amd                            fail    
 test-amd64-amd64-xl-pvh-amd                                  fail    
 test-amd64-i386-qemut-rhel6hvm-amd                           fail    
 test-amd64-i386-qemuu-rhel6hvm-amd                           fail    
 test-amd64-amd64-xl-qemut-debianhvm-amd64                    fail    
 test-amd64-i386-xl-qemut-debianhvm-amd64                     fail    
 test-amd64-amd64-xl-qemuu-debianhvm-amd64                    fail    
 test-amd64-i386-xl-qemuu-debianhvm-amd64                     fail    
 test-amd64-i386-freebsd10-amd64                              fail    
 test-amd64-amd64-xl-qemuu-ovmf-amd64                         fail    
 test-amd64-i386-xl-qemuu-ovmf-amd64                          fail    
 test-amd64-amd64-rumpuserxen-amd64                           blocked 
 test-amd64-amd64-xl-qemut-win7-amd64                         fail    
 test-amd64-i386-xl-qemut-win7-amd64                          fail    
 test-amd64-amd64-xl-qemuu-win7-amd64                         fail    
 test-amd64-i386-xl-qemuu-win7-amd64                          fail    
 test-armhf-armhf-xl-arndale                                  fail    
 test-amd64-amd64-xl-credit2                                  fail    
 test-armhf-armhf-xl-credit2                                  fail    
 test-armhf-armhf-xl-cubietruck                               fail    
 test-amd64-i386-freebsd10-i386                               fail    
 test-amd64-i386-rumpuserxen-i386                             blocked 
 test-amd64-amd64-qemuu-nested-intel                          fail    
 test-amd64-amd64-xl-pvh-intel                                fail    
 test-amd64-i386-qemut-rhel6hvm-intel                         fail    
 test-amd64-i386-qemuu-rhel6hvm-intel                         fail    
 test-amd64-amd64-libvirt                                     fail    
 test-armhf-armhf-libvirt                                     fail    
 test-amd64-i386-libvirt                                      fail    
 test-amd64-amd64-xl-multivcpu                                fail    
 test-armhf-armhf-xl-multivcpu                                fail    
 test-amd64-amd64-pair                                        fail    
 test-amd64-i386-pair                                         fail    
 test-amd64-amd64-libvirt-pair                                fail    
 test-amd64-i386-libvirt-pair                                 fail    
 test-amd64-amd64-amd64-pvgrub                                fail    
 test-amd64-amd64-i386-pvgrub                                 fail    
 test-amd64-amd64-pygrub                                      fail    
 test-armhf-armhf-libvirt-qcow2                               fail    
 test-amd64-amd64-xl-qcow2                                    fail    
 test-armhf-armhf-libvirt-raw                                 fail    
 test-amd64-i386-xl-raw                                       fail    
 test-amd64-amd64-xl-rtds                                     fail    
 test-armhf-armhf-xl-rtds                                     fail    
 test-amd64-i386-xl-qemut-winxpsp3-vcpus1                     fail    
 test-amd64-i386-xl-qemuu-winxpsp3-vcpus1                     fail    
 test-amd64-amd64-libvirt-vhd                                 fail    
 test-armhf-armhf-xl-vhd                                      fail    
 test-amd64-amd64-xl-qemut-winxpsp3                           fail    
 test-amd64-i386-xl-qemut-winxpsp3                            fail    
 test-amd64-amd64-xl-qemuu-winxpsp3                           fail    
 test-amd64-i386-xl-qemuu-winxpsp3                            fail    


------------------------------------------------------------
sg-report-flight on osstest.test-lab.xenproject.org
logs: /home/logs/logs
images: /home/logs/images

Logs, config files, etc. are available at
    http://logs.test-lab.xenproject.org/osstest/logs

Explanation of these reports, and of osstest in general, is at
    http://xenbits.xen.org/gitweb/?p=osstest.git;a=blob;f=README.email;hb=master
    http://xenbits.xen.org/gitweb/?p=osstest.git;a=blob;f=README;hb=master

Test harness code can be found at
    http://xenbits.xen.org/gitweb?p=osstest.git;a=summary


Not pushing.

(No revision log; it would be 6440 lines long.)


[-- Attachment #2: Type: text/plain, Size: 127 bytes --]

_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xen.org
https://lists.xen.org/xen-devel

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Linux 3.18 and Linux 4.1 reproducible OOM crashes
  2016-07-14  5:03 [linux-3.18 test] 97278: regressions - FAIL osstest service owner
@ 2016-07-14 12:56 ` Ian Jackson
  2016-07-14 23:34   ` Konrad Rzeszutek Wilk
  0 siblings, 1 reply; 7+ messages in thread
From: Ian Jackson @ 2016-07-14 12:56 UTC (permalink / raw)
  To: xen-devel

osstest service owner writes ("[linux-3.18 test] 97278: regressions - FAIL"):
> flight 97278 linux-3.18 real [real]
> http://logs.test-lab.xenproject.org/osstest/logs/97278/
> 
> Regressions :-(
> 
> Tests which did not succeed and are blocking,
> including tests which could not be run:
...
>  test-amd64-amd64-xl           6 xen-boot    fail REGR. vs. 96188

Loads of these.  It seems it can't boot at all.  Seems to affect amd64
dom0 kernels but not i386 or armhf.  (See flight 97279 for 4.1
results.)

It seems that the oom killer is tripping.  Obviously something is
seriously wrong.

It seems likely that this is a real bug, but whether it's in Xen or
Linux is not clear right now.  If it's in Linux it's because a similar
patch has been backported to both.

Below is serial log extract (from the job mentioned above).

Ian.


Jul 14 00:27:33.715871 [   22.813518] rc.local invoked oom-killer: gfp_mask=0x84d0, order=0, oom_score_adj=0

Jul 14 00:27:33.899843 [   22.813541] rc.local cpuset=/ mems_allowed=0

Jul 14 00:27:33.907821 [   22.813550] CPU: 0 PID: 2676 Comm: rc.local Not tainted 3.18.37 #1

Jul 14 00:27:33.907861 [   22.813556] Hardware name: Intel Corporation SandyBridge Platform/To be filled by O.E.M., BIOS S1200BT.86B.02.00.0042.050820141549 05/08/2014

Jul 14 00:27:33.923914 [   22.813565]  0000000000000000 ffff8800024b7968 ffffffff817dcab5 00000000000084d0

Jul 14 00:27:33.931995 [   22.813573]  0000000000000000 ffff8800024b79c8 ffffffff8118082f ffff8800024b7988

Jul 14 00:27:33.940021 [   22.813582]  ffffffff8114c315 ffffffff817e5ee9 0000000000000001 ffff8800024b79c8

Jul 14 00:27:33.948034 [   22.813590] Call Trace:

Jul 14 00:27:33.948065 [   22.813600]  [<ffffffff817dcab5>] dump_stack+0x7c/0x98

Jul 14 00:27:33.955838 [   22.813609]  [<ffffffff8118082f>] dump_header.isra.11+0x8f/0x1e0

Jul 14 00:27:33.963831 [   22.813616]  [<ffffffff8114c315>] ? __delayacct_freepages_end+0x45/0x50

Jul 14 00:27:33.963872 [   22.813677]  [<ffffffff817e5ee9>] ? _raw_spin_unlock_irqrestore+0x29/0x90

Jul 14 00:27:33.971868 [   22.813685]  [<ffffffff8131f93e>] ? ___ratelimit+0xae/0x160

Jul 14 00:27:33.979838 [   22.813691]  [<ffffffff81180ddc>] oom_kill_process+0x20c/0x370

Jul 14 00:27:33.987847 [   22.813699]  [<ffffffff810bbc69>] ? has_capability_noaudit+0x19/0x20

Jul 14 00:27:33.987888 [   22.813706]  [<ffffffff811812d1>] out_of_memory+0x211/0x330

Jul 14 00:27:33.995847 [   22.813714]  [<ffffffff81186331>] __alloc_pages_nodemask+0xb11/0xb50

Jul 14 00:27:34.003844 [   22.813721]  [<ffffffff81186382>] __get_free_pages+0x12/0x70

Jul 14 00:27:34.011845 [   22.813728]  [<ffffffff811863f1>] get_zeroed_page+0x11/0x20

Jul 14 00:27:34.011882 [   22.813735]  [<ffffffff811a9c31>] __pud_alloc+0x21/0x120

Jul 14 00:27:34.019841 [   22.813742]  [<ffffffff811ab312>] handle_mm_fault+0x352/0xd50

Jul 14 00:27:34.027863 [   22.813749]  [<ffffffff811a550f>] ? follow_page_mask+0x2f/0x4f0

Jul 14 00:27:34.035848 [   22.813756]  [<ffffffff811af392>] ? find_vma+0x62/0x70

Jul 14 00:27:34.035885 [   22.813763]  [<ffffffff811a5b48>] __get_user_pages+0x178/0x660

Jul 14 00:27:34.043839 [   22.813770]  [<ffffffff811a607d>] get_user_pages+0x4d/0x50

Jul 14 00:27:34.051847 [   22.813776]  [<ffffffff811db546>] copy_strings.isra.26+0x176/0x310

Jul 14 00:27:34.051886 [   22.813782]  [<ffffffff811db71a>] copy_strings_kernel+0x3a/0x60

Jul 14 00:27:34.059844 [   22.813789]  [<ffffffff811dcad8>] do_execve_common.isra.33+0x418/0x680

Jul 14 00:27:34.067846 [   22.813795]  [<ffffffff811dd014>] SyS_execve+0x24/0x30

Jul 14 00:27:34.075839 [   22.813801]  [<ffffffff817e6b19>] stub_execve+0x69/0xa0

Jul 14 00:27:34.075876 [   22.813807] Mem-Info:

Jul 14 00:27:34.083841 [   22.813810] DMA per-cpu:

Jul 14 00:27:34.083872 [   22.813813] CPU    0: hi:    0, btch:   1 usd:   0

Jul 14 00:27:34.083905 [   22.813817] CPU    1: hi:    0, btch:   1 usd:   0

Jul 14 00:27:34.091848 [   22.813821] CPU    2: hi:    0, btch:   1 usd:   0

Jul 14 00:27:34.099849 [   22.813825] CPU    3: hi:    0, btch:   1 usd:   0

Jul 14 00:27:34.099885 [   22.813829] DMA32 per-cpu:

Jul 14 00:27:34.107837 [   22.813833] CPU    0: hi:  186, btch:  31 usd:  88

Jul 14 00:27:34.107872 [   22.813837] CPU    1: hi:  186, btch:  31 usd:  72

Jul 14 00:27:34.115855 [   22.813841] CPU    2: hi:  186, btch:  31 usd:  47

Jul 14 00:27:34.115892 [   22.813845] CPU    3: hi:  186, btch:  31 usd: 171

Jul 14 00:27:34.123844 [   22.813852] active_anon:0 inactive_anon:0 isolated_anon:0

Jul 14 00:27:34.131837 [   22.813852]  active_file:0 inactive_file:0 isolated_file:0

Jul 14 00:27:34.131873 [   22.813852]  unevictable:69 dirty:9 writeback:0 unstable:0

Jul 14 00:27:34.139847 [   22.813852]  free:1036 slab_reclaimable:1887 slab_unreclaimable:4091

Jul 14 00:27:34.147839 [   22.813852]  mapped:4099 shmem:154 pagetables:706 bounce:0

Jul 14 00:27:34.155836 [   22.813852]  free_cma:0

Jul 14 00:27:34.155867 [   22.813875] DMA free:1680kB min:96kB low:120kB high:144kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:4kB isolated(anon):0kB isolated(file):0kB present:15976kB managed:15892kB mlocked:1048kB dirty:0kB writeback:0kB mapped:408kB shmem:8kB slab_reclaimable:92kB slab_unreclaimable:328kB kernel_stack:64kB pagetables:148kB unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:0 all_unreclaimable? yes

Jul 14 00:27:34.195860 [   22.813896] lowmem_reserve[]: 0 396 396 396

Jul 14 00:27:34.195898 [   22.813908] DMA32 free:2464kB min:2496kB low:3120kB high:3744kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:272kB isolated(anon):0kB isolated(file):0kB present:508308kB managed:408952kB mlocked:32180kB dirty:36kB writeback:0kB mapped:15988kB shmem:608kB slab_reclaimable:7456kB slab_unreclaimable:16036kB kernel_stack:2720kB pagetables:2676kB unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:0 all_unreclaimable? yes

Jul 14 00:27:34.243847 [   22.813929] lowmem_reserve[]: 0 0 0 0

Jul 14 00:27:34.243882 [   22.813936] DMA: 0*4kB 0*8kB 1*16kB (R) 0*32kB 0*64kB 1*128kB (R) 0*256kB 1*512kB (R) 1*1024kB (R) 0*2048kB 0*4096kB = 1680kB

Jul 14 00:27:34.251860 [   22.813956] DMA32: 0*4kB 0*8kB 1*16kB (R) 1*32kB (R) 0*64kB 1*128kB (R) 1*256kB (R) 0*512kB 0*1024kB 1*2048kB (R) 0*4096kB = 2480kB

Jul 14 00:27:34.267804 [   22.813976] 15037 total pagecache pages

Jul 14 00:27:34.267825 [   22.813980] 0 pages in swap cache

Jul 14 00:27:34.275851 [   22.813984] Swap cache stats: add 0, delete 0, find 0/0

Jul 14 00:27:34.283845 [   22.813988] Free swap  = 1949692kB

Jul 14 00:27:34.283879 [   22.813991] Total swap = 1949692kB

Jul 14 00:27:34.283908 [   22.813994] 131071 pages RAM

Jul 14 00:27:34.291912 [   22.813997] 0 pages HighMem/MovableOnly

Jul 14 00:27:34.291946 [   22.814000] 24860 pages reserved

Jul 14 00:27:34.299797 [   22.814003] [ pid ]   uid  tgid total_vm      rss nr_ptes swapents oom_score_adj name

Jul 14 00:27:34.307798 [   22.814013] [ 1316]     0  1316     9825      819      21        0         -1000 udevd

Jul 14 00:27:34.315792 [   22.814021] [ 2038]     0  2038     1084       25       7        0             0 ntpdate

Jul 14 00:27:34.323800 [   22.814027] [ 2040]     0  2040     1569      172       9        0             0 lockfile-create

Jul 14 00:27:34.331829 [   22.814034] [ 2047]     0  2047     1060      331       7        0             0 startpar

Jul 14 00:27:34.339797 [   22.814040] [ 2061]     0  2061     9270      668      24        0             0 rpcbind

Jul 14 00:27:34.347796 [   22.814047] [ 2076]   106  2076     9320      700      22        0             0 rpc.statd

Jul 14 00:27:34.355808 [   22.814054] [ 2088]     0  2088     6892       58      18        0             0 rpc.idmapd

Jul 14 00:27:34.363800 [   22.814060] [ 2131]     0  2131     1084      410       7        0             0 rc

Jul 14 00:27:34.371807 [   22.814066] [ 2137]     0  2137     3202      648      13        0             0 startpar

Jul 14 00:27:34.379805 [   22.814073] [ 2160]     0  2160    65721      788      33        0             0 rsyslogd

Jul 14 00:27:34.387805 [   22.814079] [ 2175]     0  2175     1064      394       8        0             0 acpid

Jul 14 00:27:34.395808 [   22.814085] [ 2211]     0  2211    19431     1167      41        0             0 apache2

Jul 14 00:27:34.403799 [   22.814091] [ 2214]    33  2214    91722     1032      69        0             0 apache2

Jul 14 00:27:34.411802 [   22.814098] [ 2216]    33  2216    91722     1031      69        0             0 apache2

Jul 14 00:27:34.419800 [   22.814104] [ 2422]     0  2422     4756       42      13        0             0 atd

Jul 14 00:27:34.427806 [   22.814110] [ 2423]     0  2423     6869      590      18        0             0 cron

Jul 14 00:27:34.435804 [   22.814117] [ 2606]     0  2606     8780      883      21        0             0 oxenstored

Jul 14 00:27:34.443846 [   22.814123] [ 2619]   105  2619    10560      551      24        0             0 dbus-daemon

Jul 14 00:27:34.451802 [   22.814130] [ 2624]     0  2624    13796      700      28        0         -1000 sshd

Jul 14 00:27:34.459799 [   22.814136] [ 2631]     0  2631    22682       57      19        0             0 xenconsoled

Jul 14 00:27:34.467815 [   22.814143] [ 2643]   104  2643    13313      828      26        0             0 exim4

Jul 14 00:27:34.475801 [   22.814149] [ 2652]     0  2652    60164     1974      78        0             0 qemu-system-i38

Jul 14 00:27:34.483868 [   22.814156] [ 2656]     0  2656    20648     1510      42        0             0 sshd

Jul 14 00:27:34.491859 [   22.814162] [ 2660]   107  2660     8345     1093      21        0             0 ntpd

Jul 14 00:27:34.499864 [   22.814169] [ 2673]     0  2673     1084      411       7        0             0 rc.local

Jul 14 00:27:34.507852 [   22.814175] [ 2676]     0  2676     1084       44       6        0             0 rc.local

Jul 14 00:27:34.515859 [   22.814181] [ 2677]  1000  2677    20648      946      41        0             0 sshd

Jul 14 00:27:34.523849 [   22.814188] Out of memory: Kill process 2652 (qemu-system-i38) score 3 or sacrifice child

Jul 14 00:27:34.531862 [   22.814194] Killed process 2652 (qemu-system-i38) total-vm:240656kB, anon-rss:2440kB, file-rss:5456kB

Jul 14 00:27:34.539871 [   22.822755] sshd invoked oom-killer: gfp_mask=0x2000d0, order=0, oom_score_adj=0

Jul 14 00:27:34.547800 [   22.822760] sshd cpuset=/ mems_allowed=0

Jul 14 00:27:34.555868 [   22.822764] CPU: 1 PID: 2677 Comm: sshd Not tainted 3.18.37 #1

Jul 14 00:27:34.563847 [   22.822767] Hardware name: Intel Corporation SandyBridge Platform/To be filled by O.E.M., BIOS S1200BT.86B.02.00.0042.050820141549 05/08/2014

Jul 14 00:27:34.571852 [   22.822771]  0000000000000000 ffff8800024fb990 ffffffff817dcab5 00000000002000d0

Jul 14 00:27:34.579860 [   22.822775]  0000000000000000 ffff8800024fb9f0 ffffffff8118082f ffff8800024fb9b0

Jul 14 00:27:34.587803 [   22.822779]  ffffffff8114c315 ffffffff817e5ee9 0000000000000001 ffff8800024fb9f0

Jul 14 00:27:34.595862 [   22.822783] Call Trace:

Jul 14 00:27:34.595892 [   22.822787]  [<ffffffff817dcab5>] dump_stack+0x7c/0x98

Jul 14 00:27:34.603859 [   22.822791]  [<ffffffff8118082f>] dump_header.isra.11+0x8f/0x1e0

Jul 14 00:27:34.611843 [   22.822794]  [<ffffffff8114c315>] ? __delayacct_freepages_end+0x45/0x50

Jul 14 00:27:34.619891 [   22.822798]  [<ffffffff817e5ee9>] ? _raw_spin_unlock_irqrestore+0x29/0x90

Jul 14 00:27:34.619933 [   22.822801]  [<ffffffff8131f93e>] ? ___ratelimit+0xae/0x160

Jul 14 00:27:34.627853 [   22.822804]  [<ffffffff81180ddc>] oom_kill_process+0x20c/0x370

Jul 14 00:27:34.635900 [   22.822808]  [<ffffffff810bbc69>] ? has_capability_noaudit+0x19/0x20

Jul 14 00:27:34.643851 [   22.822812]  [<ffffffff811812d1>] out_of_memory+0x211/0x330

Jul 14 00:27:34.643888 [   22.822815]  [<ffffffff81186331>] __alloc_pages_nodemask+0xb11/0xb50

Jul 14 00:27:34.651845 [   22.822820]  [<ffffffff811cb40c>] new_slab+0x2bc/0x320

Jul 14 00:27:34.660009 [   22.822823]  [<ffffffff811cc770>] __slab_alloc.constprop.69+0x440/0x570

Jul 14 00:27:34.667913 [   22.822828]  [<ffffffff81054f3e>] ? xen_leave_lazy_mmu+0xe/0x10

Jul 14 00:27:34.667952 [   22.822831]  [<ffffffff811b6721>] ? anon_vma_fork+0x71/0x140

Jul 14 00:27:34.675855 [   22.822835]  [<ffffffff811b651b>] ? anon_vma_clone+0x6b/0x200

Jul 14 00:27:34.683863 [   22.822838]  [<ffffffff811b6721>] ? anon_vma_fork+0x71/0x140

Jul 14 00:27:34.691849 [   22.822841]  [<ffffffff811cdad6>] kmem_cache_alloc+0x116/0x150

Jul 14 00:27:34.691887 [   22.822844]  [<ffffffff811b6721>] anon_vma_fork+0x71/0x140

Jul 14 00:27:34.699848 [   22.822848]  [<ffffffff810b2017>] copy_process.part.45+0x1687/0x1a50

Jul 14 00:27:34.707853 [   22.822852]  [<ffffffff810b2577>] do_fork+0xb7/0x3b0

Jul 14 00:27:34.707888 [   22.822856]  [<ffffffff811f3a34>] ? get_unused_fd_flags+0x34/0x40

Jul 14 00:27:34.715857 [   22.822860]  [<ffffffff811f3a88>] ? __fd_install+0x48/0x60

Jul 14 00:27:34.723872 [   22.822862]  [<ffffffff810b28f1>] SyS_clone+0x11/0x20

Jul 14 00:27:34.723910 [   22.822866]  [<ffffffff817e68c9>] stub_clone+0x69/0x90

Jul 14 00:27:34.731848 [   22.822868]  [<ffffffff817e6589>] ? system_call_fastpath+0x12/0x17

Jul 14 00:27:34.739823 [   22.822871] Mem-Info:

Jul 14 00:27:34.739841 [   22.822872] DMA per-cpu:

Jul 14 00:27:34.747843 [   22.822874] CPU    0: hi:    0, btch:   1 usd:   0

Jul 14 00:27:34.747879 [   22.822876] CPU    1: hi:    0, btch:   1 usd:   0

Jul 14 00:27:34.755836 [   22.822878] CPU    2: hi:    0, btch:   1 usd:   0

Jul 14 00:27:34.755870 [   22.822880] CPU    3: hi:    0, btch:   1 usd:   0

Jul 14 00:27:34.763850 [   22.822882] DMA32 per-cpu:

Jul 14 00:27:34.763881 [   22.822884] CPU    0: hi:  186, btch:  31 usd:  88

Jul 14 00:27:34.771838 [   22.822886] CPU    1: hi:  186, btch:  31 usd: 157

Jul 14 00:27:34.779837 [   22.822888] CPU    2: hi:  186, btch:  31 usd:  47

Jul 14 00:27:34.779872 [   22.822890] CPU    3: hi:  186, btch:  31 usd: 173

Jul 14 00:27:34.787884 [   22.822893] active_anon:0 inactive_anon:0 isolated_anon:0

Jul 14 00:27:34.787920 [   22.822893]  active_file:0 inactive_file:0 isolated_file:0

Jul 14 00:27:34.795844 [   22.822893]  unevictable:69 dirty:9 writeback:0 unstable:0

Jul 14 00:27:34.803834 [   22.822893]  free:1036 slab_reclaimable:1887 slab_unreclaimable:4091

Jul 14 00:27:34.811846 [   22.822893]  mapped:3101 shmem:154 pagetables:650 bounce:0

Jul 14 00:27:34.811884 [   22.822893]  free_cma:0

Jul 14 00:27:34.819842 [   22.822904] DMA free:1680kB min:96kB low:120kB high:144kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:4kB isolated(anon):0kB isolated(file):0kB present:15976kB managed:15892kB mlocked:1048kB dirty:0kB writeback:0kB mapped:336kB shmem:8kB slab_reclaimable:92kB slab_unreclaimable:328kB kernel_stack:64kB pagetables:148kB unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:0 all_unreclaimable? yes

Jul 14 00:27:34.859831 [   22.822914] lowmem_reserve[]: 0 396 396 396

Jul 14 00:27:34.859866 [   22.822920] DMA32 free:2464kB min:2496kB low:3120kB high:3744kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:272kB isolated(anon):0kB isolated(file):0kB present:508308kB managed:408952kB mlocked:32180kB dirty:36kB writeback:0kB mapped:12068kB shmem:608kB slab_reclaimable:7456kB slab_unreclaimable:16036kB kernel_stack:2720kB pagetables:2452kB unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:0 all_unreclaimable? yes

Jul 14 00:27:34.899849 [   22.822930] lowmem_reserve[]: 0 0 0 0

Jul 14 00:27:34.907838 [   22.822934] DMA: 0*4kB 0*8kB 1*16kB (R) 0*32kB 0*64kB 1*128kB (R) 0*256kB 1*512kB (R) 1*1024kB (R) 0*2048kB 0*4096kB = 1680kB

Jul 14 00:27:34.915876 [   22.822943] DMA32: 0*4kB 0*8kB 1*16kB (R) 1*32kB (R) 0*64kB 1*128kB (R) 1*256kB (R) 0*512kB 0*1024kB 1*2048kB (R) 0*4096kB = 2480kB

Jul 14 00:27:34.931842 [   22.822953] 15037 total pagecache pages

Jul 14 00:27:34.931878 [   22.822955] 0 pages in swap cache

Jul 14 00:27:34.939850 [   22.822957] Swap cache stats: add 0, delete 0, find 0/0

Jul 14 00:27:34.939888 [   22.822959] Free swap  = 1949692kB

Jul 14 00:27:34.947842 [   22.822960] Total swap = 1949692kB

Jul 14 00:27:34.947874 [   22.822962] 131071 pages RAM

Jul 14 00:27:34.947903 [   22.822963] 0 pages HighMem/MovableOnly

Jul 14 00:27:34.955868 [   22.822965] 24860 pages reserved

Jul 14 00:27:34.955900 [   22.822966] [ pid ]   uid  tgid total_vm      rss nr_ptes swapents oom_score_adj name

Jul 14 00:27:34.963860 [   22.822971] [ 1316]     0  1316     9825      819      21        0         -1000 udevd

Jul 14 00:27:34.971874 [   22.822976] [ 2038]     0  2038     1084       25       7        0             0 ntpdate

Jul 14 00:27:34.979864 [   22.822978] [ 2040]     0  2040     1569      172       9        0             0 lockfile-create

Jul 14 00:27:34.987862 [   22.822981] [ 2047]     0  2047     1060      331       7        0             0 startpar

Jul 14 00:27:34.995869 [   22.822984] [ 2061]     0  2061     9270      668      24        0             0 rpcbind

Jul 14 00:27:35.003863 [   22.822987] [ 2076]   106  2076     9320      700      22        0             0 rpc.statd

Jul 14 00:27:35.019848 [   22.822991] [ 2088]     0  2088     6892       58      18        0             0 rpc.idmapd

Jul 14 00:27:35.027853 [   22.822994] [ 2131]     0  2131     1084      410       7        0             0 rc

Jul 14 00:27:35.035843 [   22.822997] [ 2137]     0  2137     3202      648      13        0             0 startpar

Jul 14 00:27:35.043857 [   22.823000] [ 2160]     0  2160    65721      788      33        0             0 rsyslogd

Jul 14 00:27:35.051846 [   22.823003] [ 2175]     0  2175     1064      394       8        0             0 acpid

Jul 14 00:27:35.059852 [   22.823006] [ 2211]     0  2211    19431     1167      41        0             0 apache2

Jul 14 00:27:35.067848 [   22.823009] [ 2214]    33  2214    91722     1032      69        0             0 apache2

Jul 14 00:27:35.075852 [   22.823012] [ 2216]    33  2216    91722     1031      69        0             0 apache2

Jul 14 00:27:35.083847 [   22.823015] [ 2422]     0  2422     4756       42      13        0             0 atd

Jul 14 00:27:35.091857 [   22.823018] [ 2423]     0  2423     6869      590      18        0             0 cron

Jul 14 00:27:35.099857 [   22.823021] [ 2606]     0  2606     8780      883      21        0             0 oxenstored

Jul 14 00:27:35.107854 [   22.823024] [ 2619]   105  2619    10560      551      24        0             0 dbus-daemon

Jul 14 00:27:35.115866 [   22.823027] [ 2624]     0  2624    13796      700      28        0         -1000 sshd

Jul 14 00:27:35.123797 [   22.823030] [ 2631]     0  2631    22682       57      19        0             0 xenconsoled

Jul 14 00:27:35.131849 [   22.823034] [ 2643]   104  2643    13313      828      26        0             0 exim4

Jul 14 00:27:35.139844 [   22.823037] [ 2656]     0  2656    20648     1510      42        0             0 sshd

Jul 14 00:27:35.147846 [   22.823040] [ 2660]   107  2660     8345     1093      21        0             0 ntpd

Jul 14 00:27:35.155856 [   22.823043] [ 2673]     0  2673     1084      411       7        0             0 rc.local

Jul 14 00:27:35.163847 [   22.823046] [ 2676]     0  2676     1084       44       6        0             0 rc.local

Jul 14 00:27:35.171852 [   22.823049] [ 2677]  1000  2677    20648      946      41        0             0 sshd

Jul 14 00:27:35.179883 [   22.823052] Out of memory: Kill process 2656 (sshd) score 2 or sacrifice child

Jul 14 00:27:35.187860 [   22.823055] Killed process 2677 (sshd) total-vm:82592kB, anon-rss:848kB, file-rss:2936kB

Jul 14 00:27:35.195838 [   22.825444] oxenstored invoked oom-killer: gfp_mask=0x280da, order=0, oom_score_adj=0

Jul 14 00:27:35.203861 [   22.825449] oxenstored cpuset=/ mems_allowed=0

Jul 14 00:27:35.203897 [   22.825453] CPU: 2 PID: 2606 Comm: oxenstored Not tainted 3.18.37 #1

Jul 14 00:27:35.211845 [   22.825457] Hardware name: Intel Corporation SandyBridge Platform/To be filled by O.E.M., BIOS S1200BT.86B.02.00.0042.050820141549 05/08/2014

Jul 14 00:27:35.227847 [   22.825461]  0000000000000000 ffff880008dc3ad8 ffffffff817dcab5 00000000000280da

Jul 14 00:27:35.235847 [   22.825466]  0000000000000000 ffff880008dc3b38 ffffffff8118082f ffff880008dc3af8

Jul 14 00:27:35.243891 [   22.825469]  ffffffff8114c315 ffffffff817e5ee9 0000000000000001 ffff880008dc3b38

Jul 14 00:27:35.243937 [   22.825473] Call Trace:

Jul 14 00:27:35.251845 [   22.825477]  [<ffffffff817dcab5>] dump_stack+0x7c/0x98

Jul 14 00:27:35.251881 [   22.825481]  [<ffffffff8118082f>] dump_header.isra.11+0x8f/0x1e0

Jul 14 00:27:35.259846 [   22.825484]  [<ffffffff8114c315>] ? __delayacct_freepages_end+0x45/0x50

Jul 14 00:27:35.267864 [   22.825488]  [<ffffffff817e5ee9>] ? _raw_spin_unlock_irqrestore+0x29/0x90

Jul 14 00:27:35.275843 [   22.825492]  [<ffffffff8131f93e>] ? ___ratelimit+0xae/0x160

Jul 14 00:27:35.283904 [   22.825494]  [<ffffffff81180ddc>] oom_kill_process+0x20c/0x370

Jul 14 00:27:35.283942 [   22.825498]  [<ffffffff810bbc69>] ? has_capability_noaudit+0x19/0x20

Jul 14 00:27:35.291845 [   22.825501]  [<ffffffff811812d1>] out_of_memory+0x211/0x330

Jul 14 00:27:35.299881 [   22.825505]  [<ffffffff81186331>] __alloc_pages_nodemask+0xb11/0xb50

Jul 14 00:27:35.307834 [   22.825509]  [<ffffffff811ab988>] handle_mm_fault+0x9c8/0xd50

Jul 14 00:27:35.307871 [   22.825513]  [<ffffffff811ea170>] ? poll_select_copy_remaining+0x130/0x130

Jul 14 00:27:35.315942 [   22.825517]  [<ffffffff810a68dd>] __do_page_fault+0x1bd/0x590

Jul 14 00:27:35.323846 [   22.825521]  [<ffffffff[   23.782240] getty invoked oom-killer: gfp_mask=0x280da, order=0, oom_score_adj=0

Jul 14 00:27:35.331856 [   23.782247] getty cpuset=/ mems_allowed=0

Jul 14 00:27:35.339842 [   23.782252] CPU: 3 PID: 2685 Comm: getty Not tainted 3.18.37 #1

Jul 14 00:27:35.339884 [   23.782255] Hardware name: Intel Corporation SandyBridge Platform/To be filled by O.E.M., BIOS S1200BT.86B.02.00.0042.050820141549 05/08/2014

Jul 14 00:27:35.355849 [   23.782259]  0000000000000000 ffff8800158db7f8 ffffffff817dcab5 00000000000280da

Jul 14 00:27:35.363925 [   23.782263]  0000000000000000 ffff8800158db858 ffffffff8118082f ffff8800158db818

Jul 14 00:27:35.371864 [   23.782268]  ffffffff8114c315 ffffffff817e5ee9 0000000000000001 ffff8800158db858

Jul 14 00:27:35.379855 [   23.782272] Call Trace:

Jul 14 00:27:35.379886 [   23.782278]  [<ffffffff817dcab5>] dump_stack+0x7c/0x98

Jul 14 00:27:35.387850 [   23.782283]  [<ffffffff8118082f>] dump_header.isra.11+0x8f/0x1e0

Jul 14 00:27:35.387888 [   23.782287]  [<ffffffff8114c315>] ? __delayacct_freepages_end+0x45/0x50

Jul 14 00:27:35.395877 [   23.782291]  [<ffffffff817e5ee9>] ? _raw_spin_unlock_irqrestore+0x29/0x90

Jul 14 00:27:35.403858 [   23.782295]  [<ffffffff8131f93e>] ? ___ratelimit+0xae/0x160

Jul 14 00:27:35.411839 [   23.782298]  [<ffffffff81180ddc>] oom_kill_process+0x20c/0x370

Jul 14 00:27:35.419845 [   23.782301]  [<ffffffff811812d1>] out_of_memory+0x211/0x330

Jul 14 00:27:35.419883 [   23.782305]  [<ffffffff81186331>] __alloc_pages_nodemask+0xb11/0xb50

Jul 14 00:27:35.427868 [   23.782309]  [<ffffffff811ab988>] handle_mm_fault+0x9c8/0xd50

Jul 14 00:27:35.435856 [   23.782320]  [<ffffffffa009d912>] ? search_dir+0xc2/0x120 [ext4]

Jul 14 00:27:35.443840 [   23.782324]  [<ffffffff810a68dd>] __do_page_fault+0x1bd/0x590

Jul 14 00:27:35.443877 [   23.782329]  [<ffffffff81233736>] ? dquot_file_open+0x16/0x50

Jul 14 00:27:35.451860 [   23.782334]  [<ffffffffa008c953>] ? ext4_file_open+0x43/0x180 [ext4]

Jul 14 00:27:35.459849 [   23.782339]  [<ffffffffa008c910>] ? ext4_check_all_de+0xb0/0xb0 [ext4]

Jul 14 00:27:35.467884 [   23.782343]  [<ffffffff810a6cdc>] do_page_fault+0x2c/0x40

Jul 14 00:27:35.467920 [   23.782346]  [<ffffffff817e81b8>] page_fault+0x28/0x30

Jul 14 00:27:35.475841 [   23.782350]  [<ffffffff811a47a8>] ? copy_page_to_iter+0x68/0x3c0

Jul 14 00:27:35.483845 [   23.782353]  [<ffffffff8117ed88>] generic_file_read_iter+0x278/0x5d0

Jul 14 00:27:35.491842 [   23.782357]  [<ffffffff811d4ad9>] new_sync_read+0x79/0xb0

Jul 14 00:27:35.491880 [   23.782360]  [<ffffffff811d5303>] vfs_read+0x93/0x180

Jul 14 00:27:35.499846 [   23.782364]  [<ffffffff811d5e61>] SyS_read+0x41/0xb0

Jul 14 00:27:35.507832 [   23.782367]  [<ffffffff817e6589>] system_call_fastpath+0x12/0x17

Jul 14 00:27:35.507869 [   23.782370] Mem-Info:

Jul 14 00:27:35.515838 [   23.782371] DMA per-cpu:

Jul 14 00:27:35.515869 [   23.782373] CPU    0: hi:    0, btch:   1 usd:   0

Jul 14 00:27:35.523838 [   23.782375] CPU    1: hi:    0, btch:   1 usd:   0

Jul 14 00:27:35.523873 [   23.782377] CPU    2: hi:    0, btch:   1 usd:   0

Jul 14 00:27:35.531843 [   23.782379] CPU    3: hi:    0, btch:   1 usd:   0

Jul 14 00:27:35.531878 [   23.782381] DMA32 per-cpu:

Jul 14 00:27:35.539836 [   23.782382] CPU    0: hi:  186, btch:  31 usd:   0

Jul 14 00:27:35.539873 [   23.782384] CPU    1: hi:  186, btch:  31 usd:   0

Jul 14 00:27:35.547841 [   23.782386] CPU    2: hi:  186, btch:  31 usd:   0

Jul 14 00:27:35.555833 [   23.782388] CPU    3: hi:  186, btch:  31 usd:   0

Jul 14 00:27:35.555868 [   23.782392] active_anon:0 inactive_anon:0 isolated_anon:0

Jul 14 00:27:35.563857 [   23.782392]  active_file:0 inactive_file:0 isolated_file:0

Jul 14 00:27:35.563895 [   23.782392]  unevictable:1 dirty:0 writeback:18 unstable:0

Jul 14 00:27:35.571859 [   23.782392]  free:1060 slab_reclaimable:1888 slab_unreclaimable:4074

Jul 14 00:27:35.579860 [   23.782392]  mapped:2923 shmem:153 pagetables:588 bounce:0

Jul 14 00:27:35.587840 [   23.782392]  free_cma:0

Jul 14 00:27:35.587872 [   23.782403] DMA free:1668kB min:96kB low:120kB high:144kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:0kB isolated(anon):0kB isolated(file):0kB present:15976kB managed:15892kB mlocked:1044kB dirty:0kB writeback:4kB mapped:340kB shmem:8kB slab_reclaimable:92kB slab_unreclaimable:332kB kernel_stack:48kB pagetables:116kB unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:0 all_unreclaimable? yes

Jul 14 00:27:35.627846 [   23.782413] lowmem_reserve[]: 0 396 396 396

Jul 14 00:27:35.627882 [   23.782419] DMA32 free:2572kB min:2496kB low:3120kB high:3744kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:4kB isolated(anon):0kB isolated(file):0kB present:508308kB managed:408952kB mlocked:31912kB dirty:0kB writeback:68kB mapped:11352kB shmem:604kB slab_reclaimable:7460kB slab_unreclaimable:15964kB kernel_stack:2400kB pagetables:2236kB unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:36 all_unreclaimable? yes

Jul 14 00:27:35.675846 [   23.782431] lowmem_reserve[]: 0 0 0 0

Jul 14 00:27:35.675882 [   23.782433] DMA: 0*4kB 0*8kB 0*16kB 0*32kB 0*64kB 1*128kB (R) 0*256kB 1*512kB (R) 1*1024kB (R) 0*2048kB 0*4096kB = 1664kB

Jul 14 00:27:35.683866 [   23.782442] DMA32: 29*4kB (UR) 0*8kB 3*16kB (R) 0*32kB 1*64kB (R) 1*128kB (R) 1*256kB (R) 0*512kB 0*1024kB 1*2048kB (R) 0*4096kB = 2660kB

Jul 14 00:27:35.699851 [   23.782452] 15076 total pagecache pages

Jul 14 00:27:35.699885 [   23.782455] 0 pages in swap cache

Jul 14 00:27:35.707849 [   23.782465] Swap cache stats: add 0, delete 0, find 0/0

Jul 14 00:27:35.715836 [   23.782468] Free swap  = 1949692kB

Jul 14 00:27:35.715869 [   23.782472] Total swap = 1949692kB

Jul 14 00:27:35.715899 [   23.782476] 131071 pages RAM

Jul 14 00:27:35.723847 [   23.782480] 0 pages HighMem/MovableOnly

Jul 14 00:27:35.723882 [   23.782484] 24860 pages reserved

Jul 14 00:27:35.731838 [   23.782487] [ pid ]   uid  tgid total_vm      rss nr_ptes swapents oom_score_adj name

Jul 14 00:27:35.739850 [   23.782495] [ 1316]     0  1316     9825      819      21        0         -1000 udevd

Jul 14 00:27:35.747851 [   23.782501] [ 2038]     0  2038     1084       25       7        0             0 ntpdate

Jul 14 00:27:35.755844 [   23.782506] [ 2040]     0  2040     1569      172       9        0             0 lockfile-create

Jul 14 00:27:35.763856 [   23.782509] [ 2047]     0  2047     1060      331       7        0             0 startpar

Jul 14 00:27:35.771890 [   23.782512] [ 2061]     0  2061     9270      668      24        0             0 rpcbind

Jul 14 00:27:35.779829 [   23.782516] [ 2076]   106  2076     9320      700      22        0             0 rpc.statd

Jul 14 00:27:35.787856 [   23.782519] [ 2088]     0  2088     6892       58      18        0             0 rpc.idmapd

Jul 14 00:27:35.795799 [   23.782522] [ 2160]     0  2160    65721      785      33        0             0 rsyslogd

Jul 14 00:27:35.803857 [   23.782525] [ 2175]     0  2175     1064      394       8        0             0 acpid

Jul 14 00:27:35.811854 [   23.782528] [ 2211]     0  2211    19431     1167      41        0             0 apache2

Jul 14 00:27:35.820250 [   23.782532] [ 2216]    33  2216    91722     1031      69        0             0 apache2

Jul 14 00:27:35.827884 [   23.782534] [ 2422]     0  2422     4756       42      13        0             0 atd

Jul 14 00:27:35.835897 [   23.782537] [ 2423]     0  2423     6869      590      18        0             0 cron

Jul 14 00:27:35.843893 [   23.782541] [ 2606]     0  2606     8780      883      21        0             0 oxenstored

Jul 14 00:27:35.851883 [   23.782544] [ 2619]   105  2619    10560      551      24        0             0 dbus-daemon

Jul 14 00:27:35.859858 [   23.782547] [ 2624]     0  2624    13796      700      28        0         -1000 sshd

Jul 14 00:27:35.867866 [   23.782550] [ 2631]     0  2631    22682       57      19        0             0 xenconsoled

Jul 14 00:27:35.875902 [   23.782554] [ 2643]   104  2643    13313      828      26        0             0 exim4

Jul 14 00:27:35.883887 [   23.782557] [ 2660]   107  2660     8346     1093      21        0             0 ntpd

Jul 14 00:27:35.891912 [   23.782560] [ 2681]     0  2681     2650      399      10        0             0 getty

Jul 14 00:27:35.899869 [   23.782563] [ 2682]     0  2682     3166      417      12        0             0 getty

Jul 14 00:27:35.907853 [   23.782566] [ 2683]     0  2683     1063      173       7        0             0 getty

Jul 14 00:27:35.915859 [   23.782569] [ 2684]     0  2684     3166      416      12        0             0 getty

Jul 14 00:27:35.923807 [   23.782572] [ 2685]     0  2685     3560      425      12        0             0 getty

Jul 14 00:27:35.931919 [   23.782575] [ 2686]     0  2686     3166      425      11        0             0 getty

Jul 14 00:27:35.939914 [   23.782578] [ 2687]     0  2687     1591      404       8        0             0 getty

Jul 14 00:27:35.947888 [   23.782581] [ 2688]     0  2688     3166      437      12        0             0 getty

Jul 14 00:27:35.956047 [   23.782587] [ 2690]     0  2690     3166       35       9        0             0 getty

Jul 14 00:27:35.964079 [   23.782592] [ 2691]     0  2691     3166       34       9        0             0 getty

Jul 14 00:27:35.972103 [   23.782597] [ 2693]     0  2693     9825      195      19        0         -1000 udevd

Jul 14 00:27:35.980089 [   23.782601] [ 2694]     0  2694     9825      195      18        0         -1000 udevd

Jul 14 00:27:35.987933 [   23.782604] Out of memory: Kill process 2211 (apache2) score 1 or sacrifice child

Jul 14 00:27:35.996043 [   23.782607] Killed process 2216 (apache2) total-vm:366888kB, anon-rss:1556kB, file-rss:2568kB

Jul 14 00:27:36.004061 [   23.790140] getty invoked oom-killer: gfp_mask=0x200da, order=0, oom_score_adj=0

Jul 14 00:27:36.011954 [   23.790145] getty cpuset=/ mems_allowed=0

Jul 14 00:27:36.019911 [   23.790150] CPU: 0 PID: 2681 Comm: getty Not tainted 3.18.37 #1

Jul 14 00:27:36.019951 [   23.790152] Hardware name: Intel Corporation SandyBridge Platform/To be filled by O.E.M., BIOS S1200BT.86B.02.00.0042.050820141549 05/08/2014

Jul 14 00:27:36.035934 [   23.790156]  0000000000000000 ffff88001220ba78 ffffffff817dcab5 00000000000200da

Jul 14 00:27:36.043951 [   23.790161]  0000000000000000 ffff88001220bad8 ffffffff8118082f ffff88001220ba98

Jul 14 00:27:36.051973 [   23.790166]  ffffffff8114c315 ffffffff817e5ee9 0000000000000001 ffff88001220bad8

Jul 14 00:27:36.059965 [   23.790170] Call Trace:

Jul 14 00:27:36.059997 [   23.790174]  [<ffffffff817dcab5>] dump_stack+0x7c/0x98

Jul 14 00:27:36.067938 [   23.790178]  [<ffffffff8118082f>] dump_header.isra.11+0x8f/0x1e0

Jul 14 00:27:36.067977 [   23.790181]  [<ffffffff8114c315>] ? __delayacct_freepages_end+0x45/0x50

Jul 14 00:27:36.076024 [   23.790185]  [<ffffffff817e5ee9>] ? _raw_spin_unlock_irqrestore+0x29/0x90

Jul 14 00:27:36.083947 [   23.790188]  [<ffffffff8131f93e>] ? ___ratelimit+0xae/0x160

Jul 14 00:27:36.092010 [   23.790191]  [<ffffffff81180ddc>] oom_kill_process+0x20c/0x370

Jul 14 00:27:36.099947 [   23.790195]  [<ffffffff811812d1>] out_of_memory+0x211/0x330

Jul 14 00:27:36.099991 [   23.790198]  [<ffffffff81186331>] __alloc_pages_nodemask+0xb11/0xb50

Jul 14 00:27:36.107888 [   23.790202]  [<ffffffff811a9757>] do_cow_fault.isra.96+0x47/0x2a0

Jul 14 00:27:36.116087 [   23.790207]  [<ffffffff81054bf9>] ? __raw_callee_save_xen_pmd_val+0x11/0x1e

Jul 14 00:27:36.123936 [   23.790210]  [<ffffffff811ab675>] handle_mm_fault+0x6b5/0xd50

Jul 14 00:27:36.124004 [   23.790214]  [<ffffffff810a68dd>] __do_page_fault+0x1bd/0x590

Jul 14 00:27:36.132061 [   23.790217]  [<ffffffff81198939>] ? vm_mmap_pgoff+0x99/0xc0

Jul 14 00:27:36.140090 [   23.790220]  [<ffffffff810a6cdc>] do_page_fault+0x2c/0x40

Jul 14 00:27:36.148022 [   23.790224]  [<ffffffff817e81b8>] page_fault+0x28/0x30

Jul 14 00:27:36.148057 [   23.790226] Mem-Info:

Jul 14 00:27:36.148083 [   23.790227] DMA per-cpu:

Jul 14 00:27:36.155974 [   23.790229] CPU    0: hi:    0, btch:   1 usd:   0

Jul 14 00:27:36.155995 [   23.790231] CPU    1: hi:    0, btch:   1 usd:   0

Jul 14 00:27:36.164132 [   23.790233] CPU    2: hi:    0, btch:   1 usd:   0

Jul 14 00:27:36.172145 [   23.790235] CPU    3: hi:    0, btch:   1 usd:   0

Jul 14 00:27:36.172208 [   23.790237] DMA32 per-cpu:

Jul 14 00:27:36.180000 [   23.790238] CPU    0: hi:  186, btch:  31 usd:   0

Jul 14 00:27:36.180049 [   23.790241] CPU    1: hi:  186, btch:  31 usd: 128

Jul 14 00:27:36.187954 [   23.790243] CPU    2: hi:  186, btch:  31 usd:   0

Jul 14 00:27:36.187996 [   23.790245] CPU    3: hi:  186, btch:  31 usd:   0

Jul 14 00:27:36.195965 [   23.790248] active_anon:0 inactive_anon:0 isolated_anon:0

Jul 14 00:27:36.203964 [   23.790248]  active_file:0 inactive_file:0 isolated_file:0

Jul 14 00:27:36.204010 [   23.790248]  unevictable:1 dirty:4 writeback:0 unstable:0

Jul 14 00:27:36.211969 [   23.790248]  free:1012 slab_reclaimable:1888 slab_unreclaimable:4067

Jul 14 00:27:36.219948 [   23.790248]  mapped:2877 shmem:154 pagetables:521 bounce:0

Jul 14 00:27:36.227943 [   23.790248]  free_cma:0

Jul 14 00:27:36.227982 [   23.790259] DMA free:1680kB min:96kB low:120kB high:144kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:0kB isolated(anon):0kB isolated(file):0kB present:15976kB managed:15892kB mlocked:1044kB dirty:0kB writeback:0kB mapped:340kB shmem:8kB slab_reclaimable:92kB slab_unreclaimable:332kB kernel_stack:48kB pagetables:84kB unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:0 all_unreclaimable? yes

Jul 14 00:27:36.267995 [   23.790269] lowmem_reserve[]: 0 396 396 396

Jul 14 00:27:36.268052 [   23.790275] DMA32 free:2368kB min:2496kB low:3120kB high:3744kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:4kB isolated(anon):0kB isolated(file):0kB present:508308kB managed:408952kB mlocked:319
Jul 14 00:28:18.168603 <client 0x1c9b400 connected - now 1 clients>


_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xen.org
https://lists.xen.org/xen-devel

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: Linux 3.18 and Linux 4.1 reproducible OOM crashes
  2016-07-14 12:56 ` Linux 3.18 and Linux 4.1 reproducible OOM crashes Ian Jackson
@ 2016-07-14 23:34   ` Konrad Rzeszutek Wilk
  2016-07-15 10:00     ` Ian Jackson
  0 siblings, 1 reply; 7+ messages in thread
From: Konrad Rzeszutek Wilk @ 2016-07-14 23:34 UTC (permalink / raw)
  To: Ian Jackson; +Cc: xen-devel

On Thu, Jul 14, 2016 at 01:56:24PM +0100, Ian Jackson wrote:
> osstest service owner writes ("[linux-3.18 test] 97278: regressions - FAIL"):
> > flight 97278 linux-3.18 real [real]
> > http://logs.test-lab.xenproject.org/osstest/logs/97278/
> > 
> > Regressions :-(
> > 
> > Tests which did not succeed and are blocking,
> > including tests which could not be run:
> ...
> >  test-amd64-amd64-xl           6 xen-boot    fail REGR. vs. 96188
> 
> Loads of these.  It seems it can't boot at all.  Seems to affect amd64
> dom0 kernels but not i386 or armhf.  (See flight 97279 for 4.1
> results.)
> 
> It seems that the oom killer is tripping.  Obviously something is
> seriously wrong.
> 
> It seems likely that this is a real bug, but whether it's in Xen or
> Linux is not clear right now.  If it's in Linux it's because a similar
> patch has been backported to both.

Yikes! Is the tool able to narrow down the bisect that causes this?

Thanks.

_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xen.org
https://lists.xen.org/xen-devel

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: Linux 3.18 and Linux 4.1 reproducible OOM crashes
  2016-07-14 23:34   ` Konrad Rzeszutek Wilk
@ 2016-07-15 10:00     ` Ian Jackson
  2016-07-18 11:17       ` Linux 3.18 and Linux 4.1 reproducible OOM crashes under Xen Ian Jackson
  0 siblings, 1 reply; 7+ messages in thread
From: Ian Jackson @ 2016-07-15 10:00 UTC (permalink / raw)
  To: Konrad Rzeszutek Wilk; +Cc: xen-devel

Konrad Rzeszutek Wilk writes ("Re: [Xen-devel] Linux 3.18 and Linux 4.1 reproducible OOM crashes"):
> On Thu, Jul 14, 2016 at 01:56:24PM +0100, Ian Jackson wrote:
...
> > It seems likely that this is a real bug, but whether it's in Xen or
> > Linux is not clear right now.  If it's in Linux it's because a similar
> > patch has been backported to both.
> 
> Yikes! Is the tool able to narrow down the bisect that causes this?

It's working on it but it's a large range of commits.

(Also the bisector's scheduler is not working as well as it should,
which I am tinkering with to try to get it to perform better.)

Ian.

_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xen.org
https://lists.xen.org/xen-devel

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Linux 3.18 and Linux 4.1 reproducible OOM crashes under Xen
  2016-07-15 10:00     ` Ian Jackson
@ 2016-07-18 11:17       ` Ian Jackson
  2016-07-18 11:25         ` Michal Hocko
  0 siblings, 1 reply; 7+ messages in thread
From: Ian Jackson @ 2016-07-18 11:17 UTC (permalink / raw)
  To: Lukasz Odzioba, Konrad Rzeszutek Wilk
  Cc: Michal Hocko, stable, xen-devel, Andrew Morton, Linus Torvalds,
	Sasha Levin

The Xen Project CI system (`osstest') reports that the current tip of
the Linux 3.18 and Linux 4.1 stable branches have serious problems
booting Xen.

Our automatic bisector has completed its work and fingered (in 3.18)
a2d8c5147532 "mm/swap.c: flush lru pvecs on compound page arrival"
(and a backport of what seems to be the same commit to Linux 4.1).

The failure as accompanied by lots of oom killer activity.

For the full logs for a repro, see

  http://logs.test-lab.xenproject.org/osstest/logs/97435/test-amd64-amd64-xl-credit2/info.html

See particularly `serial-chardonnay0.log', near `Jul 16 14:16:32'.

More details below, including:

  - serial log extract from an earlier test failure
      (for our reference, flight 97278 job test-amd64-amd64-xl)

  - report from the osstest bisector

Thanks for your attention.

Ian.


Jul 14 00:27:33.715871 [   22.813518] rc.local invoked oom-killer: gfp_mask=0x84d0, order=0, oom_score_adj=0

Jul 14 00:27:33.899843 [   22.813541] rc.local cpuset=/ mems_allowed=0

Jul 14 00:27:33.907821 [   22.813550] CPU: 0 PID: 2676 Comm: rc.local Not tainted 3.18.37 #1

Jul 14 00:27:33.907861 [   22.813556] Hardware name: Intel Corporation SandyBridge Platform/To be filled by O.E.M., BIOS S1200BT.86B.02.00.0042.050820141549 05/08/2014

Jul 14 00:27:33.923914 [   22.813565]  0000000000000000 ffff8800024b7968 ffffffff817dcab5 00000000000084d0

Jul 14 00:27:33.931995 [   22.813573]  0000000000000000 ffff8800024b79c8 ffffffff8118082f ffff8800024b7988

Jul 14 00:27:33.940021 [   22.813582]  ffffffff8114c315 ffffffff817e5ee9 0000000000000001 ffff8800024b79c8

Jul 14 00:27:33.948034 [   22.813590] Call Trace:

Jul 14 00:27:33.948065 [   22.813600]  [<ffffffff817dcab5>] dump_stack+0x7c/0x98

Jul 14 00:27:33.955838 [   22.813609]  [<ffffffff8118082f>] dump_header.isra.11+0x8f/0x1e0

Jul 14 00:27:33.963831 [   22.813616]  [<ffffffff8114c315>] ? __delayacct_freepages_end+0x45/0x50

Jul 14 00:27:33.963872 [   22.813677]  [<ffffffff817e5ee9>] ? _raw_spin_unlock_irqrestore+0x29/0x90

Jul 14 00:27:33.971868 [   22.813685]  [<ffffffff8131f93e>] ? ___ratelimit+0xae/0x160

Jul 14 00:27:33.979838 [   22.813691]  [<ffffffff81180ddc>] oom_kill_process+0x20c/0x370

Jul 14 00:27:33.987847 [   22.813699]  [<ffffffff810bbc69>] ? has_capability_noaudit+0x19/0x20

Jul 14 00:27:33.987888 [   22.813706]  [<ffffffff811812d1>] out_of_memory+0x211/0x330

Jul 14 00:27:33.995847 [   22.813714]  [<ffffffff81186331>] __alloc_pages_nodemask+0xb11/0xb50

Jul 14 00:27:34.003844 [   22.813721]  [<ffffffff81186382>] __get_free_pages+0x12/0x70

Jul 14 00:27:34.011845 [   22.813728]  [<ffffffff811863f1>] get_zeroed_page+0x11/0x20

Jul 14 00:27:34.011882 [   22.813735]  [<ffffffff811a9c31>] __pud_alloc+0x21/0x120

Jul 14 00:27:34.019841 [   22.813742]  [<ffffffff811ab312>] handle_mm_fault+0x352/0xd50

Jul 14 00:27:34.027863 [   22.813749]  [<ffffffff811a550f>] ? follow_page_mask+0x2f/0x4f0

Jul 14 00:27:34.035848 [   22.813756]  [<ffffffff811af392>] ? find_vma+0x62/0x70

Jul 14 00:27:34.035885 [   22.813763]  [<ffffffff811a5b48>] __get_user_pages+0x178/0x660

Jul 14 00:27:34.043839 [   22.813770]  [<ffffffff811a607d>] get_user_pages+0x4d/0x50

Jul 14 00:27:34.051847 [   22.813776]  [<ffffffff811db546>] copy_strings.isra.26+0x176/0x310

Jul 14 00:27:34.051886 [   22.813782]  [<ffffffff811db71a>] copy_strings_kernel+0x3a/0x60

Jul 14 00:27:34.059844 [   22.813789]  [<ffffffff811dcad8>] do_execve_common.isra.33+0x418/0x680

Jul 14 00:27:34.067846 [   22.813795]  [<ffffffff811dd014>] SyS_execve+0x24/0x30

Jul 14 00:27:34.075839 [   22.813801]  [<ffffffff817e6b19>] stub_execve+0x69/0xa0

Jul 14 00:27:34.075876 [   22.813807] Mem-Info:

Jul 14 00:27:34.083841 [   22.813810] DMA per-cpu:

Jul 14 00:27:34.083872 [   22.813813] CPU    0: hi:    0, btch:   1 usd:   0

Jul 14 00:27:34.083905 [   22.813817] CPU    1: hi:    0, btch:   1 usd:   0

Jul 14 00:27:34.091848 [   22.813821] CPU    2: hi:    0, btch:   1 usd:   0

Jul 14 00:27:34.099849 [   22.813825] CPU    3: hi:    0, btch:   1 usd:   0

Jul 14 00:27:34.099885 [   22.813829] DMA32 per-cpu:

Jul 14 00:27:34.107837 [   22.813833] CPU    0: hi:  186, btch:  31 usd:  88

Jul 14 00:27:34.107872 [   22.813837] CPU    1: hi:  186, btch:  31 usd:  72

Jul 14 00:27:34.115855 [   22.813841] CPU    2: hi:  186, btch:  31 usd:  47

Jul 14 00:27:34.115892 [   22.813845] CPU    3: hi:  186, btch:  31 usd: 171

Jul 14 00:27:34.123844 [   22.813852] active_anon:0 inactive_anon:0 isolated_anon:0

Jul 14 00:27:34.131837 [   22.813852]  active_file:0 inactive_file:0 isolated_file:0

Jul 14 00:27:34.131873 [   22.813852]  unevictable:69 dirty:9 writeback:0 unstable:0

Jul 14 00:27:34.139847 [   22.813852]  free:1036 slab_reclaimable:1887 slab_unreclaimable:4091

Jul 14 00:27:34.147839 [   22.813852]  mapped:4099 shmem:154 pagetables:706 bounce:0

Jul 14 00:27:34.155836 [   22.813852]  free_cma:0

Jul 14 00:27:34.155867 [   22.813875] DMA free:1680kB min:96kB low:120kB high:144kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:4kB isolated(anon):0kB isolated(file):0kB present:15976kB managed:15892kB mlocked:1048kB dirty:0kB writeback:0kB mapped:408kB shmem:8kB slab_reclaimable:92kB slab_unreclaimable:328kB kernel_stack:64kB pagetables:148kB unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:0 all_unreclaimable? yes

Jul 14 00:27:34.195860 [   22.813896] lowmem_reserve[]: 0 396 396 396

Jul 14 00:27:34.195898 [   22.813908] DMA32 free:2464kB min:2496kB low:3120kB high:3744kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:272kB isolated(anon):0kB isolated(file):0kB present:508308kB managed:408952kB mlocked:32180kB dirty:36kB writeback:0kB mapped:15988kB shmem:608kB slab_reclaimable:7456kB slab_unreclaimable:16036kB kernel_stack:2720kB pagetables:2676kB unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:0 all_unreclaimable? yes

Jul 14 00:27:34.243847 [   22.813929] lowmem_reserve[]: 0 0 0 0

Jul 14 00:27:34.243882 [   22.813936] DMA: 0*4kB 0*8kB 1*16kB (R) 0*32kB 0*64kB 1*128kB (R) 0*256kB 1*512kB (R) 1*1024kB (R) 0*2048kB 0*4096kB = 1680kB

Jul 14 00:27:34.251860 [   22.813956] DMA32: 0*4kB 0*8kB 1*16kB (R) 1*32kB (R) 0*64kB 1*128kB (R) 1*256kB (R) 0*512kB 0*1024kB 1*2048kB (R) 0*4096kB = 2480kB

Jul 14 00:27:34.267804 [   22.813976] 15037 total pagecache pages

Jul 14 00:27:34.267825 [   22.813980] 0 pages in swap cache

Jul 14 00:27:34.275851 [   22.813984] Swap cache stats: add 0, delete 0, find 0/0

Jul 14 00:27:34.283845 [   22.813988] Free swap  = 1949692kB

Jul 14 00:27:34.283879 [   22.813991] Total swap = 1949692kB

Jul 14 00:27:34.283908 [   22.813994] 131071 pages RAM

Jul 14 00:27:34.291912 [   22.813997] 0 pages HighMem/MovableOnly

Jul 14 00:27:34.291946 [   22.814000] 24860 pages reserved

Jul 14 00:27:34.299797 [   22.814003] [ pid ]   uid  tgid total_vm      rss nr_ptes swapents oom_score_adj name

Jul 14 00:27:34.307798 [   22.814013] [ 1316]     0  1316     9825      819      21        0         -1000 udevd

Jul 14 00:27:34.315792 [   22.814021] [ 2038]     0  2038     1084       25       7        0             0 ntpdate

Jul 14 00:27:34.323800 [   22.814027] [ 2040]     0  2040     1569      172       9        0             0 lockfile-create

Jul 14 00:27:34.331829 [   22.814034] [ 2047]     0  2047     1060      331       7        0             0 startpar

Jul 14 00:27:34.339797 [   22.814040] [ 2061]     0  2061     9270      668      24        0             0 rpcbind

Jul 14 00:27:34.347796 [   22.814047] [ 2076]   106  2076     9320      700      22        0             0 rpc.statd

Jul 14 00:27:34.355808 [   22.814054] [ 2088]     0  2088     6892       58      18        0             0 rpc.idmapd

Jul 14 00:27:34.363800 [   22.814060] [ 2131]     0  2131     1084      410       7        0             0 rc

Jul 14 00:27:34.371807 [   22.814066] [ 2137]     0  2137     3202      648      13        0             0 startpar

Jul 14 00:27:34.379805 [   22.814073] [ 2160]     0  2160    65721      788      33        0             0 rsyslogd

Jul 14 00:27:34.387805 [   22.814079] [ 2175]     0  2175     1064      394       8        0             0 acpid

Jul 14 00:27:34.395808 [   22.814085] [ 2211]     0  2211    19431     1167      41        0             0 apache2

Jul 14 00:27:34.403799 [   22.814091] [ 2214]    33  2214    91722     1032      69        0             0 apache2

Jul 14 00:27:34.411802 [   22.814098] [ 2216]    33  2216    91722     1031      69        0             0 apache2

Jul 14 00:27:34.419800 [   22.814104] [ 2422]     0  2422     4756       42      13        0             0 atd

Jul 14 00:27:34.427806 [   22.814110] [ 2423]     0  2423     6869      590      18        0             0 cron

Jul 14 00:27:34.435804 [   22.814117] [ 2606]     0  2606     8780      883      21        0             0 oxenstored

Jul 14 00:27:34.443846 [   22.814123] [ 2619]   105  2619    10560      551      24        0             0 dbus-daemon

Jul 14 00:27:34.451802 [   22.814130] [ 2624]     0  2624    13796      700      28        0         -1000 sshd

Jul 14 00:27:34.459799 [   22.814136] [ 2631]     0  2631    22682       57      19        0             0 xenconsoled

Jul 14 00:27:34.467815 [   22.814143] [ 2643]   104  2643    13313      828      26        0             0 exim4

Jul 14 00:27:34.475801 [   22.814149] [ 2652]     0  2652    60164     1974      78        0             0 qemu-system-i38

Jul 14 00:27:34.483868 [   22.814156] [ 2656]     0  2656    20648     1510      42        0             0 sshd

Jul 14 00:27:34.491859 [   22.814162] [ 2660]   107  2660     8345     1093      21        0             0 ntpd

Jul 14 00:27:34.499864 [   22.814169] [ 2673]     0  2673     1084      411       7        0             0 rc.local

Jul 14 00:27:34.507852 [   22.814175] [ 2676]     0  2676     1084       44       6        0             0 rc.local

Jul 14 00:27:34.515859 [   22.814181] [ 2677]  1000  2677    20648      946      41        0             0 sshd

Jul 14 00:27:34.523849 [   22.814188] Out of memory: Kill process 2652 (qemu-system-i38) score 3 or sacrifice child

Jul 14 00:27:34.531862 [   22.814194] Killed process 2652 (qemu-system-i38) total-vm:240656kB, anon-rss:2440kB, file-rss:5456kB

Jul 14 00:27:34.539871 [   22.822755] sshd invoked oom-killer: gfp_mask=0x2000d0, order=0, oom_score_adj=0

Jul 14 00:27:34.547800 [   22.822760] sshd cpuset=/ mems_allowed=0

Jul 14 00:27:34.555868 [   22.822764] CPU: 1 PID: 2677 Comm: sshd Not tainted 3.18.37 #1

Jul 14 00:27:34.563847 [   22.822767] Hardware name: Intel Corporation SandyBridge Platform/To be filled by O.E.M., BIOS S1200BT.86B.02.00.0042.050820141549 05/08/2014

Jul 14 00:27:34.571852 [   22.822771]  0000000000000000 ffff8800024fb990 ffffffff817dcab5 00000000002000d0

Jul 14 00:27:34.579860 [   22.822775]  0000000000000000 ffff8800024fb9f0 ffffffff8118082f ffff8800024fb9b0

Jul 14 00:27:34.587803 [   22.822779]  ffffffff8114c315 ffffffff817e5ee9 0000000000000001 ffff8800024fb9f0

Jul 14 00:27:34.595862 [   22.822783] Call Trace:

Jul 14 00:27:34.595892 [   22.822787]  [<ffffffff817dcab5>] dump_stack+0x7c/0x98

Jul 14 00:27:34.603859 [   22.822791]  [<ffffffff8118082f>] dump_header.isra.11+0x8f/0x1e0

Jul 14 00:27:34.611843 [   22.822794]  [<ffffffff8114c315>] ? __delayacct_freepages_end+0x45/0x50

Jul 14 00:27:34.619891 [   22.822798]  [<ffffffff817e5ee9>] ? _raw_spin_unlock_irqrestore+0x29/0x90

Jul 14 00:27:34.619933 [   22.822801]  [<ffffffff8131f93e>] ? ___ratelimit+0xae/0x160

Jul 14 00:27:34.627853 [   22.822804]  [<ffffffff81180ddc>] oom_kill_process+0x20c/0x370

Jul 14 00:27:34.635900 [   22.822808]  [<ffffffff810bbc69>] ? has_capability_noaudit+0x19/0x20

Jul 14 00:27:34.643851 [   22.822812]  [<ffffffff811812d1>] out_of_memory+0x211/0x330

Jul 14 00:27:34.643888 [   22.822815]  [<ffffffff81186331>] __alloc_pages_nodemask+0xb11/0xb50

Jul 14 00:27:34.651845 [   22.822820]  [<ffffffff811cb40c>] new_slab+0x2bc/0x320

Jul 14 00:27:34.660009 [   22.822823]  [<ffffffff811cc770>] __slab_alloc.constprop.69+0x440/0x570

Jul 14 00:27:34.667913 [   22.822828]  [<ffffffff81054f3e>] ? xen_leave_lazy_mmu+0xe/0x10

Jul 14 00:27:34.667952 [   22.822831]  [<ffffffff811b6721>] ? anon_vma_fork+0x71/0x140

Jul 14 00:27:34.675855 [   22.822835]  [<ffffffff811b651b>] ? anon_vma_clone+0x6b/0x200

Jul 14 00:27:34.683863 [   22.822838]  [<ffffffff811b6721>] ? anon_vma_fork+0x71/0x140

Jul 14 00:27:34.691849 [   22.822841]  [<ffffffff811cdad6>] kmem_cache_alloc+0x116/0x150

Jul 14 00:27:34.691887 [   22.822844]  [<ffffffff811b6721>] anon_vma_fork+0x71/0x140

Jul 14 00:27:34.699848 [   22.822848]  [<ffffffff810b2017>] copy_process.part.45+0x1687/0x1a50

Jul 14 00:27:34.707853 [   22.822852]  [<ffffffff810b2577>] do_fork+0xb7/0x3b0

Jul 14 00:27:34.707888 [   22.822856]  [<ffffffff811f3a34>] ? get_unused_fd_flags+0x34/0x40

Jul 14 00:27:34.715857 [   22.822860]  [<ffffffff811f3a88>] ? __fd_install+0x48/0x60

Jul 14 00:27:34.723872 [   22.822862]  [<ffffffff810b28f1>] SyS_clone+0x11/0x20

Jul 14 00:27:34.723910 [   22.822866]  [<ffffffff817e68c9>] stub_clone+0x69/0x90

Jul 14 00:27:34.731848 [   22.822868]  [<ffffffff817e6589>] ? system_call_fastpath+0x12/0x17

Jul 14 00:27:34.739823 [   22.822871] Mem-Info:

Jul 14 00:27:34.739841 [   22.822872] DMA per-cpu:

Jul 14 00:27:34.747843 [   22.822874] CPU    0: hi:    0, btch:   1 usd:   0

Jul 14 00:27:34.747879 [   22.822876] CPU    1: hi:    0, btch:   1 usd:   0

Jul 14 00:27:34.755836 [   22.822878] CPU    2: hi:    0, btch:   1 usd:   0

Jul 14 00:27:34.755870 [   22.822880] CPU    3: hi:    0, btch:   1 usd:   0

Jul 14 00:27:34.763850 [   22.822882] DMA32 per-cpu:

Jul 14 00:27:34.763881 [   22.822884] CPU    0: hi:  186, btch:  31 usd:  88

Jul 14 00:27:34.771838 [   22.822886] CPU    1: hi:  186, btch:  31 usd: 157

Jul 14 00:27:34.779837 [   22.822888] CPU    2: hi:  186, btch:  31 usd:  47

Jul 14 00:27:34.779872 [   22.822890] CPU    3: hi:  186, btch:  31 usd: 173

Jul 14 00:27:34.787884 [   22.822893] active_anon:0 inactive_anon:0 isolated_anon:0

Jul 14 00:27:34.787920 [   22.822893]  active_file:0 inactive_file:0 isolated_file:0

Jul 14 00:27:34.795844 [   22.822893]  unevictable:69 dirty:9 writeback:0 unstable:0

Jul 14 00:27:34.803834 [   22.822893]  free:1036 slab_reclaimable:1887 slab_unreclaimable:4091

Jul 14 00:27:34.811846 [   22.822893]  mapped:3101 shmem:154 pagetables:650 bounce:0

Jul 14 00:27:34.811884 [   22.822893]  free_cma:0

Jul 14 00:27:34.819842 [   22.822904] DMA free:1680kB min:96kB low:120kB high:144kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:4kB isolated(anon):0kB isolated(file):0kB present:15976kB managed:15892kB mlocked:1048kB dirty:0kB writeback:0kB mapped:336kB shmem:8kB slab_reclaimable:92kB slab_unreclaimable:328kB kernel_stack:64kB pagetables:148kB unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:0 all_unreclaimable? yes

Jul 14 00:27:34.859831 [   22.822914] lowmem_reserve[]: 0 396 396 396

Jul 14 00:27:34.859866 [   22.822920] DMA32 free:2464kB min:2496kB low:3120kB high:3744kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:272kB isolated(anon):0kB isolated(file):0kB present:508308kB managed:408952kB mlocked:32180kB dirty:36kB writeback:0kB mapped:12068kB shmem:608kB slab_reclaimable:7456kB slab_unreclaimable:16036kB kernel_stack:2720kB pagetables:2452kB unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:0 all_unreclaimable? yes

Jul 14 00:27:34.899849 [   22.822930] lowmem_reserve[]: 0 0 0 0

Jul 14 00:27:34.907838 [   22.822934] DMA: 0*4kB 0*8kB 1*16kB (R) 0*32kB 0*64kB 1*128kB (R) 0*256kB 1*512kB (R) 1*1024kB (R) 0*2048kB 0*4096kB = 1680kB

Jul 14 00:27:34.915876 [   22.822943] DMA32: 0*4kB 0*8kB 1*16kB (R) 1*32kB (R) 0*64kB 1*128kB (R) 1*256kB (R) 0*512kB 0*1024kB 1*2048kB (R) 0*4096kB = 2480kB

Jul 14 00:27:34.931842 [   22.822953] 15037 total pagecache pages

Jul 14 00:27:34.931878 [   22.822955] 0 pages in swap cache

Jul 14 00:27:34.939850 [   22.822957] Swap cache stats: add 0, delete 0, find 0/0

Jul 14 00:27:34.939888 [   22.822959] Free swap  = 1949692kB

Jul 14 00:27:34.947842 [   22.822960] Total swap = 1949692kB

Jul 14 00:27:34.947874 [   22.822962] 131071 pages RAM

Jul 14 00:27:34.947903 [   22.822963] 0 pages HighMem/MovableOnly

Jul 14 00:27:34.955868 [   22.822965] 24860 pages reserved

Jul 14 00:27:34.955900 [   22.822966] [ pid ]   uid  tgid total_vm      rss nr_ptes swapents oom_score_adj name

Jul 14 00:27:34.963860 [   22.822971] [ 1316]     0  1316     9825      819      21        0         -1000 udevd

Jul 14 00:27:34.971874 [   22.822976] [ 2038]     0  2038     1084       25       7        0             0 ntpdate

Jul 14 00:27:34.979864 [   22.822978] [ 2040]     0  2040     1569      172       9        0             0 lockfile-create

Jul 14 00:27:34.987862 [   22.822981] [ 2047]     0  2047     1060      331       7        0             0 startpar

Jul 14 00:27:34.995869 [   22.822984] [ 2061]     0  2061     9270      668      24        0             0 rpcbind

Jul 14 00:27:35.003863 [   22.822987] [ 2076]   106  2076     9320      700      22        0             0 rpc.statd

Jul 14 00:27:35.019848 [   22.822991] [ 2088]     0  2088     6892       58      18        0             0 rpc.idmapd

Jul 14 00:27:35.027853 [   22.822994] [ 2131]     0  2131     1084      410       7        0             0 rc

Jul 14 00:27:35.035843 [   22.822997] [ 2137]     0  2137     3202      648      13        0             0 startpar

Jul 14 00:27:35.043857 [   22.823000] [ 2160]     0  2160    65721      788      33        0             0 rsyslogd

Jul 14 00:27:35.051846 [   22.823003] [ 2175]     0  2175     1064      394       8        0             0 acpid

Jul 14 00:27:35.059852 [   22.823006] [ 2211]     0  2211    19431     1167      41        0             0 apache2

Jul 14 00:27:35.067848 [   22.823009] [ 2214]    33  2214    91722     1032      69        0             0 apache2

Jul 14 00:27:35.075852 [   22.823012] [ 2216]    33  2216    91722     1031      69        0             0 apache2

Jul 14 00:27:35.083847 [   22.823015] [ 2422]     0  2422     4756       42      13        0             0 atd

Jul 14 00:27:35.091857 [   22.823018] [ 2423]     0  2423     6869      590      18        0             0 cron

Jul 14 00:27:35.099857 [   22.823021] [ 2606]     0  2606     8780      883      21        0             0 oxenstored

Jul 14 00:27:35.107854 [   22.823024] [ 2619]   105  2619    10560      551      24        0             0 dbus-daemon

Jul 14 00:27:35.115866 [   22.823027] [ 2624]     0  2624    13796      700      28        0         -1000 sshd

Jul 14 00:27:35.123797 [   22.823030] [ 2631]     0  2631    22682       57      19        0             0 xenconsoled

Jul 14 00:27:35.131849 [   22.823034] [ 2643]   104  2643    13313      828      26        0             0 exim4

Jul 14 00:27:35.139844 [   22.823037] [ 2656]     0  2656    20648     1510      42        0             0 sshd

Jul 14 00:27:35.147846 [   22.823040] [ 2660]   107  2660     8345     1093      21        0             0 ntpd

Jul 14 00:27:35.155856 [   22.823043] [ 2673]     0  2673     1084      411       7        0             0 rc.local

Jul 14 00:27:35.163847 [   22.823046] [ 2676]     0  2676     1084       44       6        0             0 rc.local

Jul 14 00:27:35.171852 [   22.823049] [ 2677]  1000  2677    20648      946      41        0             0 sshd

Jul 14 00:27:35.179883 [   22.823052] Out of memory: Kill process 2656 (sshd) score 2 or sacrifice child

Jul 14 00:27:35.187860 [   22.823055] Killed process 2677 (sshd) total-vm:82592kB, anon-rss:848kB, file-rss:2936kB

Jul 14 00:27:35.195838 [   22.825444] oxenstored invoked oom-killer: gfp_mask=0x280da, order=0, oom_score_adj=0

Jul 14 00:27:35.203861 [   22.825449] oxenstored cpuset=/ mems_allowed=0

Jul 14 00:27:35.203897 [   22.825453] CPU: 2 PID: 2606 Comm: oxenstored Not tainted 3.18.37 #1

Jul 14 00:27:35.211845 [   22.825457] Hardware name: Intel Corporation SandyBridge Platform/To be filled by O.E.M., BIOS S1200BT.86B.02.00.0042.050820141549 05/08/2014

Jul 14 00:27:35.227847 [   22.825461]  0000000000000000 ffff880008dc3ad8 ffffffff817dcab5 00000000000280da

Jul 14 00:27:35.235847 [   22.825466]  0000000000000000 ffff880008dc3b38 ffffffff8118082f ffff880008dc3af8

Jul 14 00:27:35.243891 [   22.825469]  ffffffff8114c315 ffffffff817e5ee9 0000000000000001 ffff880008dc3b38

Jul 14 00:27:35.243937 [   22.825473] Call Trace:

Jul 14 00:27:35.251845 [   22.825477]  [<ffffffff817dcab5>] dump_stack+0x7c/0x98

Jul 14 00:27:35.251881 [   22.825481]  [<ffffffff8118082f>] dump_header.isra.11+0x8f/0x1e0

Jul 14 00:27:35.259846 [   22.825484]  [<ffffffff8114c315>] ? __delayacct_freepages_end+0x45/0x50

Jul 14 00:27:35.267864 [   22.825488]  [<ffffffff817e5ee9>] ? _raw_spin_unlock_irqrestore+0x29/0x90

Jul 14 00:27:35.275843 [   22.825492]  [<ffffffff8131f93e>] ? ___ratelimit+0xae/0x160

Jul 14 00:27:35.283904 [   22.825494]  [<ffffffff81180ddc>] oom_kill_process+0x20c/0x370

Jul 14 00:27:35.283942 [   22.825498]  [<ffffffff810bbc69>] ? has_capability_noaudit+0x19/0x20

Jul 14 00:27:35.291845 [   22.825501]  [<ffffffff811812d1>] out_of_memory+0x211/0x330

Jul 14 00:27:35.299881 [   22.825505]  [<ffffffff81186331>] __alloc_pages_nodemask+0xb11/0xb50

Jul 14 00:27:35.307834 [   22.825509]  [<ffffffff811ab988>] handle_mm_fault+0x9c8/0xd50

Jul 14 00:27:35.307871 [   22.825513]  [<ffffffff811ea170>] ? poll_select_copy_remaining+0x130/0x130

Jul 14 00:27:35.315942 [   22.825517]  [<ffffffff810a68dd>] __do_page_fault+0x1bd/0x590

Jul 14 00:27:35.323846 [   22.825521]  [<ffffffff[   23.782240] getty invoked oom-killer: gfp_mask=0x280da, order=0, oom_score_adj=0

Jul 14 00:27:35.331856 [   23.782247] getty cpuset=/ mems_allowed=0

Jul 14 00:27:35.339842 [   23.782252] CPU: 3 PID: 2685 Comm: getty Not tainted 3.18.37 #1

Jul 14 00:27:35.339884 [   23.782255] Hardware name: Intel Corporation SandyBridge Platform/To be filled by O.E.M., BIOS S1200BT.86B.02.00.0042.050820141549 05/08/2014

Jul 14 00:27:35.355849 [   23.782259]  0000000000000000 ffff8800158db7f8 ffffffff817dcab5 00000000000280da

Jul 14 00:27:35.363925 [   23.782263]  0000000000000000 ffff8800158db858 ffffffff8118082f ffff8800158db818

Jul 14 00:27:35.371864 [   23.782268]  ffffffff8114c315 ffffffff817e5ee9 0000000000000001 ffff8800158db858

Jul 14 00:27:35.379855 [   23.782272] Call Trace:

Jul 14 00:27:35.379886 [   23.782278]  [<ffffffff817dcab5>] dump_stack+0x7c/0x98

Jul 14 00:27:35.387850 [   23.782283]  [<ffffffff8118082f>] dump_header.isra.11+0x8f/0x1e0

Jul 14 00:27:35.387888 [   23.782287]  [<ffffffff8114c315>] ? __delayacct_freepages_end+0x45/0x50

Jul 14 00:27:35.395877 [   23.782291]  [<ffffffff817e5ee9>] ? _raw_spin_unlock_irqrestore+0x29/0x90

Jul 14 00:27:35.403858 [   23.782295]  [<ffffffff8131f93e>] ? ___ratelimit+0xae/0x160

Jul 14 00:27:35.411839 [   23.782298]  [<ffffffff81180ddc>] oom_kill_process+0x20c/0x370

Jul 14 00:27:35.419845 [   23.782301]  [<ffffffff811812d1>] out_of_memory+0x211/0x330

Jul 14 00:27:35.419883 [   23.782305]  [<ffffffff81186331>] __alloc_pages_nodemask+0xb11/0xb50

Jul 14 00:27:35.427868 [   23.782309]  [<ffffffff811ab988>] handle_mm_fault+0x9c8/0xd50

Jul 14 00:27:35.435856 [   23.782320]  [<ffffffffa009d912>] ? search_dir+0xc2/0x120 [ext4]

Jul 14 00:27:35.443840 [   23.782324]  [<ffffffff810a68dd>] __do_page_fault+0x1bd/0x590

Jul 14 00:27:35.443877 [   23.782329]  [<ffffffff81233736>] ? dquot_file_open+0x16/0x50

Jul 14 00:27:35.451860 [   23.782334]  [<ffffffffa008c953>] ? ext4_file_open+0x43/0x180 [ext4]

Jul 14 00:27:35.459849 [   23.782339]  [<ffffffffa008c910>] ? ext4_check_all_de+0xb0/0xb0 [ext4]

Jul 14 00:27:35.467884 [   23.782343]  [<ffffffff810a6cdc>] do_page_fault+0x2c/0x40

Jul 14 00:27:35.467920 [   23.782346]  [<ffffffff817e81b8>] page_fault+0x28/0x30

Jul 14 00:27:35.475841 [   23.782350]  [<ffffffff811a47a8>] ? copy_page_to_iter+0x68/0x3c0

Jul 14 00:27:35.483845 [   23.782353]  [<ffffffff8117ed88>] generic_file_read_iter+0x278/0x5d0

Jul 14 00:27:35.491842 [   23.782357]  [<ffffffff811d4ad9>] new_sync_read+0x79/0xb0

Jul 14 00:27:35.491880 [   23.782360]  [<ffffffff811d5303>] vfs_read+0x93/0x180

Jul 14 00:27:35.499846 [   23.782364]  [<ffffffff811d5e61>] SyS_read+0x41/0xb0

Jul 14 00:27:35.507832 [   23.782367]  [<ffffffff817e6589>] system_call_fastpath+0x12/0x17

Jul 14 00:27:35.507869 [   23.782370] Mem-Info:

Jul 14 00:27:35.515838 [   23.782371] DMA per-cpu:

Jul 14 00:27:35.515869 [   23.782373] CPU    0: hi:    0, btch:   1 usd:   0

Jul 14 00:27:35.523838 [   23.782375] CPU    1: hi:    0, btch:   1 usd:   0

Jul 14 00:27:35.523873 [   23.782377] CPU    2: hi:    0, btch:   1 usd:   0

Jul 14 00:27:35.531843 [   23.782379] CPU    3: hi:    0, btch:   1 usd:   0

Jul 14 00:27:35.531878 [   23.782381] DMA32 per-cpu:

Jul 14 00:27:35.539836 [   23.782382] CPU    0: hi:  186, btch:  31 usd:   0

Jul 14 00:27:35.539873 [   23.782384] CPU    1: hi:  186, btch:  31 usd:   0

Jul 14 00:27:35.547841 [   23.782386] CPU    2: hi:  186, btch:  31 usd:   0

Jul 14 00:27:35.555833 [   23.782388] CPU    3: hi:  186, btch:  31 usd:   0

Jul 14 00:27:35.555868 [   23.782392] active_anon:0 inactive_anon:0 isolated_anon:0

Jul 14 00:27:35.563857 [   23.782392]  active_file:0 inactive_file:0 isolated_file:0

Jul 14 00:27:35.563895 [   23.782392]  unevictable:1 dirty:0 writeback:18 unstable:0

Jul 14 00:27:35.571859 [   23.782392]  free:1060 slab_reclaimable:1888 slab_unreclaimable:4074

Jul 14 00:27:35.579860 [   23.782392]  mapped:2923 shmem:153 pagetables:588 bounce:0

Jul 14 00:27:35.587840 [   23.782392]  free_cma:0

Jul 14 00:27:35.587872 [   23.782403] DMA free:1668kB min:96kB low:120kB high:144kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:0kB isolated(anon):0kB isolated(file):0kB present:15976kB managed:15892kB mlocked:1044kB dirty:0kB writeback:4kB mapped:340kB shmem:8kB slab_reclaimable:92kB slab_unreclaimable:332kB kernel_stack:48kB pagetables:116kB unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:0 all_unreclaimable? yes

Jul 14 00:27:35.627846 [   23.782413] lowmem_reserve[]: 0 396 396 396

Jul 14 00:27:35.627882 [   23.782419] DMA32 free:2572kB min:2496kB low:3120kB high:3744kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:4kB isolated(anon):0kB isolated(file):0kB present:508308kB managed:408952kB mlocked:31912kB dirty:0kB writeback:68kB mapped:11352kB shmem:604kB slab_reclaimable:7460kB slab_unreclaimable:15964kB kernel_stack:2400kB pagetables:2236kB unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:36 all_unreclaimable? yes

Jul 14 00:27:35.675846 [   23.782431] lowmem_reserve[]: 0 0 0 0

Jul 14 00:27:35.675882 [   23.782433] DMA: 0*4kB 0*8kB 0*16kB 0*32kB 0*64kB 1*128kB (R) 0*256kB 1*512kB (R) 1*1024kB (R) 0*2048kB 0*4096kB = 1664kB

Jul 14 00:27:35.683866 [   23.782442] DMA32: 29*4kB (UR) 0*8kB 3*16kB (R) 0*32kB 1*64kB (R) 1*128kB (R) 1*256kB (R) 0*512kB 0*1024kB 1*2048kB (R) 0*4096kB = 2660kB

Jul 14 00:27:35.699851 [   23.782452] 15076 total pagecache pages

Jul 14 00:27:35.699885 [   23.782455] 0 pages in swap cache

Jul 14 00:27:35.707849 [   23.782465] Swap cache stats: add 0, delete 0, find 0/0

Jul 14 00:27:35.715836 [   23.782468] Free swap  = 1949692kB

Jul 14 00:27:35.715869 [   23.782472] Total swap = 1949692kB

Jul 14 00:27:35.715899 [   23.782476] 131071 pages RAM

Jul 14 00:27:35.723847 [   23.782480] 0 pages HighMem/MovableOnly

Jul 14 00:27:35.723882 [   23.782484] 24860 pages reserved

Jul 14 00:27:35.731838 [   23.782487] [ pid ]   uid  tgid total_vm      rss nr_ptes swapents oom_score_adj name

Jul 14 00:27:35.739850 [   23.782495] [ 1316]     0  1316     9825      819      21        0         -1000 udevd

Jul 14 00:27:35.747851 [   23.782501] [ 2038]     0  2038     1084       25       7        0             0 ntpdate

Jul 14 00:27:35.755844 [   23.782506] [ 2040]     0  2040     1569      172       9        0             0 lockfile-create

Jul 14 00:27:35.763856 [   23.782509] [ 2047]     0  2047     1060      331       7        0             0 startpar

Jul 14 00:27:35.771890 [   23.782512] [ 2061]     0  2061     9270      668      24        0             0 rpcbind

Jul 14 00:27:35.779829 [   23.782516] [ 2076]   106  2076     9320      700      22        0             0 rpc.statd

Jul 14 00:27:35.787856 [   23.782519] [ 2088]     0  2088     6892       58      18        0             0 rpc.idmapd

Jul 14 00:27:35.795799 [   23.782522] [ 2160]     0  2160    65721      785      33        0             0 rsyslogd

Jul 14 00:27:35.803857 [   23.782525] [ 2175]     0  2175     1064      394       8        0             0 acpid

Jul 14 00:27:35.811854 [   23.782528] [ 2211]     0  2211    19431     1167      41        0             0 apache2

Jul 14 00:27:35.820250 [   23.782532] [ 2216]    33  2216    91722     1031      69        0             0 apache2

Jul 14 00:27:35.827884 [   23.782534] [ 2422]     0  2422     4756       42      13        0             0 atd

Jul 14 00:27:35.835897 [   23.782537] [ 2423]     0  2423     6869      590      18        0             0 cron

Jul 14 00:27:35.843893 [   23.782541] [ 2606]     0  2606     8780      883      21        0             0 oxenstored

Jul 14 00:27:35.851883 [   23.782544] [ 2619]   105  2619    10560      551      24        0             0 dbus-daemon

Jul 14 00:27:35.859858 [   23.782547] [ 2624]     0  2624    13796      700      28        0         -1000 sshd

Jul 14 00:27:35.867866 [   23.782550] [ 2631]     0  2631    22682       57      19        0             0 xenconsoled

Jul 14 00:27:35.875902 [   23.782554] [ 2643]   104  2643    13313      828      26        0             0 exim4

Jul 14 00:27:35.883887 [   23.782557] [ 2660]   107  2660     8346     1093      21        0             0 ntpd

Jul 14 00:27:35.891912 [   23.782560] [ 2681]     0  2681     2650      399      10        0             0 getty

Jul 14 00:27:35.899869 [   23.782563] [ 2682]     0  2682     3166      417      12        0             0 getty

Jul 14 00:27:35.907853 [   23.782566] [ 2683]     0  2683     1063      173       7        0             0 getty

Jul 14 00:27:35.915859 [   23.782569] [ 2684]     0  2684     3166      416      12        0             0 getty

Jul 14 00:27:35.923807 [   23.782572] [ 2685]     0  2685     3560      425      12        0             0 getty

Jul 14 00:27:35.931919 [   23.782575] [ 2686]     0  2686     3166      425      11        0             0 getty

Jul 14 00:27:35.939914 [   23.782578] [ 2687]     0  2687     1591      404       8        0             0 getty

Jul 14 00:27:35.947888 [   23.782581] [ 2688]     0  2688     3166      437      12        0             0 getty

Jul 14 00:27:35.956047 [   23.782587] [ 2690]     0  2690     3166       35       9        0             0 getty

Jul 14 00:27:35.964079 [   23.782592] [ 2691]     0  2691     3166       34       9        0             0 getty

Jul 14 00:27:35.972103 [   23.782597] [ 2693]     0  2693     9825      195      19        0         -1000 udevd

Jul 14 00:27:35.980089 [   23.782601] [ 2694]     0  2694     9825      195      18        0         -1000 udevd

Jul 14 00:27:35.987933 [   23.782604] Out of memory: Kill process 2211 (apache2) score 1 or sacrifice child

Jul 14 00:27:35.996043 [   23.782607] Killed process 2216 (apache2) total-vm:366888kB, anon-rss:1556kB, file-rss:2568kB

Jul 14 00:27:36.004061 [   23.790140] getty invoked oom-killer: gfp_mask=0x200da, order=0, oom_score_adj=0

Jul 14 00:27:36.011954 [   23.790145] getty cpuset=/ mems_allowed=0

Jul 14 00:27:36.019911 [   23.790150] CPU: 0 PID: 2681 Comm: getty Not tainted 3.18.37 #1

Jul 14 00:27:36.019951 [   23.790152] Hardware name: Intel Corporation SandyBridge Platform/To be filled by O.E.M., BIOS S1200BT.86B.02.00.0042.050820141549 05/08/2014

Jul 14 00:27:36.035934 [   23.790156]  0000000000000000 ffff88001220ba78 ffffffff817dcab5 00000000000200da

Jul 14 00:27:36.043951 [   23.790161]  0000000000000000 ffff88001220bad8 ffffffff8118082f ffff88001220ba98

Jul 14 00:27:36.051973 [   23.790166]  ffffffff8114c315 ffffffff817e5ee9 0000000000000001 ffff88001220bad8

Jul 14 00:27:36.059965 [   23.790170] Call Trace:

Jul 14 00:27:36.059997 [   23.790174]  [<ffffffff817dcab5>] dump_stack+0x7c/0x98

Jul 14 00:27:36.067938 [   23.790178]  [<ffffffff8118082f>] dump_header.isra.11+0x8f/0x1e0

Jul 14 00:27:36.067977 [   23.790181]  [<ffffffff8114c315>] ? __delayacct_freepages_end+0x45/0x50

Jul 14 00:27:36.076024 [   23.790185]  [<ffffffff817e5ee9>] ? _raw_spin_unlock_irqrestore+0x29/0x90

Jul 14 00:27:36.083947 [   23.790188]  [<ffffffff8131f93e>] ? ___ratelimit+0xae/0x160

Jul 14 00:27:36.092010 [   23.790191]  [<ffffffff81180ddc>] oom_kill_process+0x20c/0x370

Jul 14 00:27:36.099947 [   23.790195]  [<ffffffff811812d1>] out_of_memory+0x211/0x330

Jul 14 00:27:36.099991 [   23.790198]  [<ffffffff81186331>] __alloc_pages_nodemask+0xb11/0xb50

Jul 14 00:27:36.107888 [   23.790202]  [<ffffffff811a9757>] do_cow_fault.isra.96+0x47/0x2a0

Jul 14 00:27:36.116087 [   23.790207]  [<ffffffff81054bf9>] ? __raw_callee_save_xen_pmd_val+0x11/0x1e

Jul 14 00:27:36.123936 [   23.790210]  [<ffffffff811ab675>] handle_mm_fault+0x6b5/0xd50

Jul 14 00:27:36.124004 [   23.790214]  [<ffffffff810a68dd>] __do_page_fault+0x1bd/0x590

Jul 14 00:27:36.132061 [   23.790217]  [<ffffffff81198939>] ? vm_mmap_pgoff+0x99/0xc0

Jul 14 00:27:36.140090 [   23.790220]  [<ffffffff810a6cdc>] do_page_fault+0x2c/0x40

Jul 14 00:27:36.148022 [   23.790224]  [<ffffffff817e81b8>] page_fault+0x28/0x30

Jul 14 00:27:36.148057 [   23.790226] Mem-Info:

Jul 14 00:27:36.148083 [   23.790227] DMA per-cpu:

Jul 14 00:27:36.155974 [   23.790229] CPU    0: hi:    0, btch:   1 usd:   0

Jul 14 00:27:36.155995 [   23.790231] CPU    1: hi:    0, btch:   1 usd:   0

Jul 14 00:27:36.164132 [   23.790233] CPU    2: hi:    0, btch:   1 usd:   0

Jul 14 00:27:36.172145 [   23.790235] CPU    3: hi:    0, btch:   1 usd:   0

Jul 14 00:27:36.172208 [   23.790237] DMA32 per-cpu:

Jul 14 00:27:36.180000 [   23.790238] CPU    0: hi:  186, btch:  31 usd:   0

Jul 14 00:27:36.180049 [   23.790241] CPU    1: hi:  186, btch:  31 usd: 128

Jul 14 00:27:36.187954 [   23.790243] CPU    2: hi:  186, btch:  31 usd:   0

Jul 14 00:27:36.187996 [   23.790245] CPU    3: hi:  186, btch:  31 usd:   0

Jul 14 00:27:36.195965 [   23.790248] active_anon:0 inactive_anon:0 isolated_anon:0

Jul 14 00:27:36.203964 [   23.790248]  active_file:0 inactive_file:0 isolated_file:0

Jul 14 00:27:36.204010 [   23.790248]  unevictable:1 dirty:4 writeback:0 unstable:0

Jul 14 00:27:36.211969 [   23.790248]  free:1012 slab_reclaimable:1888 slab_unreclaimable:4067

Jul 14 00:27:36.219948 [   23.790248]  mapped:2877 shmem:154 pagetables:521 bounce:0

Jul 14 00:27:36.227943 [   23.790248]  free_cma:0

Jul 14 00:27:36.227982 [   23.790259] DMA free:1680kB min:96kB low:120kB high:144kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:0kB isolated(anon):0kB isolated(file):0kB present:15976kB managed:15892kB mlocked:1044kB dirty:0kB writeback:0kB mapped:340kB shmem:8kB slab_reclaimable:92kB slab_unreclaimable:332kB kernel_stack:48kB pagetables:84kB unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:0 all_unreclaimable? yes

Jul 14 00:27:36.267995 [   23.790269] lowmem_reserve[]: 0 396 396 396

Jul 14 00:27:36.268052 [   23.790275] DMA32 free:2368kB min:2496kB low:3120kB high:3744kB active_anon:0kB inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:4kB isolated(anon):0kB isolated(file):0kB present:508308kB managed:408952kB mlocked:319
Jul 14 00:28:18.168603 <client 0x1c9b400 connected - now 1 clients>



From: osstest service owner <osstest-admin@xenproject.org>
To: <xen-devel@lists.xensource.com>, <osstest-admin@xenproject.org>
Subject: [linux-3.18 bisection] complete test-amd64-amd64-xl-credit2
Date: Sat, 16 Jul 2016 14:52:57 +0000

branch xen-unstable
xenbranch xen-unstable
job test-amd64-amd64-xl-credit2
testid xen-boot

Tree: linux git://git.kernel.org/pub/scm/linux/kernel/git/stable/linux-stable.git
Tree: linuxfirmware git://xenbits.xen.org/osstest/linux-firmware.git
Tree: qemu git://xenbits.xen.org/qemu-xen-traditional.git
Tree: qemuu git://xenbits.xen.org/qemu-xen.git
Tree: xen git://xenbits.xen.org/xen.git

*** Found and reproduced problem changeset ***

  Bug is in tree:  linux git://git.kernel.org/pub/scm/linux/kernel/git/stable/linux-stable.git
  Bug introduced:  a2d8c514753276394d68414f563591f174ef86cb
  Bug not present: 8f620446135b64ca6f96cf32066a76d64e79a388
  Last fail repro: http://logs.test-lab.xenproject.org/osstest/logs/97435/


  commit a2d8c514753276394d68414f563591f174ef86cb
  Author: Lukasz Odzioba <lukasz.odzioba@intel.com>
  Date:   Fri Jun 24 14:50:01 2016 -0700
  
      mm/swap.c: flush lru pvecs on compound page arrival
      
      [ Upstream commit 8f182270dfec432e93fae14f9208a6b9af01009f ]
      
      Currently we can have compound pages held on per cpu pagevecs, which
      leads to a lot of memory unavailable for reclaim when needed.  In the
      systems with hundreads of processors it can be GBs of memory.
      
      On of the way of reproducing the problem is to not call munmap
      explicitly on all mapped regions (i.e.  after receiving SIGTERM).  After
      that some pages (with THP enabled also huge pages) may end up on
      lru_add_pvec, example below.
      
        void main() {
        #pragma omp parallel
        {
      	size_t size = 55 * 1000 * 1000; // smaller than  MEM/CPUS
      	void *p = mmap(NULL, size, PROT_READ | PROT_WRITE,
      		MAP_PRIVATE | MAP_ANONYMOUS , -1, 0);
      	if (p != MAP_FAILED)
      		memset(p, 0, size);
      	//munmap(p, size); // uncomment to make the problem go away
        }
        }
      
      When we run it with THP enabled it will leave significant amount of
      memory on lru_add_pvec.  This memory will be not reclaimed if we hit
      OOM, so when we run above program in a loop:
      
      	for i in `seq 100`; do ./a.out; done
      
      many processes (95% in my case) will be killed by OOM.
      
      The primary point of the LRU add cache is to save the zone lru_lock
      contention with a hope that more pages will belong to the same zone and
      so their addition can be batched.  The huge page is already a form of
      batched addition (it will add 512 worth of memory in one go) so skipping
      the batching seems like a safer option when compared to a potential
      excess in the caching which can be quite large and much harder to fix
      because lru_add_drain_all is way to expensive and it is not really clear
      what would be a good moment to call it.
      
      Similarly we can reproduce the problem on lru_deactivate_pvec by adding:
      madvise(p, size, MADV_FREE); after memset.
      
      This patch flushes lru pvecs on compound page arrival making the problem
      less severe - after applying it kill rate of above example drops to 0%,
      due to reducing maximum amount of memory held on pvec from 28MB (with
      THP) to 56kB per CPU.
      
      Suggested-by: Michal Hocko <mhocko@suse.com>
      Link: http://lkml.kernel.org/r/1466180198-18854-1-git-send-email-lukasz.odzioba@intel.com
      Signed-off-by: Lukasz Odzioba <lukasz.odzioba@intel.com>
      Acked-by: Michal Hocko <mhocko@suse.com>
      Cc: Kirill Shutemov <kirill.shutemov@linux.intel.com>
      Cc: Andrea Arcangeli <aarcange@redhat.com>
      Cc: Vladimir Davydov <vdavydov@parallels.com>
      Cc: Ming Li <mingli199x@qq.com>
      Cc: Minchan Kim <minchan@kernel.org>
      Cc: <stable@vger.kernel.org>
      Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
      Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
      Signed-off-by: Sasha Levin <sasha.levin@oracle.com>


For bisection revision-tuple graph see:
   http://logs.test-lab.xenproject.org/osstest/results/bisect/linux-3.18/test-amd64-amd64-xl-credit2.xen-boot.html
Revision IDs in each graph node refer, respectively, to the Trees above.

----------------------------------------
Running cs-bisection-step --graph-out=/home/logs/results/bisect/linux-3.18/test-amd64-amd64-xl-credit2.xen-boot --summary-out=tmp/97435.bisection-summary --basis-template=96188 --blessings=real,real-bisect linux-3.18 test-amd64-amd64-xl-credit2 xen-boot
Searching for failure / basis pass:
 97377 fail [host=chardonnay0] / 96188 [host=italia0] 96161 [host=baroque1] 95844 [host=godello0] 95809 [host=godello1] 95597 [host=fiano0] 95521 [host=pinot1] 95458 [host=elbling1] 95406 [host=fiano1] 94728 [host=merlot1] 94153 [host=elbling0] 94083 ok.
Failure / basis pass flights: 97377 / 94083
(tree with no url: minios)
(tree with no url: ovmf)
(tree with no url: seabios)
Tree: linux git://git.kernel.org/pub/scm/linux/kernel/git/stable/linux-stable.git
Tree: linuxfirmware git://xenbits.xen.org/osstest/linux-firmware.git
Tree: qemu git://xenbits.xen.org/qemu-xen-traditional.git
Tree: qemuu git://xenbits.xen.org/qemu-xen.git
Tree: xen git://xenbits.xen.org/xen.git
Latest 0ac0a856d986c1ab240753479f5e50fdfab82b14 c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f ea210c52abb6458e39f5365f7f2c3abb9c191c47
Basis pass 6b12ebc0ecce75d7bd3660cd85f8b47a615c2071 c530a75c1e6a472b0eb9558310b518f0dfcd8860 e4ceb77cf88bc44f0b7fe39225c49d660735f327 62b3d206425c245ed0a020390a64640d40d97471 c79fc6c4bee28b40948838a760b4aaadf6b5cd47
Generating revisions with ./adhoc-revtuple-generator  git://git.kernel.org/pub/scm/linux/kernel/git/stable/linux-stable.git#6b12ebc0ecce75d7bd3660cd85f8b47a615c2071-0ac0a856d986c1ab240753479f5e50fdfab82b14 git://xenbits.xen.org/osstest/linux-firmware.git#c530a75c1e6a472b0eb9558310b518f0dfcd8860-c530a75c1e6a472b0eb9558310b518f0dfcd8860 git://xenbits.xen.org/qemu-xen-traditional.git#e4ceb77cf88bc44f0b7fe39225c49d660735f327-6e20809727261599e8527c456eb078c0e89139a1 git://xenbits.xen.org/qemu-xen.git#62b3d206425c245ed0a020390a64640d40d97471-44a072f0de0d57c95c2212bbce02888832b7b74f git://xenbits.xen.org/xen.git#c79fc6c4bee28b40948838a760b4aaadf6b5cd47-ea210c52abb6458e39f5365f7f2c3abb9c191c47
Loaded 12048 nodes in revision graph
Searching for test results:
 94035 [host=huxelrebe0]
 94083 pass 6b12ebc0ecce75d7bd3660cd85f8b47a615c2071 c530a75c1e6a472b0eb9558310b518f0dfcd8860 e4ceb77cf88bc44f0b7fe39225c49d660735f327 62b3d206425c245ed0a020390a64640d40d97471 c79fc6c4bee28b40948838a760b4aaadf6b5cd47
 94056 [host=huxelrebe1]
 94153 [host=elbling0]
 94728 [host=merlot1]
 95406 [host=fiano1]
 95458 [host=elbling1]
 95521 [host=pinot1]
 95597 [host=fiano0]
 95809 [host=godello1]
 95844 [host=godello0]
 96161 [host=baroque1]
 96188 [host=italia0]
 97278 fail irrelevant
 97289 fail 0ac0a856d986c1ab240753479f5e50fdfab82b14 c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f ea210c52abb6458e39f5365f7f2c3abb9c191c47
 97321 pass 6b12ebc0ecce75d7bd3660cd85f8b47a615c2071 c530a75c1e6a472b0eb9558310b518f0dfcd8860 e4ceb77cf88bc44f0b7fe39225c49d660735f327 62b3d206425c245ed0a020390a64640d40d97471 c79fc6c4bee28b40948838a760b4aaadf6b5cd47
 97346 fail 0ac0a856d986c1ab240753479f5e50fdfab82b14 c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f ea210c52abb6458e39f5365f7f2c3abb9c191c47
 97357 pass f27ca140ad82b5e76282cc5b54bfb0a665520d17 c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f 22ea8ad02e465e32cd40887c750b55c3a997a288
 97352 pass b5076139991c6b12c62346d9880eec1d4227d99f c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f 87beb45e0b05be76755cac53322aae4f5b426aac
 97319 fail 0ac0a856d986c1ab240753479f5e50fdfab82b14 c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f ea210c52abb6458e39f5365f7f2c3abb9c191c47
 97389 fail e23042d05035bd64c980ea8f1d9d311972b09104 c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f 22ea8ad02e465e32cd40887c750b55c3a997a288
 97404 pass faa35ed7c7dd74a62bb58340e0ba1819ec33e4e1 c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f 22ea8ad02e465e32cd40887c750b55c3a997a288
 97423 fail a2d8c514753276394d68414f563591f174ef86cb c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f 22ea8ad02e465e32cd40887c750b55c3a997a288
 97393 pass 1b9dc6680de288cb47e0a3c1587ba69879b3c26f c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f 22ea8ad02e465e32cd40887c750b55c3a997a288
 97397 pass 4c2b0216cdf54e81f7c0e841b5bb1116701ae25b c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f 22ea8ad02e465e32cd40887c750b55c3a997a288
 97364 fail 30888a2ea001e237ae9960de877d6f4d2351d8a2 c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f 22ea8ad02e465e32cd40887c750b55c3a997a288
 97376 blocked 71c879eb92223676c4583e130f1b0ce26cddb891 c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f 22ea8ad02e465e32cd40887c750b55c3a997a288
 97413 pass 8f620446135b64ca6f96cf32066a76d64e79a388 c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f 22ea8ad02e465e32cd40887c750b55c3a997a288
 97379 fail 6d94f01566e30c87ebd42e1175ade4f648735578 c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f 22ea8ad02e465e32cd40887c750b55c3a997a288
 97399 fail a2d8c514753276394d68414f563591f174ef86cb c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f 22ea8ad02e465e32cd40887c750b55c3a997a288
 97417 fail a2d8c514753276394d68414f563591f174ef86cb c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f 22ea8ad02e465e32cd40887c750b55c3a997a288
 97421 pass 8f620446135b64ca6f96cf32066a76d64e79a388 c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f 22ea8ad02e465e32cd40887c750b55c3a997a288
 97377 fail 0ac0a856d986c1ab240753479f5e50fdfab82b14 c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f ea210c52abb6458e39f5365f7f2c3abb9c191c47
 97428 pass 8f620446135b64ca6f96cf32066a76d64e79a388 c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f 22ea8ad02e465e32cd40887c750b55c3a997a288
 97435 fail a2d8c514753276394d68414f563591f174ef86cb c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f 22ea8ad02e465e32cd40887c750b55c3a997a288
Searching for interesting versions
 Result found: flight 94083 (pass), for basis pass
 Result found: flight 97289 (fail), for basis failure
 Repro found: flight 97321 (pass), for basis pass
 Repro found: flight 97346 (fail), for basis failure
 0 revisions at 8f620446135b64ca6f96cf32066a76d64e79a388 c530a75c1e6a472b0eb9558310b518f0dfcd8860 6e20809727261599e8527c456eb078c0e89139a1 44a072f0de0d57c95c2212bbce02888832b7b74f 22ea8ad02e465e32cd40887c750b55c3a997a288
No revisions left to test, checking graph state.
 Result found: flight 97413 (pass), for last pass
 Result found: flight 97417 (fail), for first failure
 Repro found: flight 97421 (pass), for last pass
 Repro found: flight 97423 (fail), for first failure
 Repro found: flight 97428 (pass), for last pass
 Repro found: flight 97435 (fail), for first failure

*** Found and reproduced problem changeset ***

  Bug is in tree:  linux git://git.kernel.org/pub/scm/linux/kernel/git/stable/linux-stable.git
  Bug introduced:  a2d8c514753276394d68414f563591f174ef86cb
  Bug not present: 8f620446135b64ca6f96cf32066a76d64e79a388
  Last fail repro: http://logs.test-lab.xenproject.org/osstest/logs/97435/


  commit a2d8c514753276394d68414f563591f174ef86cb
  Author: Lukasz Odzioba <lukasz.odzioba@intel.com>
  Date:   Fri Jun 24 14:50:01 2016 -0700
  
      mm/swap.c: flush lru pvecs on compound page arrival
      
      [ Upstream commit 8f182270dfec432e93fae14f9208a6b9af01009f ]
      
      Currently we can have compound pages held on per cpu pagevecs, which
      leads to a lot of memory unavailable for reclaim when needed.  In the
      systems with hundreads of processors it can be GBs of memory.
      
      On of the way of reproducing the problem is to not call munmap
      explicitly on all mapped regions (i.e.  after receiving SIGTERM).  After
      that some pages (with THP enabled also huge pages) may end up on
      lru_add_pvec, example below.
      
        void main() {
        #pragma omp parallel
        {
      	size_t size = 55 * 1000 * 1000; // smaller than  MEM/CPUS
      	void *p = mmap(NULL, size, PROT_READ | PROT_WRITE,
      		MAP_PRIVATE | MAP_ANONYMOUS , -1, 0);
      	if (p != MAP_FAILED)
      		memset(p, 0, size);
      	//munmap(p, size); // uncomment to make the problem go away
        }
        }
      
      When we run it with THP enabled it will leave significant amount of
      memory on lru_add_pvec.  This memory will be not reclaimed if we hit
      OOM, so when we run above program in a loop:
      
      	for i in `seq 100`; do ./a.out; done
      
      many processes (95% in my case) will be killed by OOM.
      
      The primary point of the LRU add cache is to save the zone lru_lock
      contention with a hope that more pages will belong to the same zone and
      so their addition can be batched.  The huge page is already a form of
      batched addition (it will add 512 worth of memory in one go) so skipping
      the batching seems like a safer option when compared to a potential
      excess in the caching which can be quite large and much harder to fix
      because lru_add_drain_all is way to expensive and it is not really clear
      what would be a good moment to call it.
      
      Similarly we can reproduce the problem on lru_deactivate_pvec by adding:
      madvise(p, size, MADV_FREE); after memset.
      
      This patch flushes lru pvecs on compound page arrival making the problem
      less severe - after applying it kill rate of above example drops to 0%,
      due to reducing maximum amount of memory held on pvec from 28MB (with
      THP) to 56kB per CPU.
      
      Suggested-by: Michal Hocko <mhocko@suse.com>
      Link: http://lkml.kernel.org/r/1466180198-18854-1-git-send-email-lukasz.odzioba@intel.com
      Signed-off-by: Lukasz Odzioba <lukasz.odzioba@intel.com>
      Acked-by: Michal Hocko <mhocko@suse.com>
      Cc: Kirill Shutemov <kirill.shutemov@linux.intel.com>
      Cc: Andrea Arcangeli <aarcange@redhat.com>
      Cc: Vladimir Davydov <vdavydov@parallels.com>
      Cc: Ming Li <mingli199x@qq.com>
      Cc: Minchan Kim <minchan@kernel.org>
      Cc: <stable@vger.kernel.org>
      Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
      Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
      Signed-off-by: Sasha Levin <sasha.levin@oracle.com>

dot: graph is too large for cairo-renderer bitmaps. Scaling by 0.719964 to fit
pnmtopng: 47 colors found
Revision graph left in /home/logs/results/bisect/linux-3.18/test-amd64-amd64-xl-credit2.xen-boot.{dot,ps,png,html,svg}.
----------------------------------------
97435: tolerable ALL FAIL

flight 97435 linux-3.18 real-bisect [real]
http://logs.test-lab.xenproject.org/osstest/logs/97435/

Failures :-/ but no regressions.

Tests which did not succeed,
including tests which could not be run:
 test-amd64-amd64-xl-credit2   6 xen-boot                fail baseline untested


jobs:
 test-amd64-amd64-xl-credit2                                  fail    


------------------------------------------------------------
sg-report-flight on osstest.test-lab.xenproject.org
logs: /home/logs/logs
images: /home/logs/images

Logs, config files, etc. are available at
    http://logs.test-lab.xenproject.org/osstest/logs

Explanation of these reports, and of osstest in general, is at
    http://xenbits.xen.org/gitweb/?p=osstest.git;a=blob;f=README.email;hb=master
    http://xenbits.xen.org/gitweb/?p=osstest.git;a=blob;f=README;hb=master

Test harness code can be found at
    http://xenbits.xen.org/gitweb?p=osstest.git;a=summary

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: Linux 3.18 and Linux 4.1 reproducible OOM crashes under Xen
  2016-07-18 11:17       ` Linux 3.18 and Linux 4.1 reproducible OOM crashes under Xen Ian Jackson
@ 2016-07-18 11:25         ` Michal Hocko
  2016-07-18 11:43           ` Odzioba, Lukasz
  0 siblings, 1 reply; 7+ messages in thread
From: Michal Hocko @ 2016-07-18 11:25 UTC (permalink / raw)
  To: Ian Jackson
  Cc: Lukasz Odzioba, Konrad Rzeszutek Wilk, stable, xen-devel,
	Andrew Morton, Linus Torvalds, Sasha Levin

On Mon 18-07-16 12:17:58, Ian Jackson wrote:
> The Xen Project CI system (`osstest') reports that the current tip of
> the Linux 3.18 and Linux 4.1 stable branches have serious problems
> booting Xen.
> 
> Our automatic bisector has completed its work and fingered (in 3.18)
> a2d8c5147532 "mm/swap.c: flush lru pvecs on compound page arrival"
> (and a backport of what seems to be the same commit to Linux 4.1).

Yes, this patch has been misbackported to the stable tree (at least 4.1
and maybe others which took it over from there). See
http://lkml.kernel.org/r/20160714175521.3675e3d6@gandalf.local.home
-- 
Michal Hocko
SUSE Labs

^ permalink raw reply	[flat|nested] 7+ messages in thread

* RE: Linux 3.18 and Linux 4.1 reproducible OOM crashes under Xen
  2016-07-18 11:25         ` Michal Hocko
@ 2016-07-18 11:43           ` Odzioba, Lukasz
  0 siblings, 0 replies; 7+ messages in thread
From: Odzioba, Lukasz @ 2016-07-18 11:43 UTC (permalink / raw)
  To: Michal Hocko, Ian Jackson
  Cc: Konrad Rzeszutek Wilk, stable, xen-devel, Andrew Morton,
	Linus Torvalds, Sasha Levin

On Mon 18-07-2016 1:26 PM, Michal Hocko wrote:
> Yes, this patch has been misbackported to the stable tree (at least 4.1
> and maybe others which took it over from there). See
> http://lkml.kernel.org/r/20160714175521.3675e3d6@gandalf.local.home

Symptoms looks similar to this one:
http://marc.info/?l=linux-mm&m=146868046305014&w=2

Thanks,
Lukas

^ permalink raw reply	[flat|nested] 7+ messages in thread

end of thread, other threads:[~2016-07-18 11:43 UTC | newest]

Thread overview: 7+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2016-07-14  5:03 [linux-3.18 test] 97278: regressions - FAIL osstest service owner
2016-07-14 12:56 ` Linux 3.18 and Linux 4.1 reproducible OOM crashes Ian Jackson
2016-07-14 23:34   ` Konrad Rzeszutek Wilk
2016-07-15 10:00     ` Ian Jackson
2016-07-18 11:17       ` Linux 3.18 and Linux 4.1 reproducible OOM crashes under Xen Ian Jackson
2016-07-18 11:25         ` Michal Hocko
2016-07-18 11:43           ` Odzioba, Lukasz

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).