From: Shahar Salzman <shahar.salzman at kaminario.com>
To: spdk@lists.01.org
Subject: Re: [SPDK] Strange CI failure
Date: Wed, 30 Jan 2019 12:18:11 +0000 [thread overview]
Message-ID: <AM6PR04MB5127CD7BC86799776E53844C89900@AM6PR04MB5127.eurprd04.prod.outlook.com> (raw)
In-Reply-To: 6487C194-EC48-4ABD-93E2-A1D3B2D2C346@intel.com
[-- Attachment #1: Type: text/plain, Size: 4499 bytes --]
It looks like there are now consistent failures in iscsi and spdk-nvme-cli tests, I tried to retrigger and the failures happened again:
spdk nvme cli:
https://ci.spdk.io/spdk/builds/review/253dd179d38ac2b608f5adf1edad56e1ec6eb519.1548848568/fedora-03/build.log
iscsi:
https://ci.spdk.io/spdk/builds/review/253dd179d38ac2b608f5adf1edad56e1ec6eb519.1548848568/fedora-09/build.log
________________________________
From: Harris, James R <james.r.harris(a)intel.com>
Sent: Tuesday, January 29, 2019 6:09 PM
To: Storage Performance Development Kit; Shahar Salzman
Subject: Re: [SPDK] Strange CI failure
Thanks Shahar. For now, you can reply to your own patch on GerritHub with just the word "retrigger" - it will re-run your patch through the test pool. That will get your patch unblocked while Paul looks at the intermittent test failure.
-Jim
On 1/29/19, 8:48 AM, "SPDK on behalf of Luse, Paul E" <spdk-bounces(a)lists.01.org on behalf of paul.e.luse(a)intel.com> wrote:
Thanks! I've got a few hours of meetings coming up but here's what I see. If you can repro that'd be great, we can get a github issue up and going. If not I can look deeper into this later if someone else doesn't jump in by then with an "aha" moment :)
Starting SPDK v19.01-pre / DPDK 18.11.0 initialization...
[ DPDK EAL parameters: identify -c 0x1 -n 1 -m 0 --base-virtaddr=0x200000000000 --file-prefix=spdk0 --proc-type=auto ]
EAL: Detected 16 lcore(s)
EAL: Detected 2 NUMA nodes
EAL: Auto-detected process type: SECONDARY
EAL: Multi-process socket /var/run/dpdk/spdk0/mp_socket_835807_c029d817e596b
EAL: Probing VFIO support...
EAL: VFIO support initialized
test/nvme/nvme.sh: line 108: 835807 Segmentation fault (core dumped) $rootdir/examples/nvme/identify/identify -i 0
08:50:18 # trap - ERR
08:50:18 # print_backtrace
08:50:18 # [[ ehxBE =~ e ]]
08:50:18 # local shell_options=ehxBE
08:50:18 # set +x
========== Backtrace start: ==========
From: Shahar Salzman [mailto:shahar.salzman(a)kaminario.com]
Sent: Tuesday, January 29, 2019 8:35 AM
To: Luse, Paul E <paul.e.luse(a)intel.com>; Storage Performance Development Kit <spdk(a)lists.01.org>
Subject: Re: Strange CI failure
https://ci.spdk.io/spdk-jenkins/results/autotest-per-patch/builds/21382/archive/nvme_phy_autotest/build.log
I can copy paste it if you cannot reach the link.
________________________________
From: SPDK <spdk-bounces(a)lists.01.org<mailto:spdk-bounces(a)lists.01.org>> on behalf of Luse, Paul E <paul.e.luse(a)intel.com<mailto:paul.e.luse(a)intel.com>>
Sent: Tuesday, January 29, 2019 5:22 PM
To: Storage Performance Development Kit
Subject: Re: [SPDK] Strange CI failure
Can you send a link to the full log?
-----Original Message-----
From: SPDK [mailto:spdk-bounces(a)lists.01.org] On Behalf Of Shahar Salzman
Sent: Tuesday, January 29, 2019 8:21 AM
To: Storage Performance Development Kit <spdk(a)lists.01.org<mailto:spdk(a)lists.01.org>>
Subject: [SPDK] Strange CI failure
Hi,
I have encountered a CI failure that has nothing to do with my code.
The reason that I know it has nothing to do with it, is that the change is a gdb macro.
Do we know that this test machine is unstable?
Here is the backtrace:
========== Backtrace start: ==========
in test/nvme/nvme.sh:108 -> main()
...
103 report_test_completion "nightly_nvme_reset"
104 timing_exit reset
105 fi
106
107 timing_enter identify
=> 108 $rootdir/examples/nvme/identify/identify -i 0
109 for bdf in $(iter_pci_class_code 01 08 02); do
110 $rootdir/examples/nvme/identify/identify -r "trtype:PCIe traddr:${bdf}" -i 0
111 done
112 timing_exit identify
113
...
Shahar
_______________________________________________
SPDK mailing list
SPDK(a)lists.01.org<mailto:SPDK(a)lists.01.org>
https://lists.01.org/mailman/listinfo/spdk
_______________________________________________
SPDK mailing list
SPDK(a)lists.01.org<mailto:SPDK(a)lists.01.org>
https://lists.01.org/mailman/listinfo/spdk
_______________________________________________
SPDK mailing list
SPDK(a)lists.01.org
https://lists.01.org/mailman/listinfo/spdk
next reply other threads:[~2019-01-30 12:18 UTC|newest]
Thread overview: 10+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-01-30 12:18 Shahar Salzman [this message]
-- strict thread matches above, loose matches on Subject: below --
2019-01-31 8:47 [SPDK] Strange CI failure Shahar Salzman
2019-01-30 14:28 Howell, Seth
2019-01-29 18:29 Luse, Paul E
2019-01-29 16:09 Harris, James R
2019-01-29 16:08 Shahar Salzman
2019-01-29 15:47 Luse, Paul E
2019-01-29 15:35 Shahar Salzman
2019-01-29 15:22 Luse, Paul E
2019-01-29 15:20 Shahar Salzman
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=AM6PR04MB5127CD7BC86799776E53844C89900@AM6PR04MB5127.eurprd04.prod.outlook.com \
--to=spdk@lists.01.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.