* [xen-unstable test] 1959: FAIL
@ 2010-07-24 8:48 ` xen.org
[not found] ` <4C4B1B8A.6070403@goop.org>
0 siblings, 1 reply; 12+ messages in thread
From: xen.org @ 2010-07-24 8:48 UTC (permalink / raw)
To: xen-devel; +Cc: ian.jackson
flight 1959 xen-unstable real
tests which did not succeed:
test-amd64-amd64-pair 7 xen-boot/dst_host fail
test-amd64-amd64-pair 6 xen-boot/src_host fail
test-amd64-amd64-pv 4 xen-boot fail
test-amd64-amd64-win 4 xen-boot fail
test-amd64-amd64-xl 4 xen-boot fail
test-amd64-i386-pair 12 guest-migrate/src_host/dst_host fail
test-amd64-i386-pv 8 guest-saverestore fail
test-amd64-i386-win 5 windows-install fail
test-amd64-i386-xl 8 guest-saverestore fail
test-amd64-xcpkern-i386-pair 14 remus-start fail
test-amd64-xcpkern-i386-xl 10 guest-stop fail
test-i386-i386-pair 12 guest-migrate/src_host/dst_host fail
test-i386-i386-pv 8 guest-saverestore fail
test-i386-i386-win 5 windows-install fail
test-i386-i386-xl 8 guest-saverestore fail
test-i386-xcpkern-i386-pair 12 guest-migrate/src_host/dst_host fail
test-i386-xcpkern-i386-xl 10 guest-stop fail
version targeted for testing:
xen b0b1a6163203
jobs:
build-amd64 pass
build-amd64-oldkern pass
build-i386 pass
build-i386-oldkern pass
build-i386-xcpkern pass
test-amd64-amd64-pair fail
test-amd64-amd64-pv fail
test-amd64-amd64-win fail
test-amd64-amd64-xl fail
test-amd64-i386-pair fail
test-amd64-i386-pv fail
test-amd64-i386-win fail
test-amd64-i386-xl fail
test-amd64-xcpkern-i386-pair fail
test-amd64-xcpkern-i386-pv pass
test-amd64-xcpkern-i386-win pass
test-amd64-xcpkern-i386-xl fail
test-i386-i386-pair fail
test-i386-i386-pv fail
test-i386-i386-win fail
test-i386-i386-xl fail
test-i386-xcpkern-i386-pair fail
test-i386-xcpkern-i386-pv pass
test-i386-xcpkern-i386-win pass
test-i386-xcpkern-i386-xl fail
-------------------------------------------------------------------------------
build-amd64:
1 host-install(1) pass
2 host-build-prep pass
3 xen-build pass
linux c0a00fbeb3b16c473c0f
qemu 833e7e9a4c95739429dc
xen 21840:b0b1a6163203
-------------------------------------------------------------------------------
build-amd64-oldkern:
1 xen-build pass
linux 1023:85ca9742b8b9
qemu 833e7e9a4c95739429dc
xen 21840:b0b1a6163203
-------------------------------------------------------------------------------
build-i386:
1 host-install(1) pass
2 host-build-prep pass
3 xen-build pass
linux c0a00fbeb3b16c473c0f
qemu 833e7e9a4c95739429dc
xen 21840:b0b1a6163203
-------------------------------------------------------------------------------
build-i386-oldkern:
1 xen-build pass
linux 1023:85ca9742b8b9
qemu 833e7e9a4c95739429dc
xen 21840:b0b1a6163203
-------------------------------------------------------------------------------
build-i386-xcpkern:
1 kernel-build pass
linux 811:d1c43bf74f7c
pq_linux 773:e16c9e905eb1
-------------------------------------------------------------------------------
test-amd64-amd64-pair:
1 xen-build-check(1) pass
2 host-install/src_host(2) pass
3 host-install/dst_host(3) pass
4 xen-install/src_host pass
5 xen-install/dst_host pass
6 xen-boot/src_host fail
7 xen-boot/dst_host fail
8 capture-logs/src_host(8) pass
9 capture-logs/dst_host(9) pass
-------------------------------------------------------------------------------
test-amd64-amd64-pv:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot fail
5 capture-logs(5) pass
-------------------------------------------------------------------------------
test-amd64-amd64-win:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot fail
5 capture-logs(5) pass
-------------------------------------------------------------------------------
test-amd64-amd64-xl:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot fail
5 capture-logs(5) pass
-------------------------------------------------------------------------------
test-amd64-i386-pair:
1 xen-build-check(1) pass
2 host-install/src_host(2) pass
3 host-install/dst_host(3) pass
4 xen-install/src_host pass
5 xen-install/dst_host pass
6 xen-boot/src_host pass
7 xen-boot/dst_host pass
8 debian-install/dst_host pass
9 debian-fixup/dst_host pass
10 guests-nbd-mirror pass
11 guest-start pass
12 guest-migrate/src_host/dst_host fail
13 capture-logs/src_host(13) pass
14 capture-logs/dst_host(14) pass
-------------------------------------------------------------------------------
test-amd64-i386-pv:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore fail
9 capture-logs(9) pass
-------------------------------------------------------------------------------
test-amd64-i386-win:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 windows-install fail
6 capture-logs(6) pass
-------------------------------------------------------------------------------
test-amd64-i386-xl:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore fail
9 capture-logs(9) pass
-------------------------------------------------------------------------------
test-amd64-xcpkern-i386-pair:
1 xen-build-check(1) pass
2 host-install/src_host(2) pass
3 host-install/dst_host(3) pass
4 xen-install/src_host pass
5 xen-install/dst_host pass
6 xen-boot/src_host pass
7 xen-boot/dst_host pass
8 debian-install/dst_host pass
9 debian-fixup/dst_host pass
10 guests-nbd-mirror pass
11 guest-start pass
12 guest-migrate/src_host/dst_host pass
13 guest-migrate/dst_host/src_host pass
14 remus-start fail
15 capture-logs/src_host(15) pass
16 capture-logs/dst_host(16) pass
-------------------------------------------------------------------------------
test-amd64-xcpkern-i386-pv:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore pass
9 guest-localmigrate pass
10 guest-stop pass
11 capture-logs(11) pass
-------------------------------------------------------------------------------
test-amd64-xcpkern-i386-win:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 windows-install pass
6 guest-saverestore pass
7 guest-localmigrate pass
8 guest-stop pass
9 capture-logs(9) pass
-------------------------------------------------------------------------------
test-amd64-xcpkern-i386-xl:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore pass
9 guest-localmigrate pass
10 guest-stop fail
11 capture-logs(11) pass
-------------------------------------------------------------------------------
test-i386-i386-pair:
1 xen-build-check(1) pass
2 host-install/src_host(2) pass
3 host-install/dst_host(3) pass
4 xen-install/src_host pass
5 xen-install/dst_host pass
6 xen-boot/src_host pass
7 xen-boot/dst_host pass
8 debian-install/dst_host pass
9 debian-fixup/dst_host pass
10 guests-nbd-mirror pass
11 guest-start pass
12 guest-migrate/src_host/dst_host fail
13 capture-logs/src_host(13) pass
14 capture-logs/dst_host(14) pass
-------------------------------------------------------------------------------
test-i386-i386-pv:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore fail
9 capture-logs(9) pass
-------------------------------------------------------------------------------
test-i386-i386-win:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 windows-install fail
6 capture-logs(6) pass
-------------------------------------------------------------------------------
test-i386-i386-xl:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore fail
9 capture-logs(9) pass
-------------------------------------------------------------------------------
test-i386-xcpkern-i386-pair:
1 xen-build-check(1) pass
2 host-install/src_host(2) pass
3 host-install/dst_host(3) pass
4 xen-install/src_host pass
5 xen-install/dst_host pass
6 xen-boot/src_host pass
7 xen-boot/dst_host pass
8 debian-install/dst_host pass
9 debian-fixup/dst_host pass
10 guests-nbd-mirror pass
11 guest-start pass
12 guest-migrate/src_host/dst_host fail
13 capture-logs/src_host(13) pass
14 capture-logs/dst_host(14) pass
-------------------------------------------------------------------------------
test-i386-xcpkern-i386-pv:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore pass
9 guest-localmigrate pass
10 guest-stop pass
11 capture-logs(11) pass
-------------------------------------------------------------------------------
test-i386-xcpkern-i386-win:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 windows-install pass
6 guest-saverestore pass
7 guest-localmigrate pass
8 guest-stop pass
9 capture-logs(9) pass
-------------------------------------------------------------------------------
test-i386-xcpkern-i386-xl:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore pass
9 guest-localmigrate pass
10 guest-stop fail
11 capture-logs(11) pass
------------------------------------------------------------
sg-report-flight on woking.cam.xci-test.com
logs: /home/xc_osstest/logs
images: /home/xc_osstest/images
Logs, config files, etc. are available at
http://www.chiark.greenend.org.uk/~xensrcts/logs
Published non-staging tree is already up to date.
^ permalink raw reply [flat|nested] 12+ messages in thread
* [xen-unstable test] 1970: tolerable FAIL - PUSHED
@ 2010-07-29 8:50 xen.org
2010-07-24 8:48 ` [xen-unstable test] 1959: FAIL xen.org
0 siblings, 1 reply; 12+ messages in thread
From: xen.org @ 2010-07-29 8:50 UTC (permalink / raw)
To: xen-devel; +Cc: ian.jackson
flight 1970 xen-unstable real
Failures :-/ but no regressions.
tests which did not succeed:
test-amd64-amd64-pair 12 guest-migrate/src_host/dst_host fail never pass
test-amd64-amd64-pv 8 guest-saverestore fail blocked in 1967
test-amd64-amd64-xl 8 guest-saverestore fail blocked in 1967
test-amd64-i386-pair 12 guest-migrate/src_host/dst_host fail like 1967
test-amd64-i386-pv 8 guest-saverestore fail like 1967
test-amd64-i386-win 5 windows-install fail like 1967
test-amd64-i386-xl 8 guest-saverestore fail like 1967
test-amd64-xcpkern-i386-pair 14 remus-start fail like 1967
test-amd64-xcpkern-i386-xl 10 guest-stop fail never pass
test-i386-i386-pair 12 guest-migrate/src_host/dst_host fail never pass
test-i386-i386-pv 8 guest-saverestore fail like 1967
test-i386-i386-win 5 windows-install fail like 1967
test-i386-i386-xl 8 guest-saverestore fail like 1967
test-i386-xcpkern-i386-pair 12 guest-migrate/src_host/dst_host fail like 1967
test-i386-xcpkern-i386-xl 10 guest-stop fail never pass
version targeted for testing:
xen 2007fd03f53d
baseline version:
xen ebede381efe8
jobs:
build-amd64 pass
build-amd64-oldkern pass
build-i386 pass
build-i386-oldkern pass
build-i386-xcpkern pass
test-amd64-amd64-pair fail
test-amd64-amd64-pv fail
test-amd64-amd64-win pass
test-amd64-amd64-xl fail
test-amd64-i386-pair fail
test-amd64-i386-pv fail
test-amd64-i386-win fail
test-amd64-i386-xl fail
test-amd64-xcpkern-i386-pair fail
test-amd64-xcpkern-i386-pv pass
test-amd64-xcpkern-i386-win pass
test-amd64-xcpkern-i386-xl fail
test-i386-i386-pair fail
test-i386-i386-pv fail
test-i386-i386-win fail
test-i386-i386-xl fail
test-i386-xcpkern-i386-pair fail
test-i386-xcpkern-i386-pv pass
test-i386-xcpkern-i386-win pass
test-i386-xcpkern-i386-xl fail
-------------------------------------------------------------------------------
build-amd64:
1 host-install(1) pass
2 host-build-prep pass
3 xen-build pass
linux 78b55f90e72348e23109
qemu 833e7e9a4c95739429dc
xen 21885:2007fd03f53d
-------------------------------------------------------------------------------
build-amd64-oldkern:
1 xen-build pass
linux 1024:2893229a53a0
qemu 833e7e9a4c95739429dc
xen 21885:2007fd03f53d
-------------------------------------------------------------------------------
build-i386:
1 host-install(1) pass
2 host-build-prep pass
3 xen-build pass
linux 78b55f90e72348e23109
qemu 833e7e9a4c95739429dc
xen 21885:2007fd03f53d
-------------------------------------------------------------------------------
build-i386-oldkern:
1 xen-build pass
linux 1024:2893229a53a0
qemu 833e7e9a4c95739429dc
xen 21885:2007fd03f53d
-------------------------------------------------------------------------------
build-i386-xcpkern:
1 kernel-build pass
linux 811:d1c43bf74f7c
pq_linux 773:e16c9e905eb1
-------------------------------------------------------------------------------
test-amd64-amd64-pair:
1 xen-build-check(1) pass
2 host-install/src_host(2) pass
3 host-install/dst_host(3) pass
4 xen-install/src_host pass
5 xen-install/dst_host pass
6 xen-boot/src_host pass
7 xen-boot/dst_host pass
8 debian-install/dst_host pass
9 debian-fixup/dst_host pass
10 guests-nbd-mirror pass
11 guest-start pass
12 guest-migrate/src_host/dst_host fail
13 capture-logs/src_host(13) pass
14 capture-logs/dst_host(14) pass
-------------------------------------------------------------------------------
test-amd64-amd64-pv:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore fail
9 capture-logs(9) pass
-------------------------------------------------------------------------------
test-amd64-amd64-win:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 windows-install pass
6 guest-saverestore pass
7 guest-localmigrate pass
8 guest-stop pass
9 capture-logs(9) pass
-------------------------------------------------------------------------------
test-amd64-amd64-xl:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore fail
9 capture-logs(9) pass
-------------------------------------------------------------------------------
test-amd64-i386-pair:
1 xen-build-check(1) pass
2 host-install/src_host(2) pass
3 host-install/dst_host(3) pass
4 xen-install/src_host pass
5 xen-install/dst_host pass
6 xen-boot/src_host pass
7 xen-boot/dst_host pass
8 debian-install/dst_host pass
9 debian-fixup/dst_host pass
10 guests-nbd-mirror pass
11 guest-start pass
12 guest-migrate/src_host/dst_host fail
13 capture-logs/src_host(13) pass
14 capture-logs/dst_host(14) pass
-------------------------------------------------------------------------------
test-amd64-i386-pv:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore fail
9 capture-logs(9) pass
-------------------------------------------------------------------------------
test-amd64-i386-win:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 windows-install fail
6 capture-logs(6) pass
-------------------------------------------------------------------------------
test-amd64-i386-xl:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore fail
9 capture-logs(9) pass
-------------------------------------------------------------------------------
test-amd64-xcpkern-i386-pair:
1 xen-build-check(1) pass
2 host-install/src_host(2) pass
3 host-install/dst_host(3) pass
4 xen-install/src_host pass
5 xen-install/dst_host pass
6 xen-boot/src_host pass
7 xen-boot/dst_host pass
8 debian-install/dst_host pass
9 debian-fixup/dst_host pass
10 guests-nbd-mirror pass
11 guest-start pass
12 guest-migrate/src_host/dst_host pass
13 guest-migrate/dst_host/src_host pass
14 remus-start fail
15 capture-logs/src_host(15) pass
16 capture-logs/dst_host(16) pass
-------------------------------------------------------------------------------
test-amd64-xcpkern-i386-pv:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore pass
9 guest-localmigrate pass
10 guest-stop pass
11 capture-logs(11) pass
-------------------------------------------------------------------------------
test-amd64-xcpkern-i386-win:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 windows-install pass
6 guest-saverestore pass
7 guest-localmigrate pass
8 guest-stop pass
9 capture-logs(9) pass
-------------------------------------------------------------------------------
test-amd64-xcpkern-i386-xl:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore pass
9 guest-localmigrate pass
10 guest-stop fail
11 capture-logs(11) pass
-------------------------------------------------------------------------------
test-i386-i386-pair:
1 xen-build-check(1) pass
2 host-install/src_host(2) pass
3 host-install/dst_host(3) pass
4 xen-install/src_host pass
5 xen-install/dst_host pass
6 xen-boot/src_host pass
7 xen-boot/dst_host pass
8 debian-install/dst_host pass
9 debian-fixup/dst_host pass
10 guests-nbd-mirror pass
11 guest-start pass
12 guest-migrate/src_host/dst_host fail
13 capture-logs/src_host(13) pass
14 capture-logs/dst_host(14) pass
-------------------------------------------------------------------------------
test-i386-i386-pv:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore fail
9 capture-logs(9) pass
-------------------------------------------------------------------------------
test-i386-i386-win:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 windows-install fail
6 capture-logs(6) pass
-------------------------------------------------------------------------------
test-i386-i386-xl:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore fail
9 capture-logs(9) pass
-------------------------------------------------------------------------------
test-i386-xcpkern-i386-pair:
1 xen-build-check(1) pass
2 host-install/src_host(2) pass
3 host-install/dst_host(3) pass
4 xen-install/src_host pass
5 xen-install/dst_host pass
6 xen-boot/src_host pass
7 xen-boot/dst_host pass
8 debian-install/dst_host pass
9 debian-fixup/dst_host pass
10 guests-nbd-mirror pass
11 guest-start pass
12 guest-migrate/src_host/dst_host fail
13 capture-logs/src_host(13) pass
14 capture-logs/dst_host(14) pass
-------------------------------------------------------------------------------
test-i386-xcpkern-i386-pv:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore pass
9 guest-localmigrate pass
10 guest-stop pass
11 capture-logs(11) pass
-------------------------------------------------------------------------------
test-i386-xcpkern-i386-win:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 windows-install pass
6 guest-saverestore pass
7 guest-localmigrate pass
8 guest-stop pass
9 capture-logs(9) pass
-------------------------------------------------------------------------------
test-i386-xcpkern-i386-xl:
1 xen-build-check(1) pass
2 host-install(2) pass
3 xen-install pass
4 xen-boot pass
5 debian-install pass
6 debian-fixup pass
7 guest-start pass
8 guest-saverestore pass
9 guest-localmigrate pass
10 guest-stop fail
11 capture-logs(11) pass
------------------------------------------------------------
sg-report-flight on woking.cam.xci-test.com
logs: /home/xc_osstest/logs
images: /home/xc_osstest/images
Logs, config files, etc. are available at
http://www.chiark.greenend.org.uk/~xensrcts/logs
Test harness code can be found at
http://xenbits.xensource.com/gitweb?p=osstest.git;a=summary
Pushing revision :
+ branch=xen-unstable
+ revision=2007fd03f53d
+ cd repos/xen-unstable.hg
+ hg push -r 2007fd03f53d ssh://xen@xenbits.xensource.com/HG/xen-unstable.hg
pushing to ssh://xen@xenbits.xensource.com/HG/xen-unstable.hg
searching for changes
remote: adding changesets
remote: adding manifests
remote: adding file changes
remote: added 5 changesets with 24 changes to 24 files
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: [xen-unstable test] 1959: FAIL [and 1 more messages]
[not found] ` <19537.44367.380875.758088@mariner.uk.xensource.com>
@ 2010-07-29 16:38 ` Jeremy Fitzhardinge
2010-07-29 17:59 ` Ian Jackson
2010-07-30 4:58 ` Xu, Dongxiao
0 siblings, 2 replies; 12+ messages in thread
From: Jeremy Fitzhardinge @ 2010-07-29 16:38 UTC (permalink / raw)
To: Ian Jackson; +Cc: Ian Campbell, Xu, Dongxiao, Xen-devel
On 07/29/2010 09:33 AM, Ian Jackson wrote:
> Jeremy Fitzhardinge writes ("Re: [Xen-devel] [xen-unstable test] 1959: FAIL [and 1 more messages]"):
>> On 07/29/2010 08:30 AM, Ian Jackson wrote:
>>> Is save/restore supposed to work in pvops ? (Using your kernel for
>>> both dom0 and domU.) That would seem to be the next thing to pick off
>>> the list ...
>> Yes. IanC has been tracking down a bug where it fails after a few
>> thousand iterations (now fixed?), but aside from that its apparently OK.
> Well, I was able to reproduce the failure that the automatic test was
> getting. There's a problem with it not getting a copy of the console
> output for some reason but I was able to get this oops from the guest.
>
> Just after the restore I was able to connect to the PV console and it
> echoed a few of my CRs (before saving I had logged in on the console),
> and then it produced the oops. Now it's apparently completely wedged.
>
Ah, it looks like the netfront smartpoll stuff isn't coping with
save/restore.
J
> Ian.
>
>
>
> [ 63.681260] BUG: unable to handle kernel NULL pointer dereference at 00000010
> [ 63.681293] IP: [<c1300381>] smart_poll_function+0xbb/0xf2
> [ 63.681320] *pdpt = 000000001fee5027 *pde = 0000000000000000
> [ 63.681344] Oops: 0000 [#1] SMP
> [ 63.681362] last sysfs file: /sys/kernel/uevent_seqnum
> [ 63.681376] Modules linked in: [last unloaded: scsi_wait_scan]
> [ 63.681398]
> [ 63.681410] Pid: 5, comm: events/0 Not tainted (2.6.32.16 #1)
> [ 63.681424] EIP: 0061:[<c1300381>] EFLAGS: 00010002 CPU: 0
> [ 63.681438] EIP is at smart_poll_function+0xbb/0xf2
> [ 63.681451] EAX: 00000000 EBX: dfea8320 ECX: 00000001 EDX: 00000062
> [ 63.681465] ESI: 00000064 EDI: 00000000 EBP: df849cfc ESP: df849cdc
> [ 63.681479] DS: 007b ES: 007b FS: 00d8 GS: 0000 SS: 0069
> [ 63.681493] Process events/0 (pid: 5, ti=df848000 task=df839480 task.ti=df848000)
> [ 63.681508] Stack:
> [ 63.681516] dfeac828 00000002 dfeac828 dfea8368 dfea0068 0000002c dfeac828 c13002c6
> [ 63.681555]<0> df849d1c c1079ab8 df849d48 c2386358 c2386328 0000002c 00000000 c2386328
> [ 63.681598]<0> df849d5c c1079ce8 000028cd 88bbcad9 d3b2cae7 0000000e d3b2cae7 0000000e
> [ 63.681646] Call Trace:
> [ 63.681662] [<c13002c6>] ? smart_poll_function+0x0/0xf2
> [ 63.681683] [<c1079ab8>] ? __run_hrtimer+0xa9/0xf6
> [ 63.681701] [<c1079ce8>] ? hrtimer_interrupt+0xcd/0x1c8
> [ 63.681719] [<c102cef0>] ? xen_timer_interrupt+0x2b/0x224
> [ 63.681737] [<c102cb3f>] ? xen_force_evtchn_callback+0xf/0x14
> [ 63.681755] [<c102d2ac>] ? check_events+0x8/0xc
> [ 63.681776] [<c102d2a3>] ? xen_restore_fl_direct_end+0x0/0x1
> [ 63.681795] [<c14d1471>] ? _spin_unlock_irqrestore+0x2f/0x31
> [ 63.681814] [<c105eb9e>] ? try_to_wake_up+0x2fa/0x304
> [ 63.681832] [<c102cb3f>] ? xen_force_evtchn_callback+0xf/0x14
> [ 63.681850] [<c10a0cc4>] ? handle_IRQ_event+0x5f/0x122
> [ 63.681867] [<c10a233f>] ? handle_level_irq+0x58/0xa9
> [ 63.681886] [<c121983d>] ? __xen_evtchn_do_upcall+0xab/0x131
> [ 63.681904] [<c1219c71>] ? xen_evtchn_do_upcall+0x20/0x30
> [ 63.682179] [<c102ffe7>] ? xen_do_upcall+0x7/0xc
> [ 63.682179] [<c102007b>] ? apic_reg_write+0xa5/0x52f
> [ 63.682179] [<c1002227>] ? hypercall_page+0x227/0x1005
> [ 63.682179] [<c102cb3f>] ? xen_force_evtchn_callback+0xf/0x14
> [ 63.682179] [<c102d2ac>] ? check_events+0x8/0xc
> [ 63.682179] [<c102d26b>] ? xen_irq_enable_direct_end+0x0/0x1
> [ 63.682179] [<c105cd0d>] ? finish_task_switch+0x4f/0xa6
> [ 63.682179] [<c14d0186>] ? schedule+0x7dd/0x861
> [ 63.682179] [<c106dd9d>] ? __mod_timer+0x135/0x140
> [ 63.682179] [<c102d2ac>] ? check_events+0x8/0xc
> [ 63.682179] [<c102d2a3>] ? xen_restore_fl_direct_end+0x0/0x1
> [ 63.682179] [<c14d1471>] ? _spin_unlock_irqrestore+0x2f/0x31
> [ 63.682179] [<c107750b>] ? prepare_to_wait+0x43/0x48
> [ 63.682179] [<c10742a2>] ? worker_thread+0x94/0x1d2
> [ 63.682179] [<c10ccd2d>] ? vmstat_update+0x0/0x2f
> [ 63.682179] [<c1077357>] ? autoremove_wake_function+0x0/0x33
> [ 63.682179] [<c107420e>] ? worker_thread+0x0/0x1d2
> [ 63.682179] [<c1077120>] ? kthread+0x61/0x66
> [ 63.682179] [<c10770bf>] ? kthread+0x0/0x66
> [ 63.682179] [<c102ff97>] ? kernel_thread_helper+0x7/0x10
> [ 63.682179] Code: c6 89 d0 31 d2 f7 f6 85 d2 75 1a 85 c9 75 0c 8b 83 90 08 00 00 c6 40 10 00 eb 32 c7 83 54 45 00 00 00 00 00 00 8b 83 90 08 00 00<80> 78 10 00 74 1c 8b 4d e8 b8 00 ca 9a 3b 31 d2 f7 71 44 31 c9
> [ 63.682179] EIP: [<c1300381>] smart_poll_function+0xbb/0xf2 SS:ESP 0069:df849cdc
> [ 63.682179] CR2: 0000000000000010
> [ 63.682179] ---[ end trace 760037e75e5675c8 ]---
> [ 63.682179] Kernel panic - not syncing: Fatal exception in interrupt
> [ 63.682179] Pid: 5, comm: events/0 Tainted: G D 2.6.32.16 #1
> [ 63.682179] Call Trace:
> [ 63.682179] [<c14cf7b1>] ? printk+0xf/0x11
> [ 63.682179] [<c14cf6ee>] panic+0x39/0xed
> [ 63.682179] [<c14d2085>] oops_end+0xa1/0xb0
> [ 63.682179] [<c104b426>] no_context+0x137/0x141
> [ 63.682179] [<c104b56f>] __bad_area_nosemaphore+0x13f/0x147
> [ 63.682179] [<c104b584>] bad_area_nosemaphore+0xd/0x10
> [ 63.682179] [<c14d3275>] do_page_fault+0x1c6/0x32b
> [ 63.682179] [<c14d30af>] ? do_page_fault+0x0/0x32b
> [ 63.682179] [<c14d16c6>] error_code+0x66/0x6c
> [ 63.682179] [<c14d30af>] ? do_page_fault+0x0/0x32b
> [ 63.682179] [<c1300381>] ? smart_poll_function+0xbb/0xf2
> [ 63.682179] [<c13002c6>] ? smart_poll_function+0x0/0xf2
> [ 63.682179] [<c1079ab8>] __run_hrtimer+0xa9/0xf6
> [ 63.682179] [<c1079ce8>] hrtimer_interrupt+0xcd/0x1c8
> [ 63.682179] [<c102cef0>] xen_timer_interrupt+0x2b/0x224
> [ 63.682179] [<c102cb3f>] ? xen_force_evtchn_callback+0xf/0x14
> [ 63.682179] [<c102d2ac>] ? check_events+0x8/0xc
> [ 63.682179] [<c102d2a3>] ? xen_restore_fl_direct_end+0x0/0x1
> [ 63.682179] [<c14d1471>] ? _spin_unlock_irqrestore+0x2f/0x31
> [ 63.682179] [<c105eb9e>] ? try_to_wake_up+0x2fa/0x304
> [ 63.682179] [<c102cb3f>] ? xen_force_evtchn_callback+0xf/0x14
> [ 63.682179] [<c10a0cc4>] handle_IRQ_event+0x5f/0x122
> [ 63.682179] [<c10a233f>] handle_level_irq+0x58/0xa9
> [ 63.682179] [<c121983d>] __xen_evtchn_do_upcall+0xab/0x131
> [ 63.682179] [<c1219c71>] xen_evtchn_do_upcall+0x20/0x30
> [ 63.682179] [<c102ffe7>] xen_do_upcall+0x7/0xc
> [ 63.682179] [<c102007b>] ? apic_reg_write+0xa5/0x52f
> [ 63.682179] [<c1002227>] ? hypercall_page+0x227/0x1005
> [ 63.682179] [<c102cb3f>] ? xen_force_evtchn_callback+0xf/0x14
> [ 63.682179] [<c102d2ac>] check_events+0x8/0xc
> [ 63.682179] [<c102d26b>] ? xen_irq_enable_direct_end+0x0/0x1
> [ 63.682179] [<c105cd0d>] ? finish_task_switch+0x4f/0xa6
> [ 63.682179] [<c14d0186>] schedule+0x7dd/0x861
> [ 63.682179] [<c106dd9d>] ? __mod_timer+0x135/0x140
> [ 63.682179] [<c102d2ac>] ? check_events+0x8/0xc
> [ 63.682179] [<c102d2a3>] ? xen_restore_fl_direct_end+0x0/0x1
> [ 63.682179] [<c14d1471>] ? _spin_unlock_irqrestore+0x2f/0x31
> [ 63.682179] [<c107750b>] ? prepare_to_wait+0x43/0x48
> [ 63.682179] [<c10742a2>] worker_thread+0x94/0x1d2
> [ 63.682179] [<c10ccd2d>] ? vmstat_update+0x0/0x2f
> [ 63.682179] [<c1077357>] ? autoremove_wake_function+0x0/0x33
> [ 63.682179] [<c107420e>] ? worker_thread+0x0/0x1d2
> [ 63.682179] [<c1077120>] kthread+0x61/0x66
> [ 63.682179] [<c10770bf>] ? kthread+0x0/0x66
> [ 63.682179] [<c102ff97>] kernel_thread_helper+0x7/0x10
>
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: [xen-unstable test] 1959: FAIL [and 1 more messages]
2010-07-29 16:38 ` [xen-unstable test] 1959: FAIL [and 1 more messages] Jeremy Fitzhardinge
@ 2010-07-29 17:59 ` Ian Jackson
2010-07-30 4:58 ` Xu, Dongxiao
1 sibling, 0 replies; 12+ messages in thread
From: Ian Jackson @ 2010-07-29 17:59 UTC (permalink / raw)
To: Jeremy Fitzhardinge; +Cc: Ian Campbell, Xu, Dongxiao, Xen-devel
Jeremy Fitzhardinge writes ("Re: [Xen-devel] [xen-unstable test] 1959: FAIL [and 1 more messages]"):
> Ah, it looks like the netfront smartpoll stuff isn't coping with
> save/restore.
Is there some way I can disable it for my tests or is it going to be
fixed soon ?
Ian.
^ permalink raw reply [flat|nested] 12+ messages in thread
* RE: [xen-unstable test] 1959: FAIL [and 1 more messages]
2010-07-29 16:38 ` [xen-unstable test] 1959: FAIL [and 1 more messages] Jeremy Fitzhardinge
2010-07-29 17:59 ` Ian Jackson
@ 2010-07-30 4:58 ` Xu, Dongxiao
2010-07-30 9:18 ` Ian Campbell
1 sibling, 1 reply; 12+ messages in thread
From: Xu, Dongxiao @ 2010-07-30 4:58 UTC (permalink / raw)
To: Jeremy Fitzhardinge, Ian Jackson; +Cc: Ian Campbell, Xen-devel
Hi Ian,
Can you try this patch to see if it fixes your problem?
I tried it on my local machine, VM is pinging another host.
After save/restore/migration, the ping is still ongoing.
Thanks,
Dongxiao
From: Dongxiao Xu <dongxiao.xu@intel.com>
When saving guest, the shared ring will be set to NULL, and
in this case the polling timer interrupt should stop.
Thanks Ian J for reporting the problem.
Signed-off-by: Dongxiao Xu <dongxiao.xu@intel.com>
---
drivers/net/xen-netfront.c | 4 ++++
1 files changed, 4 insertions(+), 0 deletions(-)
diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c
index 328fe40..394bbc8 100644
--- a/drivers/net/xen-netfront.c
+++ b/drivers/net/xen-netfront.c
@@ -1340,6 +1340,10 @@ static enum hrtimer_restart smart_poll_function(struct hrtimer *timer)
np = netdev_priv(dev);
spin_lock_irqsave(&np->tx_lock, flags);
+
+ if (!np->rx.sring)
+ goto end;
+
np->smart_poll.counter++;
if (likely(netif_carrier_ok(dev))) {
--
1.6.3
Jeremy Fitzhardinge wrote:
> On 07/29/2010 09:33 AM, Ian Jackson wrote:
>> Jeremy Fitzhardinge writes ("Re: [Xen-devel] [xen-unstable test]
>> 1959: FAIL [and 1 more messages]"):
>>> On 07/29/2010 08:30 AM, Ian Jackson wrote:
>>>> Is save/restore supposed to work in pvops ? (Using your kernel for
>>>> both dom0 and domU.) That would seem to be the next thing to pick
>>>> off the list ...
>>> Yes. IanC has been tracking down a bug where it fails after a few
>>> thousand iterations (now fixed?), but aside from that its
>>> apparently OK.
>> Well, I was able to reproduce the failure that the automatic test was
>> getting. There's a problem with it not getting a copy of the console
>> output for some reason but I was able to get this oops from the
>> guest.
>>
>> Just after the restore I was able to connect to the PV console and it
>> echoed a few of my CRs (before saving I had logged in on the
>> console),
>> and then it produced the oops. Now it's apparently completely
>> wedged.
>>
>
> Ah, it looks like the netfront smartpoll stuff isn't coping with
> save/restore.
>
> J
>
>> Ian.
>>
>>
>>
>> [ 63.681260] BUG: unable to handle kernel NULL pointer dereference
>> at 00000010 [ 63.681293] IP: [<c1300381>]
>> smart_poll_function+0xbb/0xf2 [ 63.681320] *pdpt =
>> 000000001fee5027 *pde = 0000000000000000 [ 63.681344] Oops: 0000
>> [#1] SMP [ 63.681362] last sysfs file: /sys/kernel/uevent_seqnum
>> [ 63.681376] Modules linked in: [last unloaded: scsi_wait_scan] [
>> 63.681398] [ 63.681410] Pid: 5, comm: events/0 Not tainted
>> (2.6.32.16 #1) [ 63.681424] EIP: 0061:[<c1300381>] EFLAGS:
>> 00010002 CPU: 0 [ 63.681438] EIP is at
>> smart_poll_function+0xbb/0xf2 [ 63.681451] EAX: 00000000 EBX:
>> dfea8320 ECX: 00000001 EDX: 00000062 [ 63.681465] ESI: 00000064
>> EDI: 00000000 EBP: df849cfc ESP: df849cdc [ 63.681479] DS: 007b
>> ES: 007b FS: 00d8 GS: 0000 SS: 0069 [ 63.681493] Process events/0
>> (pid: 5, ti=df848000 task=df839480 task.ti=df848000) [ 63.681508]
>> Stack: [ 63.681516] dfeac828 00000002 dfeac828 dfea8368 dfea0068
>> 0000002c dfeac828 c13002c6 [ 63.681555]<0> df849d1c c1079ab8
>> df849d48 c2386358 c2386328 0000002c 00000000 c2386328 [
>> 63.681598]<0> df849d5c c1079ce8 000028cd 88bbcad9 d3b2cae7 0000000e
>> d3b2cae7 0000000e [ 63.681646] Call Trace: [ 63.681662]
>> [<c13002c6>] ? smart_poll_function+0x0/0xf2 [ 63.681683]
>> [<c1079ab8>] ? __run_hrtimer+0xa9/0xf6 [ 63.681701] [<c1079ce8>]
>> ? hrtimer_interrupt+0xcd/0x1c8 [ 63.681719] [<c102cef0>] ?
>> xen_timer_interrupt+0x2b/0x224 [ 63.681737] [<c102cb3f>] ?
>> xen_force_evtchn_callback+0xf/0x14 [ 63.681755] [<c102d2ac>] ?
>> check_events+0x8/0xc [ 63.681776] [<c102d2a3>] ?
>> xen_restore_fl_direct_end+0x0/0x1 [ 63.681795] [<c14d1471>] ?
>> _spin_unlock_irqrestore+0x2f/0x31 [ 63.681814] [<c105eb9e>] ?
>> try_to_wake_up+0x2fa/0x304 [ 63.681832] [<c102cb3f>] ?
>> xen_force_evtchn_callback+0xf/0x14 [ 63.681850] [<c10a0cc4>] ?
>> handle_IRQ_event+0x5f/0x122 [ 63.681867] [<c10a233f>] ?
>> handle_level_irq+0x58/0xa9 [ 63.681886] [<c121983d>] ?
>> __xen_evtchn_do_upcall+0xab/0x131 [ 63.681904] [<c1219c71>] ?
>> xen_evtchn_do_upcall+0x20/0x30 [ 63.682179] [<c102ffe7>] ?
>> xen_do_upcall+0x7/0xc [ 63.682179] [<c102007b>] ?
>> apic_reg_write+0xa5/0x52f [ 63.682179] [<c1002227>] ?
>> hypercall_page+0x227/0x1005 [ 63.682179] [<c102cb3f>] ?
>> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>] ?
>> check_events+0x8/0xc [ 63.682179] [<c102d26b>] ?
>> xen_irq_enable_direct_end+0x0/0x1 [ 63.682179] [<c105cd0d>] ?
>> finish_task_switch+0x4f/0xa6 [ 63.682179] [<c14d0186>] ?
>> schedule+0x7dd/0x861 [ 63.682179] [<c106dd9d>] ?
>> __mod_timer+0x135/0x140 [ 63.682179] [<c102d2ac>] ?
>> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ?
>> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ?
>> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c107750b>] ?
>> prepare_to_wait+0x43/0x48 [ 63.682179] [<c10742a2>] ?
>> worker_thread+0x94/0x1d2 [ 63.682179] [<c10ccd2d>] ?
>> vmstat_update+0x0/0x2f [ 63.682179] [<c1077357>] ?
>> autoremove_wake_function+0x0/0x33 [ 63.682179] [<c107420e>] ?
>> worker_thread+0x0/0x1d2 [ 63.682179] [<c1077120>] ?
>> kthread+0x61/0x66 [ 63.682179] [<c10770bf>] ? kthread+0x0/0x66 [
>> 63.682179] [<c102ff97>] ? kernel_thread_helper+0x7/0x10 [
>> 63.682179] Code: c6 89 d0 31 d2 f7 f6 85 d2 75 1a 85 c9 75 0c 8b 83
>> 90 08 00 00 c6 40 10 00 eb 32 c7 83 54 45 00 00 00 00 00 00 8b 83 90
>> 08 00 00<80> 78 10 00 74 1c 8b 4d e8 b8 00 ca 9a 3b 31 d2 f7 71 44
>> 31 c9 [ 63.682179] EIP: [<c1300381>] smart_poll_function+0xbb/0xf2
>> SS:ESP 0069:df849cdc [ 63.682179] CR2: 0000000000000010 [
>> 63.682179] ---[ end trace 760037e75e5675c8 ]--- [ 63.682179]
>> Kernel panic - not syncing: Fatal exception in interrupt [
>> 63.682179] Pid: 5, comm: events/0 Tainted: G D 2.6.32.16 #1
>> [ 63.682179] Call Trace: [ 63.682179] [<c14cf7b1>] ?
>> printk+0xf/0x11 [ 63.682179] [<c14cf6ee>] panic+0x39/0xed [
>> 63.682179] [<c14d2085>] oops_end+0xa1/0xb0 [ 63.682179]
>> [<c104b426>] no_context+0x137/0x141 [ 63.682179] [<c104b56f>]
>> __bad_area_nosemaphore+0x13f/0x147 [ 63.682179] [<c104b584>]
>> bad_area_nosemaphore+0xd/0x10 [ 63.682179] [<c14d3275>]
>> do_page_fault+0x1c6/0x32b [ 63.682179] [<c14d30af>] ?
>> do_page_fault+0x0/0x32b [ 63.682179] [<c14d16c6>]
>> error_code+0x66/0x6c [ 63.682179] [<c14d30af>] ?
>> do_page_fault+0x0/0x32b [ 63.682179] [<c1300381>] ?
>> smart_poll_function+0xbb/0xf2 [ 63.682179] [<c13002c6>] ?
>> smart_poll_function+0x0/0xf2 [ 63.682179] [<c1079ab8>]
>> __run_hrtimer+0xa9/0xf6 [ 63.682179] [<c1079ce8>]
>> hrtimer_interrupt+0xcd/0x1c8 [ 63.682179] [<c102cef0>]
>> xen_timer_interrupt+0x2b/0x224 [ 63.682179] [<c102cb3f>] ?
>> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>] ?
>> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ?
>> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ?
>> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c105eb9e>] ?
>> try_to_wake_up+0x2fa/0x304 [ 63.682179] [<c102cb3f>] ?
>> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c10a0cc4>]
>> handle_IRQ_event+0x5f/0x122 [ 63.682179] [<c10a233f>]
>> handle_level_irq+0x58/0xa9 [ 63.682179] [<c121983d>]
>> __xen_evtchn_do_upcall+0xab/0x131 [ 63.682179] [<c1219c71>]
>> xen_evtchn_do_upcall+0x20/0x30 [ 63.682179] [<c102ffe7>]
>> xen_do_upcall+0x7/0xc [ 63.682179] [<c102007b>] ?
>> apic_reg_write+0xa5/0x52f [ 63.682179] [<c1002227>] ?
>> hypercall_page+0x227/0x1005 [ 63.682179] [<c102cb3f>] ?
>> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>]
>> check_events+0x8/0xc [ 63.682179] [<c102d26b>] ?
>> xen_irq_enable_direct_end+0x0/0x1 [ 63.682179] [<c105cd0d>] ?
>> finish_task_switch+0x4f/0xa6 [ 63.682179] [<c14d0186>]
>> schedule+0x7dd/0x861 [ 63.682179] [<c106dd9d>] ?
>> __mod_timer+0x135/0x140 [ 63.682179] [<c102d2ac>] ?
>> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ?
>> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ?
>> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c107750b>] ?
>> prepare_to_wait+0x43/0x48 [ 63.682179] [<c10742a2>]
>> worker_thread+0x94/0x1d2 [ 63.682179] [<c10ccd2d>] ?
>> vmstat_update+0x0/0x2f [ 63.682179] [<c1077357>] ?
>> autoremove_wake_function+0x0/0x33 [ 63.682179] [<c107420e>] ?
>> worker_thread+0x0/0x1d2 [ 63.682179] [<c1077120>]
>> kthread+0x61/0x66 [ 63.682179] [<c10770bf>] ? kthread+0x0/0x66 [
>> 63.682179] [<c102ff97>] kernel_thread_helper+0x7/0x10
^ permalink raw reply related [flat|nested] 12+ messages in thread
* RE: [xen-unstable test] 1959: FAIL [and 1 more messages]
2010-07-30 4:58 ` Xu, Dongxiao
@ 2010-07-30 9:18 ` Ian Campbell
2010-07-30 14:43 ` Jeremy Fitzhardinge
2010-08-10 17:38 ` Jeremy Fitzhardinge
0 siblings, 2 replies; 12+ messages in thread
From: Ian Campbell @ 2010-07-30 9:18 UTC (permalink / raw)
To: Xu, Dongxiao; +Cc: Jeremy Fitzhardinge, Xen-devel, Ian Jackson
On Fri, 2010-07-30 at 05:58 +0100, Xu, Dongxiao wrote:
> Hi Ian,
>
> Can you try this patch to see if it fixes your problem?
>
> I tried it on my local machine, VM is pinging another host.
> After save/restore/migration, the ping is still ongoing.
>
> Thanks,
> Dongxiao
>
>
> From: Dongxiao Xu <dongxiao.xu@intel.com>
>
> When saving guest, the shared ring will be set to NULL, and
> in this case the polling timer interrupt should stop.
>
> Thanks Ian J for reporting the problem.
>
> Signed-off-by: Dongxiao Xu <dongxiao.xu@intel.com>
> ---
> drivers/net/xen-netfront.c | 4 ++++
> 1 files changed, 4 insertions(+), 0 deletions(-)
>
> diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c
> index 328fe40..394bbc8 100644
> --- a/drivers/net/xen-netfront.c
> +++ b/drivers/net/xen-netfront.c
> @@ -1340,6 +1340,10 @@ static enum hrtimer_restart smart_poll_function(struct hrtimer *timer)
> np = netdev_priv(dev);
>
> spin_lock_irqsave(&np->tx_lock, flags);
> +
> + if (!np->rx.sring)
> + goto end;
> +
Isn't there a period until the end of xennet_disconnect_backend is
reached on resume where rx.sring will still point to the old shared
ring? If so is it safe to drop through in that case?
Would it be safer to add an explicit suspend handler which stopped the
timer?
Ian.
> np->smart_poll.counter++;xennet_disconnect_backend
>
> if (likely(netif_carrier_ok(dev))) {
> --
> 1.6.3
>
>
>
> Jeremy Fitzhardinge wrote:
> > On 07/29/2010 09:33 AM, Ian Jackson wrote:
> >> Jeremy Fitzhardinge writes ("Re: [Xen-devel] [xen-unstable test]
> >> 1959: FAIL [and 1 more messages]"):
> >>> On 07/29/2010 08:30 AM, Ian Jackson wrote:
> >>>> Is save/restore supposed to work in pvops ? (Using your kernel for
> >>>> both dom0 and domU.) That would seem to be the next thing to pick
> >>>> off the list ...
> >>> Yes. IanC has been tracking down a bug where it fails after a few
> >>> thousand iterations (now fixed?), but aside from that its
> >>> apparently OK.
> >> Well, I was able to reproduce the failure that the automatic test was
> >> getting. There's a problem with it not getting a copy of the console
> >> output for some reason but I was able to get this oops from the
> >> guest.
> >>
> >> Just after the restore I was able to connect to the PV console and it
> >> echoed a few of my CRs (before saving I had logged in on the
> >> console),
> >> and then it produced the oops. Now it's apparently completely
> >> wedged.
> >>
> >
> > Ah, it looks like the netfront smartpoll stuff isn't coping with
> > save/restore.
> >
> > J
> >
> >> Ian.
> >>
> >>
> >>
> >> [ 63.681260] BUG: unable to handle kernel NULL pointer dereference
> >> at 00000010 [ 63.681293] IP: [<c1300381>]
> >> smart_poll_function+0xbb/0xf2 [ 63.681320] *pdpt =
> >> 000000001fee5027 *pde = 0000000000000000 [ 63.681344] Oops: 0000
> >> [#1] SMP [ 63.681362] last sysfs file: /sys/kernel/uevent_seqnum
> >> [ 63.681376] Modules linked in: [last unloaded: scsi_wait_scan] [
> >> 63.681398] [ 63.681410] Pid: 5, comm: events/0 Not tainted
> >> (2.6.32.16 #1) [ 63.681424] EIP: 0061:[<c1300381>] EFLAGS:
> >> 00010002 CPU: 0 [ 63.681438] EIP is at
> >> smart_poll_function+0xbb/0xf2 [ 63.681451] EAX: 00000000 EBX:
> >> dfea8320 ECX: 00000001 EDX: 00000062 [ 63.681465] ESI: 00000064
> >> EDI: 00000000 EBP: df849cfc ESP: df849cdc [ 63.681479] DS: 007b
> >> ES: 007b FS: 00d8 GS: 0000 SS: 0069 [ 63.681493] Process events/0
> >> (pid: 5, ti=df848000 task=df839480 task.ti=df848000) [ 63.681508]
> >> Stack: [ 63.681516] dfeac828 00000002 dfeac828 dfea8368 dfea0068
> >> 0000002c dfeac828 c13002c6 [ 63.681555]<0> df849d1c c1079ab8
> >> df849d48 c2386358 c2386328 0000002c 00000000 c2386328 [
> >> 63.681598]<0> df849d5c c1079ce8 000028cd 88bbcad9 d3b2cae7 0000000e
> >> d3b2cae7 0000000e [ 63.681646] Call Trace: [ 63.681662]
> >> [<c13002c6>] ? smart_poll_function+0x0/0xf2 [ 63.681683]
> >> [<c1079ab8>] ? __run_hrtimer+0xa9/0xf6 [ 63.681701] [<c1079ce8>]
> >> ? hrtimer_interrupt+0xcd/0x1c8 [ 63.681719] [<c102cef0>] ?
> >> xen_timer_interrupt+0x2b/0x224 [ 63.681737] [<c102cb3f>] ?
> >> xen_force_evtchn_callback+0xf/0x14 [ 63.681755] [<c102d2ac>] ?
> >> check_events+0x8/0xc [ 63.681776] [<c102d2a3>] ?
> >> xen_restore_fl_direct_end+0x0/0x1 [ 63.681795] [<c14d1471>] ?
> >> _spin_unlock_irqrestore+0x2f/0x31 [ 63.681814] [<c105eb9e>] ?
> >> try_to_wake_up+0x2fa/0x304 [ 63.681832] [<c102cb3f>] ?
> >> xen_force_evtchn_callback+0xf/0x14 [ 63.681850] [<c10a0cc4>] ?
> >> handle_IRQ_event+0x5f/0x122 [ 63.681867] [<c10a233f>] ?
> >> handle_level_irq+0x58/0xa9 [ 63.681886] [<c121983d>] ?
> >> __xen_evtchn_do_upcall+0xab/0x131 [ 63.681904] [<c1219c71>] ?
> >> xen_evtchn_do_upcall+0x20/0x30 [ 63.682179] [<c102ffe7>] ?
> >> xen_do_upcall+0x7/0xc [ 63.682179] [<c102007b>] ?
> >> apic_reg_write+0xa5/0x52f [ 63.682179] [<c1002227>] ?
> >> hypercall_page+0x227/0x1005 [ 63.682179] [<c102cb3f>] ?
> >> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>] ?
> >> check_events+0x8/0xc [ 63.682179] [<c102d26b>] ?
> >> xen_irq_enable_direct_end+0x0/0x1 [ 63.682179] [<c105cd0d>] ?
> >> finish_task_switch+0x4f/0xa6 [ 63.682179] [<c14d0186>] ?
> >> schedule+0x7dd/0x861 [ 63.682179] [<c106dd9d>] ?
> >> __mod_timer+0x135/0x140 [ 63.682179] [<c102d2ac>] ?
> >> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ?
> >> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ?
> >> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c107750b>] ?
> >> prepare_to_wait+0x43/0x48 [ 63.682179] [<c10742a2>] ?
> >> worker_thread+0x94/0x1d2 [ 63.682179] [<c10ccd2d>] ?
> >> vmstat_update+0x0/0x2f [ 63.682179] [<c1077357>] ?
> >> autoremove_wake_function+0x0/0x33 [ 63.682179] [<c107420e>] ?
> >> worker_thread+0x0/0x1d2 [ 63.682179] [<c1077120>] ?
> >> kthread+0x61/0x66 [ 63.682179] [<c10770bf>] ? kthread+0x0/0x66 [
> >> 63.682179] [<c102ff97>] ? kernel_thread_helper+0x7/0x10 [
> >> 63.682179] Code: c6 89 d0 31 d2 f7 f6 85 d2 75 1a 85 c9 75 0c 8b 83
> >> 90 08 00 00 c6 40 10 00 eb 32 c7 83 54 45 00 00 00 00 00 00 8b 83 90
> >> 08 00 00<80> 78 10 00 74 1c 8b 4d e8 b8 00 ca 9a 3b 31 d2 f7 71 44
> >> 31 c9 [ 63.682179] EIP: [<c1300381>] smart_poll_function+0xbb/0xf2
> >> SS:ESP 0069:df849cdc [ 63.682179] CR2: 0000000000000010 [
> >> 63.682179] ---[ end trace 760037e75e5675c8 ]--- [ 63.682179]
> >> Kernel panic - not syncing: Fatal exception in interrupt [
> >> 63.682179] Pid: 5, comm: events/0 Tainted: G D 2.6.32.16 #1
> >> [ 63.682179] Call Trace: [ 63.682179] [<c14cf7b1>] ?
> >> printk+0xf/0x11 [ 63.682179] [<c14cf6ee>] panic+0x39/0xed [
> >> 63.682179] [<c14d2085>] oops_end+0xa1/0xb0 [ 63.682179]
> >> [<c104b426>] no_context+0x137/0x141 [ 63.682179] [<c104b56f>]
> >> __bad_area_nosemaphore+0x13f/0x147 [ 63.682179] [<c104b584>]
> >> bad_area_nosemaphore+0xd/0x10 [ 63.682179] [<c14d3275>]
> >> do_page_fault+0x1c6/0x32b [ 63.682179] [<c14d30af>] ?
> >> do_page_fault+0x0/0x32b [ 63.682179] [<c14d16c6>]
> >> error_code+0x66/0x6c [ 63.682179] [<c14d30af>] ?
> >> do_page_fault+0x0/0x32b [ 63.682179] [<c1300381>] ?
> >> smart_poll_function+0xbb/0xf2 [ 63.682179] [<c13002c6>] ?
> >> smart_poll_function+0x0/0xf2 [ 63.682179] [<c1079ab8>]
> >> __run_hrtimer+0xa9/0xf6 [ 63.682179] [<c1079ce8>]
> >> hrtimer_interrupt+0xcd/0x1c8 [ 63.682179] [<c102cef0>]
> >> xen_timer_interrupt+0x2b/0x224 [ 63.682179] [<c102cb3f>] ?
> >> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>] ?
> >> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ?
> >> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ?
> >> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c105eb9e>] ?
> >> try_to_wake_up+0x2fa/0x304 [ 63.682179] [<c102cb3f>] ?
> >> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c10a0cc4>]
> >> handle_IRQ_event+0x5f/0x122 [ 63.682179] [<c10a233f>]
> >> handle_level_irq+0x58/0xa9 [ 63.682179] [<c121983d>]
> >> __xen_evtchn_do_upcall+0xab/0x131 [ 63.682179] [<c1219c71>]
> >> xen_evtchn_do_upcall+0x20/0x30 [ 63.682179] [<c102ffe7>]
> >> xen_do_upcall+0x7/0xc [ 63.682179] [<c102007b>] ?
> >> apic_reg_write+0xa5/0x52f [ 63.682179] [<c1002227>] ?
> >> hypercall_page+0x227/0x1005 [ 63.682179] [<c102cb3f>] ?
> >> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>]
> >> check_events+0x8/0xc [ 63.682179] [<c102d26b>] ?
> >> xen_irq_enable_direct_end+0x0/0x1 [ 63.682179] [<c105cd0d>] ?
> >> finish_task_switch+0x4f/0xa6 [ 63.682179] [<c14d0186>]
> >> schedule+0x7dd/0x861 [ 63.682179] [<c106dd9d>] ?
> >> __mod_timer+0x135/0x140 [ 63.682179] [<c102d2ac>] ?
> >> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ?
> >> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ?
> >> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c107750b>] ?
> >> prepare_to_wait+0x43/0x48 [ 63.682179] [<c10742a2>]
> >> worker_thread+0x94/0x1d2 [ 63.682179] [<c10ccd2d>] ?
> >> vmstat_update+0x0/0x2f [ 63.682179] [<c1077357>] ?
> >> autoremove_wake_function+0x0/0x33 [ 63.682179] [<c107420e>] ?
> >> worker_thread+0x0/0x1d2 [ 63.682179] [<c1077120>]
> >> kthread+0x61/0x66 [ 63.682179] [<c10770bf>] ? kthread+0x0/0x66 [
> >> 63.682179] [<c102ff97>] kernel_thread_helper+0x7/0x10
>
^ permalink raw reply [flat|nested] 12+ messages in thread
* RE: [xen-unstable test] 1959: FAIL [and 1 more messages]
2010-07-30 9:18 ` Ian Campbell
@ 2010-07-30 14:43 ` Jeremy Fitzhardinge
2010-08-10 17:38 ` Jeremy Fitzhardinge
1 sibling, 0 replies; 12+ messages in thread
From: Jeremy Fitzhardinge @ 2010-07-30 14:43 UTC (permalink / raw)
To: Ian Campbell, Xu, Dongxiao; +Cc: Xen-devel, Ian Jackson
I think its safer to explicitly stop the timer. Might there also be a race with a pending timer when setting up the new one on reconnect?
J
"Ian Campbell" <Ian.Campbell@eu.citrix.com> wrote:
>On Fri, 2010-07-30 at 05:58 +0100, Xu, Dongxiao wrote:
>> Hi Ian,
>>
>> Can you try this patch to see if it fixes your problem?
>>
>> I tried it on my local machine, VM is pinging another host.
>> After save/restore/migration, the ping is still ongoing.
>>
>> Thanks,
>> Dongxiao
>>
>>
>> From: Dongxiao Xu <dongxiao.xu@intel.com>
>>
>> When saving guest, the shared ring will be set to NULL, and
>> in this case the polling timer interrupt should stop.
>>
>> Thanks Ian J for reporting the problem.
>>
>> Signed-off-by: Dongxiao Xu <dongxiao.xu@intel.com>
>> ---
>> drivers/net/xen-netfront.c | 4 ++++
>> 1 files changed, 4 insertions(+), 0 deletions(-)
>>
>> diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c
>> index 328fe40..394bbc8 100644
>> --- a/drivers/net/xen-netfront.c
>> +++ b/drivers/net/xen-netfront.c
>> @@ -1340,6 +1340,10 @@ static enum hrtimer_restart smart_poll_function(struct hrtimer *timer)
>> np = netdev_priv(dev);
>>
>> spin_lock_irqsave(&np->tx_lock, flags);
>> +
>> + if (!np->rx.sring)
>> + goto end;
>> +
>
>Isn't there a period until the end of xennet_disconnect_backend is
>reached on resume where rx.sring will still point to the old shared
>ring? If so is it safe to drop through in that case?
>
>Would it be safer to add an explicit suspend handler which stopped the
>timer?
>
>Ian.
>
>> np->smart_poll.counter++;xennet_disconnect_backend
>>
>> if (likely(netif_carrier_ok(dev))) {
>> --
>> 1.6.3
>>
>>
>>
>> Jeremy Fitzhardinge wrote:
>> > On 07/29/2010 09:33 AM, Ian Jackson wrote:
>> >> Jeremy Fitzhardinge writes ("Re: [Xen-devel] [xen-unstable test]
>> >> 1959: FAIL [and 1 more messages]"):
>> >>> On 07/29/2010 08:30 AM, Ian Jackson wrote:
>> >>>> Is save/restore supposed to work in pvops ? (Using your kernel for
>> >>>> both dom0 and domU.) That would seem to be the next thing to pick
>> >>>> off the list ...
>> >>> Yes. IanC has been tracking down a bug where it fails after a few
>> >>> thousand iterations (now fixed?), but aside from that its
>> >>> apparently OK.
>> >> Well, I was able to reproduce the failure that the automatic test was
>> >> getting. There's a problem with it not getting a copy of the console
>> >> output for some reason but I was able to get this oops from the
>> >> guest.
>> >>
>> >> Just after the restore I was able to connect to the PV console and it
>> >> echoed a few of my CRs (before saving I had logged in on the
>> >> console),
>> >> and then it produced the oops. Now it's apparently completely
>> >> wedged.
>> >>
>> >
>> > Ah, it looks like the netfront smartpoll stuff isn't coping with
>> > save/restore.
>> >
>> > J
>> >
>> >> Ian.
>> >>
>> >>
>> >>
>> >> [ 63.681260] BUG: unable to handle kernel NULL pointer dereference
>> >> at 00000010 [ 63.681293] IP: [<c1300381>]
>> >> smart_poll_function+0xbb/0xf2 [ 63.681320] *pdpt =
>> >> 000000001fee5027 *pde = 0000000000000000 [ 63.681344] Oops: 0000
>> >> [#1] SMP [ 63.681362] last sysfs file: /sys/kernel/uevent_seqnum
>> >> [ 63.681376] Modules linked in: [last unloaded: scsi_wait_scan] [
>> >> 63.681398] [ 63.681410] Pid: 5, comm: events/0 Not tainted
>> >> (2.6.32.16 #1) [ 63.681424] EIP: 0061:[<c1300381>] EFLAGS:
>> >> 00010002 CPU: 0 [ 63.681438] EIP is at
>> >> smart_poll_function+0xbb/0xf2 [ 63.681451] EAX: 00000000 EBX:
>> >> dfea8320 ECX: 00000001 EDX: 00000062 [ 63.681465] ESI: 00000064
>> >> EDI: 00000000 EBP: df849cfc ESP: df849cdc [ 63.681479] DS: 007b
>> >> ES: 007b FS: 00d8 GS: 0000 SS: 0069 [ 63.681493] Process events/0
>> >> (pid: 5, ti=df848000 task=df839480 task.ti=df848000) [ 63.681508]
>> >> Stack: [ 63.681516] dfeac828 00000002 dfeac828 dfea8368 dfea0068
>> >> 0000002c dfeac828 c13002c6 [ 63.681555]<0> df849d1c c1079ab8
>> >> df849d48 c2386358 c2386328 0000002c 00000000 c2386328 [
>> >> 63.681598]<0> df849d5c c1079ce8 000028cd 88bbcad9 d3b2cae7 0000000e
>> >> d3b2cae7 0000000e [ 63.681646] Call Trace: [ 63.681662]
>> >> [<c13002c6>] ? smart_poll_function+0x0/0xf2 [ 63.681683]
>> >> [<c1079ab8>] ? __run_hrtimer+0xa9/0xf6 [ 63.681701] [<c1079ce8>]
>> >> ? hrtimer_interrupt+0xcd/0x1c8 [ 63.681719] [<c102cef0>] ?
>> >> xen_timer_interrupt+0x2b/0x224 [ 63.681737] [<c102cb3f>] ?
>> >> xen_force_evtchn_callback+0xf/0x14 [ 63.681755] [<c102d2ac>] ?
>> >> check_events+0x8/0xc [ 63.681776] [<c102d2a3>] ?
>> >> xen_restore_fl_direct_end+0x0/0x1 [ 63.681795] [<c14d1471>] ?
>> >> _spin_unlock_irqrestore+0x2f/0x31 [ 63.681814] [<c105eb9e>] ?
>> >> try_to_wake_up+0x2fa/0x304 [ 63.681832] [<c102cb3f>] ?
>> >> xen_force_evtchn_callback+0xf/0x14 [ 63.681850] [<c10a0cc4>] ?
>> >> handle_IRQ_event+0x5f/0x122 [ 63.681867] [<c10a233f>] ?
>> >> handle_level_irq+0x58/0xa9 [ 63.681886] [<c121983d>] ?
>> >> __xen_evtchn_do_upcall+0xab/0x131 [ 63.681904] [<c1219c71>] ?
>> >> xen_evtchn_do_upcall+0x20/0x30 [ 63.682179] [<c102ffe7>] ?
>> >> xen_do_upcall+0x7/0xc [ 63.682179] [<c102007b>] ?
>> >> apic_reg_write+0xa5/0x52f [ 63.682179] [<c1002227>] ?
>> >> hypercall_page+0x227/0x1005 [ 63.682179] [<c102cb3f>] ?
>> >> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>] ?
>> >> check_events+0x8/0xc [ 63.682179] [<c102d26b>] ?
>> >> xen_irq_enable_direct_end+0x0/0x1 [ 63.682179] [<c105cd0d>] ?
>> >> finish_task_switch+0x4f/0xa6 [ 63.682179] [<c14d0186>] ?
>> >> schedule+0x7dd/0x861 [ 63.682179] [<c106dd9d>] ?
>> >> __mod_timer+0x135/0x140 [ 63.682179] [<c102d2ac>] ?
>> >> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ?
>> >> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ?
>> >> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c107750b>] ?
>> >> prepare_to_wait+0x43/0x48 [ 63.682179] [<c10742a2>] ?
>> >> worker_thread+0x94/0x1d2 [ 63.682179] [<c10ccd2d>] ?
>> >> vmstat_update+0x0/0x2f [ 63.682179] [<c1077357>] ?
>> >> autoremove_wake_function+0x0/0x33 [ 63.682179] [<c107420e>] ?
>> >> worker_thread+0x0/0x1d2 [ 63.682179] [<c1077120>] ?
>> >> kthread+0x61/0x66 [ 63.682179] [<c10770bf>] ? kthread+0x0/0x66 [
>> >> 63.682179] [<c102ff97>] ? kernel_thread_helper+0x7/0x10 [
>> >> 63.682179] Code: c6 89 d0 31 d2 f7 f6 85 d2 75 1a 85 c9 75 0c 8b 83
>> >> 90 08 00 00 c6 40 10 00 eb 32 c7 83 54 45 00 00 00 00 00 00 8b 83 90
>> >> 08 00 00<80> 78 10 00 74 1c 8b 4d e8 b8 00 ca 9a 3b 31 d2 f7 71 44
>> >> 31 c9 [ 63.682179] EIP: [<c1300381>] smart_poll_function+0xbb/0xf2
>> >> SS:ESP 0069:df849cdc [ 63.682179] CR2: 0000000000000010 [
>> >> 63.682179] ---[ end trace 760037e75e5675c8 ]--- [ 63.682179]
>> >> Kernel panic - not syncing: Fatal exception in interrupt [
>> >> 63.682179] Pid: 5, comm: events/0 Tainted: G D 2.6.32.16 #1
>> >> [ 63.682179] Call Trace: [ 63.682179] [<c14cf7b1>] ?
>> >> printk+0xf/0x11 [ 63.682179] [<c14cf6ee>] panic+0x39/0xed [
>> >> 63.682179] [<c14d2085>] oops_end+0xa1/0xb0 [ 63.682179]
>> >> [<c104b426>] no_context+0x137/0x141 [ 63.682179] [<c104b56f>]
>> >> __bad_area_nosemaphore+0x13f/0x147 [ 63.682179] [<c104b584>]
>> >> bad_area_nosemaphore+0xd/0x10 [ 63.682179] [<c14d3275>]
>> >> do_page_fault+0x1c6/0x32b [ 63.682179] [<c14d30af>] ?
>> >> do_page_fault+0x0/0x32b [ 63.682179] [<c14d16c6>]
>> >> error_code+0x66/0x6c [ 63.682179] [<c14d30af>] ?
>> >> do_page_fault+0x0/0x32b [ 63.682179] [<c1300381>] ?
>> >> smart_poll_function+0xbb/0xf2 [ 63.682179] [<c13002c6>] ?
>> >> smart_poll_function+0x0/0xf2 [ 63.682179] [<c1079ab8>]
>> >> __run_hrtimer+0xa9/0xf6 [ 63.682179] [<c1079ce8>]
>> >> hrtimer_interrupt+0xcd/0x1c8 [ 63.682179] [<c102cef0>]
>> >> xen_timer_interrupt+0x2b/0x224 [ 63.682179] [<c102cb3f>] ?
>> >> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>] ?
>> >> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ?
>> >> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ?
>> >> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c105eb9e>] ?
>> >> try_to_wake_up+0x2fa/0x304 [ 63.682179] [<c102cb3f>] ?
>> >> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c10a0cc4>]
>> >> handle_IRQ_event+0x5f/0x122 [ 63.682179] [<c10a233f>]
>> >> handle_level_irq+0x58/0xa9 [ 63.682179] [<c121983d>]
>> >> __xen_evtchn_do_upcall+0xab/0x131 [ 63.682179] [<c1219c71>]
>> >> xen_evtchn_do_upcall+0x20/0x30 [ 63.682179] [<c102ffe7>]
>> >> xen_do_upcall+0x7/0xc [ 63.682179] [<c102007b>] ?
>> >> apic_reg_write+0xa5/0x52f [ 63.682179] [<c1002227>] ?
>> >> hypercall_page+0x227/0x1005 [ 63.682179] [<c102cb3f>] ?
>> >> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>]
>> >> check_events+0x8/0xc [ 63.682179] [<c102d26b>] ?
>> >> xen_irq_enable_direct_end+0x0/0x1 [ 63.682179] [<c105cd0d>] ?
>> >> finish_task_switch+0x4f/0xa6 [ 63.682179] [<c14d0186>]
>> >> schedule+0x7dd/0x861 [ 63.682179] [<c106dd9d>] ?
>> >> __mod_timer+0x135/0x140 [ 63.682179] [<c102d2ac>] ?
>> >> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ?
>> >> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ?
>> >> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c107750b>] ?
>> >> prepare_to_wait+0x43/0x48 [ 63.682179] [<c10742a2>]
>> >> worker_thread+0x94/0x1d2 [ 63.682179] [<c10ccd2d>] ?
>> >> vmstat_update+0x0/0x2f [ 63.682179] [<c1077357>] ?
>> >> autoremove_wake_function+0x0/0x33 [ 63.682179] [<c107420e>] ?
>> >> worker_thread+0x0/0x1d2 [ 63.682179] [<c1077120>]
>> >> kthread+0x61/0x66 [ 63.682179] [<c10770bf>] ? kthread+0x0/0x66 [
>> >> 63.682179] [<c102ff97>] kernel_thread_helper+0x7/0x10
>>
>
>
--
Sent from my Android phone with K-9 Mail. Please excuse my brevity.
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: [xen-unstable test] 1959: FAIL [and 1 more messages]
2010-07-30 9:18 ` Ian Campbell
2010-07-30 14:43 ` Jeremy Fitzhardinge
@ 2010-08-10 17:38 ` Jeremy Fitzhardinge
2010-08-11 3:19 ` Xu, Dongxiao
1 sibling, 1 reply; 12+ messages in thread
From: Jeremy Fitzhardinge @ 2010-08-10 17:38 UTC (permalink / raw)
To: Xu, Dongxiao; +Cc: Ian Campbell, Xen-devel, Ian Jackson
On 07/30/2010 02:18 AM, Ian Campbell wrote:
>> diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c
>> index 328fe40..394bbc8 100644
>> --- a/drivers/net/xen-netfront.c
>> +++ b/drivers/net/xen-netfront.c
>> @@ -1340,6 +1340,10 @@ static enum hrtimer_restart smart_poll_function(struct hrtimer *timer)
>> np = netdev_priv(dev);
>>
>> spin_lock_irqsave(&np->tx_lock, flags);
>> +
>> + if (!np->rx.sring)
>> + goto end;
>> +
> Isn't there a period until the end of xennet_disconnect_backend is
> reached on resume where rx.sring will still point to the old shared
> ring? If so is it safe to drop through in that case?
>
> Would it be safer to add an explicit suspend handler which stopped the
> timer?
Dongxiao, do you have a comment/updated patch for this? I'm going to
revert the smartpoll stuff in the meantime, because its causing tests to
fail.
Thanks,
J
> Ian.
>
>> np->smart_poll.counter++;xennet_disconnect_backend
>>
>> if (likely(netif_carrier_ok(dev))) {
>> --
>> 1.6.3
>>
>>
>>
>> Jeremy Fitzhardinge wrote:
>>> On 07/29/2010 09:33 AM, Ian Jackson wrote:
>>>> Jeremy Fitzhardinge writes ("Re: [Xen-devel] [xen-unstable test]
>>>> 1959: FAIL [and 1 more messages]"):
>>>>> On 07/29/2010 08:30 AM, Ian Jackson wrote:
>>>>>> Is save/restore supposed to work in pvops ? (Using your kernel for
>>>>>> both dom0 and domU.) That would seem to be the next thing to pick
>>>>>> off the list ...
>>>>> Yes. IanC has been tracking down a bug where it fails after a few
>>>>> thousand iterations (now fixed?), but aside from that its
>>>>> apparently OK.
>>>> Well, I was able to reproduce the failure that the automatic test was
>>>> getting. There's a problem with it not getting a copy of the console
>>>> output for some reason but I was able to get this oops from the
>>>> guest.
>>>>
>>>> Just after the restore I was able to connect to the PV console and it
>>>> echoed a few of my CRs (before saving I had logged in on the
>>>> console),
>>>> and then it produced the oops. Now it's apparently completely
>>>> wedged.
>>>>
>>> Ah, it looks like the netfront smartpoll stuff isn't coping with
>>> save/restore.
>>>
>>> J
>>>
>>>> Ian.
>>>>
>>>>
>>>>
>>>> [ 63.681260] BUG: unable to handle kernel NULL pointer dereference
>>>> at 00000010 [ 63.681293] IP: [<c1300381>]
>>>> smart_poll_function+0xbb/0xf2 [ 63.681320] *pdpt =
>>>> 000000001fee5027 *pde = 0000000000000000 [ 63.681344] Oops: 0000
>>>> [#1] SMP [ 63.681362] last sysfs file: /sys/kernel/uevent_seqnum
>>>> [ 63.681376] Modules linked in: [last unloaded: scsi_wait_scan] [
>>>> 63.681398] [ 63.681410] Pid: 5, comm: events/0 Not tainted
>>>> (2.6.32.16 #1) [ 63.681424] EIP: 0061:[<c1300381>] EFLAGS:
>>>> 00010002 CPU: 0 [ 63.681438] EIP is at
>>>> smart_poll_function+0xbb/0xf2 [ 63.681451] EAX: 00000000 EBX:
>>>> dfea8320 ECX: 00000001 EDX: 00000062 [ 63.681465] ESI: 00000064
>>>> EDI: 00000000 EBP: df849cfc ESP: df849cdc [ 63.681479] DS: 007b
>>>> ES: 007b FS: 00d8 GS: 0000 SS: 0069 [ 63.681493] Process events/0
>>>> (pid: 5, ti=df848000 task=df839480 task.ti=df848000) [ 63.681508]
>>>> Stack: [ 63.681516] dfeac828 00000002 dfeac828 dfea8368 dfea0068
>>>> 0000002c dfeac828 c13002c6 [ 63.681555]<0> df849d1c c1079ab8
>>>> df849d48 c2386358 c2386328 0000002c 00000000 c2386328 [
>>>> 63.681598]<0> df849d5c c1079ce8 000028cd 88bbcad9 d3b2cae7 0000000e
>>>> d3b2cae7 0000000e [ 63.681646] Call Trace: [ 63.681662]
>>>> [<c13002c6>] ? smart_poll_function+0x0/0xf2 [ 63.681683]
>>>> [<c1079ab8>] ? __run_hrtimer+0xa9/0xf6 [ 63.681701] [<c1079ce8>]
>>>> ? hrtimer_interrupt+0xcd/0x1c8 [ 63.681719] [<c102cef0>] ?
>>>> xen_timer_interrupt+0x2b/0x224 [ 63.681737] [<c102cb3f>] ?
>>>> xen_force_evtchn_callback+0xf/0x14 [ 63.681755] [<c102d2ac>] ?
>>>> check_events+0x8/0xc [ 63.681776] [<c102d2a3>] ?
>>>> xen_restore_fl_direct_end+0x0/0x1 [ 63.681795] [<c14d1471>] ?
>>>> _spin_unlock_irqrestore+0x2f/0x31 [ 63.681814] [<c105eb9e>] ?
>>>> try_to_wake_up+0x2fa/0x304 [ 63.681832] [<c102cb3f>] ?
>>>> xen_force_evtchn_callback+0xf/0x14 [ 63.681850] [<c10a0cc4>] ?
>>>> handle_IRQ_event+0x5f/0x122 [ 63.681867] [<c10a233f>] ?
>>>> handle_level_irq+0x58/0xa9 [ 63.681886] [<c121983d>] ?
>>>> __xen_evtchn_do_upcall+0xab/0x131 [ 63.681904] [<c1219c71>] ?
>>>> xen_evtchn_do_upcall+0x20/0x30 [ 63.682179] [<c102ffe7>] ?
>>>> xen_do_upcall+0x7/0xc [ 63.682179] [<c102007b>] ?
>>>> apic_reg_write+0xa5/0x52f [ 63.682179] [<c1002227>] ?
>>>> hypercall_page+0x227/0x1005 [ 63.682179] [<c102cb3f>] ?
>>>> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>] ?
>>>> check_events+0x8/0xc [ 63.682179] [<c102d26b>] ?
>>>> xen_irq_enable_direct_end+0x0/0x1 [ 63.682179] [<c105cd0d>] ?
>>>> finish_task_switch+0x4f/0xa6 [ 63.682179] [<c14d0186>] ?
>>>> schedule+0x7dd/0x861 [ 63.682179] [<c106dd9d>] ?
>>>> __mod_timer+0x135/0x140 [ 63.682179] [<c102d2ac>] ?
>>>> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ?
>>>> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ?
>>>> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c107750b>] ?
>>>> prepare_to_wait+0x43/0x48 [ 63.682179] [<c10742a2>] ?
>>>> worker_thread+0x94/0x1d2 [ 63.682179] [<c10ccd2d>] ?
>>>> vmstat_update+0x0/0x2f [ 63.682179] [<c1077357>] ?
>>>> autoremove_wake_function+0x0/0x33 [ 63.682179] [<c107420e>] ?
>>>> worker_thread+0x0/0x1d2 [ 63.682179] [<c1077120>] ?
>>>> kthread+0x61/0x66 [ 63.682179] [<c10770bf>] ? kthread+0x0/0x66 [
>>>> 63.682179] [<c102ff97>] ? kernel_thread_helper+0x7/0x10 [
>>>> 63.682179] Code: c6 89 d0 31 d2 f7 f6 85 d2 75 1a 85 c9 75 0c 8b 83
>>>> 90 08 00 00 c6 40 10 00 eb 32 c7 83 54 45 00 00 00 00 00 00 8b 83 90
>>>> 08 00 00<80> 78 10 00 74 1c 8b 4d e8 b8 00 ca 9a 3b 31 d2 f7 71 44
>>>> 31 c9 [ 63.682179] EIP: [<c1300381>] smart_poll_function+0xbb/0xf2
>>>> SS:ESP 0069:df849cdc [ 63.682179] CR2: 0000000000000010 [
>>>> 63.682179] ---[ end trace 760037e75e5675c8 ]--- [ 63.682179]
>>>> Kernel panic - not syncing: Fatal exception in interrupt [
>>>> 63.682179] Pid: 5, comm: events/0 Tainted: G D 2.6.32.16 #1
>>>> [ 63.682179] Call Trace: [ 63.682179] [<c14cf7b1>] ?
>>>> printk+0xf/0x11 [ 63.682179] [<c14cf6ee>] panic+0x39/0xed [
>>>> 63.682179] [<c14d2085>] oops_end+0xa1/0xb0 [ 63.682179]
>>>> [<c104b426>] no_context+0x137/0x141 [ 63.682179] [<c104b56f>]
>>>> __bad_area_nosemaphore+0x13f/0x147 [ 63.682179] [<c104b584>]
>>>> bad_area_nosemaphore+0xd/0x10 [ 63.682179] [<c14d3275>]
>>>> do_page_fault+0x1c6/0x32b [ 63.682179] [<c14d30af>] ?
>>>> do_page_fault+0x0/0x32b [ 63.682179] [<c14d16c6>]
>>>> error_code+0x66/0x6c [ 63.682179] [<c14d30af>] ?
>>>> do_page_fault+0x0/0x32b [ 63.682179] [<c1300381>] ?
>>>> smart_poll_function+0xbb/0xf2 [ 63.682179] [<c13002c6>] ?
>>>> smart_poll_function+0x0/0xf2 [ 63.682179] [<c1079ab8>]
>>>> __run_hrtimer+0xa9/0xf6 [ 63.682179] [<c1079ce8>]
>>>> hrtimer_interrupt+0xcd/0x1c8 [ 63.682179] [<c102cef0>]
>>>> xen_timer_interrupt+0x2b/0x224 [ 63.682179] [<c102cb3f>] ?
>>>> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>] ?
>>>> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ?
>>>> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ?
>>>> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c105eb9e>] ?
>>>> try_to_wake_up+0x2fa/0x304 [ 63.682179] [<c102cb3f>] ?
>>>> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c10a0cc4>]
>>>> handle_IRQ_event+0x5f/0x122 [ 63.682179] [<c10a233f>]
>>>> handle_level_irq+0x58/0xa9 [ 63.682179] [<c121983d>]
>>>> __xen_evtchn_do_upcall+0xab/0x131 [ 63.682179] [<c1219c71>]
>>>> xen_evtchn_do_upcall+0x20/0x30 [ 63.682179] [<c102ffe7>]
>>>> xen_do_upcall+0x7/0xc [ 63.682179] [<c102007b>] ?
>>>> apic_reg_write+0xa5/0x52f [ 63.682179] [<c1002227>] ?
>>>> hypercall_page+0x227/0x1005 [ 63.682179] [<c102cb3f>] ?
>>>> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>]
>>>> check_events+0x8/0xc [ 63.682179] [<c102d26b>] ?
>>>> xen_irq_enable_direct_end+0x0/0x1 [ 63.682179] [<c105cd0d>] ?
>>>> finish_task_switch+0x4f/0xa6 [ 63.682179] [<c14d0186>]
>>>> schedule+0x7dd/0x861 [ 63.682179] [<c106dd9d>] ?
>>>> __mod_timer+0x135/0x140 [ 63.682179] [<c102d2ac>] ?
>>>> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ?
>>>> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ?
>>>> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c107750b>] ?
>>>> prepare_to_wait+0x43/0x48 [ 63.682179] [<c10742a2>]
>>>> worker_thread+0x94/0x1d2 [ 63.682179] [<c10ccd2d>] ?
>>>> vmstat_update+0x0/0x2f [ 63.682179] [<c1077357>] ?
>>>> autoremove_wake_function+0x0/0x33 [ 63.682179] [<c107420e>] ?
>>>> worker_thread+0x0/0x1d2 [ 63.682179] [<c1077120>]
>>>> kthread+0x61/0x66 [ 63.682179] [<c10770bf>] ? kthread+0x0/0x66 [
>>>> 63.682179] [<c102ff97>] kernel_thread_helper+0x7/0x10
>
^ permalink raw reply [flat|nested] 12+ messages in thread
* RE: [xen-unstable test] 1959: FAIL [and 1 more messages]
2010-08-10 17:38 ` Jeremy Fitzhardinge
@ 2010-08-11 3:19 ` Xu, Dongxiao
2010-08-11 15:01 ` Ian Campbell
0 siblings, 1 reply; 12+ messages in thread
From: Xu, Dongxiao @ 2010-08-11 3:19 UTC (permalink / raw)
To: Jeremy Fitzhardinge; +Cc: Ian Campbell, Xen-devel, Ian Jackson
[-- Attachment #1: Type: text/plain, Size: 9687 bytes --]
Hi Jeremy,
Sorry for the late response, recently I was spinning on some other tasks.
The updated patch is attached, which explicitly disable hrtimer when VM suspends.
Thanks,
Dongxiao
________________________________________
From: Jeremy Fitzhardinge [jeremy@goop.org]
Sent: Tuesday, August 10, 2010 10:38 AM
To: Xu, Dongxiao
Cc: Ian Campbell; Ian Jackson; Xen-devel@lists.xensource.com
Subject: Re: [Xen-devel] [xen-unstable test] 1959: FAIL [and 1 more messages]
On 07/30/2010 02:18 AM, Ian Campbell wrote:
>> diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c
>> index 328fe40..394bbc8 100644
>> --- a/drivers/net/xen-netfront.c
>> +++ b/drivers/net/xen-netfront.c
>> @@ -1340,6 +1340,10 @@ static enum hrtimer_restart smart_poll_function(struct hrtimer *timer)
>> np = netdev_priv(dev);
>>
>> spin_lock_irqsave(&np->tx_lock, flags);
>> +
>> + if (!np->rx.sring)
>> + goto end;
>> +
> Isn't there a period until the end of xennet_disconnect_backend is
> reached on resume where rx.sring will still point to the old shared
> ring? If so is it safe to drop through in that case?
>
> Would it be safer to add an explicit suspend handler which stopped the
> timer?
Dongxiao, do you have a comment/updated patch for this? I'm going to
revert the smartpoll stuff in the meantime, because its causing tests to
fail.
Thanks,
J
> Ian.
>
>> np->smart_poll.counter++;xennet_disconnect_backend
>>
>> if (likely(netif_carrier_ok(dev))) {
>> --
>> 1.6.3
>>
>>
>>
>> Jeremy Fitzhardinge wrote:
>>> On 07/29/2010 09:33 AM, Ian Jackson wrote:
>>>> Jeremy Fitzhardinge writes ("Re: [Xen-devel] [xen-unstable test]
>>>> 1959: FAIL [and 1 more messages]"):
>>>>> On 07/29/2010 08:30 AM, Ian Jackson wrote:
>>>>>> Is save/restore supposed to work in pvops ? (Using your kernel for
>>>>>> both dom0 and domU.) That would seem to be the next thing to pick
>>>>>> off the list ...
>>>>> Yes. IanC has been tracking down a bug where it fails after a few
>>>>> thousand iterations (now fixed?), but aside from that its
>>>>> apparently OK.
>>>> Well, I was able to reproduce the failure that the automatic test was
>>>> getting. There's a problem with it not getting a copy of the console
>>>> output for some reason but I was able to get this oops from the
>>>> guest.
>>>>
>>>> Just after the restore I was able to connect to the PV console and it
>>>> echoed a few of my CRs (before saving I had logged in on the
>>>> console),
>>>> and then it produced the oops. Now it's apparently completely
>>>> wedged.
>>>>
>>> Ah, it looks like the netfront smartpoll stuff isn't coping with
>>> save/restore.
>>>
>>> J
>>>
>>>> Ian.
>>>>
>>>>
>>>>
>>>> [ 63.681260] BUG: unable to handle kernel NULL pointer dereference
>>>> at 00000010 [ 63.681293] IP: [<c1300381>]
>>>> smart_poll_function+0xbb/0xf2 [ 63.681320] *pdpt =
>>>> 000000001fee5027 *pde = 0000000000000000 [ 63.681344] Oops: 0000
>>>> [#1] SMP [ 63.681362] last sysfs file: /sys/kernel/uevent_seqnum
>>>> [ 63.681376] Modules linked in: [last unloaded: scsi_wait_scan] [
>>>> 63.681398] [ 63.681410] Pid: 5, comm: events/0 Not tainted
>>>> (2.6.32.16 #1) [ 63.681424] EIP: 0061:[<c1300381>] EFLAGS:
>>>> 00010002 CPU: 0 [ 63.681438] EIP is at
>>>> smart_poll_function+0xbb/0xf2 [ 63.681451] EAX: 00000000 EBX:
>>>> dfea8320 ECX: 00000001 EDX: 00000062 [ 63.681465] ESI: 00000064
>>>> EDI: 00000000 EBP: df849cfc ESP: df849cdc [ 63.681479] DS: 007b
>>>> ES: 007b FS: 00d8 GS: 0000 SS: 0069 [ 63.681493] Process events/0
>>>> (pid: 5, ti=df848000 task=df839480 task.ti=df848000) [ 63.681508]
>>>> Stack: [ 63.681516] dfeac828 00000002 dfeac828 dfea8368 dfea0068
>>>> 0000002c dfeac828 c13002c6 [ 63.681555]<0> df849d1c c1079ab8
>>>> df849d48 c2386358 c2386328 0000002c 00000000 c2386328 [
>>>> 63.681598]<0> df849d5c c1079ce8 000028cd 88bbcad9 d3b2cae7 0000000e
>>>> d3b2cae7 0000000e [ 63.681646] Call Trace: [ 63.681662]
>>>> [<c13002c6>] ? smart_poll_function+0x0/0xf2 [ 63.681683]
>>>> [<c1079ab8>] ? __run_hrtimer+0xa9/0xf6 [ 63.681701] [<c1079ce8>]
>>>> ? hrtimer_interrupt+0xcd/0x1c8 [ 63.681719] [<c102cef0>] ?
>>>> xen_timer_interrupt+0x2b/0x224 [ 63.681737] [<c102cb3f>] ?
>>>> xen_force_evtchn_callback+0xf/0x14 [ 63.681755] [<c102d2ac>] ?
>>>> check_events+0x8/0xc [ 63.681776] [<c102d2a3>] ?
>>>> xen_restore_fl_direct_end+0x0/0x1 [ 63.681795] [<c14d1471>] ?
>>>> _spin_unlock_irqrestore+0x2f/0x31 [ 63.681814] [<c105eb9e>] ?
>>>> try_to_wake_up+0x2fa/0x304 [ 63.681832] [<c102cb3f>] ?
>>>> xen_force_evtchn_callback+0xf/0x14 [ 63.681850] [<c10a0cc4>] ?
>>>> handle_IRQ_event+0x5f/0x122 [ 63.681867] [<c10a233f>] ?
>>>> handle_level_irq+0x58/0xa9 [ 63.681886] [<c121983d>] ?
>>>> __xen_evtchn_do_upcall+0xab/0x131 [ 63.681904] [<c1219c71>] ?
>>>> xen_evtchn_do_upcall+0x20/0x30 [ 63.682179] [<c102ffe7>] ?
>>>> xen_do_upcall+0x7/0xc [ 63.682179] [<c102007b>] ?
>>>> apic_reg_write+0xa5/0x52f [ 63.682179] [<c1002227>] ?
>>>> hypercall_page+0x227/0x1005 [ 63.682179] [<c102cb3f>] ?
>>>> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>] ?
>>>> check_events+0x8/0xc [ 63.682179] [<c102d26b>] ?
>>>> xen_irq_enable_direct_end+0x0/0x1 [ 63.682179] [<c105cd0d>] ?
>>>> finish_task_switch+0x4f/0xa6 [ 63.682179] [<c14d0186>] ?
>>>> schedule+0x7dd/0x861 [ 63.682179] [<c106dd9d>] ?
>>>> __mod_timer+0x135/0x140 [ 63.682179] [<c102d2ac>] ?
>>>> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ?
>>>> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ?
>>>> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c107750b>] ?
>>>> prepare_to_wait+0x43/0x48 [ 63.682179] [<c10742a2>] ?
>>>> worker_thread+0x94/0x1d2 [ 63.682179] [<c10ccd2d>] ?
>>>> vmstat_update+0x0/0x2f [ 63.682179] [<c1077357>] ?
>>>> autoremove_wake_function+0x0/0x33 [ 63.682179] [<c107420e>] ?
>>>> worker_thread+0x0/0x1d2 [ 63.682179] [<c1077120>] ?
>>>> kthread+0x61/0x66 [ 63.682179] [<c10770bf>] ? kthread+0x0/0x66 [
>>>> 63.682179] [<c102ff97>] ? kernel_thread_helper+0x7/0x10 [
>>>> 63.682179] Code: c6 89 d0 31 d2 f7 f6 85 d2 75 1a 85 c9 75 0c 8b 83
>>>> 90 08 00 00 c6 40 10 00 eb 32 c7 83 54 45 00 00 00 00 00 00 8b 83 90
>>>> 08 00 00<80> 78 10 00 74 1c 8b 4d e8 b8 00 ca 9a 3b 31 d2 f7 71 44
>>>> 31 c9 [ 63.682179] EIP: [<c1300381>] smart_poll_function+0xbb/0xf2
>>>> SS:ESP 0069:df849cdc [ 63.682179] CR2: 0000000000000010 [
>>>> 63.682179] ---[ end trace 760037e75e5675c8 ]--- [ 63.682179]
>>>> Kernel panic - not syncing: Fatal exception in interrupt [
>>>> 63.682179] Pid: 5, comm: events/0 Tainted: G D 2.6.32.16 #1
>>>> [ 63.682179] Call Trace: [ 63.682179] [<c14cf7b1>] ?
>>>> printk+0xf/0x11 [ 63.682179] [<c14cf6ee>] panic+0x39/0xed [
>>>> 63.682179] [<c14d2085>] oops_end+0xa1/0xb0 [ 63.682179]
>>>> [<c104b426>] no_context+0x137/0x141 [ 63.682179] [<c104b56f>]
>>>> __bad_area_nosemaphore+0x13f/0x147 [ 63.682179] [<c104b584>]
>>>> bad_area_nosemaphore+0xd/0x10 [ 63.682179] [<c14d3275>]
>>>> do_page_fault+0x1c6/0x32b [ 63.682179] [<c14d30af>] ?
>>>> do_page_fault+0x0/0x32b [ 63.682179] [<c14d16c6>]
>>>> error_code+0x66/0x6c [ 63.682179] [<c14d30af>] ?
>>>> do_page_fault+0x0/0x32b [ 63.682179] [<c1300381>] ?
>>>> smart_poll_function+0xbb/0xf2 [ 63.682179] [<c13002c6>] ?
>>>> smart_poll_function+0x0/0xf2 [ 63.682179] [<c1079ab8>]
>>>> __run_hrtimer+0xa9/0xf6 [ 63.682179] [<c1079ce8>]
>>>> hrtimer_interrupt+0xcd/0x1c8 [ 63.682179] [<c102cef0>]
>>>> xen_timer_interrupt+0x2b/0x224 [ 63.682179] [<c102cb3f>] ?
>>>> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>] ?
>>>> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ?
>>>> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ?
>>>> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c105eb9e>] ?
>>>> try_to_wake_up+0x2fa/0x304 [ 63.682179] [<c102cb3f>] ?
>>>> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c10a0cc4>]
>>>> handle_IRQ_event+0x5f/0x122 [ 63.682179] [<c10a233f>]
>>>> handle_level_irq+0x58/0xa9 [ 63.682179] [<c121983d>]
>>>> __xen_evtchn_do_upcall+0xab/0x131 [ 63.682179] [<c1219c71>]
>>>> xen_evtchn_do_upcall+0x20/0x30 [ 63.682179] [<c102ffe7>]
>>>> xen_do_upcall+0x7/0xc [ 63.682179] [<c102007b>] ?
>>>> apic_reg_write+0xa5/0x52f [ 63.682179] [<c1002227>] ?
>>>> hypercall_page+0x227/0x1005 [ 63.682179] [<c102cb3f>] ?
>>>> xen_force_evtchn_callback+0xf/0x14 [ 63.682179] [<c102d2ac>]
>>>> check_events+0x8/0xc [ 63.682179] [<c102d26b>] ?
>>>> xen_irq_enable_direct_end+0x0/0x1 [ 63.682179] [<c105cd0d>] ?
>>>> finish_task_switch+0x4f/0xa6 [ 63.682179] [<c14d0186>]
>>>> schedule+0x7dd/0x861 [ 63.682179] [<c106dd9d>] ?
>>>> __mod_timer+0x135/0x140 [ 63.682179] [<c102d2ac>] ?
>>>> check_events+0x8/0xc [ 63.682179] [<c102d2a3>] ?
>>>> xen_restore_fl_direct_end+0x0/0x1 [ 63.682179] [<c14d1471>] ?
>>>> _spin_unlock_irqrestore+0x2f/0x31 [ 63.682179] [<c107750b>] ?
>>>> prepare_to_wait+0x43/0x48 [ 63.682179] [<c10742a2>]
>>>> worker_thread+0x94/0x1d2 [ 63.682179] [<c10ccd2d>] ?
>>>> vmstat_update+0x0/0x2f [ 63.682179] [<c1077357>] ?
>>>> autoremove_wake_function+0x0/0x33 [ 63.682179] [<c107420e>] ?
>>>> worker_thread+0x0/0x1d2 [ 63.682179] [<c1077120>]
>>>> kthread+0x61/0x66 [ 63.682179] [<c10770bf>] ? kthread+0x0/0x66 [
>>>> 63.682179] [<c102ff97>] kernel_thread_helper+0x7/0x10
>
[-- Attachment #2: 0001-Netfront-Fix-save-restore-after-enabled-smart-poll-f.patch --]
[-- Type: text/plain, Size: 1878 bytes --]
From 791fc8ed6e2888af3bd398f22562776c757cc4ac Mon Sep 17 00:00:00 2001
From: Dongxiao Xu <dongxiao.xu@intel.com>
Date: Wed, 11 Aug 2010 11:06:06 +0800
Subject: [PATCH] Netfront: Fix save/restore after enabled smart poll feature
When s/r the guest, the shared ring will be set to NULL, and
in this case the polling timer interrupt should stop.
This fix includes the two parts:
1) Stop hrtimer when guest suspends.
2) Add check to avoid NULL pointer dereference.
Thanks Ian J for reporting the problem.
Signed-off-by: Dongxiao Xu <dongxiao.xu@intel.com>
---
drivers/net/xen-netfront.c | 13 +++++++++++++
1 files changed, 13 insertions(+), 0 deletions(-)
diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c
index 328fe40..eaea132 100644
--- a/drivers/net/xen-netfront.c
+++ b/drivers/net/xen-netfront.c
@@ -1289,6 +1289,14 @@ static void xennet_disconnect_backend(struct netfront_info *info)
info->rx.sring = NULL;
}
+static int netfront_suspend(struct xenbus_device *dev, pm_message_t state)
+{
+ struct netfront_info *info = dev_get_drvdata(&dev->dev);
+ struct hrtimer *timer = &info->smart_poll.timer;
+ hrtimer_cancel(timer);
+ return 0;
+}
+
/**
* We are reconnecting to the backend, due to a suspend/resume, or a backend
* driver restart. We tear down our netif structure and recreate it, but
@@ -1340,6 +1348,10 @@ static enum hrtimer_restart smart_poll_function(struct hrtimer *timer)
np = netdev_priv(dev);
spin_lock_irqsave(&np->tx_lock, flags);
+
+ if (!np->rx.sring)
+ goto end;
+
np->smart_poll.counter++;
if (likely(netif_carrier_ok(dev))) {
@@ -1910,6 +1922,7 @@ static struct xenbus_driver netfront_driver = {
.ids = netfront_ids,
.probe = netfront_probe,
.remove = __devexit_p(xennet_remove),
+ .suspend = netfront_suspend,
.resume = netfront_resume,
.otherend_changed = netback_changed,
};
--
1.6.3
[-- Attachment #3: Type: text/plain, Size: 138 bytes --]
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xensource.com
http://lists.xensource.com/xen-devel
^ permalink raw reply related [flat|nested] 12+ messages in thread
* RE: [xen-unstable test] 1959: FAIL [and 1 more messages]
2010-08-11 3:19 ` Xu, Dongxiao
@ 2010-08-11 15:01 ` Ian Campbell
2010-08-12 0:37 ` Xu, Dongxiao
0 siblings, 1 reply; 12+ messages in thread
From: Ian Campbell @ 2010-08-11 15:01 UTC (permalink / raw)
To: Xu, Dongxiao; +Cc: Jeremy Fitzhardinge, Xen-devel, Ian Jackson
On Wed, 2010-08-11 at 04:19 +0100, Xu, Dongxiao wrote:
> From 791fc8ed6e2888af3bd398f22562776c757cc4ac Mon Sep 17 00:00:00 2001
> From: Dongxiao Xu <dongxiao.xu@intel.com>
> Date: Wed, 11 Aug 2010 11:06:06 +0800
> Subject: [PATCH] Netfront: Fix save/restore after enabled smart poll
> feature
>
> When s/r the guest, the shared ring will be set to NULL, and
> in this case the polling timer interrupt should stop.
>
> This fix includes the two parts:
> 1) Stop hrtimer when guest suspends.
> 2) Add check to avoid NULL pointer dereference.
Looks good to me but do you really need 2) given 1)?
Ian.
>
> Thanks Ian J for reporting the problem.
>
> Signed-off-by: Dongxiao Xu <dongxiao.xu@intel.com>
> ---
> drivers/net/xen-netfront.c | 13 +++++++++++++
> 1 files changed, 13 insertions(+), 0 deletions(-)
>
> diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c
> index 328fe40..eaea132 100644
> --- a/drivers/net/xen-netfront.c
> +++ b/drivers/net/xen-netfront.c
> @@ -1289,6 +1289,14 @@ static void xennet_disconnect_backend(struct
> netfront_info *info)
> info->rx.sring = NULL;
> }
>
> +static int netfront_suspend(struct xenbus_device *dev, pm_message_t
> state)
> +{
> + struct netfront_info *info = dev_get_drvdata(&dev->dev);
> + struct hrtimer *timer = &info->smart_poll.timer;
> + hrtimer_cancel(timer);
> + return 0;
> +}
> +
> /**
> * We are reconnecting to the backend, due to a suspend/resume, or a
> backend
> * driver restart. We tear down our netif structure and recreate it,
> but
> @@ -1340,6 +1348,10 @@ static enum hrtimer_restart
> smart_poll_function(struct hrtimer *timer)
> np = netdev_priv(dev);
>
> spin_lock_irqsave(&np->tx_lock, flags);
> +
> + if (!np->rx.sring)
> + goto end;
> +
> np->smart_poll.counter++;
>
> if (likely(netif_carrier_ok(dev))) {
> @@ -1910,6 +1922,7 @@ static struct xenbus_driver netfront_driver = {
> .ids = netfront_ids,
> .probe = netfront_probe,
> .remove = __devexit_p(xennet_remove),
> + .suspend = netfront_suspend,
> .resume = netfront_resume,
> .otherend_changed = netback_changed,
> };
^ permalink raw reply [flat|nested] 12+ messages in thread
* RE: [xen-unstable test] 1959: FAIL [and 1 more messages]
2010-08-11 15:01 ` Ian Campbell
@ 2010-08-12 0:37 ` Xu, Dongxiao
2010-08-13 7:16 ` Ian Campbell
0 siblings, 1 reply; 12+ messages in thread
From: Xu, Dongxiao @ 2010-08-12 0:37 UTC (permalink / raw)
To: Ian Campbell; +Cc: Jeremy Fitzhardinge, Xen-devel, Ian Jackson
Hi Ian,
Currently logic (2) is not needed to make s/r work, since sring is set to NULL after resume. However I think this check could avoid error if later any code sets sring to NULL in other places. So I still place it there.
Thanks,
Dongxiao
________________________________________
From: Ian Campbell [Ian.Campbell@eu.citrix.com]
Sent: Wednesday, August 11, 2010 8:01 AM
To: Xu, Dongxiao
Cc: Jeremy Fitzhardinge; Ian Jackson; Xen-devel@lists.xensource.com
Subject: RE: [Xen-devel] [xen-unstable test] 1959: FAIL [and 1 more messages]
On Wed, 2010-08-11 at 04:19 +0100, Xu, Dongxiao wrote:
> From 791fc8ed6e2888af3bd398f22562776c757cc4ac Mon Sep 17 00:00:00 2001
> From: Dongxiao Xu <dongxiao.xu@intel.com>
> Date: Wed, 11 Aug 2010 11:06:06 +0800
> Subject: [PATCH] Netfront: Fix save/restore after enabled smart poll
> feature
>
> When s/r the guest, the shared ring will be set to NULL, and
> in this case the polling timer interrupt should stop.
>
> This fix includes the two parts:
> 1) Stop hrtimer when guest suspends.
> 2) Add check to avoid NULL pointer dereference.
Looks good to me but do you really need 2) given 1)?
Ian.
>
> Thanks Ian J for reporting the problem.
>
> Signed-off-by: Dongxiao Xu <dongxiao.xu@intel.com>
> ---
> drivers/net/xen-netfront.c | 13 +++++++++++++
> 1 files changed, 13 insertions(+), 0 deletions(-)
>
> diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c
> index 328fe40..eaea132 100644
> --- a/drivers/net/xen-netfront.c
> +++ b/drivers/net/xen-netfront.c
> @@ -1289,6 +1289,14 @@ static void xennet_disconnect_backend(struct
> netfront_info *info)
> info->rx.sring = NULL;
> }
>
> +static int netfront_suspend(struct xenbus_device *dev, pm_message_t
> state)
> +{
> + struct netfront_info *info = dev_get_drvdata(&dev->dev);
> + struct hrtimer *timer = &info->smart_poll.timer;
> + hrtimer_cancel(timer);
> + return 0;
> +}
> +
> /**
> * We are reconnecting to the backend, due to a suspend/resume, or a
> backend
> * driver restart. We tear down our netif structure and recreate it,
> but
> @@ -1340,6 +1348,10 @@ static enum hrtimer_restart
> smart_poll_function(struct hrtimer *timer)
> np = netdev_priv(dev);
>
> spin_lock_irqsave(&np->tx_lock, flags);
> +
> + if (!np->rx.sring)
> + goto end;
> +
> np->smart_poll.counter++;
>
> if (likely(netif_carrier_ok(dev))) {
> @@ -1910,6 +1922,7 @@ static struct xenbus_driver netfront_driver = {
> .ids = netfront_ids,
> .probe = netfront_probe,
> .remove = __devexit_p(xennet_remove),
> + .suspend = netfront_suspend,
> .resume = netfront_resume,
> .otherend_changed = netback_changed,
> };
^ permalink raw reply [flat|nested] 12+ messages in thread
* RE: [xen-unstable test] 1959: FAIL [and 1 more messages]
2010-08-12 0:37 ` Xu, Dongxiao
@ 2010-08-13 7:16 ` Ian Campbell
0 siblings, 0 replies; 12+ messages in thread
From: Ian Campbell @ 2010-08-13 7:16 UTC (permalink / raw)
To: Xu, Dongxiao; +Cc: Jeremy Fitzhardinge, Xen-devel, Ian Jackson
On Thu, 2010-08-12 at 01:37 +0100, Xu, Dongxiao wrote:
> Hi Ian,
>
> Currently logic (2) is not needed to make s/r work, since sring is set
> to NULL after resume. However I think this check could avoid error if
> later any code sets sring to NULL in other places. So I still place it
> there.
OK, that's fair enough.
Ian.
>
> Thanks,
> Dongxiao
>
> ________________________________________
> From: Ian Campbell [Ian.Campbell@eu.citrix.com]
> Sent: Wednesday, August 11, 2010 8:01 AM
> To: Xu, Dongxiao
> Cc: Jeremy Fitzhardinge; Ian Jackson; Xen-devel@lists.xensource.com
> Subject: RE: [Xen-devel] [xen-unstable test] 1959: FAIL [and 1 more messages]
>
> On Wed, 2010-08-11 at 04:19 +0100, Xu, Dongxiao wrote:
> > From 791fc8ed6e2888af3bd398f22562776c757cc4ac Mon Sep 17 00:00:00 2001
> > From: Dongxiao Xu <dongxiao.xu@intel.com>
> > Date: Wed, 11 Aug 2010 11:06:06 +0800
> > Subject: [PATCH] Netfront: Fix save/restore after enabled smart poll
> > feature
> >
> > When s/r the guest, the shared ring will be set to NULL, and
> > in this case the polling timer interrupt should stop.
> >
> > This fix includes the two parts:
> > 1) Stop hrtimer when guest suspends.
> > 2) Add check to avoid NULL pointer dereference.
>
> Looks good to me but do you really need 2) given 1)?
>
> Ian.
>
>
> >
> > Thanks Ian J for reporting the problem.
> >
> > Signed-off-by: Dongxiao Xu <dongxiao.xu@intel.com>
> > ---
> > drivers/net/xen-netfront.c | 13 +++++++++++++
> > 1 files changed, 13 insertions(+), 0 deletions(-)
> >
> > diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c
> > index 328fe40..eaea132 100644
> > --- a/drivers/net/xen-netfront.c
> > +++ b/drivers/net/xen-netfront.c
> > @@ -1289,6 +1289,14 @@ static void xennet_disconnect_backend(struct
> > netfront_info *info)
> > info->rx.sring = NULL;
> > }
> >
> > +static int netfront_suspend(struct xenbus_device *dev, pm_message_t
> > state)
> > +{
> > + struct netfront_info *info = dev_get_drvdata(&dev->dev);
> > + struct hrtimer *timer = &info->smart_poll.timer;
> > + hrtimer_cancel(timer);
> > + return 0;
> > +}
> > +
> > /**
> > * We are reconnecting to the backend, due to a suspend/resume, or a
> > backend
> > * driver restart. We tear down our netif structure and recreate it,
> > but
> > @@ -1340,6 +1348,10 @@ static enum hrtimer_restart
> > smart_poll_function(struct hrtimer *timer)
> > np = netdev_priv(dev);
> >
> > spin_lock_irqsave(&np->tx_lock, flags);
> > +
> > + if (!np->rx.sring)
> > + goto end;
> > +
> > np->smart_poll.counter++;
> >
> > if (likely(netif_carrier_ok(dev))) {
> > @@ -1910,6 +1922,7 @@ static struct xenbus_driver netfront_driver = {
> > .ids = netfront_ids,
> > .probe = netfront_probe,
> > .remove = __devexit_p(xennet_remove),
> > + .suspend = netfront_suspend,
> > .resume = netfront_resume,
> > .otherend_changed = netback_changed,
> > };
^ permalink raw reply [flat|nested] 12+ messages in thread
end of thread, other threads:[~2010-08-13 7:16 UTC | newest]
Thread overview: 12+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2010-07-29 8:50 [xen-unstable test] 1970: tolerable FAIL - PUSHED xen.org
2010-07-24 8:48 ` [xen-unstable test] 1959: FAIL xen.org
[not found] ` <4C4B1B8A.6070403@goop.org>
[not found] ` <19533.23953.645991.205290@mariner.uk.xensource.com>
[not found] ` <4C4DC4AF.2090901@goop.org>
[not found] ` <19535.145.995122.44883@mariner.uk.xensource.com>
[not found] ` <4C4F063C.9030301@goop.org>
[not found] ` <19535.2235.492925.965460@mariner.uk.xensource.com>
[not found] ` <4C4F0B18.5080800@goop.org>
[not found] ` <19535.3062.112242.870153@mariner.uk.xensource.com>
[not found] ` <4C4F0FDB.2010607@goop.org>
[not found] ` <19537.40568.143464.400709@mariner.uk.xensource.com>
[not found] ` <4C51A084.5010903@goop.org>
[not found] ` <19537.44367.380875.758088@mariner.uk.xensource.com>
2010-07-29 16:38 ` [xen-unstable test] 1959: FAIL [and 1 more messages] Jeremy Fitzhardinge
2010-07-29 17:59 ` Ian Jackson
2010-07-30 4:58 ` Xu, Dongxiao
2010-07-30 9:18 ` Ian Campbell
2010-07-30 14:43 ` Jeremy Fitzhardinge
2010-08-10 17:38 ` Jeremy Fitzhardinge
2010-08-11 3:19 ` Xu, Dongxiao
2010-08-11 15:01 ` Ian Campbell
2010-08-12 0:37 ` Xu, Dongxiao
2010-08-13 7:16 ` Ian Campbell
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.