All of lore.kernel.org
 help / color / mirror / Atom feed
* xfs_repair - segemnation fault | git error
@ 2012-08-10  8:23 Velimir Galic
  2012-08-10 22:53 ` Dave Chinner
  2012-08-11 10:57 ` Michael Monnerie
  0 siblings, 2 replies; 4+ messages in thread
From: Velimir Galic @ 2012-08-10  8:23 UTC (permalink / raw)
  To: xfs


[-- Attachment #1.1: Type: text/plain, Size: 1410 bytes --]

hi Guys,

I hope you are able to help me, because i think that i
tried everything that i know off to repair my raid after a disk failure.

Systemspecs:

   - HP N40L
      - CPU: AMD Turion II @ 1,5GHZ
      - 8GB ECC RAM
      - Disks:
         - 250GB | OS
         - 4x2GB | RAID 5 | Softraid
      - OS: Openmediavault @ 3.0.20 (last version)


Problem:

In my raid a disk was fault so mdadm kicked the disk out of the raid. I put
a replacement in the box and started a resync of the raid and than
is happened, a second disk got lost in the raid. The raidsync wasn't
finished and now i'm fighting 3 weeks or more i don't no any more :-( to
repair the file system.

I tried a lot of different versions of xfs_repair (2.9.4, 2.9.8, 2.10.2,
3.0.4, 3.1.2, 3.1.6, 3.1,8), but every precompiled version got an
"segmentation fault" or hung at phase 3 with 100% cpu load for few day. I
also tried different distros like ubuntu, debian, redhat, gentoo (yes, i'm
a little bit desperate :-) )  Than i tried the git version and at first it
looked good but than i got an another error but i don't understand what the
problem is.

tried options without any luck: xfs_repair -v -P | xfs_repair -v -P -m 6144
| xfs_repair -v -P -L

Metadump of the filesystem isn't possible at this moment, but i could
bootup a live cd and give it a try with it. I'm pretty stuck and don't no
what to do any more.


with kind regards
Veli

[-- Attachment #1.2: Type: text/html, Size: 1789 bytes --]

[-- Attachment #2: md127.metadump.err --]
[-- Type: application/octet-stream, Size: 3536 bytes --]

root@hpbox:~# xfs_metadump /dev/md127 /media/localhdd/md127.metadump
*** glibc detected *** xfs_db: double free or corruption (!prev): 0x0000000001ccb000 ***
======= Backtrace: =========
/lib/libc.so.6(+0x71bd6)[0x7f5e34b55bd6]
/lib/libc.so.6(cfree+0x6c)[0x7f5e34b5a94c]
xfs_db[0x41a8e3]
xfs_db[0x41c3ec]
xfs_db[0x41e2bc]
xfs_db[0x41c152]
xfs_db[0x41dfd4]
xfs_db[0x41c152]
xfs_db[0x41eedb]
xfs_db[0x418d08]
/lib/libc.so.6(__libc_start_main+0xfd)[0x7f5e34b02c8d]
xfs_db[0x4027d9]
======= Memory map: ========
00400000-00475000 r-xp 00000000 08:41 2338                               /usr/sbin/xfs_db
00675000-00676000 rw-p 00075000 08:41 2338                               /usr/sbin/xfs_db
00676000-00689000 rw-p 00000000 00:00 0 
01c83000-01f1d000 rw-p 00000000 00:00 0                                  [heap]
7f5e30000000-7f5e30021000 rw-p 00000000 00:00 0 
7f5e30021000-7f5e34000000 ---p 00000000 00:00 0 
7f5e348c7000-7f5e348dd000 r-xp 00000000 08:41 21105                      /lib/libgcc_s.so.1
7f5e348dd000-7f5e34adc000 ---p 00016000 08:41 21105                      /lib/libgcc_s.so.1
7f5e34adc000-7f5e34add000 rw-p 00015000 08:41 21105                      /lib/libgcc_s.so.1
7f5e34ae4000-7f5e34c3d000 r-xp 00000000 08:41 3222                       /lib/libc-2.11.3.so
7f5e34c3d000-7f5e34e3c000 ---p 00159000 08:41 3222                       /lib/libc-2.11.3.so
7f5e34e3c000-7f5e34e40000 r--p 00158000 08:41 3222                       /lib/libc-2.11.3.so
7f5e34e40000-7f5e34e41000 rw-p 0015c000 08:41 3222                       /lib/libc-2.11.3.so
7f5e34e41000-7f5e34e46000 rw-p 00000000 00:00 0 
7f5e34e46000-7f5e34e5d000 r-xp 00000000 08:41 3217                       /lib/libpthread-2.11.3.so
7f5e34e5d000-7f5e3505c000 ---p 00017000 08:41 3217                       /lib/libpthread-2.11.3.so
7f5e3505c000-7f5e3505d000 r--p 00016000 08:41 3217                       /lib/libpthread-2.11.3.so
7f5e3505d000-7f5e3505e000 rw-p 00017000 08:41 3217                       /lib/libpthread-2.11.3.so
7f5e3505e000-7f5e35062000 rw-p 00000000 00:00 0 
7f5e35062000-7f5e35069000 r-xp 00000000 08:41 3219                       /lib/librt-2.11.3.so
7f5e35069000-7f5e35268000 ---p 00007000 08:41 3219                       /lib/librt-2.11.3.so
7f5e35268000-7f5e35269000 r--p 00006000 08:41 3219                       /lib/librt-2.11.3.so
7f5e35269000-7f5e3526a000 rw-p 00007000 08:41 3219                       /lib/librt-2.11.3.so
7f5e3526a000-7f5e3526e000 r-xp 00000000 08:41 21195                      /lib/libuuid.so.1.3.0
7f5e3526e000-7f5e3546d000 ---p 00004000 08:41 21195                      /lib/libuuid.so.1.3.0
7f5e3546d000-7f5e3546e000 rw-p 00003000 08:41 21195                      /lib/libuuid.so.1.3.0
7f5e3546e000-7f5e3548c000 r-xp 00000000 08:41 3218                       /lib/ld-2.11.3.so
7f5e354c7000-7f5e35509000 rw-p 00000000 00:00 0 
7f5e35509000-7f5e3567e000 r--p 00000000 08:41 2104                       /usr/lib/locale/locale-archive
7f5e3567e000-7f5e35682000 rw-p 00000000 00:00 0 
7f5e35689000-7f5e3568b000 rw-p 00000000 00:00 0 
7f5e3568b000-7f5e3568c000 r--p 0001d000 08:41 3218                       /lib/ld-2.11.3.so
7f5e3568c000-7f5e3568d000 rw-p 0001e000 08:41 3218                       /lib/ld-2.11.3.so
7f5e3568d000-7f5e3568e000 rw-p 00000000 00:00 0 
7fff33b52000-7fff33b73000 rw-p 00000000 00:00 0                          [stack]
7fff33bff000-7fff33c00000 r-xp 00000000 00:00 0                          [vdso]
ffffffffff600000-ffffffffff601000 r-xp 00000000 00:00 0                  [vsyscall]
Aborted

[-- Attachment #3: Type: text/plain, Size: 121 bytes --]

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: xfs_repair - segemnation fault | git error
  2012-08-10  8:23 xfs_repair - segemnation fault | git error Velimir Galic
@ 2012-08-10 22:53 ` Dave Chinner
  2012-08-11 10:57 ` Michael Monnerie
  1 sibling, 0 replies; 4+ messages in thread
From: Dave Chinner @ 2012-08-10 22:53 UTC (permalink / raw)
  To: Velimir Galic; +Cc: xfs

On Fri, Aug 10, 2012 at 10:23:03AM +0200, Velimir Galic wrote:
> hi Guys,
> 
> I hope you are able to help me, because i think that i
> tried everything that i know off to repair my raid after a disk failure.
> 
> Systemspecs:
> 
>    - HP N40L
>       - CPU: AMD Turion II @ 1,5GHZ
>       - 8GB ECC RAM
>       - Disks:
>          - 250GB | OS
>          - 4x2GB | RAID 5 | Softraid
>       - OS: Openmediavault @ 3.0.20 (last version)
> 
> 
> Problem:
> 
> In my raid a disk was fault so mdadm kicked the disk out of the raid. I put
> a replacement in the box and started a resync of the raid and than
> is happened, a second disk got lost in the raid. The raidsync wasn't

Ok, so you've lost a good chunk of the filesystem then? I don't lik
eto say it, but double disk failures tend to result in unrecoverable
data loss. Given the contents of this email, I know you don't have
backups to restore from, so you might be left with only bits and
pieces even if a repair can be made to run successfully.

> finished and now i'm fighting 3 weeks or more i don't no any more :-( to
> repair the file system.
> 
> I tried a lot of different versions of xfs_repair (2.9.4, 2.9.8, 2.10.2,
> 3.0.4, 3.1.2, 3.1.6, 3.1,8), but every precompiled version got an
> "segmentation fault" or hung at phase 3 with 100% cpu load for few day. I
> also tried different distros like ubuntu, debian, redhat, gentoo (yes, i'm
> a little bit desperate :-) )  Than i tried the git version and at first it
> looked good but than i got an another error but i don't understand what the
> problem is.

Perhaps you should tell us whatthe error is. i.e. attach the output
of xfs_repair when it fails.

> 
> tried options without any luck: xfs_repair -v -P | xfs_repair -v -P -m 6144
> | xfs_repair -v -P -L
> 
> Metadump of the filesystem isn't possible at this moment,

Why not?

> but i could
> bootup a live cd and give it a try with it. I'm pretty stuck and don't no
> what to do any more.

Run your compiled, unstripped xfs-repair binary under gdb and when
it segfaults, dump of the stack trace. That will at least tell use
where it is failing....

Cheers,

Dave.
-- 
Dave Chinner
david@fromorbit.com

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: xfs_repair - segemnation fault | git error
  2012-08-10  8:23 xfs_repair - segemnation fault | git error Velimir Galic
  2012-08-10 22:53 ` Dave Chinner
@ 2012-08-11 10:57 ` Michael Monnerie
  2012-08-13 11:58   ` Velimir Galic
  1 sibling, 1 reply; 4+ messages in thread
From: Michael Monnerie @ 2012-08-11 10:57 UTC (permalink / raw)
  To: xfs; +Cc: Velimir Galic


[-- Attachment #1.1: Type: text/plain, Size: 878 bytes --]

Am Freitag, 10. August 2012, 10:23:03 schrieb Velimir Galic:
> a second disk got lost in the raid. The raidsync wasn't finished and
> now i'm fighting 3 weeks or more i don't no any more :-( to repair
> the file system. 

How far did the recovery go when it broke? If it was on 10% when the 2nd 
disk broke, you can stop to try to get anything. If it was 90% finished, 
you might have a chance...

> 4x2GB | RAID 5 | Softraid

This is a 6GB net RAID 5 device? Which devices?

> I tried a lot of different versions of xfs_repair

I hope you got a copy of all disks before you even tried to repair? 
Otherwise it might be too late now, when repair wrote something on that 
device.

-- 
mit freundlichen Grüssen,
Michael Monnerie, Ing. BSc

it-management Internet Services: Protéger
http://proteger.at [gesprochen: Prot-e-schee]
Tel: +43 660 / 415 6531

[-- Attachment #1.2: This is a digitally signed message part. --]
[-- Type: application/pgp-signature, Size: 198 bytes --]

[-- Attachment #2: Type: text/plain, Size: 121 bytes --]

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: xfs_repair - segemnation fault | git error
  2012-08-11 10:57 ` Michael Monnerie
@ 2012-08-13 11:58   ` Velimir Galic
  0 siblings, 0 replies; 4+ messages in thread
From: Velimir Galic @ 2012-08-13 11:58 UTC (permalink / raw)
  To: Michael Monnerie; +Cc: xfs


[-- Attachment #1.1: Type: text/plain, Size: 1256 bytes --]

hi,

Thanks for the fast replay and the effort!

2012/8/11 Michael Monnerie <michael.monnerie@is.it-management.at>

> Am Freitag, 10. August 2012, 10:23:03 schrieb Velimir Galic:
> > a second disk got lost in the raid. The raidsync wasn't finished and
> > now i'm fighting 3 weeks or more i don't no any more :-( to repair
> > the file system.
>
> How far did the recovery go when it broke? If it was on 10% when the 2nd
> disk broke, you can stop to try to get anything. If it was 90% finished,
> you might have a chance...
>

that's what I thought but hoped for that isn't a problem. The sync failed
at approx. 20% or beneath.


> > 4x2GB | RAID 5 | Softraid
>
> This is a 6GB net RAID 5 device? Which devices?
>

WDC WD20EARS-22MVWB0


>
> > I tried a lot of different versions of xfs_repair
>
> I hope you got a copy of all disks before you even tried to repair?
> Otherwise it might be too late now, when repair wrote something on that
> device.
>
>
Nope, I wasn't clever enough to do that, but it isn't hopeless because I
have a
3 months old backup form my important files. So I have to life with the 3
months lost, I hope is not to much.


Thanks for trying to help, but I stop her with trying to recover anything
and start over.

Shit happens! :-)

[-- Attachment #1.2: Type: text/html, Size: 2155 bytes --]

[-- Attachment #2: Type: text/plain, Size: 121 bytes --]

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2012-08-13 11:58 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2012-08-10  8:23 xfs_repair - segemnation fault | git error Velimir Galic
2012-08-10 22:53 ` Dave Chinner
2012-08-11 10:57 ` Michael Monnerie
2012-08-13 11:58   ` Velimir Galic

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.