All of lore.kernel.org
 help / color / mirror / Atom feed
* corrupt xfs log
@ 2017-08-18 11:56 Ingard -
  2017-08-18 12:02 ` Bill O'Donnell
  0 siblings, 1 reply; 16+ messages in thread
From: Ingard - @ 2017-08-18 11:56 UTC (permalink / raw)
  To: linux-xfs

After a server crash we've encountered a corrupt xfs filesystem. When
trying to mount said filesystem normally the system hangs.
This was initially on a ubuntu trusty server with 3.13 kernel with
xfsprogs 3.1.9

We've installed a newer kernel (4.4.0-92) and compiled xfsprogs v
4.12.0 from source. We're still not able to mount the filesystem (and
replay the log) normally.
We are able to mount it -o ro,norecovery, but we're reluctant to do
xfs_repair -L without trying everything we can first. The filesystem
is browsable albeit a few paths which gives an error : "Structure
needs cleaning"

Does anyone have any advice as to how we might recover/repair the
corrupt log so we can replay it? Or is xfs_repair -L the only way
forward?


Excerpt from kern.log:
2017-08-17T13:40:41.122121+02:00 dn-238 kernel: [  294.300347] XFS
(sdd1): Mounting V4 filesystem in no-recovery mode. Filesystem will be
inconsistent.

2017-08-17T17:04:54.794194+02:00 dn-238 kernel: [12548.400260] XFS
(sdd1): Metadata corruption detected at xfs_inode_buf_verify+0x6f/0xd0
[xfs], xfs_inode block 0x81c9c210
2017-08-17T17:04:54.794216+02:00 dn-238 kernel: [12548.400342] XFS
(sdd1): Unmount and run xfs_repair
2017-08-17T17:04:54.794218+02:00 dn-238 kernel: [12548.400374] XFS
(sdd1): First 64 bytes of corrupted metadata buffer:
2017-08-17T17:04:54.794220+02:00 dn-238 kernel: [12548.400418]
ffff880171fff000: 3f 1a 33 54 5b 55 85 0b 7c f5 c6 d5 cf 51 47 41
?.3T[U..|....QGA
2017-08-17T17:04:54.794222+02:00 dn-238 kernel: [12548.400473]
ffff880171fff010: 97 ba ba 03 5c e4 02 7a e6 bc fb 5d f1 72 db c1
....\..z...].r..
2017-08-17T17:04:54.794223+02:00 dn-238 kernel: [12548.400527]
ffff880171fff020: c8 ad 3a 76 c7 e4 20 92 88 a2 35 0c 1f 36 cf b5
..:v.. ...5..6..
2017-08-17T17:04:54.794226+02:00 dn-238 kernel: [12548.400581]
ffff880171fff030: 8a bc 42 75 86 50 a0 a2 be 2c 2d 99 96 2d e1 ee
..Bu.P...,-..-..

kind regards
ingard

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: corrupt xfs log
  2017-08-18 11:56 corrupt xfs log Ingard -
@ 2017-08-18 12:02 ` Bill O'Donnell
  2017-08-18 12:17   ` Brian Foster
  2017-08-18 13:43   ` Ingard -
  0 siblings, 2 replies; 16+ messages in thread
From: Bill O'Donnell @ 2017-08-18 12:02 UTC (permalink / raw)
  To: Ingard -; +Cc: linux-xfs

On Fri, Aug 18, 2017 at 01:56:31PM +0200, Ingard - wrote:
> After a server crash we've encountered a corrupt xfs filesystem. When
> trying to mount said filesystem normally the system hangs.
> This was initially on a ubuntu trusty server with 3.13 kernel with
> xfsprogs 3.1.9
> 
> We've installed a newer kernel (4.4.0-92) and compiled xfsprogs v
> 4.12.0 from source. We're still not able to mount the filesystem (and
> replay the log) normally.
> We are able to mount it -o ro,norecovery, but we're reluctant to do
> xfs_repair -L without trying everything we can first. The filesystem
> is browsable albeit a few paths which gives an error : "Structure
> needs cleaning"
> 
> Does anyone have any advice as to how we might recover/repair the
> corrupt log so we can replay it? Or is xfs_repair -L the only way
> forward?

Can you try xfs_repair -n (only scans the fs and reports what repairs
would be made)?

Thanks-
Bill


> 
> 
> Excerpt from kern.log:
> 2017-08-17T13:40:41.122121+02:00 dn-238 kernel: [  294.300347] XFS
> (sdd1): Mounting V4 filesystem in no-recovery mode. Filesystem will be
> inconsistent.
> 
> 2017-08-17T17:04:54.794194+02:00 dn-238 kernel: [12548.400260] XFS
> (sdd1): Metadata corruption detected at xfs_inode_buf_verify+0x6f/0xd0
> [xfs], xfs_inode block 0x81c9c210
> 2017-08-17T17:04:54.794216+02:00 dn-238 kernel: [12548.400342] XFS
> (sdd1): Unmount and run xfs_repair
> 2017-08-17T17:04:54.794218+02:00 dn-238 kernel: [12548.400374] XFS
> (sdd1): First 64 bytes of corrupted metadata buffer:
> 2017-08-17T17:04:54.794220+02:00 dn-238 kernel: [12548.400418]
> ffff880171fff000: 3f 1a 33 54 5b 55 85 0b 7c f5 c6 d5 cf 51 47 41
> ?.3T[U..|....QGA
> 2017-08-17T17:04:54.794222+02:00 dn-238 kernel: [12548.400473]
> ffff880171fff010: 97 ba ba 03 5c e4 02 7a e6 bc fb 5d f1 72 db c1
> ....\..z...].r..
> 2017-08-17T17:04:54.794223+02:00 dn-238 kernel: [12548.400527]
> ffff880171fff020: c8 ad 3a 76 c7 e4 20 92 88 a2 35 0c 1f 36 cf b5
> ..:v.. ...5..6..
> 2017-08-17T17:04:54.794226+02:00 dn-238 kernel: [12548.400581]
> ffff880171fff030: 8a bc 42 75 86 50 a0 a2 be 2c 2d 99 96 2d e1 ee
> ..Bu.P...,-..-..
> 
> kind regards
> ingard
> --
> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: corrupt xfs log
  2017-08-18 12:02 ` Bill O'Donnell
@ 2017-08-18 12:17   ` Brian Foster
  2017-08-21 12:08     ` Ingard -
  2017-08-18 13:43   ` Ingard -
  1 sibling, 1 reply; 16+ messages in thread
From: Brian Foster @ 2017-08-18 12:17 UTC (permalink / raw)
  To: Bill O'Donnell; +Cc: Ingard -, linux-xfs

On Fri, Aug 18, 2017 at 07:02:24AM -0500, Bill O'Donnell wrote:
> On Fri, Aug 18, 2017 at 01:56:31PM +0200, Ingard - wrote:
> > After a server crash we've encountered a corrupt xfs filesystem. When
> > trying to mount said filesystem normally the system hangs.
> > This was initially on a ubuntu trusty server with 3.13 kernel with
> > xfsprogs 3.1.9
> > 
> > We've installed a newer kernel (4.4.0-92) and compiled xfsprogs v
> > 4.12.0 from source. We're still not able to mount the filesystem (and
> > replay the log) normally.
> > We are able to mount it -o ro,norecovery, but we're reluctant to do
> > xfs_repair -L without trying everything we can first. The filesystem
> > is browsable albeit a few paths which gives an error : "Structure
> > needs cleaning"
> > 
> > Does anyone have any advice as to how we might recover/repair the
> > corrupt log so we can replay it? Or is xfs_repair -L the only way
> > forward?
> 
> Can you try xfs_repair -n (only scans the fs and reports what repairs
> would be made)?
> 

An xfs_metadump of the fs might be useful as well. Then we can see if we
can reproduce the mount hang on latest kernels and if so, potentially
try and root cause it.

Brian

> Thanks-
> Bill
> 
> 
> > 
> > 
> > Excerpt from kern.log:
> > 2017-08-17T13:40:41.122121+02:00 dn-238 kernel: [  294.300347] XFS
> > (sdd1): Mounting V4 filesystem in no-recovery mode. Filesystem will be
> > inconsistent.
> > 
> > 2017-08-17T17:04:54.794194+02:00 dn-238 kernel: [12548.400260] XFS
> > (sdd1): Metadata corruption detected at xfs_inode_buf_verify+0x6f/0xd0
> > [xfs], xfs_inode block 0x81c9c210
> > 2017-08-17T17:04:54.794216+02:00 dn-238 kernel: [12548.400342] XFS
> > (sdd1): Unmount and run xfs_repair
> > 2017-08-17T17:04:54.794218+02:00 dn-238 kernel: [12548.400374] XFS
> > (sdd1): First 64 bytes of corrupted metadata buffer:
> > 2017-08-17T17:04:54.794220+02:00 dn-238 kernel: [12548.400418]
> > ffff880171fff000: 3f 1a 33 54 5b 55 85 0b 7c f5 c6 d5 cf 51 47 41
> > ?.3T[U..|....QGA
> > 2017-08-17T17:04:54.794222+02:00 dn-238 kernel: [12548.400473]
> > ffff880171fff010: 97 ba ba 03 5c e4 02 7a e6 bc fb 5d f1 72 db c1
> > ....\..z...].r..
> > 2017-08-17T17:04:54.794223+02:00 dn-238 kernel: [12548.400527]
> > ffff880171fff020: c8 ad 3a 76 c7 e4 20 92 88 a2 35 0c 1f 36 cf b5
> > ..:v.. ...5..6..
> > 2017-08-17T17:04:54.794226+02:00 dn-238 kernel: [12548.400581]
> > ffff880171fff030: 8a bc 42 75 86 50 a0 a2 be 2c 2d 99 96 2d e1 ee
> > ..Bu.P...,-..-..
> > 
> > kind regards
> > ingard
> > --
> > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> > the body of a message to majordomo@vger.kernel.org
> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: corrupt xfs log
  2017-08-18 12:02 ` Bill O'Donnell
  2017-08-18 12:17   ` Brian Foster
@ 2017-08-18 13:43   ` Ingard -
  1 sibling, 0 replies; 16+ messages in thread
From: Ingard - @ 2017-08-18 13:43 UTC (permalink / raw)
  To: Bill O'Donnell; +Cc: linux-xfs

[-- Attachment #1: Type: text/plain, Size: 35 bytes --]

output from xfs_repair -n attached

[-- Attachment #2: xfs_repair.txt --]
[-- Type: text/plain, Size: 44810 bytes --]

$ xfs_repair -n /dev/sdd1
Phase 1 - find and verify superblock...
        - reporting progress in intervals of 15 minutes
Phase 2 - using internal log
        - zero log...
ALERT: The filesystem has valuable metadata changes in a log which is being
ignored because the -n option was used.  Expect spurious inconsistencies
which may be resolved by first mounting the filesystem to replay the log.
        - scan filesystem freespace and inode maps...
Metadata corruption detected at xfs_allocbt block 0xb000cbff8/0x1000
btree block 22/105855 is suspect, error -117
bad magic # 0x530939a7 in btbno block 22/105855
agf_btreeblks 390, counted 389 in ag 22
agi unlinked bucket 60 is 298941500 in ag 14 (inode=60428483644)
Metadata corruption detected at xfs_allocbt block 0x29000059e0/0x1000
btree block 82/8124 is suspect, error -117
bad magic # 0xd36c598e in btbno block 82/8124
agf_btreeblks 403, counted 402 in ag 82
sb_icount 64, counted 79473472
sb_ifree 61, counted 2941
sb_fdblocks 23439345295, counted 1093782972
        - 14:09:48: scanning filesystem freespace - 88 of 88 allocation groups done
        - found root inode chunk
Phase 3 - for each AG...
        - scan (but don't clear) agi unlinked lists...
        - 14:09:48: scanning agi unlinked lists - 88 of 88 allocation groups done
        - process known inodes and perform inode discovery...
        - agno = 0
        - agno = 15
        - agno = 60
        - agno = 45
        - agno = 75
        - agno = 30
^C
14:09:54 [root@dn-238[DP]:/mnt] $
14:09:54 [root@dn-238[DP]:/mnt] $
14:09:54 [root@dn-238[DP]:/mnt] $
14:09:54 [root@dn-238[DP]:/mnt] $
14:09:55 [root@dn-238[DP]:/mnt] $
14:09:55 [root@dn-238[DP]:/mnt] $ xfs_repair -n /dev/sdd1 |tee xfsrepair.txt
Phase 1 - find and verify superblock...
        - reporting progress in intervals of 15 minutes
Phase 2 - using internal log
        - zero log...
ALERT: The filesystem has valuable metadata changes in a log which is being
ignored because the -n option was used.  Expect spurious inconsistencies
which may be resolved by first mounting the filesystem to replay the log.
        - scan filesystem freespace and inode maps...
Metadata corruption detected at xfs_allocbt block 0xb000cbff8/0x1000
btree block 22/105855 is suspect, error -117
bad magic # 0x530939a7 in btbno block 22/105855
agf_btreeblks 390, counted 389 in ag 22
agi unlinked bucket 60 is 298941500 in ag 14 (inode=60428483644)
Metadata corruption detected at xfs_allocbt block 0x29000059e0/0x1000
btree block 82/8124 is suspect, error -117
bad magic # 0xd36c598e in btbno block 82/8124
agf_btreeblks 403, counted 402 in ag 82
sb_icount 64, counted 79473472
sb_ifree 61, counted 2941
sb_fdblocks 23439345295, counted 1093782972
        - 14:10:30: scanning filesystem freespace - 88 of 88 allocation groups done
        - found root inode chunk
Phase 3 - for each AG...
        - scan (but don't clear) agi unlinked lists...
        - 14:10:30: scanning agi unlinked lists - 88 of 88 allocation groups done
        - process known inodes and perform inode discovery...
        - agno = 0
        - agno = 15
        - agno = 30
        - agno = 45
        - agno = 60
        - agno = 75
        - agno = 31
        - agno = 46
        - agno = 61
        - agno = 16
        - agno = 76
        - agno = 1
        - agno = 47
        - agno = 32
        - agno = 17
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
        - agno = 62
        - agno = 77
bad magic number 0x3f1a on inode 4354967584
bad version number 0x5b on inode 4354967584
bad (negative) size -4743366169855139346 on inode 4354967584
bad magic number 0xf749 on inode 4354967585
bad version number 0x19 on inode 4354967585
bad (negative) size -4925401835952624351 on inode 4354967585
bad magic number 0xefc6 on inode 4354967586
bad version number 0xffffffc0 on inode 4354967586
bad (negative) size -6663211212969851433 on inode 4354967586
bad magic number 0x93db on inode 4354967587
bad version number 0x5f on inode 4354967587
bad inode format in inode 4354967587
bad magic number 0x54da on inode 4354967588
bad version number 0xffffff86 on inode 4354967588
bad inode format in inode 4354967588
bad magic number 0x5893 on inode 4354967589
bad version number 0x0 on inode 4354967589
bad inode format in inode 4354967589
bad magic number 0x309f on inode 4354967590
bad version number 0xffffffc2 on inode 4354967590
bad (negative) size -7620939567173939239 on inode 4354967590
bad magic number 0x9fce on inode 4354967591
bad version number 0x38 on inode 4354967591
bad inode format in inode 4354967591
bad magic number 0xb59f on inode 4354967592
bad version number 0x4 on inode 4354967592
bad inode format in inode 4354967592
bad magic number 0x53b4 on inode 4354967593
bad version number 0xfffffff4 on inode 4354967593
bad (negative) size -6181984635646857242 on inode 4354967593
bad magic number 0x7c77 on inode 4354967594
bad version number 0x3e on inode 4354967594
bad (negative) size -5208493334119488045 on inode 4354967594
bad magic number 0x6698 on inode 4354967595
bad version number 0x51 on inode 4354967595
bad inode format in inode 4354967595
bad magic number 0xe872 on inode 4354967596
bad version number 0xffffffa1 on inode 4354967596
bad (negative) size -4217771995784516902 on inode 4354967596
bad magic number 0x2c98 on inode 4354967597
bad version number 0xffffff8e on inode 4354967597
bad (negative) size -1829068493719510754 on inode 4354967597
bad magic number 0x5ead on inode 4354967598
bad version number 0xffffffa3 on inode 4354967598
bad (negative) size -3586320840011193807 on inode 4354967598
bad magic number 0xf5ac on inode 4354967599
bad version number 0x4d on inode 4354967599
bad (negative) size -7822733889952692968 on inode 4354967599
bad magic number 0x5c7e on inode 4354967600
bad version number 0x4e on inode 4354967600
bad (negative) size -8116057573661653832 on inode 4354967600
bad magic number 0x57c5 on inode 4354967601
bad version number 0xffffffe5 on inode 4354967601
bad inode format in inode 4354967601
bad magic number 0xcc0e on inode 4354967602
bad version number 0x74 on inode 4354967602
bad (negative) size -7175657235803173389 on inode 4354967602
bad magic number 0x17a3 on inode 4354967603
bad version number 0x3b on inode 4354967603
bad inode format in inode 4354967603
bad magic number 0xf085 on inode 4354967604
bad version number 0xffffffd3 on inode 4354967604
bad (negative) size -4829046689272463617 on inode 4354967604
bad magic number 0xd47a on inode 4354967605
bad version number 0xe on inode 4354967605
bad inode format in inode 4354967605
bad magic number 0xcb77 on inode 4354967606
bad version number 0xfffffff6 on inode 4354967606
bad (negative) size -4813752938159903980 on inode 4354967606
bad magic number 0xfb56 on inode 4354967607
bad version number 0x0 on inode 4354967607
bad (negative) size -5002825908862883310 on inode 4354967607
bad magic number 0xf74f on inode 4354967608
bad version number 0x31 on inode 4354967608
bad (negative) size -2804153168172737527 on inode 4354967608
bad magic number 0xa92e on inode 4354967609
bad version number 0xffffffad on inode 4354967609
bad (negative) size -3138139371176466993 on inode 4354967609
bad magic number 0x75e5 on inode 4354967610
bad version number 0x66 on inode 4354967610
bad inode format in inode 4354967610
bad magic number 0x6b8d on inode 4354967611
bad version number 0x26 on inode 4354967611
bad inode format in inode 4354967611
bad magic number 0xf803 on inode 4354967612
bad version number 0x7f on inode 4354967612
bad inode format in inode 4354967612
bad magic number 0xaa2f on inode 4354967613
bad version number 0x31 on inode 4354967613
bad (negative) size -8402135599746543486 on inode 4354967613
bad magic number 0xbd27 on inode 4354967614
bad version number 0xffffffec on inode 4354967614
bad inode format in inode 4354967614
bad magic number 0x27bd on inode 4354967615
bad version number 0x1f on inode 4354967615
bad inode format in inode 4354967615
bad magic number 0x3f1a on inode 4354967584, would reset magic number
bad version number 0x5b on inode 4354967584, would reset version number
bad (negative) size -4743366169855139346 on inode 4354967584
would have cleared inode 4354967584
bad magic number 0xf749 on inode 4354967585, would reset magic number
bad version number 0x19 on inode 4354967585, would reset version number
bad (negative) size -4925401835952624351 on inode 4354967585
would have cleared inode 4354967585
bad magic number 0xefc6 on inode 4354967586, would reset magic number
bad version number 0xffffffc0 on inode 4354967586, would reset version number
bad (negative) size -6663211212969851433 on inode 4354967586
would have cleared inode 4354967586
bad magic number 0x93db on inode 4354967587, would reset magic number
bad version number 0x5f on inode 4354967587, would reset version number
bad inode format in inode 4354967587
would have cleared inode 4354967587
bad magic number 0x54da on inode 4354967588, would reset magic number
bad version number 0xffffff86 on inode 4354967588, would reset version number
bad inode format in inode 4354967588
would have cleared inode 4354967588
bad magic number 0x5893 on inode 4354967589, would reset magic number
bad version number 0x0 on inode 4354967589, would reset version number
bad inode format in inode 4354967589
would have cleared inode 4354967589
bad magic number 0x309f on inode 4354967590, would reset magic number
bad version number 0xffffffc2 on inode 4354967590, would reset version number
bad (negative) size -7620939567173939239 on inode 4354967590
would have cleared inode 4354967590
bad magic number 0x9fce on inode 4354967591, would reset magic number
bad version number 0x38 on inode 4354967591, would reset version number
bad inode format in inode 4354967591
would have cleared inode 4354967591
bad magic number 0xb59f on inode 4354967592, would reset magic number
bad version number 0x4 on inode 4354967592, would reset version number
bad inode format in inode 4354967592
would have cleared inode 4354967592
bad magic number 0x53b4 on inode 4354967593, would reset magic number
bad version number 0xfffffff4 on inode 4354967593, would reset version number
bad (negative) size -6181984635646857242 on inode 4354967593
would have cleared inode 4354967593
bad magic number 0x7c77 on inode 4354967594, would reset magic number
bad version number 0x3e on inode 4354967594, would reset version number
bad (negative) size -5208493334119488045 on inode 4354967594
would have cleared inode 4354967594
bad magic number 0x6698 on inode 4354967595, would reset magic number
bad version number 0x51 on inode 4354967595, would reset version number
bad inode format in inode 4354967595
would have cleared inode 4354967595
bad magic number 0xe872 on inode 4354967596, would reset magic number
bad version number 0xffffffa1 on inode 4354967596, would reset version number
bad (negative) size -4217771995784516902 on inode 4354967596
would have cleared inode 4354967596
bad magic number 0x2c98 on inode 4354967597, would reset magic number
bad version number 0xffffff8e on inode 4354967597, would reset version number
bad (negative) size -1829068493719510754 on inode 4354967597
would have cleared inode 4354967597
bad magic number 0x5ead on inode 4354967598, would reset magic number
bad version number 0xffffffa3 on inode 4354967598, would reset version number
bad (negative) size -3586320840011193807 on inode 4354967598
would have cleared inode 4354967598
bad magic number 0xf5ac on inode 4354967599, would reset magic number
bad version number 0x4d on inode 4354967599, would reset version number
bad (negative) size -7822733889952692968 on inode 4354967599
would have cleared inode 4354967599
bad magic number 0x5c7e on inode 4354967600, would reset magic number
bad version number 0x4e on inode 4354967600, would reset version number
bad (negative) size -8116057573661653832 on inode 4354967600
would have cleared inode 4354967600
bad magic number 0x57c5 on inode 4354967601, would reset magic number
bad version number 0xffffffe5 on inode 4354967601, would reset version number
bad inode format in inode 4354967601
would have cleared inode 4354967601
bad magic number 0xcc0e on inode 4354967602, would reset magic number
bad version number 0x74 on inode 4354967602, would reset version number
bad (negative) size -7175657235803173389 on inode 4354967602
would have cleared inode 4354967602
bad magic number 0x17a3 on inode 4354967603, would reset magic number
bad version number 0x3b on inode 4354967603, would reset version number
bad inode format in inode 4354967603
would have cleared inode 4354967603
bad magic number 0xf085 on inode 4354967604, would reset magic number
bad version number 0xffffffd3 on inode 4354967604, would reset version number
bad (negative) size -4829046689272463617 on inode 4354967604
would have cleared inode 4354967604
bad magic number 0xd47a on inode 4354967605, would reset magic number
bad version number 0xe on inode 4354967605, would reset version number
bad inode format in inode 4354967605
would have cleared inode 4354967605
bad magic number 0xcb77 on inode 4354967606, would reset magic number
bad version number 0xfffffff6 on inode 4354967606, would reset version number
bad (negative) size -4813752938159903980 on inode 4354967606
would have cleared inode 4354967606
bad magic number 0xfb56 on inode 4354967607, would reset magic number
bad version number 0x0 on inode 4354967607, would reset version number
bad (negative) size -5002825908862883310 on inode 4354967607
would have cleared inode 4354967607
bad magic number 0xf74f on inode 4354967608, would reset magic number
bad version number 0x31 on inode 4354967608, would reset version number
bad (negative) size -2804153168172737527 on inode 4354967608
would have cleared inode 4354967608
bad magic number 0xa92e on inode 4354967609, would reset magic number
bad version number 0xffffffad on inode 4354967609, would reset version number
bad (negative) size -3138139371176466993 on inode 4354967609
would have cleared inode 4354967609
bad magic number 0x75e5 on inode 4354967610, would reset magic number
bad version number 0x66 on inode 4354967610, would reset version number
bad inode format in inode 4354967610
would have cleared inode 4354967610
bad magic number 0x6b8d on inode 4354967611, would reset magic number
bad version number 0x26 on inode 4354967611, would reset version number
bad inode format in inode 4354967611
would have cleared inode 4354967611
bad magic number 0xf803 on inode 4354967612, would reset magic number
bad version number 0x7f on inode 4354967612, would reset version number
bad inode format in inode 4354967612
would have cleared inode 4354967612
bad magic number 0xaa2f on inode 4354967613, would reset magic number
bad version number 0x31 on inode 4354967613, would reset version number
bad (negative) size -8402135599746543486 on inode 4354967613
would have cleared inode 4354967613
bad magic number 0xbd27 on inode 4354967614, would reset magic number
bad version number 0xffffffec on inode 4354967614, would reset version number
bad inode format in inode 4354967614
would have cleared inode 4354967614
bad magic number 0x27bd on inode 4354967615, would reset magic number
bad version number 0x1f on inode 4354967615, would reset version number
bad inode format in inode 4354967615
would have cleared inode 4354967615
        - agno = 2
        - agno = 33
        - agno = 48
        - agno = 63
        - agno = 18
        - agno = 78
        - agno = 49
        - agno = 34
        - agno = 3
        - agno = 19
        - agno = 64
        - agno = 35
        - agno = 50
        - agno = 79
        - agno = 20
        - agno = 65
        - agno = 4
        - agno = 36
        - agno = 51
        - agno = 21
        - agno = 66
        - agno = 80
        - agno = 5
        - agno = 37
        - agno = 52
        - agno = 22
        - agno = 67
        - agno = 38
        - agno = 53
        - agno = 81
        - agno = 6
        - agno = 23
        - agno = 68
        - agno = 39
        - agno = 54
        - agno = 82
        - agno = 7
        - agno = 24
        - agno = 69
        - agno = 40
        - agno = 55
        - agno = 25
        - agno = 8
        - agno = 83
        - agno = 41
        - agno = 70
        - agno = 56
        - agno = 26
        - agno = 42
        - agno = 57
        - agno = 71
        - agno = 9
        - agno = 84
        - agno = 27
        - agno = 43
        - agno = 58
        - agno = 72
        - agno = 10
        - agno = 85
        - agno = 44
        - agno = 28
        - agno = 59
        - agno = 73
        - agno = 11
        - agno = 86
        - agno = 29
        - agno = 74
        - agno = 12
        - agno = 87
        - agno = 13
        - agno = 14
        - 14:32:06: process known inodes and inode discovery - 79473472 of 64 inodes done
        - process newly discovered inodes...
        - 14:32:06: process newly discovered inodes - 88 of 88 allocation groups done
Phase 4 - check for duplicate blocks...
        - setting up duplicate extent list...
        - 14:32:09: setting up duplicate extent list - 88 of 88 allocation groups done
        - check for inodes claiming duplicate blocks...
        - agno = 0
        - agno = 30
        - agno = 60
        - agno = 15
        - agno = 75
        - agno = 45
entry "102" at block 3 offset 160 in directory inode 128854679580 references free inode 4354967595
        would clear inode number in entry at offset 160...
        - agno = 31
        - agno = 46
entry "123" at block 0 offset 1088 in directory inode 257750135845 references free inode 4354967607
        would clear inode number in entry at offset 1088...
        - agno = 16
        - agno = 61
        - agno = 76
        - agno = 1
entry "755" at block 0 offset 256 in directory inode 133160166402 references free inode 4354967608
        would clear inode number in entry at offset 256...
entry "335" at block 0 offset 800 in directory inode 133160166402 references free inode 4354967612
        would clear inode number in entry at offset 800...
        - agno = 32
        - agno = 47
entry "553" at block 0 offset 2112 in directory inode 68723563568 references free inode 4354967601
        would clear inode number in entry at offset 2112...
entry "478" at block 3 offset 704 in directory inode 262041560107 references free inode 4354967588
        would clear inode number in entry at offset 704...
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
        - agno = 17
        - agno = 62
entry "868747126f598190569cc83b2028d3c3" in shortform directory 4295693332 references free inode 4354967584
would have junked entry "868747126f598190569cc83b2028d3c3" in directory inode 4295693332
would have corrected i8 count in directory 4295693332 from 3 to 2
entry "88e7499849585ec20b4b95d014632d62" at block 0 offset 144 in directory inode 4307964959 references free inode 4354967606
        would clear inode number in entry at offset 144...
entry "a22e4ffdf3bcf47d316a7f20c4b65927" in shortform directory 4310804523 references free inode 4354967597
would have junked entry "a22e4ffdf3bcf47d316a7f20c4b65927" in directory inode 4310804523
would have corrected i8 count in directory 4310804523 from 2 to 1
entry "aff7afd9850b2b80fbaa45e4e261150c" in shortform directory 4318433342 references free inode 4354967585
would have junked entry "aff7afd9850b2b80fbaa45e4e261150c" in directory inode 4318433342
would have corrected i8 count in directory 4318433342 from 2 to 1
entry "8d7e5ad25683e71cd1dfe4949a754bc5" in shortform directory 4318560261 references free inode 4354967615
would have junked entry "8d7e5ad25683e71cd1dfe4949a754bc5" in directory inode 4318560261
would have corrected i8 count in directory 4318560261 from 2 to 1
        - agno = 77
entry "acfbd6d039cc61f057ab7fd49d59770f" at block 0 offset 48 in directory inode 4344747031 references free inode 4354967594
        would clear inode number in entry at offset 48...
entry "b872391b03d655726a8ceb9e6a3b3b14" at block 0 offset 48 in directory inode 4354966589 references free inode 4354967592
        would clear inode number in entry at offset 48...
entry "6d5dffea5cd56f445fb0b425ee93691b" in shortform directory 4354967571 references free inode 4354967604
would have junked entry "6d5dffea5cd56f445fb0b425ee93691b" in directory inode 4354967571
would have corrected i8 count in directory 4354967571 from 2 to 1
bad magic number 0x3f1a on inode 4354967584, would reset magic number
bad version number 0x5b on inode 4354967584, would reset version number
bad (negative) size -4743366169855139346 on inode 4354967584
would have cleared inode 4354967584
bad magic number 0xf749 on inode 4354967585, would reset magic number
bad version number 0x19 on inode 4354967585, would reset version number
bad (negative) size -4925401835952624351 on inode 4354967585
would have cleared inode 4354967585
bad magic number 0xefc6 on inode 4354967586, would reset magic number
bad version number 0xffffffc0 on inode 4354967586, would reset version number
bad (negative) size -6663211212969851433 on inode 4354967586
would have cleared inode 4354967586
bad magic number 0x93db on inode 4354967587, would reset magic number
bad version number 0x5f on inode 4354967587, would reset version number
bad inode format in inode 4354967587
would have cleared inode 4354967587
bad magic number 0x54da on inode 4354967588, would reset magic number
bad version number 0xffffff86 on inode 4354967588, would reset version number
bad inode format in inode 4354967588
would have cleared inode 4354967588
bad magic number 0x5893 on inode 4354967589, would reset magic number
bad version number 0x0 on inode 4354967589, would reset version number
bad inode format in inode 4354967589
would have cleared inode 4354967589
bad magic number 0x309f on inode 4354967590, would reset magic number
bad version number 0xffffffc2 on inode 4354967590, would reset version number
bad (negative) size -7620939567173939239 on inode 4354967590
would have cleared inode 4354967590
bad magic number 0x9fce on inode 4354967591, would reset magic number
bad version number 0x38 on inode 4354967591, would reset version number
bad inode format in inode 4354967591
would have cleared inode 4354967591
bad magic number 0xb59f on inode 4354967592, would reset magic number
bad version number 0x4 on inode 4354967592, would reset version number
bad inode format in inode 4354967592
would have cleared inode 4354967592
bad magic number 0x53b4 on inode 4354967593, would reset magic number
bad version number 0xfffffff4 on inode 4354967593, would reset version number
bad (negative) size -6181984635646857242 on inode 4354967593
would have cleared inode 4354967593
bad magic number 0x7c77 on inode 4354967594, would reset magic number
bad version number 0x3e on inode 4354967594, would reset version number
bad (negative) size -5208493334119488045 on inode 4354967594
would have cleared inode 4354967594
bad magic number 0x6698 on inode 4354967595, would reset magic number
bad version number 0x51 on inode 4354967595, would reset version number
bad inode format in inode 4354967595
would have cleared inode 4354967595
bad magic number 0xe872 on inode 4354967596, would reset magic number
bad version number 0xffffffa1 on inode 4354967596, would reset version number
bad (negative) size -4217771995784516902 on inode 4354967596
would have cleared inode 4354967596
bad magic number 0x2c98 on inode 4354967597, would reset magic number
bad version number 0xffffff8e on inode 4354967597, would reset version number
bad (negative) size -1829068493719510754 on inode 4354967597
would have cleared inode 4354967597
bad magic number 0x5ead on inode 4354967598, would reset magic number
bad version number 0xffffffa3 on inode 4354967598, would reset version number
bad (negative) size -3586320840011193807 on inode 4354967598
would have cleared inode 4354967598
bad magic number 0xf5ac on inode 4354967599, would reset magic number
bad version number 0x4d on inode 4354967599, would reset version number
bad (negative) size -7822733889952692968 on inode 4354967599
would have cleared inode 4354967599
bad magic number 0x5c7e on inode 4354967600, would reset magic number
bad version number 0x4e on inode 4354967600, would reset version number
bad (negative) size -8116057573661653832 on inode 4354967600
would have cleared inode 4354967600
bad magic number 0x57c5 on inode 4354967601, would reset magic number
bad version number 0xffffffe5 on inode 4354967601, would reset version number
bad inode format in inode 4354967601
would have cleared inode 4354967601
bad magic number 0xcc0e on inode 4354967602, would reset magic number
bad version number 0x74 on inode 4354967602, would reset version number
bad (negative) size -7175657235803173389 on inode 4354967602
would have cleared inode 4354967602
bad magic number 0x17a3 on inode 4354967603, would reset magic number
bad version number 0x3b on inode 4354967603, would reset version number
bad inode format in inode 4354967603
would have cleared inode 4354967603
bad magic number 0xf085 on inode 4354967604, would reset magic number
bad version number 0xffffffd3 on inode 4354967604, would reset version number
bad (negative) size -4829046689272463617 on inode 4354967604
would have cleared inode 4354967604
bad magic number 0xd47a on inode 4354967605, would reset magic number
bad version number 0xe on inode 4354967605, would reset version number
bad inode format in inode 4354967605
would have cleared inode 4354967605
bad magic number 0xcb77 on inode 4354967606, would reset magic number
bad version number 0xfffffff6 on inode 4354967606, would reset version number
bad (negative) size -4813752938159903980 on inode 4354967606
would have cleared inode 4354967606
bad magic number 0xfb56 on inode 4354967607, would reset magic number
bad version number 0x0 on inode 4354967607, would reset version number
bad (negative) size -5002825908862883310 on inode 4354967607
would have cleared inode 4354967607
bad magic number 0xf74f on inode 4354967608, would reset magic number
bad version number 0x31 on inode 4354967608, would reset version number
bad (negative) size -2804153168172737527 on inode 4354967608
would have cleared inode 4354967608
bad magic number 0xa92e on inode 4354967609, would reset magic number
bad version number 0xffffffad on inode 4354967609, would reset version number
bad (negative) size -3138139371176466993 on inode 4354967609
would have cleared inode 4354967609
bad magic number 0x75e5 on inode 4354967610, would reset magic number
bad version number 0x66 on inode 4354967610, would reset version number
bad inode format in inode 4354967610
would have cleared inode 4354967610
bad magic number 0x6b8d on inode 4354967611, would reset magic number
bad version number 0x26 on inode 4354967611, would reset version number
bad inode format in inode 4354967611
would have cleared inode 4354967611
bad magic number 0xf803 on inode 4354967612, would reset magic number
bad version number 0x7f on inode 4354967612, would reset version number
bad inode format in inode 4354967612
would have cleared inode 4354967612
bad magic number 0xaa2f on inode 4354967613, would reset magic number
bad version number 0x31 on inode 4354967613, would reset version number
bad (negative) size -8402135599746543486 on inode 4354967613
would have cleared inode 4354967613
bad magic number 0xbd27 on inode 4354967614, would reset magic number
bad version number 0xffffffec on inode 4354967614, would reset version number
bad inode format in inode 4354967614
would have cleared inode 4354967614
bad magic number 0x27bd on inode 4354967615, would reset magic number
bad version number 0x1f on inode 4354967615, would reset version number
bad inode format in inode 4354967615
would have cleared inode 4354967615
        - agno = 2
entry "831" at block 2 offset 1680 in directory inode 201865261100 references free inode 4354967605
        would clear inode number in entry at offset 1680...
        - agno = 33
        - agno = 48

        - agno = 18
        - agno = 63
entry "503" at block 0 offset 3584 in directory inode 330718645273 references free inode 4354967613
        would clear inode number in entry at offset 3584...
        - agno = 78
        - agno = 34
        - agno = 49
        - agno = 3
        - agno = 19
entry "125" at block 0 offset 2160 in directory inode 270592950284 references free inode 4354967596
        would clear inode number in entry at offset 2160...
entry "746" at block 0 offset 2448 in directory inode 270649053207 references free inode 4354967611
        would clear inode number in entry at offset 2448...
        - agno = 64
        - agno = 35
        - agno = 50
        - agno = 79
        - agno = 20
        - agno = 65
        - agno = 4
        - agno = 36
        - agno = 51
        - agno = 21
        - agno = 66
        - agno = 80
        - agno = 5
entry "937" at block 0 offset 944 in directory inode 154623186993 references free inode 4354967610
        would clear inode number in entry at offset 944...
        - agno = 37
        - agno = 52
        - agno = 22
        - agno = 67
        - agno = 38
        - agno = 53
        - agno = 81
        - agno = 6
        - agno = 23
        - agno = 68
        - agno = 39
entry "612" at block 1 offset 1840 in directory inode 227634536482 references free inode 4354967599
        would clear inode number in entry at offset 1840...
        - agno = 54
        - agno = 7
        - agno = 82
        - agno = 24
entry "361" at block 1 offset 272 in directory inode 231928297598 references free inode 4354967591
        would clear inode number in entry at offset 272...
entry "875" at block 0 offset 3840 in directory inode 231928298526 references free inode 4354967593
        would clear inode number in entry at offset 3840...
entry "592" at block 0 offset 3728 in directory inode 292065894452 references free inode 4354967600
        would clear inode number in entry at offset 3728...
        - agno = 69
        - agno = 40
        - agno = 55
entry "922" at block 0 offset 3216 in directory inode 103079291917 references free inode 4354967589
        would clear inode number in entry at offset 3216...
entry "561" at block 0 offset 944 in directory inode 103079473192 references free inode 4354967609
        would clear inode number in entry at offset 944...
entry "033" at block 0 offset 544 in directory inode 103136944171 references free inode 4354967590
        would clear inode number in entry at offset 544...
        - agno = 8
        - agno = 25
entry "702" at block 0 offset 960 in directory inode 352256730114 references free inode 4354967586
        would clear inode number in entry at offset 960...
entry "965" at block 0 offset 1168 in directory inode 352256730114 references free inode 4354967587
        would clear inode number in entry at offset 1168...
        - agno = 83
        - agno = 41
        - agno = 70
entry "510" at block 1 offset 336 in directory inode 236351051817 references free inode 4354967602
        would clear inode number in entry at offset 336...
        - agno = 56
        - agno = 26
        - agno = 42
        - agno = 57
        - agno = 9
        - agno = 71
        - agno = 84
        - agno = 27
        - agno = 43
        - agno = 58
        - agno = 72
        - agno = 10
        - agno = 85
        - agno = 44
        - agno = 28
        - agno = 59
        - agno = 73
        - agno = 11
        - agno = 86
        - agno = 29
        - agno = 74
        - agno = 12
entry "990" at block 1 offset 224 in directory inode 369367197707 references free inode 4354967598
        would clear inode number in entry at offset 224...
        - agno = 87
        - agno = 13
        - agno = 14
entry "479" at block 3 offset 144 in directory inode 60193580050 references free inode 4354967603
        would clear inode number in entry at offset 144...
        - 14:53:41: check for inodes claiming duplicate blocks - 79473472 of 64 inodes done
No modify flag set, skipping phase 5
Phase 6 - check inode connectivity...
        - traversing filesystem ...
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
Metadata corruption detected at xfs_inode block 0x81c9c210/0x2000
entry "868747126f598190569cc83b2028d3c3" in shortform directory inode 4295693332 points to free inode 4354967584
would junk entry
would fix i8count in inode 4295693332
entry "88e7499849585ec20b4b95d014632d62" in directory inode 4307964959 points to free inode 4354967606, would junk entry
bad hash table for directory inode 4307964959 (no data entry): would rebuild
entry "a22e4ffdf3bcf47d316a7f20c4b65927" in shortform directory inode 4310804523 points to free inode 4354967597
would junk entry
would fix i8count in inode 4310804523
entry "aff7afd9850b2b80fbaa45e4e261150c" in shortform directory inode 4318433342 points to free inode 4354967585
would junk entry
would fix i8count in inode 4318433342
entry "8d7e5ad25683e71cd1dfe4949a754bc5" in shortform directory inode 4318560261 points to free inode 4354967615
would junk entry
would fix i8count in inode 4318560261
entry "acfbd6d039cc61f057ab7fd49d59770f" in directory inode 4344747031 points to free inode 4354967594, would junk entry
bad hash table for directory inode 4344747031 (no data entry): would rebuild
entry "b872391b03d655726a8ceb9e6a3b3b14" in directory inode 4354966589 points to free inode 4354967592, would junk entry
bad hash table for directory inode 4354966589 (no data entry): would rebuild
entry "6d5dffea5cd56f445fb0b425ee93691b" in shortform directory inode 4354967571 points to free inode 4354967604
would junk entry
would fix i8count in inode 4354967571
entry "479" in directory inode 60193580050 points to free inode 4354967603, would junk entry
entry "553" in directory inode 68723563568 points to free inode 4354967601, would junk entry
bad hash table for directory inode 68723563568 (no data entry): would rebuild
entry "922" in directory inode 103079291917 points to free inode 4354967589, would junk entry
bad hash table for directory inode 103079291917 (no data entry): would rebuild
entry "561" in directory inode 103079473192 points to free inode 4354967609, would junk entry
bad hash table for directory inode 103079473192 (no data entry): would rebuild
entry "033" in directory inode 103136944171 points to free inode 4354967590, would junk entry
bad hash table for directory inode 103136944171 (no data entry): would rebuild
entry "102" in directory inode 128854679580 points to free inode 4354967595, would junk entry
entry "755" in directory inode 133160166402 points to free inode 4354967608, would junk entry
entry "335" in directory inode 133160166402 points to free inode 4354967612, would junk entry
entry "937" in directory inode 154623186993 points to free inode 4354967610, would junk entry
bad hash table for directory inode 154623186993 (no data entry): would rebuild
entry "831" in directory inode 201865261100 points to free inode 4354967605, would junk entry

entry "612" in directory inode 227634536482 points to free inode 4354967599, would junk entry
entry "361" in directory inode 231928297598 points to free inode 4354967591, would junk entry
bad hash table for directory inode 231928297598 (no data entry): would rebuild
entry "875" in directory inode 231928298526 points to free inode 4354967593, would junk entry
entry "510" in directory inode 236351051817 points to free inode 4354967602, would junk entry
entry "123" in directory inode 257750135845 points to free inode 4354967607, would junk entry
entry "478" in directory inode 262041560107 points to free inode 4354967588, would junk entry
entry "125" in directory inode 270592950284 points to free inode 4354967596, would junk entry
entry "746" in directory inode 270649053207 points to free inode 4354967611, would junk entry
entry "592" in directory inode 292065894452 points to free inode 4354967600, would junk entry
bad hash table for directory inode 292065894452 (no data entry): would rebuild
entry "503" in directory inode 330718645273 points to free inode 4354967613, would junk entry
entry "702" in directory inode 352256730114 points to free inode 4354967586, would junk entry
entry "965" in directory inode 352256730114 points to free inode 4354967587, would junk entry
entry "990" in directory inode 369367197707 points to free inode 4354967598, would junk entry
bad hash table for directory inode 369367197707 (no data entry): would rebuild
        - traversal finished ...
        - moving disconnected inodes to lost+found ...
disconnected inode 4354971706, would move to lost+found
disconnected inode 4355534875, would move to lost+found
disconnected inode 4356321307, would move to lost+found
disconnected inode 4356453437, would move to lost+found
disconnected inode 4365067280, would move to lost+found
disconnected inode 4367228968, would move to lost+found
disconnected inode 60428483644, would move to lost+found
Phase 7 - verify link counts...
would have reset inode 103079291917 nlinks from 406 to 405
would have reset inode 68723563568 nlinks from 314 to 313
would have reset inode 154623186993 nlinks from 146 to 145
would have reset inode 60193580050 nlinks from 982 to 981
would have reset inode 128854679580 nlinks from 971 to 970
would have reset inode 103079473192 nlinks from 127 to 126
would have reset inode 60428483644 nlinks from 0 to 1
would have reset inode 133160166402 nlinks from 619 to 617
would have reset inode 231928297598 nlinks from 377 to 376
would have reset inode 227634536482 nlinks from 979 to 978
would have reset inode 262041560107 nlinks from 954 to 953
would have reset inode 270592950284 nlinks from 792 to 791
would have reset inode 292065894452 nlinks from 353 to 352
would have reset inode 231928298526 nlinks from 916 to 915
would have reset inode 330718645273 nlinks from 556 to 555
would have reset inode 369367197707 nlinks from 443 to 442
would have reset inode 270649053207 nlinks from 951 to 950
would have reset inode 201865261100 nlinks from 715 to 714
would have reset inode 257750135845 nlinks from 1001 to 1000
would have reset inode 103136944171 nlinks from 56 to 55
would have reset inode 236351051817 nlinks from 662 to 661
would have reset inode 352256730114 nlinks from 1001 to 999
        - 15:22:58: verify and correct link counts - 88 of 88 allocation groups done
No modify flag set, skipping filesystem flush and exiting.
15:23:00 [root@dn-238[DP]:/mnt] $
15:23:00 [root@dn-238[DP]:/mnt] $
15:23:00 [root@dn-238[DP]:/mnt] $
15:30:30 [root@dn-238[DP]:/mnt] $
15:32:26 [root@dn-238[DP]:/mnt] $
15:32:27 [root@dn-238[DP]:/mnt] $
15:32:27 [root@dn-238[DP]:/mnt] $
15:32:30 [root@dn-238[DP]:/mnt] $
15:32:30 [root@dn-238[DP]:/mnt] $

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: corrupt xfs log
  2017-08-18 12:17   ` Brian Foster
@ 2017-08-21 12:08     ` Ingard -
  2017-08-21 15:51       ` Brian Foster
  0 siblings, 1 reply; 16+ messages in thread
From: Ingard - @ 2017-08-21 12:08 UTC (permalink / raw)
  To: Brian Foster; +Cc: Bill O'Donnell, linux-xfs

On Fri, Aug 18, 2017 at 2:17 PM, Brian Foster <bfoster@redhat.com> wrote:
> On Fri, Aug 18, 2017 at 07:02:24AM -0500, Bill O'Donnell wrote:
>> On Fri, Aug 18, 2017 at 01:56:31PM +0200, Ingard - wrote:
>> > After a server crash we've encountered a corrupt xfs filesystem. When
>> > trying to mount said filesystem normally the system hangs.
>> > This was initially on a ubuntu trusty server with 3.13 kernel with
>> > xfsprogs 3.1.9
>> >
>> > We've installed a newer kernel (4.4.0-92) and compiled xfsprogs v
>> > 4.12.0 from source. We're still not able to mount the filesystem (and
>> > replay the log) normally.
>> > We are able to mount it -o ro,norecovery, but we're reluctant to do
>> > xfs_repair -L without trying everything we can first. The filesystem
>> > is browsable albeit a few paths which gives an error : "Structure
>> > needs cleaning"
>> >
>> > Does anyone have any advice as to how we might recover/repair the
>> > corrupt log so we can replay it? Or is xfs_repair -L the only way
>> > forward?
>>
>> Can you try xfs_repair -n (only scans the fs and reports what repairs
>> would be made)?
>>
>
> An xfs_metadump of the fs might be useful as well. Then we can see if we
> can reproduce the mount hang on latest kernels and if so, potentially
> try and root cause it.
>
> Brian

Here is a link for the metadump :
https://www.jottacloud.com/p/ingardme/95ec2e45ba80431d962345981d38bdff
And the repair -n output :
https://www.jottacloud.com/p/ingardme/0205c6ca6f7e495ebcda5f255b96f63d

kind regards
ingard

>
>> Thanks-
>> Bill
>>
>>
>> >
>> >
>> > Excerpt from kern.log:
>> > 2017-08-17T13:40:41.122121+02:00 dn-238 kernel: [  294.300347] XFS
>> > (sdd1): Mounting V4 filesystem in no-recovery mode. Filesystem will be
>> > inconsistent.
>> >
>> > 2017-08-17T17:04:54.794194+02:00 dn-238 kernel: [12548.400260] XFS
>> > (sdd1): Metadata corruption detected at xfs_inode_buf_verify+0x6f/0xd0
>> > [xfs], xfs_inode block 0x81c9c210
>> > 2017-08-17T17:04:54.794216+02:00 dn-238 kernel: [12548.400342] XFS
>> > (sdd1): Unmount and run xfs_repair
>> > 2017-08-17T17:04:54.794218+02:00 dn-238 kernel: [12548.400374] XFS
>> > (sdd1): First 64 bytes of corrupted metadata buffer:
>> > 2017-08-17T17:04:54.794220+02:00 dn-238 kernel: [12548.400418]
>> > ffff880171fff000: 3f 1a 33 54 5b 55 85 0b 7c f5 c6 d5 cf 51 47 41
>> > ?.3T[U..|....QGA
>> > 2017-08-17T17:04:54.794222+02:00 dn-238 kernel: [12548.400473]
>> > ffff880171fff010: 97 ba ba 03 5c e4 02 7a e6 bc fb 5d f1 72 db c1
>> > ....\..z...].r..
>> > 2017-08-17T17:04:54.794223+02:00 dn-238 kernel: [12548.400527]
>> > ffff880171fff020: c8 ad 3a 76 c7 e4 20 92 88 a2 35 0c 1f 36 cf b5
>> > ..:v.. ...5..6..
>> > 2017-08-17T17:04:54.794226+02:00 dn-238 kernel: [12548.400581]
>> > ffff880171fff030: 8a bc 42 75 86 50 a0 a2 be 2c 2d 99 96 2d e1 ee
>> > ..Bu.P...,-..-..
>> >
>> > kind regards
>> > ingard
>> > --
>> > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
>> > the body of a message to majordomo@vger.kernel.org
>> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> --
>> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: corrupt xfs log
  2017-08-21 12:08     ` Ingard -
@ 2017-08-21 15:51       ` Brian Foster
  2017-08-21 20:24         ` Ingard -
  0 siblings, 1 reply; 16+ messages in thread
From: Brian Foster @ 2017-08-21 15:51 UTC (permalink / raw)
  To: Ingard -; +Cc: Bill O'Donnell, linux-xfs

On Mon, Aug 21, 2017 at 02:08:43PM +0200, Ingard - wrote:
> On Fri, Aug 18, 2017 at 2:17 PM, Brian Foster <bfoster@redhat.com> wrote:
> > On Fri, Aug 18, 2017 at 07:02:24AM -0500, Bill O'Donnell wrote:
> >> On Fri, Aug 18, 2017 at 01:56:31PM +0200, Ingard - wrote:
> >> > After a server crash we've encountered a corrupt xfs filesystem. When
> >> > trying to mount said filesystem normally the system hangs.
> >> > This was initially on a ubuntu trusty server with 3.13 kernel with
> >> > xfsprogs 3.1.9
> >> >
> >> > We've installed a newer kernel (4.4.0-92) and compiled xfsprogs v
> >> > 4.12.0 from source. We're still not able to mount the filesystem (and
> >> > replay the log) normally.
> >> > We are able to mount it -o ro,norecovery, but we're reluctant to do
> >> > xfs_repair -L without trying everything we can first. The filesystem
> >> > is browsable albeit a few paths which gives an error : "Structure
> >> > needs cleaning"
> >> >
> >> > Does anyone have any advice as to how we might recover/repair the
> >> > corrupt log so we can replay it? Or is xfs_repair -L the only way
> >> > forward?
> >>
> >> Can you try xfs_repair -n (only scans the fs and reports what repairs
> >> would be made)?
> >>
> >
> > An xfs_metadump of the fs might be useful as well. Then we can see if we
> > can reproduce the mount hang on latest kernels and if so, potentially
> > try and root cause it.
> >
> > Brian
> 
> Here is a link for the metadump :
> https://www.jottacloud.com/p/ingardme/95ec2e45ba80431d962345981d38bdff

This points to a 29GB image file, apparently uncompressed..? Could you
upload a compressed file? Thanks.

Brian

> And the repair -n output :
> https://www.jottacloud.com/p/ingardme/0205c6ca6f7e495ebcda5f255b96f63d
> 
> kind regards
> ingard
> 
> >
> >> Thanks-
> >> Bill
> >>
> >>
> >> >
> >> >
> >> > Excerpt from kern.log:
> >> > 2017-08-17T13:40:41.122121+02:00 dn-238 kernel: [  294.300347] XFS
> >> > (sdd1): Mounting V4 filesystem in no-recovery mode. Filesystem will be
> >> > inconsistent.
> >> >
> >> > 2017-08-17T17:04:54.794194+02:00 dn-238 kernel: [12548.400260] XFS
> >> > (sdd1): Metadata corruption detected at xfs_inode_buf_verify+0x6f/0xd0
> >> > [xfs], xfs_inode block 0x81c9c210
> >> > 2017-08-17T17:04:54.794216+02:00 dn-238 kernel: [12548.400342] XFS
> >> > (sdd1): Unmount and run xfs_repair
> >> > 2017-08-17T17:04:54.794218+02:00 dn-238 kernel: [12548.400374] XFS
> >> > (sdd1): First 64 bytes of corrupted metadata buffer:
> >> > 2017-08-17T17:04:54.794220+02:00 dn-238 kernel: [12548.400418]
> >> > ffff880171fff000: 3f 1a 33 54 5b 55 85 0b 7c f5 c6 d5 cf 51 47 41
> >> > ?.3T[U..|....QGA
> >> > 2017-08-17T17:04:54.794222+02:00 dn-238 kernel: [12548.400473]
> >> > ffff880171fff010: 97 ba ba 03 5c e4 02 7a e6 bc fb 5d f1 72 db c1
> >> > ....\..z...].r..
> >> > 2017-08-17T17:04:54.794223+02:00 dn-238 kernel: [12548.400527]
> >> > ffff880171fff020: c8 ad 3a 76 c7 e4 20 92 88 a2 35 0c 1f 36 cf b5
> >> > ..:v.. ...5..6..
> >> > 2017-08-17T17:04:54.794226+02:00 dn-238 kernel: [12548.400581]
> >> > ffff880171fff030: 8a bc 42 75 86 50 a0 a2 be 2c 2d 99 96 2d e1 ee
> >> > ..Bu.P...,-..-..
> >> >
> >> > kind regards
> >> > ingard
> >> > --
> >> > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> >> > the body of a message to majordomo@vger.kernel.org
> >> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >> --
> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> >> the body of a message to majordomo@vger.kernel.org
> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: corrupt xfs log
  2017-08-21 15:51       ` Brian Foster
@ 2017-08-21 20:24         ` Ingard -
  2017-08-28  8:56           ` Ingard -
  2017-08-30 14:58           ` Brian Foster
  0 siblings, 2 replies; 16+ messages in thread
From: Ingard - @ 2017-08-21 20:24 UTC (permalink / raw)
  To: Brian Foster; +Cc: Bill O'Donnell, linux-xfs

On Mon, Aug 21, 2017 at 5:51 PM, Brian Foster <bfoster@redhat.com> wrote:
> On Mon, Aug 21, 2017 at 02:08:43PM +0200, Ingard - wrote:
>> On Fri, Aug 18, 2017 at 2:17 PM, Brian Foster <bfoster@redhat.com> wrote:
>> > On Fri, Aug 18, 2017 at 07:02:24AM -0500, Bill O'Donnell wrote:
>> >> On Fri, Aug 18, 2017 at 01:56:31PM +0200, Ingard - wrote:
>> >> > After a server crash we've encountered a corrupt xfs filesystem. When
>> >> > trying to mount said filesystem normally the system hangs.
>> >> > This was initially on a ubuntu trusty server with 3.13 kernel with
>> >> > xfsprogs 3.1.9
>> >> >
>> >> > We've installed a newer kernel (4.4.0-92) and compiled xfsprogs v
>> >> > 4.12.0 from source. We're still not able to mount the filesystem (and
>> >> > replay the log) normally.
>> >> > We are able to mount it -o ro,norecovery, but we're reluctant to do
>> >> > xfs_repair -L without trying everything we can first. The filesystem
>> >> > is browsable albeit a few paths which gives an error : "Structure
>> >> > needs cleaning"
>> >> >
>> >> > Does anyone have any advice as to how we might recover/repair the
>> >> > corrupt log so we can replay it? Or is xfs_repair -L the only way
>> >> > forward?
>> >>
>> >> Can you try xfs_repair -n (only scans the fs and reports what repairs
>> >> would be made)?
>> >>
>> >
>> > An xfs_metadump of the fs might be useful as well. Then we can see if we
>> > can reproduce the mount hang on latest kernels and if so, potentially
>> > try and root cause it.
>> >
>> > Brian
>>
>> Here is a link for the metadump :
>> https://www.jottacloud.com/p/ingardme/95ec2e45ba80431d962345981d38bdff
>
> This points to a 29GB image file, apparently uncompressed..? Could you
> upload a compressed file? Thanks.

Hi. Sorry about that. Didnt realize the output would be compressable.
Here is a link to the compressed tgz (6G)
https://www.jottacloud.com/p/ingardme/cac6939649e14b98b928647f5222a2ae

>
> Brian
>
>> And the repair -n output :
>> https://www.jottacloud.com/p/ingardme/0205c6ca6f7e495ebcda5f255b96f63d
>>
>> kind regards
>> ingard
>>
>> >
>> >> Thanks-
>> >> Bill
>> >>
>> >>
>> >> >
>> >> >
>> >> > Excerpt from kern.log:
>> >> > 2017-08-17T13:40:41.122121+02:00 dn-238 kernel: [  294.300347] XFS
>> >> > (sdd1): Mounting V4 filesystem in no-recovery mode. Filesystem will be
>> >> > inconsistent.
>> >> >
>> >> > 2017-08-17T17:04:54.794194+02:00 dn-238 kernel: [12548.400260] XFS
>> >> > (sdd1): Metadata corruption detected at xfs_inode_buf_verify+0x6f/0xd0
>> >> > [xfs], xfs_inode block 0x81c9c210
>> >> > 2017-08-17T17:04:54.794216+02:00 dn-238 kernel: [12548.400342] XFS
>> >> > (sdd1): Unmount and run xfs_repair
>> >> > 2017-08-17T17:04:54.794218+02:00 dn-238 kernel: [12548.400374] XFS
>> >> > (sdd1): First 64 bytes of corrupted metadata buffer:
>> >> > 2017-08-17T17:04:54.794220+02:00 dn-238 kernel: [12548.400418]
>> >> > ffff880171fff000: 3f 1a 33 54 5b 55 85 0b 7c f5 c6 d5 cf 51 47 41
>> >> > ?.3T[U..|....QGA
>> >> > 2017-08-17T17:04:54.794222+02:00 dn-238 kernel: [12548.400473]
>> >> > ffff880171fff010: 97 ba ba 03 5c e4 02 7a e6 bc fb 5d f1 72 db c1
>> >> > ....\..z...].r..
>> >> > 2017-08-17T17:04:54.794223+02:00 dn-238 kernel: [12548.400527]
>> >> > ffff880171fff020: c8 ad 3a 76 c7 e4 20 92 88 a2 35 0c 1f 36 cf b5
>> >> > ..:v.. ...5..6..
>> >> > 2017-08-17T17:04:54.794226+02:00 dn-238 kernel: [12548.400581]
>> >> > ffff880171fff030: 8a bc 42 75 86 50 a0 a2 be 2c 2d 99 96 2d e1 ee
>> >> > ..Bu.P...,-..-..
>> >> >
>> >> > kind regards
>> >> > ingard
>> >> > --
>> >> > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
>> >> > the body of a message to majordomo@vger.kernel.org
>> >> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> >> --
>> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
>> >> the body of a message to majordomo@vger.kernel.org
>> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> --
>> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: corrupt xfs log
  2017-08-21 20:24         ` Ingard -
@ 2017-08-28  8:56           ` Ingard -
  2017-08-28 10:59             ` Brian Foster
  2017-08-30 14:58           ` Brian Foster
  1 sibling, 1 reply; 16+ messages in thread
From: Ingard - @ 2017-08-28  8:56 UTC (permalink / raw)
  To: Brian Foster; +Cc: Bill O'Donnell, linux-xfs

On Mon, Aug 21, 2017 at 10:24 PM, Ingard - <ingard1@gmail.com> wrote:
> On Mon, Aug 21, 2017 at 5:51 PM, Brian Foster <bfoster@redhat.com> wrote:
>> On Mon, Aug 21, 2017 at 02:08:43PM +0200, Ingard - wrote:
>>> On Fri, Aug 18, 2017 at 2:17 PM, Brian Foster <bfoster@redhat.com> wrote:
>>> > On Fri, Aug 18, 2017 at 07:02:24AM -0500, Bill O'Donnell wrote:
>>> >> On Fri, Aug 18, 2017 at 01:56:31PM +0200, Ingard - wrote:
>>> >> > After a server crash we've encountered a corrupt xfs filesystem. When
>>> >> > trying to mount said filesystem normally the system hangs.
>>> >> > This was initially on a ubuntu trusty server with 3.13 kernel with
>>> >> > xfsprogs 3.1.9
>>> >> >
>>> >> > We've installed a newer kernel (4.4.0-92) and compiled xfsprogs v
>>> >> > 4.12.0 from source. We're still not able to mount the filesystem (and
>>> >> > replay the log) normally.
>>> >> > We are able to mount it -o ro,norecovery, but we're reluctant to do
>>> >> > xfs_repair -L without trying everything we can first. The filesystem
>>> >> > is browsable albeit a few paths which gives an error : "Structure
>>> >> > needs cleaning"
>>> >> >
>>> >> > Does anyone have any advice as to how we might recover/repair the
>>> >> > corrupt log so we can replay it? Or is xfs_repair -L the only way
>>> >> > forward?
>>> >>
>>> >> Can you try xfs_repair -n (only scans the fs and reports what repairs
>>> >> would be made)?
>>> >>
>>> >
>>> > An xfs_metadump of the fs might be useful as well. Then we can see if we
>>> > can reproduce the mount hang on latest kernels and if so, potentially
>>> > try and root cause it.
>>> >
>>> > Brian
>>>
>>> Here is a link for the metadump :
>>> https://www.jottacloud.com/p/ingardme/95ec2e45ba80431d962345981d38bdff
>>
>> This points to a 29GB image file, apparently uncompressed..? Could you
>> upload a compressed file? Thanks.
>
> Hi. Sorry about that. Didnt realize the output would be compressable.
> Here is a link to the compressed tgz (6G)
> https://www.jottacloud.com/p/ingardme/cac6939649e14b98b928647f5222a2ae

Hi
Did you have a chance to look at the log/metadump ?

ingard
>
>>
>> Brian
>>
>>> And the repair -n output :
>>> https://www.jottacloud.com/p/ingardme/0205c6ca6f7e495ebcda5f255b96f63d
>>>
>>> kind regards
>>> ingard
>>>
>>> >
>>> >> Thanks-
>>> >> Bill
>>> >>
>>> >>
>>> >> >
>>> >> >
>>> >> > Excerpt from kern.log:
>>> >> > 2017-08-17T13:40:41.122121+02:00 dn-238 kernel: [  294.300347] XFS
>>> >> > (sdd1): Mounting V4 filesystem in no-recovery mode. Filesystem will be
>>> >> > inconsistent.
>>> >> >
>>> >> > 2017-08-17T17:04:54.794194+02:00 dn-238 kernel: [12548.400260] XFS
>>> >> > (sdd1): Metadata corruption detected at xfs_inode_buf_verify+0x6f/0xd0
>>> >> > [xfs], xfs_inode block 0x81c9c210
>>> >> > 2017-08-17T17:04:54.794216+02:00 dn-238 kernel: [12548.400342] XFS
>>> >> > (sdd1): Unmount and run xfs_repair
>>> >> > 2017-08-17T17:04:54.794218+02:00 dn-238 kernel: [12548.400374] XFS
>>> >> > (sdd1): First 64 bytes of corrupted metadata buffer:
>>> >> > 2017-08-17T17:04:54.794220+02:00 dn-238 kernel: [12548.400418]
>>> >> > ffff880171fff000: 3f 1a 33 54 5b 55 85 0b 7c f5 c6 d5 cf 51 47 41
>>> >> > ?.3T[U..|....QGA
>>> >> > 2017-08-17T17:04:54.794222+02:00 dn-238 kernel: [12548.400473]
>>> >> > ffff880171fff010: 97 ba ba 03 5c e4 02 7a e6 bc fb 5d f1 72 db c1
>>> >> > ....\..z...].r..
>>> >> > 2017-08-17T17:04:54.794223+02:00 dn-238 kernel: [12548.400527]
>>> >> > ffff880171fff020: c8 ad 3a 76 c7 e4 20 92 88 a2 35 0c 1f 36 cf b5
>>> >> > ..:v.. ...5..6..
>>> >> > 2017-08-17T17:04:54.794226+02:00 dn-238 kernel: [12548.400581]
>>> >> > ffff880171fff030: 8a bc 42 75 86 50 a0 a2 be 2c 2d 99 96 2d e1 ee
>>> >> > ..Bu.P...,-..-..
>>> >> >
>>> >> > kind regards
>>> >> > ingard
>>> >> > --
>>> >> > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
>>> >> > the body of a message to majordomo@vger.kernel.org
>>> >> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>> >> --
>>> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
>>> >> the body of a message to majordomo@vger.kernel.org
>>> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
>>> the body of a message to majordomo@vger.kernel.org
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: corrupt xfs log
  2017-08-28  8:56           ` Ingard -
@ 2017-08-28 10:59             ` Brian Foster
  0 siblings, 0 replies; 16+ messages in thread
From: Brian Foster @ 2017-08-28 10:59 UTC (permalink / raw)
  To: Ingard -; +Cc: Bill O'Donnell, linux-xfs

On Mon, Aug 28, 2017 at 10:56:13AM +0200, Ingard - wrote:
> On Mon, Aug 21, 2017 at 10:24 PM, Ingard - <ingard1@gmail.com> wrote:
> > On Mon, Aug 21, 2017 at 5:51 PM, Brian Foster <bfoster@redhat.com> wrote:
> >> On Mon, Aug 21, 2017 at 02:08:43PM +0200, Ingard - wrote:
> >>> On Fri, Aug 18, 2017 at 2:17 PM, Brian Foster <bfoster@redhat.com> wrote:
> >>> > On Fri, Aug 18, 2017 at 07:02:24AM -0500, Bill O'Donnell wrote:
> >>> >> On Fri, Aug 18, 2017 at 01:56:31PM +0200, Ingard - wrote:
> >>> >> > After a server crash we've encountered a corrupt xfs filesystem. When
> >>> >> > trying to mount said filesystem normally the system hangs.
> >>> >> > This was initially on a ubuntu trusty server with 3.13 kernel with
> >>> >> > xfsprogs 3.1.9
> >>> >> >
> >>> >> > We've installed a newer kernel (4.4.0-92) and compiled xfsprogs v
> >>> >> > 4.12.0 from source. We're still not able to mount the filesystem (and
> >>> >> > replay the log) normally.
> >>> >> > We are able to mount it -o ro,norecovery, but we're reluctant to do
> >>> >> > xfs_repair -L without trying everything we can first. The filesystem
> >>> >> > is browsable albeit a few paths which gives an error : "Structure
> >>> >> > needs cleaning"
> >>> >> >
> >>> >> > Does anyone have any advice as to how we might recover/repair the
> >>> >> > corrupt log so we can replay it? Or is xfs_repair -L the only way
> >>> >> > forward?
> >>> >>
> >>> >> Can you try xfs_repair -n (only scans the fs and reports what repairs
> >>> >> would be made)?
> >>> >>
> >>> >
> >>> > An xfs_metadump of the fs might be useful as well. Then we can see if we
> >>> > can reproduce the mount hang on latest kernels and if so, potentially
> >>> > try and root cause it.
> >>> >
> >>> > Brian
> >>>
> >>> Here is a link for the metadump :
> >>> https://www.jottacloud.com/p/ingardme/95ec2e45ba80431d962345981d38bdff
> >>
> >> This points to a 29GB image file, apparently uncompressed..? Could you
> >> upload a compressed file? Thanks.
> >
> > Hi. Sorry about that. Didnt realize the output would be compressable.
> > Here is a link to the compressed tgz (6G)
> > https://www.jottacloud.com/p/ingardme/cac6939649e14b98b928647f5222a2ae
> 
> Hi
> Did you have a chance to look at the log/metadump ?
> 

Not yet. I have it downloaded somewhere but I haven't got to it yet.
It's on the todo list, hopefully soon.

Brian

> ingard
> >
> >>
> >> Brian
> >>
> >>> And the repair -n output :
> >>> https://www.jottacloud.com/p/ingardme/0205c6ca6f7e495ebcda5f255b96f63d
> >>>
> >>> kind regards
> >>> ingard
> >>>
> >>> >
> >>> >> Thanks-
> >>> >> Bill
> >>> >>
> >>> >>
> >>> >> >
> >>> >> >
> >>> >> > Excerpt from kern.log:
> >>> >> > 2017-08-17T13:40:41.122121+02:00 dn-238 kernel: [  294.300347] XFS
> >>> >> > (sdd1): Mounting V4 filesystem in no-recovery mode. Filesystem will be
> >>> >> > inconsistent.
> >>> >> >
> >>> >> > 2017-08-17T17:04:54.794194+02:00 dn-238 kernel: [12548.400260] XFS
> >>> >> > (sdd1): Metadata corruption detected at xfs_inode_buf_verify+0x6f/0xd0
> >>> >> > [xfs], xfs_inode block 0x81c9c210
> >>> >> > 2017-08-17T17:04:54.794216+02:00 dn-238 kernel: [12548.400342] XFS
> >>> >> > (sdd1): Unmount and run xfs_repair
> >>> >> > 2017-08-17T17:04:54.794218+02:00 dn-238 kernel: [12548.400374] XFS
> >>> >> > (sdd1): First 64 bytes of corrupted metadata buffer:
> >>> >> > 2017-08-17T17:04:54.794220+02:00 dn-238 kernel: [12548.400418]
> >>> >> > ffff880171fff000: 3f 1a 33 54 5b 55 85 0b 7c f5 c6 d5 cf 51 47 41
> >>> >> > ?.3T[U..|....QGA
> >>> >> > 2017-08-17T17:04:54.794222+02:00 dn-238 kernel: [12548.400473]
> >>> >> > ffff880171fff010: 97 ba ba 03 5c e4 02 7a e6 bc fb 5d f1 72 db c1
> >>> >> > ....\..z...].r..
> >>> >> > 2017-08-17T17:04:54.794223+02:00 dn-238 kernel: [12548.400527]
> >>> >> > ffff880171fff020: c8 ad 3a 76 c7 e4 20 92 88 a2 35 0c 1f 36 cf b5
> >>> >> > ..:v.. ...5..6..
> >>> >> > 2017-08-17T17:04:54.794226+02:00 dn-238 kernel: [12548.400581]
> >>> >> > ffff880171fff030: 8a bc 42 75 86 50 a0 a2 be 2c 2d 99 96 2d e1 ee
> >>> >> > ..Bu.P...,-..-..
> >>> >> >
> >>> >> > kind regards
> >>> >> > ingard
> >>> >> > --
> >>> >> > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> >>> >> > the body of a message to majordomo@vger.kernel.org
> >>> >> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >>> >> --
> >>> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> >>> >> the body of a message to majordomo@vger.kernel.org
> >>> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >>> --
> >>> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> >>> the body of a message to majordomo@vger.kernel.org
> >>> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: corrupt xfs log
  2017-08-21 20:24         ` Ingard -
  2017-08-28  8:56           ` Ingard -
@ 2017-08-30 14:58           ` Brian Foster
  2017-08-31  7:27             ` Ingard -
  1 sibling, 1 reply; 16+ messages in thread
From: Brian Foster @ 2017-08-30 14:58 UTC (permalink / raw)
  To: Ingard -; +Cc: Bill O'Donnell, linux-xfs

[-- Attachment #1: Type: text/plain, Size: 5803 bytes --]

On Mon, Aug 21, 2017 at 10:24:32PM +0200, Ingard - wrote:
> On Mon, Aug 21, 2017 at 5:51 PM, Brian Foster <bfoster@redhat.com> wrote:
> > On Mon, Aug 21, 2017 at 02:08:43PM +0200, Ingard - wrote:
> >> On Fri, Aug 18, 2017 at 2:17 PM, Brian Foster <bfoster@redhat.com> wrote:
> >> > On Fri, Aug 18, 2017 at 07:02:24AM -0500, Bill O'Donnell wrote:
> >> >> On Fri, Aug 18, 2017 at 01:56:31PM +0200, Ingard - wrote:
> >> >> > After a server crash we've encountered a corrupt xfs filesystem. When
> >> >> > trying to mount said filesystem normally the system hangs.
> >> >> > This was initially on a ubuntu trusty server with 3.13 kernel with
> >> >> > xfsprogs 3.1.9
> >> >> >
> >> >> > We've installed a newer kernel (4.4.0-92) and compiled xfsprogs v
> >> >> > 4.12.0 from source. We're still not able to mount the filesystem (and
> >> >> > replay the log) normally.
> >> >> > We are able to mount it -o ro,norecovery, but we're reluctant to do
> >> >> > xfs_repair -L without trying everything we can first. The filesystem
> >> >> > is browsable albeit a few paths which gives an error : "Structure
> >> >> > needs cleaning"
> >> >> >
> >> >> > Does anyone have any advice as to how we might recover/repair the
> >> >> > corrupt log so we can replay it? Or is xfs_repair -L the only way
> >> >> > forward?
> >> >>
> >> >> Can you try xfs_repair -n (only scans the fs and reports what repairs
> >> >> would be made)?
> >> >>
> >> >
> >> > An xfs_metadump of the fs might be useful as well. Then we can see if we
> >> > can reproduce the mount hang on latest kernels and if so, potentially
> >> > try and root cause it.
> >> >
> >> > Brian
> >>
> >> Here is a link for the metadump :
> >> https://www.jottacloud.com/p/ingardme/95ec2e45ba80431d962345981d38bdff
> >
> > This points to a 29GB image file, apparently uncompressed..? Could you
> > upload a compressed file? Thanks.
> 
> Hi. Sorry about that. Didnt realize the output would be compressable.
> Here is a link to the compressed tgz (6G)
> https://www.jottacloud.com/p/ingardme/cac6939649e14b98b928647f5222a2ae
> 

I finally played around with this image a bit. Note that mount does not
hang on latest kernels. Instead, log recovery emits a torn write message
due to a bad crc at the head of the log and then ultimately fails due to
a bad crc at the tail of the log. I ran a couple experiments to skip the
bad crc records and/or to completely ignore all bad crc's and both still
either fail to mount (due to other corruption) or continue to show
corruption in the recovered fs. 

It's not clear to me what would have caused this corruption or log
state. Have you encountered any corruption before? If not, is this kind
of crash or unclean shutdown of the server an uncommon event?

That aside, I think the best course of action is to run 'xfs_repair -L'
on the fs. I ran a v4.12 version against the metadump image and it
successfully repaired the fs. I've attached the repair output for
reference, but I would recommend to first restore your metadump to a
temporary location, attempt to repair that and examine the results
before repairing the original fs. Note that the metadump will not have
any file content, but will represent which files might be cleared, moved
to lost+found, etc.

Brian

> >
> > Brian
> >
> >> And the repair -n output :
> >> https://www.jottacloud.com/p/ingardme/0205c6ca6f7e495ebcda5f255b96f63d
> >>
> >> kind regards
> >> ingard
> >>
> >> >
> >> >> Thanks-
> >> >> Bill
> >> >>
> >> >>
> >> >> >
> >> >> >
> >> >> > Excerpt from kern.log:
> >> >> > 2017-08-17T13:40:41.122121+02:00 dn-238 kernel: [  294.300347] XFS
> >> >> > (sdd1): Mounting V4 filesystem in no-recovery mode. Filesystem will be
> >> >> > inconsistent.
> >> >> >
> >> >> > 2017-08-17T17:04:54.794194+02:00 dn-238 kernel: [12548.400260] XFS
> >> >> > (sdd1): Metadata corruption detected at xfs_inode_buf_verify+0x6f/0xd0
> >> >> > [xfs], xfs_inode block 0x81c9c210
> >> >> > 2017-08-17T17:04:54.794216+02:00 dn-238 kernel: [12548.400342] XFS
> >> >> > (sdd1): Unmount and run xfs_repair
> >> >> > 2017-08-17T17:04:54.794218+02:00 dn-238 kernel: [12548.400374] XFS
> >> >> > (sdd1): First 64 bytes of corrupted metadata buffer:
> >> >> > 2017-08-17T17:04:54.794220+02:00 dn-238 kernel: [12548.400418]
> >> >> > ffff880171fff000: 3f 1a 33 54 5b 55 85 0b 7c f5 c6 d5 cf 51 47 41
> >> >> > ?.3T[U..|....QGA
> >> >> > 2017-08-17T17:04:54.794222+02:00 dn-238 kernel: [12548.400473]
> >> >> > ffff880171fff010: 97 ba ba 03 5c e4 02 7a e6 bc fb 5d f1 72 db c1
> >> >> > ....\..z...].r..
> >> >> > 2017-08-17T17:04:54.794223+02:00 dn-238 kernel: [12548.400527]
> >> >> > ffff880171fff020: c8 ad 3a 76 c7 e4 20 92 88 a2 35 0c 1f 36 cf b5
> >> >> > ..:v.. ...5..6..
> >> >> > 2017-08-17T17:04:54.794226+02:00 dn-238 kernel: [12548.400581]
> >> >> > ffff880171fff030: 8a bc 42 75 86 50 a0 a2 be 2c 2d 99 96 2d e1 ee
> >> >> > ..Bu.P...,-..-..
> >> >> >
> >> >> > kind regards
> >> >> > ingard
> >> >> > --
> >> >> > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> >> >> > the body of a message to majordomo@vger.kernel.org
> >> >> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >> >> --
> >> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> >> >> the body of a message to majordomo@vger.kernel.org
> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >> --
> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> >> the body of a message to majordomo@vger.kernel.org
> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

[-- Attachment #2: xfs_repair.out.gz --]
[-- Type: application/gzip, Size: 4457 bytes --]

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: corrupt xfs log
  2017-08-30 14:58           ` Brian Foster
@ 2017-08-31  7:27             ` Ingard -
  2017-08-31 10:20               ` Brian Foster
  0 siblings, 1 reply; 16+ messages in thread
From: Ingard - @ 2017-08-31  7:27 UTC (permalink / raw)
  To: Brian Foster; +Cc: Bill O'Donnell, linux-xfs

On Wed, Aug 30, 2017 at 4:58 PM, Brian Foster <bfoster@redhat.com> wrote:
> On Mon, Aug 21, 2017 at 10:24:32PM +0200, Ingard - wrote:
>> On Mon, Aug 21, 2017 at 5:51 PM, Brian Foster <bfoster@redhat.com> wrote:
>> > On Mon, Aug 21, 2017 at 02:08:43PM +0200, Ingard - wrote:
>> >> On Fri, Aug 18, 2017 at 2:17 PM, Brian Foster <bfoster@redhat.com> wrote:
>> >> > On Fri, Aug 18, 2017 at 07:02:24AM -0500, Bill O'Donnell wrote:
>> >> >> On Fri, Aug 18, 2017 at 01:56:31PM +0200, Ingard - wrote:
>> >> >> > After a server crash we've encountered a corrupt xfs filesystem. When
>> >> >> > trying to mount said filesystem normally the system hangs.
>> >> >> > This was initially on a ubuntu trusty server with 3.13 kernel with
>> >> >> > xfsprogs 3.1.9
>> >> >> >
>> >> >> > We've installed a newer kernel (4.4.0-92) and compiled xfsprogs v
>> >> >> > 4.12.0 from source. We're still not able to mount the filesystem (and
>> >> >> > replay the log) normally.
>> >> >> > We are able to mount it -o ro,norecovery, but we're reluctant to do
>> >> >> > xfs_repair -L without trying everything we can first. The filesystem
>> >> >> > is browsable albeit a few paths which gives an error : "Structure
>> >> >> > needs cleaning"
>> >> >> >
>> >> >> > Does anyone have any advice as to how we might recover/repair the
>> >> >> > corrupt log so we can replay it? Or is xfs_repair -L the only way
>> >> >> > forward?
>> >> >>
>> >> >> Can you try xfs_repair -n (only scans the fs and reports what repairs
>> >> >> would be made)?
>> >> >>
>> >> >
>> >> > An xfs_metadump of the fs might be useful as well. Then we can see if we
>> >> > can reproduce the mount hang on latest kernels and if so, potentially
>> >> > try and root cause it.
>> >> >
>> >> > Brian
>> >>
>> >> Here is a link for the metadump :
>> >> https://www.jottacloud.com/p/ingardme/95ec2e45ba80431d962345981d38bdff
>> >
>> > This points to a 29GB image file, apparently uncompressed..? Could you
>> > upload a compressed file? Thanks.
>>
>> Hi. Sorry about that. Didnt realize the output would be compressable.
>> Here is a link to the compressed tgz (6G)
>> https://www.jottacloud.com/p/ingardme/cac6939649e14b98b928647f5222a2ae
>>
>
> I finally played around with this image a bit. Note that mount does not
> hang on latest kernels. Instead, log recovery emits a torn write message
> due to a bad crc at the head of the log and then ultimately fails due to
> a bad crc at the tail of the log. I ran a couple experiments to skip the
> bad crc records and/or to completely ignore all bad crc's and both still
> either fail to mount (due to other corruption) or continue to show
> corruption in the recovered fs.
>
> It's not clear to me what would have caused this corruption or log
> state. Have you encountered any corruption before? If not, is this kind
> of crash or unclean shutdown of the server an uncommon event?
We failed to notice the log messages of corrupt fs at first. After a
few days of these messages the filesystem got shut down due to
excessive? corruption.
At that point we tried to reboot normally, but ended up with having to
do a hard reset of the server.
It is not clear to us either why the corruption happened in the first
place either. The underlying raid has been in optimal state the whole
time

>
> That aside, I think the best course of action is to run 'xfs_repair -L'
> on the fs. I ran a v4.12 version against the metadump image and it
> successfully repaired the fs. I've attached the repair output for
> reference, but I would recommend to first restore your metadump to a
> temporary location, attempt to repair that and examine the results
> before repairing the original fs. Note that the metadump will not have
> any file content, but will represent which files might be cleared, moved
> to lost+found, etc.
Ok. Thanks for looking into it. We'll proceed with the suggested
course of action.

ingard
>
> Brian
>
>> >
>> > Brian
>> >
>> >> And the repair -n output :
>> >> https://www.jottacloud.com/p/ingardme/0205c6ca6f7e495ebcda5f255b96f63d
>> >>
>> >> kind regards
>> >> ingard
>> >>
>> >> >
>> >> >> Thanks-
>> >> >> Bill
>> >> >>
>> >> >>
>> >> >> >
>> >> >> >
>> >> >> > Excerpt from kern.log:
>> >> >> > 2017-08-17T13:40:41.122121+02:00 dn-238 kernel: [  294.300347] XFS
>> >> >> > (sdd1): Mounting V4 filesystem in no-recovery mode. Filesystem will be
>> >> >> > inconsistent.
>> >> >> >
>> >> >> > 2017-08-17T17:04:54.794194+02:00 dn-238 kernel: [12548.400260] XFS
>> >> >> > (sdd1): Metadata corruption detected at xfs_inode_buf_verify+0x6f/0xd0
>> >> >> > [xfs], xfs_inode block 0x81c9c210
>> >> >> > 2017-08-17T17:04:54.794216+02:00 dn-238 kernel: [12548.400342] XFS
>> >> >> > (sdd1): Unmount and run xfs_repair
>> >> >> > 2017-08-17T17:04:54.794218+02:00 dn-238 kernel: [12548.400374] XFS
>> >> >> > (sdd1): First 64 bytes of corrupted metadata buffer:
>> >> >> > 2017-08-17T17:04:54.794220+02:00 dn-238 kernel: [12548.400418]
>> >> >> > ffff880171fff000: 3f 1a 33 54 5b 55 85 0b 7c f5 c6 d5 cf 51 47 41
>> >> >> > ?.3T[U..|....QGA
>> >> >> > 2017-08-17T17:04:54.794222+02:00 dn-238 kernel: [12548.400473]
>> >> >> > ffff880171fff010: 97 ba ba 03 5c e4 02 7a e6 bc fb 5d f1 72 db c1
>> >> >> > ....\..z...].r..
>> >> >> > 2017-08-17T17:04:54.794223+02:00 dn-238 kernel: [12548.400527]
>> >> >> > ffff880171fff020: c8 ad 3a 76 c7 e4 20 92 88 a2 35 0c 1f 36 cf b5
>> >> >> > ..:v.. ...5..6..
>> >> >> > 2017-08-17T17:04:54.794226+02:00 dn-238 kernel: [12548.400581]
>> >> >> > ffff880171fff030: 8a bc 42 75 86 50 a0 a2 be 2c 2d 99 96 2d e1 ee
>> >> >> > ..Bu.P...,-..-..
>> >> >> >
>> >> >> > kind regards
>> >> >> > ingard
>> >> >> > --
>> >> >> > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
>> >> >> > the body of a message to majordomo@vger.kernel.org
>> >> >> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> >> >> --
>> >> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
>> >> >> the body of a message to majordomo@vger.kernel.org
>> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> >> --
>> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
>> >> the body of a message to majordomo@vger.kernel.org
>> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> --
>> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: corrupt xfs log
  2017-08-31  7:27             ` Ingard -
@ 2017-08-31 10:20               ` Brian Foster
  2017-09-01  6:48                 ` Ingard -
  0 siblings, 1 reply; 16+ messages in thread
From: Brian Foster @ 2017-08-31 10:20 UTC (permalink / raw)
  To: Ingard -; +Cc: Bill O'Donnell, linux-xfs

On Thu, Aug 31, 2017 at 09:27:52AM +0200, Ingard - wrote:
> On Wed, Aug 30, 2017 at 4:58 PM, Brian Foster <bfoster@redhat.com> wrote:
> > On Mon, Aug 21, 2017 at 10:24:32PM +0200, Ingard - wrote:
> >> On Mon, Aug 21, 2017 at 5:51 PM, Brian Foster <bfoster@redhat.com> wrote:
> >> > On Mon, Aug 21, 2017 at 02:08:43PM +0200, Ingard - wrote:
> >> >> On Fri, Aug 18, 2017 at 2:17 PM, Brian Foster <bfoster@redhat.com> wrote:
> >> >> > On Fri, Aug 18, 2017 at 07:02:24AM -0500, Bill O'Donnell wrote:
> >> >> >> On Fri, Aug 18, 2017 at 01:56:31PM +0200, Ingard - wrote:
> >> >> >> > After a server crash we've encountered a corrupt xfs filesystem. When
> >> >> >> > trying to mount said filesystem normally the system hangs.
> >> >> >> > This was initially on a ubuntu trusty server with 3.13 kernel with
> >> >> >> > xfsprogs 3.1.9
> >> >> >> >
> >> >> >> > We've installed a newer kernel (4.4.0-92) and compiled xfsprogs v
> >> >> >> > 4.12.0 from source. We're still not able to mount the filesystem (and
> >> >> >> > replay the log) normally.
> >> >> >> > We are able to mount it -o ro,norecovery, but we're reluctant to do
> >> >> >> > xfs_repair -L without trying everything we can first. The filesystem
> >> >> >> > is browsable albeit a few paths which gives an error : "Structure
> >> >> >> > needs cleaning"
> >> >> >> >
> >> >> >> > Does anyone have any advice as to how we might recover/repair the
> >> >> >> > corrupt log so we can replay it? Or is xfs_repair -L the only way
> >> >> >> > forward?
> >> >> >>
> >> >> >> Can you try xfs_repair -n (only scans the fs and reports what repairs
> >> >> >> would be made)?
> >> >> >>
> >> >> >
> >> >> > An xfs_metadump of the fs might be useful as well. Then we can see if we
> >> >> > can reproduce the mount hang on latest kernels and if so, potentially
> >> >> > try and root cause it.
> >> >> >
> >> >> > Brian
> >> >>
> >> >> Here is a link for the metadump :
> >> >> https://www.jottacloud.com/p/ingardme/95ec2e45ba80431d962345981d38bdff
> >> >
> >> > This points to a 29GB image file, apparently uncompressed..? Could you
> >> > upload a compressed file? Thanks.
> >>
> >> Hi. Sorry about that. Didnt realize the output would be compressable.
> >> Here is a link to the compressed tgz (6G)
> >> https://www.jottacloud.com/p/ingardme/cac6939649e14b98b928647f5222a2ae
> >>
> >
> > I finally played around with this image a bit. Note that mount does not
> > hang on latest kernels. Instead, log recovery emits a torn write message
> > due to a bad crc at the head of the log and then ultimately fails due to
> > a bad crc at the tail of the log. I ran a couple experiments to skip the
> > bad crc records and/or to completely ignore all bad crc's and both still
> > either fail to mount (due to other corruption) or continue to show
> > corruption in the recovered fs.
> >
> > It's not clear to me what would have caused this corruption or log
> > state. Have you encountered any corruption before? If not, is this kind
> > of crash or unclean shutdown of the server an uncommon event?
> We failed to notice the log messages of corrupt fs at first. After a
> few days of these messages the filesystem got shut down due to
> excessive? corruption.
> At that point we tried to reboot normally, but ended up with having to
> do a hard reset of the server.
> It is not clear to us either why the corruption happened in the first
> place either. The underlying raid has been in optimal state the whole
> time
> 

Ok, so corruption was the first problem. If the filesystem shutdown with
something other than a log I/O error, chances are the log was flushed at
that time. It is strange that log records end up corrupted, though not
terribly out of the ordinary for the mount to ultimately fail if
recovery stumbled over existing on-disk corruption, for instance.
An xfs_repair was probably a foregone conclusion given the corruption
started on disk, anyways.

Brian

> >
> > That aside, I think the best course of action is to run 'xfs_repair -L'
> > on the fs. I ran a v4.12 version against the metadump image and it
> > successfully repaired the fs. I've attached the repair output for
> > reference, but I would recommend to first restore your metadump to a
> > temporary location, attempt to repair that and examine the results
> > before repairing the original fs. Note that the metadump will not have
> > any file content, but will represent which files might be cleared, moved
> > to lost+found, etc.
> Ok. Thanks for looking into it. We'll proceed with the suggested
> course of action.
> 
> ingard
> >
> > Brian
> >
> >> >
> >> > Brian
> >> >
> >> >> And the repair -n output :
> >> >> https://www.jottacloud.com/p/ingardme/0205c6ca6f7e495ebcda5f255b96f63d
> >> >>
> >> >> kind regards
> >> >> ingard
> >> >>
> >> >> >
> >> >> >> Thanks-
> >> >> >> Bill
> >> >> >>
> >> >> >>
> >> >> >> >
> >> >> >> >
> >> >> >> > Excerpt from kern.log:
> >> >> >> > 2017-08-17T13:40:41.122121+02:00 dn-238 kernel: [  294.300347] XFS
> >> >> >> > (sdd1): Mounting V4 filesystem in no-recovery mode. Filesystem will be
> >> >> >> > inconsistent.
> >> >> >> >
> >> >> >> > 2017-08-17T17:04:54.794194+02:00 dn-238 kernel: [12548.400260] XFS
> >> >> >> > (sdd1): Metadata corruption detected at xfs_inode_buf_verify+0x6f/0xd0
> >> >> >> > [xfs], xfs_inode block 0x81c9c210
> >> >> >> > 2017-08-17T17:04:54.794216+02:00 dn-238 kernel: [12548.400342] XFS
> >> >> >> > (sdd1): Unmount and run xfs_repair
> >> >> >> > 2017-08-17T17:04:54.794218+02:00 dn-238 kernel: [12548.400374] XFS
> >> >> >> > (sdd1): First 64 bytes of corrupted metadata buffer:
> >> >> >> > 2017-08-17T17:04:54.794220+02:00 dn-238 kernel: [12548.400418]
> >> >> >> > ffff880171fff000: 3f 1a 33 54 5b 55 85 0b 7c f5 c6 d5 cf 51 47 41
> >> >> >> > ?.3T[U..|....QGA
> >> >> >> > 2017-08-17T17:04:54.794222+02:00 dn-238 kernel: [12548.400473]
> >> >> >> > ffff880171fff010: 97 ba ba 03 5c e4 02 7a e6 bc fb 5d f1 72 db c1
> >> >> >> > ....\..z...].r..
> >> >> >> > 2017-08-17T17:04:54.794223+02:00 dn-238 kernel: [12548.400527]
> >> >> >> > ffff880171fff020: c8 ad 3a 76 c7 e4 20 92 88 a2 35 0c 1f 36 cf b5
> >> >> >> > ..:v.. ...5..6..
> >> >> >> > 2017-08-17T17:04:54.794226+02:00 dn-238 kernel: [12548.400581]
> >> >> >> > ffff880171fff030: 8a bc 42 75 86 50 a0 a2 be 2c 2d 99 96 2d e1 ee
> >> >> >> > ..Bu.P...,-..-..
> >> >> >> >
> >> >> >> > kind regards
> >> >> >> > ingard
> >> >> >> > --
> >> >> >> > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> >> >> >> > the body of a message to majordomo@vger.kernel.org
> >> >> >> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >> >> >> --
> >> >> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> >> >> >> the body of a message to majordomo@vger.kernel.org
> >> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >> >> --
> >> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> >> >> the body of a message to majordomo@vger.kernel.org
> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >> --
> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> >> the body of a message to majordomo@vger.kernel.org
> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: corrupt xfs log
  2017-08-31 10:20               ` Brian Foster
@ 2017-09-01  6:48                 ` Ingard -
  2017-09-01 11:33                   ` Brian Foster
  0 siblings, 1 reply; 16+ messages in thread
From: Ingard - @ 2017-09-01  6:48 UTC (permalink / raw)
  To: Brian Foster; +Cc: Bill O'Donnell, linux-xfs

On Thu, Aug 31, 2017 at 12:20 PM, Brian Foster <bfoster@redhat.com> wrote:
> On Thu, Aug 31, 2017 at 09:27:52AM +0200, Ingard - wrote:
>> On Wed, Aug 30, 2017 at 4:58 PM, Brian Foster <bfoster@redhat.com> wrote:
>> > On Mon, Aug 21, 2017 at 10:24:32PM +0200, Ingard - wrote:
>> >> On Mon, Aug 21, 2017 at 5:51 PM, Brian Foster <bfoster@redhat.com> wrote:
>> >> > On Mon, Aug 21, 2017 at 02:08:43PM +0200, Ingard - wrote:
>> >> >> On Fri, Aug 18, 2017 at 2:17 PM, Brian Foster <bfoster@redhat.com> wrote:
>> >> >> > On Fri, Aug 18, 2017 at 07:02:24AM -0500, Bill O'Donnell wrote:
>> >> >> >> On Fri, Aug 18, 2017 at 01:56:31PM +0200, Ingard - wrote:
>> >> >> >> > After a server crash we've encountered a corrupt xfs filesystem. When
>> >> >> >> > trying to mount said filesystem normally the system hangs.
>> >> >> >> > This was initially on a ubuntu trusty server with 3.13 kernel with
>> >> >> >> > xfsprogs 3.1.9
>> >> >> >> >
>> >> >> >> > We've installed a newer kernel (4.4.0-92) and compiled xfsprogs v
>> >> >> >> > 4.12.0 from source. We're still not able to mount the filesystem (and
>> >> >> >> > replay the log) normally.
>> >> >> >> > We are able to mount it -o ro,norecovery, but we're reluctant to do
>> >> >> >> > xfs_repair -L without trying everything we can first. The filesystem
>> >> >> >> > is browsable albeit a few paths which gives an error : "Structure
>> >> >> >> > needs cleaning"
>> >> >> >> >
>> >> >> >> > Does anyone have any advice as to how we might recover/repair the
>> >> >> >> > corrupt log so we can replay it? Or is xfs_repair -L the only way
>> >> >> >> > forward?
>> >> >> >>
>> >> >> >> Can you try xfs_repair -n (only scans the fs and reports what repairs
>> >> >> >> would be made)?
>> >> >> >>
>> >> >> >
>> >> >> > An xfs_metadump of the fs might be useful as well. Then we can see if we
>> >> >> > can reproduce the mount hang on latest kernels and if so, potentially
>> >> >> > try and root cause it.
>> >> >> >
>> >> >> > Brian
>> >> >>
>> >> >> Here is a link for the metadump :
>> >> >> https://www.jottacloud.com/p/ingardme/95ec2e45ba80431d962345981d38bdff
>> >> >
>> >> > This points to a 29GB image file, apparently uncompressed..? Could you
>> >> > upload a compressed file? Thanks.
>> >>
>> >> Hi. Sorry about that. Didnt realize the output would be compressable.
>> >> Here is a link to the compressed tgz (6G)
>> >> https://www.jottacloud.com/p/ingardme/cac6939649e14b98b928647f5222a2ae
>> >>
>> >
>> > I finally played around with this image a bit. Note that mount does not
>> > hang on latest kernels. Instead, log recovery emits a torn write message
>> > due to a bad crc at the head of the log and then ultimately fails due to
>> > a bad crc at the tail of the log. I ran a couple experiments to skip the
>> > bad crc records and/or to completely ignore all bad crc's and both still
>> > either fail to mount (due to other corruption) or continue to show
>> > corruption in the recovered fs.
>> >
>> > It's not clear to me what would have caused this corruption or log
>> > state. Have you encountered any corruption before? If not, is this kind
>> > of crash or unclean shutdown of the server an uncommon event?
>> We failed to notice the log messages of corrupt fs at first. After a
>> few days of these messages the filesystem got shut down due to
>> excessive? corruption.
>> At that point we tried to reboot normally, but ended up with having to
>> do a hard reset of the server.
>> It is not clear to us either why the corruption happened in the first
>> place either. The underlying raid has been in optimal state the whole
>> time
>>
>
> Ok, so corruption was the first problem. If the filesystem shutdown with
> something other than a log I/O error, chances are the log was flushed at
> that time. It is strange that log records end up corrupted, though not
> terribly out of the ordinary for the mount to ultimately fail if
> recovery stumbled over existing on-disk corruption, for instance.
> An xfs_repair was probably a foregone conclusion given the corruption
> started on disk, anyways.

Out of curiosity, how long did the xfs_mdrestore command run ? I'm
pushing 20ish hours now and noticed the following in kern.log :
2017-09-01T08:47:23.414139+02:00 dn-238 kernel: [1278740.983304] XFS:
xfs_mdrestore(5176) possible memory allocation deadlock size 37136 in
kmem_alloc (mode:0x2400240)

ingard

>
> Brian
>
>> >
>> > That aside, I think the best course of action is to run 'xfs_repair -L'
>> > on the fs. I ran a v4.12 version against the metadump image and it
>> > successfully repaired the fs. I've attached the repair output for
>> > reference, but I would recommend to first restore your metadump to a
>> > temporary location, attempt to repair that and examine the results
>> > before repairing the original fs. Note that the metadump will not have
>> > any file content, but will represent which files might be cleared, moved
>> > to lost+found, etc.
>> Ok. Thanks for looking into it. We'll proceed with the suggested
>> course of action.
>>
>> ingard
>> >
>> > Brian
>> >
>> >> >
>> >> > Brian
>> >> >
>> >> >> And the repair -n output :
>> >> >> https://www.jottacloud.com/p/ingardme/0205c6ca6f7e495ebcda5f255b96f63d
>> >> >>
>> >> >> kind regards
>> >> >> ingard
>> >> >>
>> >> >> >
>> >> >> >> Thanks-
>> >> >> >> Bill
>> >> >> >>
>> >> >> >>
>> >> >> >> >
>> >> >> >> >
>> >> >> >> > Excerpt from kern.log:
>> >> >> >> > 2017-08-17T13:40:41.122121+02:00 dn-238 kernel: [  294.300347] XFS
>> >> >> >> > (sdd1): Mounting V4 filesystem in no-recovery mode. Filesystem will be
>> >> >> >> > inconsistent.
>> >> >> >> >
>> >> >> >> > 2017-08-17T17:04:54.794194+02:00 dn-238 kernel: [12548.400260] XFS
>> >> >> >> > (sdd1): Metadata corruption detected at xfs_inode_buf_verify+0x6f/0xd0
>> >> >> >> > [xfs], xfs_inode block 0x81c9c210
>> >> >> >> > 2017-08-17T17:04:54.794216+02:00 dn-238 kernel: [12548.400342] XFS
>> >> >> >> > (sdd1): Unmount and run xfs_repair
>> >> >> >> > 2017-08-17T17:04:54.794218+02:00 dn-238 kernel: [12548.400374] XFS
>> >> >> >> > (sdd1): First 64 bytes of corrupted metadata buffer:
>> >> >> >> > 2017-08-17T17:04:54.794220+02:00 dn-238 kernel: [12548.400418]
>> >> >> >> > ffff880171fff000: 3f 1a 33 54 5b 55 85 0b 7c f5 c6 d5 cf 51 47 41
>> >> >> >> > ?.3T[U..|....QGA
>> >> >> >> > 2017-08-17T17:04:54.794222+02:00 dn-238 kernel: [12548.400473]
>> >> >> >> > ffff880171fff010: 97 ba ba 03 5c e4 02 7a e6 bc fb 5d f1 72 db c1
>> >> >> >> > ....\..z...].r..
>> >> >> >> > 2017-08-17T17:04:54.794223+02:00 dn-238 kernel: [12548.400527]
>> >> >> >> > ffff880171fff020: c8 ad 3a 76 c7 e4 20 92 88 a2 35 0c 1f 36 cf b5
>> >> >> >> > ..:v.. ...5..6..
>> >> >> >> > 2017-08-17T17:04:54.794226+02:00 dn-238 kernel: [12548.400581]
>> >> >> >> > ffff880171fff030: 8a bc 42 75 86 50 a0 a2 be 2c 2d 99 96 2d e1 ee
>> >> >> >> > ..Bu.P...,-..-..
>> >> >> >> >
>> >> >> >> > kind regards
>> >> >> >> > ingard
>> >> >> >> > --
>> >> >> >> > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
>> >> >> >> > the body of a message to majordomo@vger.kernel.org
>> >> >> >> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> >> >> >> --
>> >> >> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
>> >> >> >> the body of a message to majordomo@vger.kernel.org
>> >> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> >> >> --
>> >> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
>> >> >> the body of a message to majordomo@vger.kernel.org
>> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> >> --
>> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
>> >> the body of a message to majordomo@vger.kernel.org
>> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> --
>> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: corrupt xfs log
  2017-09-01  6:48                 ` Ingard -
@ 2017-09-01 11:33                   ` Brian Foster
  2017-09-01 15:11                     ` Darrick J. Wong
  0 siblings, 1 reply; 16+ messages in thread
From: Brian Foster @ 2017-09-01 11:33 UTC (permalink / raw)
  To: Ingard -; +Cc: Bill O'Donnell, linux-xfs

On Fri, Sep 01, 2017 at 08:48:03AM +0200, Ingard - wrote:
> On Thu, Aug 31, 2017 at 12:20 PM, Brian Foster <bfoster@redhat.com> wrote:
> > On Thu, Aug 31, 2017 at 09:27:52AM +0200, Ingard - wrote:
> >> On Wed, Aug 30, 2017 at 4:58 PM, Brian Foster <bfoster@redhat.com> wrote:
> >> > On Mon, Aug 21, 2017 at 10:24:32PM +0200, Ingard - wrote:
> >> >> On Mon, Aug 21, 2017 at 5:51 PM, Brian Foster <bfoster@redhat.com> wrote:
> >> >> > On Mon, Aug 21, 2017 at 02:08:43PM +0200, Ingard - wrote:
> >> >> >> On Fri, Aug 18, 2017 at 2:17 PM, Brian Foster <bfoster@redhat.com> wrote:
> >> >> >> > On Fri, Aug 18, 2017 at 07:02:24AM -0500, Bill O'Donnell wrote:
> >> >> >> >> On Fri, Aug 18, 2017 at 01:56:31PM +0200, Ingard - wrote:
> >> >> >> >> > After a server crash we've encountered a corrupt xfs filesystem. When
> >> >> >> >> > trying to mount said filesystem normally the system hangs.
> >> >> >> >> > This was initially on a ubuntu trusty server with 3.13 kernel with
> >> >> >> >> > xfsprogs 3.1.9
> >> >> >> >> >
> >> >> >> >> > We've installed a newer kernel (4.4.0-92) and compiled xfsprogs v
> >> >> >> >> > 4.12.0 from source. We're still not able to mount the filesystem (and
> >> >> >> >> > replay the log) normally.
> >> >> >> >> > We are able to mount it -o ro,norecovery, but we're reluctant to do
> >> >> >> >> > xfs_repair -L without trying everything we can first. The filesystem
> >> >> >> >> > is browsable albeit a few paths which gives an error : "Structure
> >> >> >> >> > needs cleaning"
> >> >> >> >> >
> >> >> >> >> > Does anyone have any advice as to how we might recover/repair the
> >> >> >> >> > corrupt log so we can replay it? Or is xfs_repair -L the only way
> >> >> >> >> > forward?
> >> >> >> >>
> >> >> >> >> Can you try xfs_repair -n (only scans the fs and reports what repairs
> >> >> >> >> would be made)?
> >> >> >> >>
> >> >> >> >
> >> >> >> > An xfs_metadump of the fs might be useful as well. Then we can see if we
> >> >> >> > can reproduce the mount hang on latest kernels and if so, potentially
> >> >> >> > try and root cause it.
> >> >> >> >
> >> >> >> > Brian
> >> >> >>
> >> >> >> Here is a link for the metadump :
> >> >> >> https://www.jottacloud.com/p/ingardme/95ec2e45ba80431d962345981d38bdff
> >> >> >
> >> >> > This points to a 29GB image file, apparently uncompressed..? Could you
> >> >> > upload a compressed file? Thanks.
> >> >>
> >> >> Hi. Sorry about that. Didnt realize the output would be compressable.
> >> >> Here is a link to the compressed tgz (6G)
> >> >> https://www.jottacloud.com/p/ingardme/cac6939649e14b98b928647f5222a2ae
> >> >>
> >> >
> >> > I finally played around with this image a bit. Note that mount does not
> >> > hang on latest kernels. Instead, log recovery emits a torn write message
> >> > due to a bad crc at the head of the log and then ultimately fails due to
> >> > a bad crc at the tail of the log. I ran a couple experiments to skip the
> >> > bad crc records and/or to completely ignore all bad crc's and both still
> >> > either fail to mount (due to other corruption) or continue to show
> >> > corruption in the recovered fs.
> >> >
> >> > It's not clear to me what would have caused this corruption or log
> >> > state. Have you encountered any corruption before? If not, is this kind
> >> > of crash or unclean shutdown of the server an uncommon event?
> >> We failed to notice the log messages of corrupt fs at first. After a
> >> few days of these messages the filesystem got shut down due to
> >> excessive? corruption.
> >> At that point we tried to reboot normally, but ended up with having to
> >> do a hard reset of the server.
> >> It is not clear to us either why the corruption happened in the first
> >> place either. The underlying raid has been in optimal state the whole
> >> time
> >>
> >
> > Ok, so corruption was the first problem. If the filesystem shutdown with
> > something other than a log I/O error, chances are the log was flushed at
> > that time. It is strange that log records end up corrupted, though not
> > terribly out of the ordinary for the mount to ultimately fail if
> > recovery stumbled over existing on-disk corruption, for instance.
> > An xfs_repair was probably a foregone conclusion given the corruption
> > started on disk, anyways.
> 
> Out of curiosity, how long did the xfs_mdrestore command run ? I'm
> pushing 20ish hours now and noticed the following in kern.log :
> 2017-09-01T08:47:23.414139+02:00 dn-238 kernel: [1278740.983304] XFS:
> xfs_mdrestore(5176) possible memory allocation deadlock size 37136 in
> kmem_alloc (mode:0x2400240)
> 

Heh. It certainly wasn't quick since it had to restore ~30GB or so of
metadata, but it didn't take that long. If I had to guess, I'd say it
restored within an hour.

It seems like you're running into the in-core extent list problem, which
causes pain for highly sparse or fragmented files because we store the
entire extent list in memory. An fiemap of the restored image I have
lying around shows over 1.5m extents. :/ You may need a box with more
RAM (I had 32GB) or otherwise find another large enough block device to
use the metadump. :/ If you had to bypass that step, you could at least
run 'xfs_repair -n' on the original fs to see whether repair runs to
completion in your environment.

Brian

> ingard
> 
> >
> > Brian
> >
> >> >
> >> > That aside, I think the best course of action is to run 'xfs_repair -L'
> >> > on the fs. I ran a v4.12 version against the metadump image and it
> >> > successfully repaired the fs. I've attached the repair output for
> >> > reference, but I would recommend to first restore your metadump to a
> >> > temporary location, attempt to repair that and examine the results
> >> > before repairing the original fs. Note that the metadump will not have
> >> > any file content, but will represent which files might be cleared, moved
> >> > to lost+found, etc.
> >> Ok. Thanks for looking into it. We'll proceed with the suggested
> >> course of action.
> >>
> >> ingard
> >> >
> >> > Brian
> >> >
> >> >> >
> >> >> > Brian
> >> >> >
> >> >> >> And the repair -n output :
> >> >> >> https://www.jottacloud.com/p/ingardme/0205c6ca6f7e495ebcda5f255b96f63d
> >> >> >>
> >> >> >> kind regards
> >> >> >> ingard
> >> >> >>
> >> >> >> >
> >> >> >> >> Thanks-
> >> >> >> >> Bill
> >> >> >> >>
> >> >> >> >>
> >> >> >> >> >
> >> >> >> >> >
> >> >> >> >> > Excerpt from kern.log:
> >> >> >> >> > 2017-08-17T13:40:41.122121+02:00 dn-238 kernel: [  294.300347] XFS
> >> >> >> >> > (sdd1): Mounting V4 filesystem in no-recovery mode. Filesystem will be
> >> >> >> >> > inconsistent.
> >> >> >> >> >
> >> >> >> >> > 2017-08-17T17:04:54.794194+02:00 dn-238 kernel: [12548.400260] XFS
> >> >> >> >> > (sdd1): Metadata corruption detected at xfs_inode_buf_verify+0x6f/0xd0
> >> >> >> >> > [xfs], xfs_inode block 0x81c9c210
> >> >> >> >> > 2017-08-17T17:04:54.794216+02:00 dn-238 kernel: [12548.400342] XFS
> >> >> >> >> > (sdd1): Unmount and run xfs_repair
> >> >> >> >> > 2017-08-17T17:04:54.794218+02:00 dn-238 kernel: [12548.400374] XFS
> >> >> >> >> > (sdd1): First 64 bytes of corrupted metadata buffer:
> >> >> >> >> > 2017-08-17T17:04:54.794220+02:00 dn-238 kernel: [12548.400418]
> >> >> >> >> > ffff880171fff000: 3f 1a 33 54 5b 55 85 0b 7c f5 c6 d5 cf 51 47 41
> >> >> >> >> > ?.3T[U..|....QGA
> >> >> >> >> > 2017-08-17T17:04:54.794222+02:00 dn-238 kernel: [12548.400473]
> >> >> >> >> > ffff880171fff010: 97 ba ba 03 5c e4 02 7a e6 bc fb 5d f1 72 db c1
> >> >> >> >> > ....\..z...].r..
> >> >> >> >> > 2017-08-17T17:04:54.794223+02:00 dn-238 kernel: [12548.400527]
> >> >> >> >> > ffff880171fff020: c8 ad 3a 76 c7 e4 20 92 88 a2 35 0c 1f 36 cf b5
> >> >> >> >> > ..:v.. ...5..6..
> >> >> >> >> > 2017-08-17T17:04:54.794226+02:00 dn-238 kernel: [12548.400581]
> >> >> >> >> > ffff880171fff030: 8a bc 42 75 86 50 a0 a2 be 2c 2d 99 96 2d e1 ee
> >> >> >> >> > ..Bu.P...,-..-..
> >> >> >> >> >
> >> >> >> >> > kind regards
> >> >> >> >> > ingard
> >> >> >> >> > --
> >> >> >> >> > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> >> >> >> >> > the body of a message to majordomo@vger.kernel.org
> >> >> >> >> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >> >> >> >> --
> >> >> >> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> >> >> >> >> the body of a message to majordomo@vger.kernel.org
> >> >> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >> >> >> --
> >> >> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> >> >> >> the body of a message to majordomo@vger.kernel.org
> >> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >> >> --
> >> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> >> >> the body of a message to majordomo@vger.kernel.org
> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >> --
> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> >> the body of a message to majordomo@vger.kernel.org
> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: corrupt xfs log
  2017-09-01 11:33                   ` Brian Foster
@ 2017-09-01 15:11                     ` Darrick J. Wong
  2017-09-01 16:26                       ` Brian Foster
  0 siblings, 1 reply; 16+ messages in thread
From: Darrick J. Wong @ 2017-09-01 15:11 UTC (permalink / raw)
  To: Brian Foster; +Cc: Ingard -, Bill O'Donnell, linux-xfs

On Fri, Sep 01, 2017 at 07:33:07AM -0400, Brian Foster wrote:
> On Fri, Sep 01, 2017 at 08:48:03AM +0200, Ingard - wrote:
> > On Thu, Aug 31, 2017 at 12:20 PM, Brian Foster <bfoster@redhat.com> wrote:
> > > On Thu, Aug 31, 2017 at 09:27:52AM +0200, Ingard - wrote:
> > >> On Wed, Aug 30, 2017 at 4:58 PM, Brian Foster <bfoster@redhat.com> wrote:
> > >> > On Mon, Aug 21, 2017 at 10:24:32PM +0200, Ingard - wrote:
> > >> >> On Mon, Aug 21, 2017 at 5:51 PM, Brian Foster <bfoster@redhat.com> wrote:
> > >> >> > On Mon, Aug 21, 2017 at 02:08:43PM +0200, Ingard - wrote:
> > >> >> >> On Fri, Aug 18, 2017 at 2:17 PM, Brian Foster <bfoster@redhat.com> wrote:
> > >> >> >> > On Fri, Aug 18, 2017 at 07:02:24AM -0500, Bill O'Donnell wrote:
> > >> >> >> >> On Fri, Aug 18, 2017 at 01:56:31PM +0200, Ingard - wrote:
> > >> >> >> >> > After a server crash we've encountered a corrupt xfs filesystem. When
> > >> >> >> >> > trying to mount said filesystem normally the system hangs.
> > >> >> >> >> > This was initially on a ubuntu trusty server with 3.13 kernel with
> > >> >> >> >> > xfsprogs 3.1.9
> > >> >> >> >> >
> > >> >> >> >> > We've installed a newer kernel (4.4.0-92) and compiled xfsprogs v
> > >> >> >> >> > 4.12.0 from source. We're still not able to mount the filesystem (and
> > >> >> >> >> > replay the log) normally.
> > >> >> >> >> > We are able to mount it -o ro,norecovery, but we're reluctant to do
> > >> >> >> >> > xfs_repair -L without trying everything we can first. The filesystem
> > >> >> >> >> > is browsable albeit a few paths which gives an error : "Structure
> > >> >> >> >> > needs cleaning"
> > >> >> >> >> >
> > >> >> >> >> > Does anyone have any advice as to how we might recover/repair the
> > >> >> >> >> > corrupt log so we can replay it? Or is xfs_repair -L the only way
> > >> >> >> >> > forward?
> > >> >> >> >>
> > >> >> >> >> Can you try xfs_repair -n (only scans the fs and reports what repairs
> > >> >> >> >> would be made)?
> > >> >> >> >>
> > >> >> >> >
> > >> >> >> > An xfs_metadump of the fs might be useful as well. Then we can see if we
> > >> >> >> > can reproduce the mount hang on latest kernels and if so, potentially
> > >> >> >> > try and root cause it.
> > >> >> >> >
> > >> >> >> > Brian
> > >> >> >>
> > >> >> >> Here is a link for the metadump :
> > >> >> >> https://www.jottacloud.com/p/ingardme/95ec2e45ba80431d962345981d38bdff
> > >> >> >
> > >> >> > This points to a 29GB image file, apparently uncompressed..? Could you
> > >> >> > upload a compressed file? Thanks.
> > >> >>
> > >> >> Hi. Sorry about that. Didnt realize the output would be compressable.
> > >> >> Here is a link to the compressed tgz (6G)
> > >> >> https://www.jottacloud.com/p/ingardme/cac6939649e14b98b928647f5222a2ae
> > >> >>
> > >> >
> > >> > I finally played around with this image a bit. Note that mount does not
> > >> > hang on latest kernels. Instead, log recovery emits a torn write message
> > >> > due to a bad crc at the head of the log and then ultimately fails due to
> > >> > a bad crc at the tail of the log. I ran a couple experiments to skip the
> > >> > bad crc records and/or to completely ignore all bad crc's and both still
> > >> > either fail to mount (due to other corruption) or continue to show
> > >> > corruption in the recovered fs.
> > >> >
> > >> > It's not clear to me what would have caused this corruption or log
> > >> > state. Have you encountered any corruption before? If not, is this kind
> > >> > of crash or unclean shutdown of the server an uncommon event?
> > >> We failed to notice the log messages of corrupt fs at first. After a
> > >> few days of these messages the filesystem got shut down due to
> > >> excessive? corruption.
> > >> At that point we tried to reboot normally, but ended up with having to
> > >> do a hard reset of the server.
> > >> It is not clear to us either why the corruption happened in the first
> > >> place either. The underlying raid has been in optimal state the whole
> > >> time
> > >>
> > >
> > > Ok, so corruption was the first problem. If the filesystem shutdown with
> > > something other than a log I/O error, chances are the log was flushed at
> > > that time. It is strange that log records end up corrupted, though not
> > > terribly out of the ordinary for the mount to ultimately fail if
> > > recovery stumbled over existing on-disk corruption, for instance.
> > > An xfs_repair was probably a foregone conclusion given the corruption
> > > started on disk, anyways.
> > 
> > Out of curiosity, how long did the xfs_mdrestore command run ? I'm
> > pushing 20ish hours now and noticed the following in kern.log :
> > 2017-09-01T08:47:23.414139+02:00 dn-238 kernel: [1278740.983304] XFS:
> > xfs_mdrestore(5176) possible memory allocation deadlock size 37136 in
> > kmem_alloc (mode:0x2400240)
> > 
> 
> Heh. It certainly wasn't quick since it had to restore ~30GB or so of
> metadata, but it didn't take that long. If I had to guess, I'd say it
> restored within an hour.
> 
> It seems like you're running into the in-core extent list problem, which
> causes pain for highly sparse or fragmented files because we store the
> entire extent list in memory. An fiemap of the restored image I have
> lying around shows over 1.5m extents. :/ You may need a box with more
> RAM (I had 32GB) or otherwise find another large enough block device to
> use the metadump. :/ If you had to bypass that step, you could at least
> run 'xfs_repair -n' on the original fs to see whether repair runs to
> completion in your environment.

/me wonders if it'd help if mdrestore had a command line arg to
fallocate the target file beforehand (lots of wasted space but fewer
extents) or set an extent size hint (only useful if metadata isn't
fragmented) but yes we should fix the incore extent cache memory usage
problem.

--D

> 
> Brian
> 
> > ingard
> > 
> > >
> > > Brian
> > >
> > >> >
> > >> > That aside, I think the best course of action is to run 'xfs_repair -L'
> > >> > on the fs. I ran a v4.12 version against the metadump image and it
> > >> > successfully repaired the fs. I've attached the repair output for
> > >> > reference, but I would recommend to first restore your metadump to a
> > >> > temporary location, attempt to repair that and examine the results
> > >> > before repairing the original fs. Note that the metadump will not have
> > >> > any file content, but will represent which files might be cleared, moved
> > >> > to lost+found, etc.
> > >> Ok. Thanks for looking into it. We'll proceed with the suggested
> > >> course of action.
> > >>
> > >> ingard
> > >> >
> > >> > Brian
> > >> >
> > >> >> >
> > >> >> > Brian
> > >> >> >
> > >> >> >> And the repair -n output :
> > >> >> >> https://www.jottacloud.com/p/ingardme/0205c6ca6f7e495ebcda5f255b96f63d
> > >> >> >>
> > >> >> >> kind regards
> > >> >> >> ingard
> > >> >> >>
> > >> >> >> >
> > >> >> >> >> Thanks-
> > >> >> >> >> Bill
> > >> >> >> >>
> > >> >> >> >>
> > >> >> >> >> >
> > >> >> >> >> >
> > >> >> >> >> > Excerpt from kern.log:
> > >> >> >> >> > 2017-08-17T13:40:41.122121+02:00 dn-238 kernel: [  294.300347] XFS
> > >> >> >> >> > (sdd1): Mounting V4 filesystem in no-recovery mode. Filesystem will be
> > >> >> >> >> > inconsistent.
> > >> >> >> >> >
> > >> >> >> >> > 2017-08-17T17:04:54.794194+02:00 dn-238 kernel: [12548.400260] XFS
> > >> >> >> >> > (sdd1): Metadata corruption detected at xfs_inode_buf_verify+0x6f/0xd0
> > >> >> >> >> > [xfs], xfs_inode block 0x81c9c210
> > >> >> >> >> > 2017-08-17T17:04:54.794216+02:00 dn-238 kernel: [12548.400342] XFS
> > >> >> >> >> > (sdd1): Unmount and run xfs_repair
> > >> >> >> >> > 2017-08-17T17:04:54.794218+02:00 dn-238 kernel: [12548.400374] XFS
> > >> >> >> >> > (sdd1): First 64 bytes of corrupted metadata buffer:
> > >> >> >> >> > 2017-08-17T17:04:54.794220+02:00 dn-238 kernel: [12548.400418]
> > >> >> >> >> > ffff880171fff000: 3f 1a 33 54 5b 55 85 0b 7c f5 c6 d5 cf 51 47 41
> > >> >> >> >> > ?.3T[U..|....QGA
> > >> >> >> >> > 2017-08-17T17:04:54.794222+02:00 dn-238 kernel: [12548.400473]
> > >> >> >> >> > ffff880171fff010: 97 ba ba 03 5c e4 02 7a e6 bc fb 5d f1 72 db c1
> > >> >> >> >> > ....\..z...].r..
> > >> >> >> >> > 2017-08-17T17:04:54.794223+02:00 dn-238 kernel: [12548.400527]
> > >> >> >> >> > ffff880171fff020: c8 ad 3a 76 c7 e4 20 92 88 a2 35 0c 1f 36 cf b5
> > >> >> >> >> > ..:v.. ...5..6..
> > >> >> >> >> > 2017-08-17T17:04:54.794226+02:00 dn-238 kernel: [12548.400581]
> > >> >> >> >> > ffff880171fff030: 8a bc 42 75 86 50 a0 a2 be 2c 2d 99 96 2d e1 ee
> > >> >> >> >> > ..Bu.P...,-..-..
> > >> >> >> >> >
> > >> >> >> >> > kind regards
> > >> >> >> >> > ingard
> > >> >> >> >> > --
> > >> >> >> >> > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> > >> >> >> >> > the body of a message to majordomo@vger.kernel.org
> > >> >> >> >> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > >> >> >> >> --
> > >> >> >> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> > >> >> >> >> the body of a message to majordomo@vger.kernel.org
> > >> >> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > >> >> >> --
> > >> >> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> > >> >> >> the body of a message to majordomo@vger.kernel.org
> > >> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > >> >> --
> > >> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> > >> >> the body of a message to majordomo@vger.kernel.org
> > >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > >> --
> > >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> > >> the body of a message to majordomo@vger.kernel.org
> > >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > --
> > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> > the body of a message to majordomo@vger.kernel.org
> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: corrupt xfs log
  2017-09-01 15:11                     ` Darrick J. Wong
@ 2017-09-01 16:26                       ` Brian Foster
  0 siblings, 0 replies; 16+ messages in thread
From: Brian Foster @ 2017-09-01 16:26 UTC (permalink / raw)
  To: Darrick J. Wong; +Cc: Ingard -, Bill O'Donnell, linux-xfs

On Fri, Sep 01, 2017 at 08:11:19AM -0700, Darrick J. Wong wrote:
> On Fri, Sep 01, 2017 at 07:33:07AM -0400, Brian Foster wrote:
> > On Fri, Sep 01, 2017 at 08:48:03AM +0200, Ingard - wrote:
> > > On Thu, Aug 31, 2017 at 12:20 PM, Brian Foster <bfoster@redhat.com> wrote:
> > > > On Thu, Aug 31, 2017 at 09:27:52AM +0200, Ingard - wrote:
> > > >> On Wed, Aug 30, 2017 at 4:58 PM, Brian Foster <bfoster@redhat.com> wrote:
> > > >> > On Mon, Aug 21, 2017 at 10:24:32PM +0200, Ingard - wrote:
> > > >> >> On Mon, Aug 21, 2017 at 5:51 PM, Brian Foster <bfoster@redhat.com> wrote:
> > > >> >> > On Mon, Aug 21, 2017 at 02:08:43PM +0200, Ingard - wrote:
> > > >> >> >> On Fri, Aug 18, 2017 at 2:17 PM, Brian Foster <bfoster@redhat.com> wrote:
> > > >> >> >> > On Fri, Aug 18, 2017 at 07:02:24AM -0500, Bill O'Donnell wrote:
> > > >> >> >> >> On Fri, Aug 18, 2017 at 01:56:31PM +0200, Ingard - wrote:
> > > >> >> >> >> > After a server crash we've encountered a corrupt xfs filesystem. When
> > > >> >> >> >> > trying to mount said filesystem normally the system hangs.
> > > >> >> >> >> > This was initially on a ubuntu trusty server with 3.13 kernel with
> > > >> >> >> >> > xfsprogs 3.1.9
> > > >> >> >> >> >
> > > >> >> >> >> > We've installed a newer kernel (4.4.0-92) and compiled xfsprogs v
> > > >> >> >> >> > 4.12.0 from source. We're still not able to mount the filesystem (and
> > > >> >> >> >> > replay the log) normally.
> > > >> >> >> >> > We are able to mount it -o ro,norecovery, but we're reluctant to do
> > > >> >> >> >> > xfs_repair -L without trying everything we can first. The filesystem
> > > >> >> >> >> > is browsable albeit a few paths which gives an error : "Structure
> > > >> >> >> >> > needs cleaning"
> > > >> >> >> >> >
> > > >> >> >> >> > Does anyone have any advice as to how we might recover/repair the
> > > >> >> >> >> > corrupt log so we can replay it? Or is xfs_repair -L the only way
> > > >> >> >> >> > forward?
> > > >> >> >> >>
> > > >> >> >> >> Can you try xfs_repair -n (only scans the fs and reports what repairs
> > > >> >> >> >> would be made)?
> > > >> >> >> >>
> > > >> >> >> >
> > > >> >> >> > An xfs_metadump of the fs might be useful as well. Then we can see if we
> > > >> >> >> > can reproduce the mount hang on latest kernels and if so, potentially
> > > >> >> >> > try and root cause it.
> > > >> >> >> >
> > > >> >> >> > Brian
> > > >> >> >>
> > > >> >> >> Here is a link for the metadump :
> > > >> >> >> https://www.jottacloud.com/p/ingardme/95ec2e45ba80431d962345981d38bdff
> > > >> >> >
> > > >> >> > This points to a 29GB image file, apparently uncompressed..? Could you
> > > >> >> > upload a compressed file? Thanks.
> > > >> >>
> > > >> >> Hi. Sorry about that. Didnt realize the output would be compressable.
> > > >> >> Here is a link to the compressed tgz (6G)
> > > >> >> https://www.jottacloud.com/p/ingardme/cac6939649e14b98b928647f5222a2ae
> > > >> >>
> > > >> >
> > > >> > I finally played around with this image a bit. Note that mount does not
> > > >> > hang on latest kernels. Instead, log recovery emits a torn write message
> > > >> > due to a bad crc at the head of the log and then ultimately fails due to
> > > >> > a bad crc at the tail of the log. I ran a couple experiments to skip the
> > > >> > bad crc records and/or to completely ignore all bad crc's and both still
> > > >> > either fail to mount (due to other corruption) or continue to show
> > > >> > corruption in the recovered fs.
> > > >> >
> > > >> > It's not clear to me what would have caused this corruption or log
> > > >> > state. Have you encountered any corruption before? If not, is this kind
> > > >> > of crash or unclean shutdown of the server an uncommon event?
> > > >> We failed to notice the log messages of corrupt fs at first. After a
> > > >> few days of these messages the filesystem got shut down due to
> > > >> excessive? corruption.
> > > >> At that point we tried to reboot normally, but ended up with having to
> > > >> do a hard reset of the server.
> > > >> It is not clear to us either why the corruption happened in the first
> > > >> place either. The underlying raid has been in optimal state the whole
> > > >> time
> > > >>
> > > >
> > > > Ok, so corruption was the first problem. If the filesystem shutdown with
> > > > something other than a log I/O error, chances are the log was flushed at
> > > > that time. It is strange that log records end up corrupted, though not
> > > > terribly out of the ordinary for the mount to ultimately fail if
> > > > recovery stumbled over existing on-disk corruption, for instance.
> > > > An xfs_repair was probably a foregone conclusion given the corruption
> > > > started on disk, anyways.
> > > 
> > > Out of curiosity, how long did the xfs_mdrestore command run ? I'm
> > > pushing 20ish hours now and noticed the following in kern.log :
> > > 2017-09-01T08:47:23.414139+02:00 dn-238 kernel: [1278740.983304] XFS:
> > > xfs_mdrestore(5176) possible memory allocation deadlock size 37136 in
> > > kmem_alloc (mode:0x2400240)
> > > 
> > 
> > Heh. It certainly wasn't quick since it had to restore ~30GB or so of
> > metadata, but it didn't take that long. If I had to guess, I'd say it
> > restored within an hour.
> > 
> > It seems like you're running into the in-core extent list problem, which
> > causes pain for highly sparse or fragmented files because we store the
> > entire extent list in memory. An fiemap of the restored image I have
> > lying around shows over 1.5m extents. :/ You may need a box with more
> > RAM (I had 32GB) or otherwise find another large enough block device to
> > use the metadump. :/ If you had to bypass that step, you could at least
> > run 'xfs_repair -n' on the original fs to see whether repair runs to
> > completion in your environment.
> 
> /me wonders if it'd help if mdrestore had a command line arg to
> fallocate the target file beforehand (lots of wasted space but fewer
> extents) or set an extent size hint (only useful if metadata isn't
> fragmented) but yes we should fix the incore extent cache memory usage
> problem.
> 

That's an interesting idea. I suppose one could set a really large
extent size hint on the directory to try and achieve a similar outcome
as an initial fallocate.

Eh, though looking at the restored image, the original fs is 88TB (with
only 4TB free space), so that might not be so useful in this particular
case. Heh, I have another metadump lying around of a significantly
smaller fs with what looks like about double the amount of metadata
based on the metadump size and extent count. :P There may be a middle
ground here where an fallocate could be useful.

Brian

> --D
> 
> > 
> > Brian
> > 
> > > ingard
> > > 
> > > >
> > > > Brian
> > > >
> > > >> >
> > > >> > That aside, I think the best course of action is to run 'xfs_repair -L'
> > > >> > on the fs. I ran a v4.12 version against the metadump image and it
> > > >> > successfully repaired the fs. I've attached the repair output for
> > > >> > reference, but I would recommend to first restore your metadump to a
> > > >> > temporary location, attempt to repair that and examine the results
> > > >> > before repairing the original fs. Note that the metadump will not have
> > > >> > any file content, but will represent which files might be cleared, moved
> > > >> > to lost+found, etc.
> > > >> Ok. Thanks for looking into it. We'll proceed with the suggested
> > > >> course of action.
> > > >>
> > > >> ingard
> > > >> >
> > > >> > Brian
> > > >> >
> > > >> >> >
> > > >> >> > Brian
> > > >> >> >
> > > >> >> >> And the repair -n output :
> > > >> >> >> https://www.jottacloud.com/p/ingardme/0205c6ca6f7e495ebcda5f255b96f63d
> > > >> >> >>
> > > >> >> >> kind regards
> > > >> >> >> ingard
> > > >> >> >>
> > > >> >> >> >
> > > >> >> >> >> Thanks-
> > > >> >> >> >> Bill
> > > >> >> >> >>
> > > >> >> >> >>
> > > >> >> >> >> >
> > > >> >> >> >> >
> > > >> >> >> >> > Excerpt from kern.log:
> > > >> >> >> >> > 2017-08-17T13:40:41.122121+02:00 dn-238 kernel: [  294.300347] XFS
> > > >> >> >> >> > (sdd1): Mounting V4 filesystem in no-recovery mode. Filesystem will be
> > > >> >> >> >> > inconsistent.
> > > >> >> >> >> >
> > > >> >> >> >> > 2017-08-17T17:04:54.794194+02:00 dn-238 kernel: [12548.400260] XFS
> > > >> >> >> >> > (sdd1): Metadata corruption detected at xfs_inode_buf_verify+0x6f/0xd0
> > > >> >> >> >> > [xfs], xfs_inode block 0x81c9c210
> > > >> >> >> >> > 2017-08-17T17:04:54.794216+02:00 dn-238 kernel: [12548.400342] XFS
> > > >> >> >> >> > (sdd1): Unmount and run xfs_repair
> > > >> >> >> >> > 2017-08-17T17:04:54.794218+02:00 dn-238 kernel: [12548.400374] XFS
> > > >> >> >> >> > (sdd1): First 64 bytes of corrupted metadata buffer:
> > > >> >> >> >> > 2017-08-17T17:04:54.794220+02:00 dn-238 kernel: [12548.400418]
> > > >> >> >> >> > ffff880171fff000: 3f 1a 33 54 5b 55 85 0b 7c f5 c6 d5 cf 51 47 41
> > > >> >> >> >> > ?.3T[U..|....QGA
> > > >> >> >> >> > 2017-08-17T17:04:54.794222+02:00 dn-238 kernel: [12548.400473]
> > > >> >> >> >> > ffff880171fff010: 97 ba ba 03 5c e4 02 7a e6 bc fb 5d f1 72 db c1
> > > >> >> >> >> > ....\..z...].r..
> > > >> >> >> >> > 2017-08-17T17:04:54.794223+02:00 dn-238 kernel: [12548.400527]
> > > >> >> >> >> > ffff880171fff020: c8 ad 3a 76 c7 e4 20 92 88 a2 35 0c 1f 36 cf b5
> > > >> >> >> >> > ..:v.. ...5..6..
> > > >> >> >> >> > 2017-08-17T17:04:54.794226+02:00 dn-238 kernel: [12548.400581]
> > > >> >> >> >> > ffff880171fff030: 8a bc 42 75 86 50 a0 a2 be 2c 2d 99 96 2d e1 ee
> > > >> >> >> >> > ..Bu.P...,-..-..
> > > >> >> >> >> >
> > > >> >> >> >> > kind regards
> > > >> >> >> >> > ingard
> > > >> >> >> >> > --
> > > >> >> >> >> > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> > > >> >> >> >> > the body of a message to majordomo@vger.kernel.org
> > > >> >> >> >> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > > >> >> >> >> --
> > > >> >> >> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> > > >> >> >> >> the body of a message to majordomo@vger.kernel.org
> > > >> >> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > > >> >> >> --
> > > >> >> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> > > >> >> >> the body of a message to majordomo@vger.kernel.org
> > > >> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > > >> >> --
> > > >> >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> > > >> >> the body of a message to majordomo@vger.kernel.org
> > > >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > > >> --
> > > >> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> > > >> the body of a message to majordomo@vger.kernel.org
> > > >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > > --
> > > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> > > the body of a message to majordomo@vger.kernel.org
> > > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > --
> > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> > the body of a message to majordomo@vger.kernel.org
> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 16+ messages in thread

end of thread, other threads:[~2017-09-01 16:26 UTC | newest]

Thread overview: 16+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2017-08-18 11:56 corrupt xfs log Ingard -
2017-08-18 12:02 ` Bill O'Donnell
2017-08-18 12:17   ` Brian Foster
2017-08-21 12:08     ` Ingard -
2017-08-21 15:51       ` Brian Foster
2017-08-21 20:24         ` Ingard -
2017-08-28  8:56           ` Ingard -
2017-08-28 10:59             ` Brian Foster
2017-08-30 14:58           ` Brian Foster
2017-08-31  7:27             ` Ingard -
2017-08-31 10:20               ` Brian Foster
2017-09-01  6:48                 ` Ingard -
2017-09-01 11:33                   ` Brian Foster
2017-09-01 15:11                     ` Darrick J. Wong
2017-09-01 16:26                       ` Brian Foster
2017-08-18 13:43   ` Ingard -

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.