All of lore.kernel.org
 help / color / mirror / Atom feed
* xfs corruption
@ 2015-09-03 11:09 Danny Shavit
  2015-09-03 13:22 ` Eric Sandeen
  0 siblings, 1 reply; 18+ messages in thread
From: Danny Shavit @ 2015-09-03 11:09 UTC (permalink / raw)
  To: xfs; +Cc: Alex Lyakas


[-- Attachment #1.1: Type: text/plain, Size: 632 bytes --]

Hi Dave,

We couple of more xfs corruption that we would like to share:

1. This is an interesting one, since xfs reported corruption but when
running xfs_repair, no error was found.
Attached is the kernel log section regarding the corruption (6458).
Does xfs_repair explicitly read data from the disk? In such case it might
be a memory corruption. Are you familiar with such cases?

2. xfs corruption occurred suddenly with no apparent external event.
Attached are xfs_repair and kernel logs are.
Xfs dump can be found in:
https://zadarastorage-public.s3.amazonaws.com/xfs/82.metadump.gz




-- 
Thanks,
Danny Shavit
ZadaraStorage

[-- Attachment #1.2: Type: text/html, Size: 1015 bytes --]

[-- Attachment #2: 6458-kernel.log --]
[-- Type: application/octet-stream, Size: 2688 bytes --]

The XFS volumes then entered a corrupted state:

Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.307743] XFS (dm-39): Internal error xfs_allocbt_verify at line 330 of file /mnt/share/builds/14.11--3.8.13-030813-generic/2015-04-29_10-45-42--14.11-1601-124/src/zadara-btrfs/fs/xfs/xfs_alloc_btree.c.  Caller 0xffffffffa064e9ce
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.307743]
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.314446] Pid: 25231, comm: kworker/0:0H Tainted: GF       W  O 3.8.13-030813-generic #201305111843
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.314449] Call Trace:
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.314487]  [<ffffffffa0631baf>] xfs_error_report+0x3f/0x50 [xfs]
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.314502]  [<ffffffffa064e9ce>] ? xfs_allocbt_read_verify+0xe/0x10 [xfs]
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.314514]  [<ffffffffa0631c1e>] xfs_corruption_error+0x5e/0x90 [xfs]
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.314528]  [<ffffffffa064e862>] xfs_allocbt_verify+0x92/0x1e0 [xfs]
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.314540]  [<ffffffffa064e9ce>] ? xfs_allocbt_read_verify+0xe/0x10 [xfs]
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.314547]  [<ffffffff810135aa>] ? __switch_to+0x12a/0x4a0
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.314551]  [<ffffffff81096cd8>] ? set_next_entity+0xa8/0xc0
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.314566]  [<ffffffffa064e9ce>] xfs_allocbt_read_verify+0xe/0x10 [xfs]
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.315251]  [<ffffffffa062f48f>] xfs_buf_iodone_work+0x3f/0xa0 [xfs]
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.315255]  [<ffffffff81078b81>] process_one_work+0x141/0x490
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.315257]  [<ffffffff81079b48>] worker_thread+0x168/0x400
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.315259]  [<ffffffff810799e0>] ? manage_workers+0x120/0x120
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.315262]  [<ffffffff8107f050>] kthread+0xc0/0xd0
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.315265]  [<ffffffff8107ef90>] ? flush_kthread_worker+0xb0/0xb0
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.315270]  [<ffffffff816f61ec>] ret_from_fork+0x7c/0xb0
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.315273]  [<ffffffff8107ef90>] ? flush_kthread_worker+0xb0/0xb0
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.315275] XFS (dm-39): Corruption detected. Unmount and run xfs_repair
Aug 27 01:01:34 vsa-0000014e-vc-0 kernel: [3507105.316706] XFS (dm-39): metadata I/O error: block 0x41a6eff8 ("xfs_trans_read_buf_map") error 117 numblks 8

[-- Attachment #3: 6442-82-xfs_repair.log --]
[-- Type: application/octet-stream, Size: 6009 bytes --]

root@vsa-00000110-vc-0:~# xfs_repair /dev/dm-82
Phase 1 - find and verify superblock...
Phase 2 - using internal log
        - zero log...
ERROR: The filesystem has valuable metadata changes in a log which needs to
be replayed.  Mount the filesystem to replay the log, and unmount it before
re-running xfs_repair.  If you are unable to mount the filesystem, then use
the -L option to destroy the log and attempt a repair.
Note that destroying the log may cause corruption -- please attempt a mount
of the filesystem before doing this.
root@vsa-00000110-vc-0:~# xfs_repair -L /dev/dm-82
Phase 1 - find and verify superblock...
Phase 2 - using internal log
        - zero log...
ALERT: The filesystem has valuable metadata changes in a log which is being
destroyed because the -L option was used.
        - scan filesystem freespace and inode maps...
agi unlinked bucket 1 is 12580353 in ag 3 (inode=213906945)
sb_icount 1226496, counted 1227776
sb_ifree 292180, counted 297082
sb_fdblocks 31182739, counted 55158044
        - found root inode chunk
Phase 3 - for each AG...
        - scan and clear agi unlinked lists...
        - process known inodes and perform inode discovery...
        - agno = 0
        - agno = 1
        - agno = 2
7f8d22a2c700: Badness in key lookup (length)
bp=(bno 84932992, len 16384 bytes) key=(bno 84932992, len 8192 bytes)
        - agno = 3
bad magic # 0xeabb123a in inode 213906945 (data fork) bmbt block 13369242
bad data fork in inode 213906945
cleared inode 213906945
clearing forw/back pointers in block 0 for attributes in inode 213906953
bad attribute leaf magic # 0xbc6c for dir ino 213906953
problem with attribute contents in inode 213906953
clearing inode 213906953 attributes
correcting nblocks for inode 213906953, was 66 - counted 65
clearing forw/back pointers in block 0 for attributes in inode 213906954
bad attribute leaf magic # 0xde72 for dir ino 213906954
problem with attribute contents in inode 213906954
clearing inode 213906954 attributes
correcting nblocks for inode 213906954, was 2 - counted 1
clearing forw/back pointers in block 0 for attributes in inode 213906960
bad attribute leaf magic # 0xd0eb for dir ino 213906960
problem with attribute contents in inode 213906960
clearing inode 213906960 attributes
correcting nblocks for inode 213906960, was 4 - counted 3
clearing forw/back pointers in block 0 for attributes in inode 213906961
bad attribute leaf magic # 0xb876 for dir ino 213906961
problem with attribute contents in inode 213906961
clearing inode 213906961 attributes
correcting nblocks for inode 213906961, was 5 - counted 4
        - agno = 4
        - agno = 5
clearing forw/back pointers in block 0 for attributes in inode 347235105
bad attribute leaf magic # 0xb033 for dir ino 347235105
problem with attribute contents in inode 347235105
clearing inode 347235105 attributes
correcting nblocks for inode 347235105, was 9 - counted 8
clearing forw/back pointers in block 0 for attributes in inode 347235106
bad attribute leaf magic # 0xe13 for dir ino 347235106
problem with attribute contents in inode 347235106
clearing inode 347235106 attributes
correcting nblocks for inode 347235106, was 9 - counted 8
        - agno = 6
        - agno = 7
clearing forw/back pointers in block 0 for attributes in inode 478759702
bad attribute leaf magic # 0xa065 for dir ino 478759702
problem with attribute contents in inode 478759702
clearing inode 478759702 attributes
correcting nblocks for inode 478759702, was 1561 - counted 1560
        - agno = 8
        - agno = 9
        - agno = 10
        - agno = 11
        - agno = 12
        - agno = 13
        - agno = 14
        - agno = 15
        - agno = 16
        - agno = 17
        - agno = 18
        - agno = 19
        - agno = 20
        - agno = 21
        - agno = 22
        - agno = 23
        - agno = 24
        - process newly discovered inodes...
Phase 4 - check for duplicate blocks...
        - setting up duplicate extent list...
        - check for inodes claiming duplicate blocks...
        - agno = 0
        - agno = 1
        - agno = 2
        - agno = 3
bad magic # 0x58465342 in inode 213906953 (data fork) bmbt block 0
bad data fork in inode 213906953
cleared inode 213906953
bad attribute format 1 in inode 213906954, resetting value
bad attribute format 1 in inode 213906960, resetting value
bad attribute format 1 in inode 213906961, resetting value
        - agno = 4
        - agno = 5
bad attribute format 1 in inode 347235105, resetting value
bad attribute format 1 in inode 347235106, resetting value
        - agno = 6
        - agno = 7
bad magic # 0x58465342 in inode 478759702 (data fork) bmbt block 0
bad data fork in inode 478759702
cleared inode 478759702
        - agno = 8
        - agno = 9
        - agno = 10
        - agno = 11
        - agno = 12
        - agno = 13
        - agno = 14
        - agno = 15
        - agno = 16
        - agno = 17
        - agno = 18
        - agno = 19
        - agno = 20
        - agno = 21
        - agno = 22
        - agno = 23
        - agno = 24
Phase 5 - rebuild AG headers and trees...
        - reset superblock...
7f8d24478740: Badness in key lookup (length)
bp=(bno 0, len 4096 bytes) key=(bno 0, len 512 bytes)
Phase 6 - check inode connectivity...
        - resetting contents of realtime bitmap and summary inodes
        - traversing filesystem ...
entry "3247.png" in directory inode 201326924 points to free inode 213906953
bad hash table for directory inode 201326924 (no data entry): rebuilding
rebuilding directory inode 201326924
entry "0251050.NWB" in directory inode 469762366 points to free inode 478759702
rebuilding directory inode 469762366
        - traversal finished ...
        - moving disconnected inodes to lost+found ...
Phase 7 - verify and correct link counts...
done
root@vsa-00000110-vc-0:~# echo $?
0
root@vsa-00000110-vc-0:~# crm_mon
Connection to the CIB terminated
Reconnecting...root@vsa-00000110-vc-0:~# less /var/log/kern.log
root@vsa-00000110-vc-0:~#

[-- Attachment #4: dm-82-kernel.log --]
[-- Type: application/octet-stream, Size: 2549 bytes --]

Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.685353] ffff88010ec36000: ea bb 12 3a 5f 44 01 a8 b9 2a 80 10 b3 a7 d5 af  ...:_D...*......
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.686568] XFS (dm-82): Internal error xfs_bmbt_verify at line 747 of file /mnt/share/builds/14.11--3.8.13-030813-generic/2015-06-17_03-30-37--14.11-1601-129/src/zadara-btrfs/fs/xfs/xfs_bmap_btree.c.  Caller 0xffffffffa07779ee
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.686568] 
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.689393] Pid: 17063, comm: kworker/0:1H Tainted: GF       W  O 3.8.13-030813-generic #201305111843
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.689395] Call Trace:
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.689443]  [<ffffffffa0746baf>] xfs_error_report+0x3f/0x50 [xfs]
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.689491]  [<ffffffffa07779ee>] ? xfs_bmbt_read_verify+0xe/0x10 [xfs]
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.689503]  [<ffffffffa0746c1e>] xfs_corruption_error+0x5e/0x90 [xfs]
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.689517]  [<ffffffffa0777867>] xfs_bmbt_verify+0x77/0x1e0 [xfs]
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.689535]  [<ffffffffa07779ee>] ? xfs_bmbt_read_verify+0xe/0x10 [xfs]
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.689548]  [<ffffffffa07779ee>] xfs_bmbt_read_verify+0xe/0x10 [xfs]
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.689558]  [<ffffffffa074448f>] xfs_buf_iodone_work+0x3f/0xa0 [xfs]
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.689564]  [<ffffffff81078b81>] process_one_work+0x141/0x490
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.689566]  [<ffffffff81079b48>] worker_thread+0x168/0x400
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.689569]  [<ffffffff810799e0>] ? manage_workers+0x120/0x120
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.689571]  [<ffffffff8107f050>] kthread+0xc0/0xd0
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.689574]  [<ffffffff8107ef90>] ? flush_kthread_worker+0xb0/0xb0
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.689579]  [<ffffffff816f61ec>] ret_from_fork+0x7c/0xb0
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.689582]  [<ffffffff8107ef90>] ? flush_kthread_worker+0xb0/0xb0
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.689584] XFS (dm-82): Corruption detected. Unmount and run xfs_repair
Aug 22 23:24:48 vsa-00000110-vc-0 kernel: [4194599.690508] XFS (dm-82): metadata I/O error: block 0x50ffb50 ("xfs_trans_read_buf_map") error 117 numblks 8

[-- Attachment #5: Type: text/plain, Size: 121 bytes --]

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 18+ messages in thread
* XFS Corruption
@ 2016-02-24  6:12 fangchen sun
  2016-02-24 22:23 ` Eric Sandeen
  0 siblings, 1 reply; 18+ messages in thread
From: fangchen sun @ 2016-02-24  6:12 UTC (permalink / raw)
  To: xfs


[-- Attachment #1.1: Type: text/plain, Size: 2534 bytes --]

Dear all:

I have a ceph object storage cluster, and choose XFS as the underlying file
system.
I recently ran into a problem that sometimes the function "setxattr()"
 failed,
I can only umount the disk and repair it with "xfs_repair".

os: centos 6.5
kernel version: 2.6.32

the log for dmesg command:
[41796028.532225] Pid: 1438740, comm: ceph-osd Not tainted
2.6.32-925.431.23.3.letv.el6.x86_64 #1
[41796028.532227] Call Trace:
[41796028.532255]  [<ffffffffa01e1e5f>] ? xfs_error_report+0x3f/0x50 [xfs]
[41796028.532276]  [<ffffffffa01d506a>] ? xfs_da_read_buf+0x2a/0x30 [xfs]
[41796028.532296]  [<ffffffffa01e1ece>] ? xfs_corruption_error+0x5e/0x90
[xfs]
[41796028.532316]  [<ffffffffa01d4f4c>] ? xfs_da_do_buf+0x6cc/0x770 [xfs]
[41796028.532335]  [<ffffffffa01d506a>] ? xfs_da_read_buf+0x2a/0x30 [xfs]
[41796028.532359]  [<ffffffffa0206fc7>] ? kmem_zone_alloc+0x77/0xf0 [xfs]
[41796028.532380]  [<ffffffffa01d506a>] ? xfs_da_read_buf+0x2a/0x30 [xfs]
[41796028.532399]  [<ffffffffa01bc481>] ? xfs_attr_leaf_addname+0x61/0x3d0
[xfs]
[41796028.532426]  [<ffffffffa01bc481>] ? xfs_attr_leaf_addname+0x61/0x3d0
[xfs]
[41796028.532455]  [<ffffffffa01ff187>] ? xfs_trans_add_item+0x57/0x70 [xfs]
[41796028.532476]  [<ffffffffa01cc208>] ? xfs_bmbt_get_all+0x18/0x20 [xfs]
[41796028.532495]  [<ffffffffa01bcbb4>] ? xfs_attr_set_int+0x3c4/0x510 [xfs]
[41796028.532517]  [<ffffffffa01d4f5b>] ? xfs_da_do_buf+0x6db/0x770 [xfs]
[41796028.532536]  [<ffffffffa01bcd81>] ? xfs_attr_set+0x81/0x90 [xfs]
[41796028.532560]  [<ffffffffa0216cc3>] ? __xfs_xattr_set+0x43/0x60 [xfs]
[41796028.532584]  [<ffffffffa0216d31>] ? xfs_xattr_user_set+0x11/0x20 [xfs]
[41796028.532592]  [<ffffffff811aee92>] ? generic_setxattr+0xa2/0xb0
[41796028.532596]  [<ffffffff811b134e>] ? __vfs_setxattr_noperm+0x4e/0x160
[41796028.532600]  [<ffffffff81196b77>] ? inode_permission+0xa7/0x100
[41796028.532604]  [<ffffffff811b151c>] ? vfs_setxattr+0xbc/0xc0
[41796028.532607]  [<ffffffff811b15f0>] ? setxattr+0xd0/0x150
[41796028.532612]  [<ffffffff8105af80>] ? __dequeue_entity+0x30/0x50
[41796028.532617]  [<ffffffff8100988e>] ? __switch_to+0x26e/0x320
[41796028.532621]  [<ffffffff8118aec0>] ? __sb_start_write+0x80/0x120
[41796028.532626]  [<ffffffff8152912e>] ? thread_return+0x4e/0x760
[41796028.532630]  [<ffffffff811b171d>] ? sys_fsetxattr+0xad/0xd0
[41796028.532633]  [<ffffffff8100b072>] ? system_call_fastpath+0x16/0x1b
[41796028.532636] XFS (sdi1): Corruption detected. Unmount and run
xfs_repair

Any comments will be much appreciated!

Best Regards!
sunspot

[-- Attachment #1.2: Type: text/html, Size: 3620 bytes --]

[-- Attachment #2: Type: text/plain, Size: 121 bytes --]

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 18+ messages in thread
* XFS corruption
@ 2014-12-21 11:42 Alex Lyakas
  2014-12-21 18:13 ` Eric Sandeen
  0 siblings, 1 reply; 18+ messages in thread
From: Alex Lyakas @ 2014-12-21 11:42 UTC (permalink / raw)
  To: xfs


[-- Attachment #1.1: Type: text/plain, Size: 7724 bytes --]

Greetings,
we encountered XFS corruption:

kernel: [774772.852316] ffff8801018c5000: 05 d1 fd 01 fd ff 2f ec 2f 8d 82 6a 81 fe c2 0f  .....././..j....     
kernel: [774772.854820] XFS (dm-72): Internal error xfs_bmbt_verify at line 747 of file /mnt/share/builds/14.09--3.8.13-030813-generic/2014-11-30_15-47-58--14.09-1419-28/src/zadara-btrfs/fs/xfs/xfs_bmap_btree.c.  Caller 0xffffffffa077b6be
kernel: [774772.854820]                                                                                         
kernel: [774772.860766] Pid: 14643, comm: kworker/0:0H Tainted: GF       W  O 3.8.13-030813-generic #20130511184
kernel: [774772.860771] Call Trace:                                                                             
kernel: [774772.860909]  [<ffffffffa074abaf>] xfs_error_report+0x3f/0x50 [xfs]                                  
kernel: [774772.860961]  [<ffffffffa077b6be>] ? xfs_bmbt_read_verify+0xe/0x10 [xfs]                             
kernel: [774772.860985]  [<ffffffffa074ac1e>] xfs_corruption_error+0x5e/0x90 [xfs]                              
kernel: [774772.861014]  [<ffffffffa077b537>] xfs_bmbt_verify+0x77/0x1e0 [xfs]                                  
kernel: [774772.861047]  [<ffffffffa077b6be>] ? xfs_bmbt_read_verify+0xe/0x10 [xfs]                             
kernel: [774772.861077]  [<ffffffff810135aa>] ? __switch_to+0x12a/0x4a0                                         
kernel: [774772.861129]  [<ffffffff81096cd8>] ? set_next_entity+0xa8/0xc0                                       
kernel: [774772.861145]  [<ffffffffa077b6be>] xfs_bmbt_read_verify+0xe/0x10 [xfs]                               
kernel: [774772.861157]  [<ffffffffa074848f>] xfs_buf_iodone_work+0x3f/0xa0 [xfs]                               
kernel: [774772.861161]  [<ffffffff81078b81>] process_one_work+0x141/0x490                                      
kernel: [774772.861164]  [<ffffffff81079b48>] worker_thread+0x168/0x400                                         
kernel: [774772.861166]  [<ffffffff810799e0>] ? manage_workers+0x120/0x120                                      
kernel: [774772.861170]  [<ffffffff8107f050>] kthread+0xc0/0xd0                                                 
kernel: [774772.861172]  [<ffffffff8107ef90>] ? flush_kthread_worker+0xb0/0xb0                                  
kernel: [774772.861193]  [<ffffffff816f61ec>] ret_from_fork+0x7c/0xb0                                           
kernel: [774772.861199]  [<ffffffff8107ef90>] ? flush_kthread_worker+0xb0/0xb0                                  
kernel: [774772.861318] XFS (dm-72): Corruption detected. Unmount and run xfs_repair                            
kernel: [774772.863449] XFS (dm-72): metadata I/O error: block 0x2434e3e8 ("xfs_trans_read_buf_map") error 117 numblks 8

All the corruption reports were for the same block 0x2434e3e8, which according to the code is simply disk address (xfs_daddr_t) 607445992. So there was only one block corrupted.

Some time later, XFS crashed with:
[813114.622928] NULL pointer dereference[813114.622928]  at 0000000000000008
[813114.622928] IP: [<ffffffffa077bad9>] xfs_bmbt_get_all+0x9/0x20 [xfs]
[813114.622928] PGD 0 
[813114.622928] Oops: 0000 [#1] SMP 
[813114.622928] CPU 2 
[813114.622928] Pid: 31120, comm: smbd Tainted: GF       W  O 3.8.13-030813-generic #201305111843 Bochs Bochs
[813114.622928] RIP: 0010:[<ffffffffa077bad9>]  [<ffffffffa077bad9>] xfs_bmbt_get_all+0x9/0x20 [xfs]
[813114.622928] RSP: 0018:ffff88010a193798  EFLAGS: 00010297
[813114.622928] RAX: 0000000000000964 RBX: ffff880180fa9c38 RCX: ffffa5a5a5a5a5a5
[813114.622928] RDX: ffff88010a193898 RSI: ffff88010a193898 RDI: 0000000000000000
[813114.622928] RBP: ffff88010a1937f8 R08: ffff88010a193898 R09: ffff88010a1938b8
[813114.622928] R10: ffffea0005de0940 R11: 0000000000004d0e R12: ffff88010a1938dc
[813114.622928] R13: ffff88010a1938e0 R14: ffff88010a193898 R15: ffff88010a1938b8
[813114.622928] FS:  00007eff2dc7e700(0000) GS:ffff88021fd00000(0000) knlGS:0000000000000000
[813114.622928] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[813114.622928] CR2: 0000000000000008 CR3: 0000000109574000 CR4: 00000000001406e0
[813114.622928] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
[813114.622928] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
[813114.622928] Process smbd (pid: 31120, threadinfo ffff88010a192000, task ffff88011687ae80)
[813114.622928] Stack:
[813114.622928]  ffff88010a1937f8 ffffffffa076f85a ffffffffffffffff 0000000000000000
[813114.622928]  ffffffff816ec509 000000000a193830 ffffffff816ed31d ffff88010a193898
[813114.622928]  ffff880180fa9c00 0000000000000000 ffff88010a1938dc ffff88010a1938e0
[813114.622928] Call Trace:
[813114.622928]  [<ffffffffa076f85a>] ? xfs_bmap_search_multi_extents+0xaa/0x110 [xfs]
[813114.622928]  [<ffffffff816ec509>] ? schedule+0x29/0x70
[813114.622928]  [<ffffffff816ed31d>] ? rwsem_down_failed_common+0xcd/0x170
[813114.622928]  [<ffffffffa076f92e>] xfs_bmap_search_extents+0x6e/0xf0 [xfs]
[813114.622928]  [<ffffffffa0778d6c>] xfs_bmapi_read+0xfc/0x2f0 [xfs]
[813114.622928]  [<ffffffffa0792a49>] ? xfs_ilock_map_shared+0x49/0x60 [xfs]
[813114.622928]  [<ffffffffa07459a8>] __xfs_get_blocks+0xe8/0x550 [xfs]
[813114.622928]  [<ffffffff8135d8c4>] ? call_rwsem_down_read_failed+0x14/0x30
[813114.622928]  [<ffffffffa0745e41>] xfs_get_blocks+0x11/0x20 [xfs]
[813114.622928]  [<ffffffff811d05b7>] block_read_full_page+0x127/0x360
[813114.622928]  [<ffffffffa0745e30>] ? xfs_get_blocks_direct+0x20/0x20 [xfs]
[813114.622928]  [<ffffffff811d9b0f>] do_mpage_readpage+0x35f/0x550
[813114.622928]  [<ffffffff816f1025>] ? do_async_page_fault+0x35/0x90
[813114.622928]  [<ffffffff816edd48>] ? async_page_fault+0x28/0x30
[813114.622928]  [<ffffffff811d9d4f>] mpage_readpage+0x4f/0x70
[813114.622928]  [<ffffffffa0745e30>] ? xfs_get_blocks_direct+0x20/0x20 [xfs]
[813114.622928]  [<ffffffff81134da8>] ? file_read_actor+0x68/0x160
[813114.622928]  [<ffffffff81134e04>] ? file_read_actor+0xc4/0x160
[813114.622928]  [<ffffffff81354bfe>] ? radix_tree_lookup_slot+0xe/0x10
[813114.622928]  [<ffffffffa07451b8>] xfs_vm_readpage+0x18/0x20 [xfs]
[813114.622928]  [<ffffffff811364ad>] do_generic_file_read.constprop.31+0x10d/0x440
[813114.622928]  [<ffffffff811374d1>] generic_file_aio_read+0xe1/0x220
[813114.622928]  [<ffffffffa074fb98>] xfs_file_aio_read+0x1c8/0x330 [xfs]
[813114.622928]  [<ffffffff8119ad93>] do_sync_read+0xa3/0xe0
[813114.622928]  [<ffffffff8119b4d0>] vfs_read+0xb0/0x180
[813114.622928]  [<ffffffff8119b77a>] sys_pread64+0x9a/0xa0
[813114.622928]  [<ffffffff816f629d>] system_call_fastpath+0x1a/0x1f
[813114.622928] Code: d8 4c 8b 65 e0 4c 8b 6d e8 4c 8b 75 f0 4c 8b 7d f8 c9 c3 66 66 66 66 66 66 2e 0f 1f 84 00 00 00 00 00 0f 1f 44 00 00 55 48 89 f2 <48> 8b 77 08 48 8b 3f 48 89 e5 e8 48 f8 ff ff 5d c3 66 0f 1f 44 
[813114.622928] RIP  [<ffffffffa077bad9>] xfs_bmbt_get_all+0x9/0x20 [xfs]
[813114.622928]  RSP <ffff88010a193798>
[813114.622928] CR2: 0000000000000008
[813114.721138] ---[ end trace cce2a358d4050d3d ]---

We are running XFS based on kernel 3.8.13, with our changes for large-block discard in https://github.com/zadarastorage/zadara-xfs-pushback.

We analyzed several suspects, but all of them fall on disk addresses not near the corrupted disk address. I realize that running somewhat outdated kernel + our changes within XFSs, points back at us, but this is first time we see XFS corruption after about a year of this code being exercised. So posting here, just in case this is a known issue.

I must point out that xfs_repair was able to fix this, which was awesome!

Thanks,
Alex.


[-- Attachment #1.2: Type: text/html, Size: 13688 bytes --]

[-- Attachment #2: Type: text/plain, Size: 121 bytes --]

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 18+ messages in thread

end of thread, other threads:[~2016-02-24 22:24 UTC | newest]

Thread overview: 18+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2015-09-03 11:09 xfs corruption Danny Shavit
2015-09-03 13:22 ` Eric Sandeen
2015-09-03 14:26   ` Danny Shavit
2015-09-03 14:55     ` Eric Sandeen
2015-09-03 16:14       ` Eric Sandeen
2015-09-06 10:19         ` Alex Lyakas
2015-09-06 21:56           ` Eric Sandeen
2015-09-07  8:30             ` Alex Lyakas
  -- strict thread matches above, loose matches on Subject: below --
2016-02-24  6:12 XFS Corruption fangchen sun
2016-02-24 22:23 ` Eric Sandeen
2014-12-21 11:42 XFS corruption Alex Lyakas
2014-12-21 18:13 ` Eric Sandeen
2014-12-21 23:08   ` Dave Chinner
2014-12-22 10:09     ` Alex Lyakas
2014-12-22 14:42     ` Brian Foster
2014-12-23  0:39       ` Dave Chinner
2014-12-23  9:57         ` Alex Lyakas
2014-12-23 20:36           ` Dave Chinner

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.