From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from relay.sgi.com (relay1.corp.sgi.com [137.38.102.111]) by oss.sgi.com (Postfix) with ESMTP id 4DD0E7F84 for ; Wed, 18 Feb 2015 08:58:54 -0600 (CST) Received: from cuda.sgi.com (cuda2.sgi.com [192.48.176.25]) by relay1.corp.sgi.com (Postfix) with ESMTP id 3D1268F8035 for ; Wed, 18 Feb 2015 06:58:51 -0800 (PST) Received: from mx1.redhat.com (mx1.redhat.com [209.132.183.28]) by cuda.sgi.com with ESMTP id 41ou4In0cLRKCzaX (version=TLSv1 cipher=AES256-SHA bits=256 verify=NO) for ; Wed, 18 Feb 2015 06:58:50 -0800 (PST) Date: Wed, 18 Feb 2015 09:58:44 -0500 From: Brian Foster Subject: Re: XFS: Internal error XFS_WANT_CORRUPTED_GOTO at line 990 of file fs/xfs/xfs_ialloc.c Message-ID: <20150218145844.GA62927@bfoster.bfoster> References: <20150216141039.GA48651@bfoster.bfoster> MIME-Version: 1.0 Content-Disposition: inline In-Reply-To: List-Id: XFS Filesystem from SGI List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit Errors-To: xfs-bounces@oss.sgi.com Sender: xfs-bounces@oss.sgi.com To: Pablo Silva Cc: xfs@oss.sgi.com On Wed, Feb 18, 2015 at 11:13:23AM -0300, Pablo Silva wrote: > Hi Brian! > > Thanks for your time!, i've just to run xfs_repair -n the result you can > see --> http://pastebin.centos.org/16106/ and as you can see there was > problems, but why this problems occurs?, we have xfs partition used for one > KVM where running Amanda with Centos 7, perhaps it was a bad configuration > inside KVM or perhaps a bug?, the host server is running centos 6 > Can you elaborate? You have an isolated XFS filesystem that is accessed by the host system, the kvm guest, or both? How is the fs used? Also, can you answer my other questions related to how this occurred? Are there any other errors in the logs that precede the one below? I assume the fs shut down at the point of the error below. Has it been mounted since? If not, that might be worth a try to replay the log. If the log is dirty and replayed, that should be indicated by the output in the log at mount time. You'll also want to re-run xfs_repair -n in that case. Brian > [root@vtl ~]# uname -a > Linux vtl.areaprod.b2b 2.6.32-504.8.1.el6.x86_64 #1 SMP Wed Jan 28 21:11:36 > UTC 2015 x86_64 x86_64 x86_64 GNU/Linux > > > But the KVM running centos 7 > > Linux amanda 3.10.0-123.8.1.el7.x86_64 #1 SMP Mon Sep 22 19:06:58 UTC 2014 > x86_64 x86_64 x86_64 GNU/Linux > > The KVM config is this --> http://pastebin.centos.org/16111/ > > And /etc/fstab is this---> http://pastebin.centos.org/16116/ > > > Thanks in advance, for any hint. > > -Pablo > > > On Mon, Feb 16, 2015 at 11:10 AM, Brian Foster wrote: > > > On Fri, Feb 13, 2015 at 03:44:57PM -0300, Pablo Silva wrote: > > > Hi ! > > > > > > We have a server with centos 6.6, kernel version: > > > 2.6.32-431.17.1.el6.x86_64, where we got the following message: > > > > > > Feb 12 19:22:15 vtl kernel: > > > Feb 12 19:22:15 vtl kernel: Pid: 3502, comm: touch Not tainted > > > 2.6.32-431.17.1.el6.x86_64 #1 > > > Feb 12 19:22:15 vtl kernel: Call Trace: > > > Feb 12 19:22:15 vtl kernel: [] ? > > > xfs_error_report+0x3f/0x50 [xfs] > > > Feb 12 19:22:15 vtl kernel: [] ? xfs_ialloc+0x60/0x6e0 > > [xfs] > > > Feb 12 19:22:15 vtl kernel: [] ? > > xfs_dialloc+0x43e/0x850 [xfs] > > > Feb 12 19:22:15 vtl kernel: [] ? xfs_ialloc+0x60/0x6e0 > > [xfs] > > > Feb 12 19:22:15 vtl kernel: [] ? > > > kmem_zone_zalloc+0x3a/0x50 [xfs] > > > Feb 12 19:22:15 vtl kernel: [] ? > > > xfs_dir_ialloc+0x74/0x2b0 [xfs] > > > Feb 12 19:22:15 vtl kernel: [] ? > > xfs_create+0x440/0x640 [xfs] > > > Feb 12 19:22:15 vtl kernel: [] ? > > xfs_vn_mknod+0xad/0x1c0 [xfs] > > > Feb 12 19:22:15 vtl kernel: [] ? > > xfs_vn_create+0x10/0x20 [xfs] > > > Feb 12 19:22:15 vtl kernel: [] ? vfs_create+0xe6/0x110 > > > Feb 12 19:22:15 vtl kernel: [] ? > > do_filp_open+0xa8e/0xd20 > > > Feb 12 19:22:15 vtl kernel: [] ? alloc_fd+0x92/0x160 > > > Feb 12 19:22:15 vtl kernel: XFS: Internal error > > > XFS_WANT_CORRUPTED_GOTO at line 990 of file fs/xfs/xfs_ialloc.c. > > > Caller 0xffffffffa0422980 > > > > > > > /* > > * None left in the last group, search the whole AG > > */ > > error = xfs_inobt_lookup(cur, 0, XFS_LOOKUP_GE, &i); > > if (error) > > goto error0; > > XFS_WANT_CORRUPTED_GOTO(i == 1, error0); > > > > for (;;) { > > error = xfs_inobt_get_rec(cur, &rec, &i); > > if (error) > > goto error0; > > XFS_WANT_CORRUPTED_GOTO(i == 1, error0); > > if (rec.ir_freecount > 0) > > break; > > error = xfs_btree_increment(cur, 0, &i); > > if (error) > > goto error0; > > ---> XFS_WANT_CORRUPTED_GOTO(i == 1, error0); > > } > > > > That corresponds to the check above. This code is part of the inode > > allocator where we expect an AG to have free inodes and we're doing a > > brute force search for a record. Apparently we go off the AG or some > > other problem occurs before we find a free inode record. > > > > Does 'xfs_repair -n' report any problems with this fs? Have there been > > any other storage errors reported in the logs? Is the problem > > reproducible or was it a one off occurrence? > > > > Brian > > > > > I can't find more information for this..., perhaps a bug or other > > > thing ..., welcome any hint for to research.. > > > > > > Thanks in advance! > > > > > > -Pablo > > > > > _______________________________________________ > > > xfs mailing list > > > xfs@oss.sgi.com > > > http://oss.sgi.com/mailman/listinfo/xfs > > > > _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs