From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-1.1 required=3.0 tests=DKIM_SIGNED,DKIM_VALID, DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI,SPF_HELO_NONE, SPF_PASS,URIBL_BLOCKED autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id E9012C433DF for ; Thu, 25 Jun 2020 17:43:59 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 9161A20781 for ; Thu, 25 Jun 2020 17:43:59 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (1024-bit key) header.d=tu-berlin.de header.i=@tu-berlin.de header.b="UL/C32hL" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S2406894AbgFYRn7 (ORCPT ); Thu, 25 Jun 2020 13:43:59 -0400 Received: from exchange.tu-berlin.de ([130.149.7.70]:40055 "EHLO exchange.tu-berlin.de" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S2406883AbgFYRn6 (ORCPT ); Thu, 25 Jun 2020 13:43:58 -0400 Received: from SPMA-04.tubit.win.tu-berlin.de (localhost.localdomain [127.0.0.1]) by localhost (Email Security Appliance) with SMTP id 6FCEF9739CC_EF4E25AB; Thu, 25 Jun 2020 17:43:54 +0000 (GMT) Received: from exchange.tu-berlin.de (exchange.tu-berlin.de [130.149.7.70]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-SHA384 (256/256 bits)) (Client CN "exchange.tu-berlin.de", Issuer "DFN-Verein Global Issuing CA" (not verified)) by SPMA-04.tubit.win.tu-berlin.de (Sophos Email Appliance) with ESMTPS id F05BC9739B5_EF4E259F; Thu, 25 Jun 2020 17:43:53 +0000 (GMT) Received: from ex-01.tubit.win.tu-berlin.de (130.149.7.70) by ex-mbx-07.tubit.win.tu-berlin.de (130.149.6.158) with Microsoft SMTP Server (TLS) id 15.0.1395.4; Thu, 25 Jun 2020 19:43:53 +0200 Received: from ex-02.tubit.win.tu-berlin.de (172.26.35.185) by ex-01.tubit.win.tu-berlin.de (172.26.35.184) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.529.5; Thu, 25 Jun 2020 19:43:53 +0200 Received: from ex-02.tubit.win.tu-berlin.de ([172.26.26.142]) by ex-02.tubit.win.tu-berlin.de ([172.26.26.142]) with mapi id 15.02.0529.008; Thu, 25 Jun 2020 19:43:53 +0200 From: "Kraus, Sebastian" To: "J. Bruce Fields" CC: "linux-nfs@vger.kernel.org" Subject: Re: Strange segmentation violations of rpc.gssd in Debian Buster Thread-Topic: Strange segmentation violations of rpc.gssd in Debian Buster Thread-Index: AQHWSxgxwlq3b/mulkebnqBRRSeMOQ== Date: Thu, 25 Jun 2020 17:43:53 +0000 Message-ID: <406fe972135846dc8a23b60be59b0590@tu-berlin.de> References: <20200619220434.GB1594@fieldses.org> <28a44712b25c4420909360bd813f8bfd@tu-berlin.de> <20200620170316.GH1514@fieldses.org> <5c45562c90404838944ee71a1d926c74@tu-berlin.de>,<20200622223628.GC11051@fieldses.org> In-Reply-To: <20200622223628.GC11051@fieldses.org> Accept-Language: en-US, de-DE Content-Language: en-US X-MS-Has-Attach: X-MS-TNEF-Correlator: x-originating-ip: [130.149.19.173] Content-Type: text/plain; charset="iso-8859-1" Content-Transfer-Encoding: quoted-printable MIME-Version: 1.0 X-PMWin-Version: 4.0.1, Antivirus-Engine: 3.77.1, Antivirus-Data: 5.76 X-PureMessage: [Scanned] X-SASI-RCODE: 200 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=tu-berlin.de; h=from:to:cc:subject:date:message-id:references:in-reply-to:content-type:content-transfer-encoding:mime-version; s=dkim-tub; bh=/U3QhIcvvuOPS3N4KY3poPCiXXrWoKiN3DXrcGb7HMQ=; b=UL/C32hLRlSrSI/EL/2ALCxYQIkV52PhuWMN7zJV0TO00jYZESiWppbFoTa92rx1r8jDneCV/Xod4BTI/HziPA7//1+XRKGhW8GqMT4PA98849QK+SCfSjNvJBCrxu4gwGwPsBFxX6ZO0Taai4MCctUT7cyPZMLNc14MFoPOOwo= Sender: linux-nfs-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-nfs@vger.kernel.org Dear Bruce, I got the following stack and back trace: root@all:~# coredumpctl debug PID: 6356 (rpc.gssd) UID: 0 (root) GID: 0 (root) Signal: 11 (SEGV) Timestamp: Thu 2020-06-25 11:46:08 CEST (3h 4min ago) Command Line: /usr/sbin/rpc.gssd -vvvvvvv -rrrrrrr -t 3600 -T 10 Executable: /usr/sbin/rpc.gssd Control Group: /system.slice/rpc-gssd.service Unit: rpc-gssd.service Slice: system.slice Boot ID: XXXXXXXXXXXXXXXXXXXXXXXXXXX Machine ID: YYYYYYYYYYYYYYYYYYYYYYYYYYYY Hostname: XYZ Storage: /var/lib/systemd/coredump/core.rpc\x2egssd.0.7f31136228274a= f0a1a855b91ad1e75c.6356.1593078368000000.lz4 Message: Process 6356 (rpc.gssd) of user 0 dumped core. =20 Stack trace of thread 14174: #0 0x000056233fff038e n/a (rpc.gssd) #1 0x000056233fff09f8 n/a (rpc.gssd) #2 0x000056233fff0b92 n/a (rpc.gssd) #3 0x000056233fff13b3 n/a (rpc.gssd) #4 0x00007fb2eb8dbfa3 start_thread (libpthread.so.0) #5 0x00007fb2eb80c4cf __clone (libc.so.6) =20 Stack trace of thread 6356: #0 0x00007fb2eb801819 __GI___poll (libc.so.6) #1 0x00007fb2eb6e7207 send_dg (libresolv.so.2) #2 0x00007fb2eb6e4c43 __GI___res_context_query (libresolv.= so.2) #3 0x00007fb2eb6bf536 __GI__nss_dns_gethostbyaddr2_r (libn= ss_dns.so.2) #4 0x00007fb2eb6bf823 _nss_dns_gethostbyaddr_r (libnss_dns= .so.2) #5 0x00007fb2eb81dee2 __gethostbyaddr_r (libc.so.6) #6 0x00007fb2eb8267d5 gni_host_inet_name (libc.so.6) #7 0x000056233ffef455 n/a (rpc.gssd) #8 0x000056233ffef82c n/a (rpc.gssd) #9 0x000056233fff01d0 n/a (rpc.gssd) #10 0x00007fb2ebab49ba n/a (libevent-2.1.so.6) #11 0x00007fb2ebab5537 event_base_loop (libevent-2.1.so.6) #12 0x000056233ffedeaa n/a (rpc.gssd) #13 0x00007fb2eb73709b __libc_start_main (libc.so.6) #14 0x000056233ffee03a n/a (rpc.gssd) GNU gdb (Debian 8.2.1-2+b3) 8.2.1 [...] Reading symbols from /usr/sbin/rpc.gssd...(no debugging symbols found)...do= ne. [New LWP 14174] [New LWP 6356] [Thread debugging using libthread_db enabled] Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1". Core was generated by `/usr/sbin/rpc.gssd -vvvvvvv -rrrrrrr -t 3600 -T 10'. Program terminated with signal SIGSEGV, Segmentation fault. #0 0x000056233fff038e in ?? () [Current thread is 1 (Thread 0x7fb2eaeba700 (LWP 14174))] (gdb) bt #0 0x000056233fff038e in ?? () #1 0x000056233fff09f8 in ?? () #2 0x000056233fff0b92 in ?? () #3 0x000056233fff13b3 in ?? () #4 0x00007fb2eb8dbfa3 in start_thread (arg=3D) at pthread_c= reate.c:486 #5 0x00007fb2eb80c4cf in clone () at ../sysdeps/unix/sysv/linux/x86_64/clo= ne.S:95 (gdb) quit I am not an expert in analyzing stack and backtraces. Is there anything mea= ningful, you are able to extract from the trace? As far as I see, thread 14174 caused the segmentation violation just after = its birth on clone.=20 Please correct me, if I am in error. Seems Debian Buster does not ship any dedicated package with debug symbols = for the rpc.gssd executable.=20 So far, I was not able to find such a package. What's your opinon about the trace? Best and Thanks Sebastian _____________________________ Sebastian Kraus Team IT am Institut f=FCr Chemie Geb=E4ude C, Stra=DFe des 17. Juni 115, Raum C7 Technische Universit=E4t Berlin Fakult=E4t II Institut f=FCr Chemie Sekretariat C3 Stra=DFe des 17. Juni 135 10623 Berlin Email: sebastian.kraus@tu-berlin.de ________________________________________ From: linux-nfs-owner@vger.kernel.org on = behalf of J. Bruce Fields Sent: Tuesday, June 23, 2020 00:36 To: Kraus, Sebastian Cc: linux-nfs@vger.kernel.org Subject: Re: RPC Pipefs: Frequent parsing errors in client database On Sat, Jun 20, 2020 at 09:08:55PM +0000, Kraus, Sebastian wrote: > Hi Bruce, > > >> But I think it'd be more useful to stay focused on the segfaults. > > is it a clever idea to analyze core dumps? Or are there other much better= debugging techniques w.r.t. RPC daemons? If we could at least get a backtrace out of the core dump that could be useful. > I now do more tests while fiddling around with the time-out parameters "-= T" and "-t" on the command line of rpc.gssd. > > There are several things I do not really understand about the trace shown= below: > > 1) How can it be useful that the rpc.gssd daemon tries to parse the info = file although it knows about its absence beforehand? It doesn't know beforehand, in the scenarios I described. > 2) Why are there two identifiers clnt36e and clnt36f being used for the s= ame client? This is actually happening on an NFS server, the rpc client in question is the callback client used to do things like send delegation recalls back to the NFS client. I'm not sure why two different callback clients are being created here, but there's nothing inherently weird about that. > 3) What does the in "inotify event for clntdir (nfsd4_cb/clnt36e) - e= v->wd (600) ev->name () ev->mask (0x00008000)" mean? Off the top of my head, I don't know, we'd probably need to look through header files or inotify man pages for the definitions of those masks. --b.