From: "J. Bruce Fields" <bfields@fieldses.org>
To: "Kraus, Sebastian" <sebastian.kraus@tu-berlin.de>
Cc: "linux-nfs@vger.kernel.org" <linux-nfs@vger.kernel.org>
Subject: Re: Strange segmentation violations of rpc.gssd in Debian Buster
Date: Thu, 25 Jun 2020 16:14:20 -0400 [thread overview]
Message-ID: <20200625201420.GC6605@fieldses.org> (raw)
In-Reply-To: <406fe972135846dc8a23b60be59b0590@tu-berlin.de>
On Thu, Jun 25, 2020 at 05:43:53PM +0000, Kraus, Sebastian wrote:
> Dear Bruce,
> I got the following stack and back trace:
>
> root@all:~# coredumpctl debug
> PID: 6356 (rpc.gssd)
> UID: 0 (root)
> GID: 0 (root)
> Signal: 11 (SEGV)
> Timestamp: Thu 2020-06-25 11:46:08 CEST (3h 4min ago)
> Command Line: /usr/sbin/rpc.gssd -vvvvvvv -rrrrrrr -t 3600 -T 10
> Executable: /usr/sbin/rpc.gssd
> Control Group: /system.slice/rpc-gssd.service
> Unit: rpc-gssd.service
> Slice: system.slice
> Boot ID: XXXXXXXXXXXXXXXXXXXXXXXXXXX
> Machine ID: YYYYYYYYYYYYYYYYYYYYYYYYYYYY
> Hostname: XYZ
> Storage: /var/lib/systemd/coredump/core.rpc\x2egssd.0.7f31136228274af0a1a855b91ad1e75c.6356.1593078368000000.lz4
> Message: Process 6356 (rpc.gssd) of user 0 dumped core.
>
> Stack trace of thread 14174:
> #0 0x000056233fff038e n/a (rpc.gssd)
> #1 0x000056233fff09f8 n/a (rpc.gssd)
> #2 0x000056233fff0b92 n/a (rpc.gssd)
> #3 0x000056233fff13b3 n/a (rpc.gssd)
> #4 0x00007fb2eb8dbfa3 start_thread (libpthread.so.0)
> #5 0x00007fb2eb80c4cf __clone (libc.so.6)
>
> Stack trace of thread 6356:
> #0 0x00007fb2eb801819 __GI___poll (libc.so.6)
> #1 0x00007fb2eb6e7207 send_dg (libresolv.so.2)
> #2 0x00007fb2eb6e4c43 __GI___res_context_query (libresolv.so.2)
> #3 0x00007fb2eb6bf536 __GI__nss_dns_gethostbyaddr2_r (libnss_dns.so.2)
> #4 0x00007fb2eb6bf823 _nss_dns_gethostbyaddr_r (libnss_dns.so.2)
> #5 0x00007fb2eb81dee2 __gethostbyaddr_r (libc.so.6)
> #6 0x00007fb2eb8267d5 gni_host_inet_name (libc.so.6)
> #7 0x000056233ffef455 n/a (rpc.gssd)
> #8 0x000056233ffef82c n/a (rpc.gssd)
> #9 0x000056233fff01d0 n/a (rpc.gssd)
> #10 0x00007fb2ebab49ba n/a (libevent-2.1.so.6)
> #11 0x00007fb2ebab5537 event_base_loop (libevent-2.1.so.6)
> #12 0x000056233ffedeaa n/a (rpc.gssd)
> #13 0x00007fb2eb73709b __libc_start_main (libc.so.6)
> #14 0x000056233ffee03a n/a (rpc.gssd)
>
> GNU gdb (Debian 8.2.1-2+b3) 8.2.1
> [...]
> Reading symbols from /usr/sbin/rpc.gssd...(no debugging symbols found)...done.
> [New LWP 14174]
> [New LWP 6356]
> [Thread debugging using libthread_db enabled]
> Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1".
> Core was generated by `/usr/sbin/rpc.gssd -vvvvvvv -rrrrrrr -t 3600 -T 10'.
> Program terminated with signal SIGSEGV, Segmentation fault.
> #0 0x000056233fff038e in ?? ()
> [Current thread is 1 (Thread 0x7fb2eaeba700 (LWP 14174))]
> (gdb) bt
> #0 0x000056233fff038e in ?? ()
> #1 0x000056233fff09f8 in ?? ()
> #2 0x000056233fff0b92 in ?? ()
> #3 0x000056233fff13b3 in ?? ()
> #4 0x00007fb2eb8dbfa3 in start_thread (arg=<optimized out>) at pthread_create.c:486
> #5 0x00007fb2eb80c4cf in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:95
> (gdb) quit
>
>
> I am not an expert in analyzing stack and backtraces. Is there anything meaningful, you are able to extract from the trace?
> As far as I see, thread 14174 caused the segmentation violation just after its birth on clone.
> Please correct me, if I am in error.
> Seems Debian Buster does not ship any dedicated package with debug symbols for the rpc.gssd executable.
Have you reported a debian bug? They might know how to get a good trace
out of it.
--b.
> So far, I was not able to find such a package.
> What's your opinon about the trace?
>
>
> Best and Thanks
> Sebastian
>
> _____________________________
> Sebastian Kraus
> Team IT am Institut für Chemie
> Gebäude C, Straße des 17. Juni 115, Raum C7
>
> Technische Universität Berlin
> Fakultät II
> Institut für Chemie
> Sekretariat C3
> Straße des 17. Juni 135
> 10623 Berlin
>
> Email: sebastian.kraus@tu-berlin.de
>
> ________________________________________
> From: linux-nfs-owner@vger.kernel.org <linux-nfs-owner@vger.kernel.org> on behalf of J. Bruce Fields <bfields@fieldses.org>
> Sent: Tuesday, June 23, 2020 00:36
> To: Kraus, Sebastian
> Cc: linux-nfs@vger.kernel.org
> Subject: Re: RPC Pipefs: Frequent parsing errors in client database
>
> On Sat, Jun 20, 2020 at 09:08:55PM +0000, Kraus, Sebastian wrote:
> > Hi Bruce,
> >
> > >> But I think it'd be more useful to stay focused on the segfaults.
> >
> > is it a clever idea to analyze core dumps? Or are there other much better debugging techniques w.r.t. RPC daemons?
>
> If we could at least get a backtrace out of the core dump that could be
> useful.
>
> > I now do more tests while fiddling around with the time-out parameters "-T" and "-t" on the command line of rpc.gssd.
> >
> > There are several things I do not really understand about the trace shown below:
> >
> > 1) How can it be useful that the rpc.gssd daemon tries to parse the info file although it knows about its absence beforehand?
>
> It doesn't know beforehand, in the scenarios I described.
>
> > 2) Why are there two identifiers clnt36e and clnt36f being used for the same client?
>
> This is actually happening on an NFS server, the rpc client in question
> is the callback client used to do things like send delegation recalls
> back to the NFS client.
>
> I'm not sure why two different callback clients are being created here,
> but there's nothing inherently weird about that.
>
> > 3) What does the <?> in "inotify event for clntdir (nfsd4_cb/clnt36e) - ev->wd (600) ev->name (<?>) ev->mask (0x00008000)" mean?
>
> Off the top of my head, I don't know, we'd probably need to look through
> header files or inotify man pages for the definitions of those masks.
>
> --b.
next prev parent reply other threads:[~2020-06-25 20:14 UTC|newest]
Thread overview: 21+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-06-19 21:24 RPC Pipefs: Frequent parsing errors in client database Kraus, Sebastian
2020-06-19 22:04 ` J. Bruce Fields
2020-06-20 11:35 ` Kraus, Sebastian
2020-06-20 17:03 ` J. Bruce Fields
2020-06-20 21:08 ` Kraus, Sebastian
2020-06-22 22:36 ` J. Bruce Fields
2020-06-25 17:43 ` Strange segmentation violations of rpc.gssd in Debian Buster Kraus, Sebastian
2020-06-25 20:14 ` J. Bruce Fields [this message]
2020-06-25 21:44 ` Doug Nazar
2020-06-26 12:31 ` Kraus, Sebastian
2020-06-26 17:23 ` Doug Nazar
2020-06-26 19:46 ` J. Bruce Fields
2020-06-26 20:15 ` Doug Nazar
2020-06-26 21:02 ` J. Bruce Fields
2020-06-26 21:30 ` [PATCH v2] " Doug Nazar
2020-06-26 21:44 ` J. Bruce Fields
2020-06-29 5:39 ` Kraus, Sebastian
2020-06-29 14:09 ` Doug Nazar
2020-07-01 7:39 ` Kraus, Sebastian
2020-07-01 8:13 ` [PATCH v2] " Peter Eriksson
2020-07-01 18:45 ` [PATCH v2] " Doug Nazar
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20200625201420.GC6605@fieldses.org \
--to=bfields@fieldses.org \
--cc=linux-nfs@vger.kernel.org \
--cc=sebastian.kraus@tu-berlin.de \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).