From: "J. Bruce Fields" <bfields@fieldses.org>
To: Doug Nazar <nazard@nazar.ca>
Cc: "Kraus, Sebastian" <sebastian.kraus@tu-berlin.de>,
"linux-nfs@vger.kernel.org" <linux-nfs@vger.kernel.org>,
Steve Dickson <steved@redhat.com>,
Olga Kornievskaia <aglo@umich.edu>
Subject: Re: Strange segmentation violations of rpc.gssd in Debian Buster
Date: Fri, 26 Jun 2020 17:02:43 -0400 [thread overview]
Message-ID: <20200626210243.GD11850@fieldses.org> (raw)
In-Reply-To: <3eb80b1f-e4d3-e87c-aacd-34dc28637948@nazar.ca>
On Fri, Jun 26, 2020 at 04:15:46PM -0400, Doug Nazar wrote:
> On 2020-06-26 15:46, J. Bruce Fields wrote:
> >So, yeah, either a reference count or a deep copy is probably all that's
> >needed, in alloc_upcall_info() and at the end of handle_krb5_upcall().
>
> Slightly more complex, to handle the error cases & event tear-down
> but this seems to work. I'm not sure how much of a hot spot this is
> so I just went with a global mutex. Strangely there was an existing
> unused mutex & thread_started flag.
>
> Survives basic testing but I don't have a reproducer. Maybe blocking
> access to the kdc. If I get time I'll try to setup a test
> environment.
Thanks, looks good. The only thing I wonder about:
> @@ -359,9 +357,37 @@ out:
> free(port);
> }
>
> +/* Actually frees clp and fields that might be used from other
> + * threads if was last reference.
> + */
> +static void
> +gssd_free_client(struct clnt_info *clp)
> +{
> + int refcnt;
> +
> + pthread_mutex_lock(&clp_lock);
> + refcnt = --clp->refcount;
> + pthread_mutex_unlock(&clp_lock);
> + if (refcnt > 0)
> + return;
> +
> + printerr(3, "freeing client %s\n", clp->relpath);
> +
> + free(clp->relpath);
> + free(clp->servicename);
> + free(clp->servername);
> + free(clp->protocol);
> + free(clp);
> +}
> +
> +/* Called when removing from clnt_list to tear down event handling.
> + * Will then free clp if was last reference.
> + */
> static void
> gssd_destroy_client(struct clnt_info *clp)
> {
> + printerr(3, "destroying client %s\n", clp->relpath);
> +
> if (clp->krb5_fd >= 0) {
> close(clp->krb5_fd);
Unless I'm missing something--an upcall thread could still be using this
file descriptor.
If we're particularly unlucky, we could do a new open in a moment and
reuse this file descriptor number, and then then writes in do_downcall()
could end up going to some other random file.
I think we want these closes done by gssd_free_client() in the !refcnt
case?
--b.
> event_del(&clp->krb5_ev);
> @@ -373,11 +399,7 @@ gssd_destroy_client(struct clnt_info *clp)
> }
>
> inotify_rm_watch(inotify_fd, clp->wd);
> - free(clp->relpath);
> - free(clp->servicename);
> - free(clp->servername);
> - free(clp->protocol);
> - free(clp);
> + gssd_free_client(clp);
> }
>
> static void gssd_scan(void);
> @@ -416,11 +438,21 @@ static struct clnt_upcall_info *alloc_upcall_info(struct clnt_info *clp)
> info = malloc(sizeof(struct clnt_upcall_info));
> if (info == NULL)
> return NULL;
> +
> + pthread_mutex_lock(&clp_lock);
> + clp->refcount++;
> + pthread_mutex_unlock(&clp_lock);
> info->clp = clp;
>
> return info;
> }
>
> +void free_upcall_info(struct clnt_upcall_info *info)
> +{
> + gssd_free_client(info->clp);
> + free(info);
> +}
> +
> /* For each upcall read the upcall info into the buffer, then create a
> * thread in a detached state so that resources are released back into
> * the system without the need for a join.
> @@ -438,13 +470,13 @@ gssd_clnt_gssd_cb(int UNUSED(fd), short UNUSED(which), void *data)
> info->lbuflen = read(clp->gssd_fd, info->lbuf, sizeof(info->lbuf));
> if (info->lbuflen <= 0 || info->lbuf[info->lbuflen-1] != '\n') {
> printerr(0, "WARNING: %s: failed reading request\n", __func__);
> - free(info);
> + free_upcall_info(info);
> return;
> }
> info->lbuf[info->lbuflen-1] = 0;
>
> if (start_upcall_thread(handle_gssd_upcall, info))
> - free(info);
> + free_upcall_info(info);
> }
>
> static void
> @@ -461,12 +493,12 @@ gssd_clnt_krb5_cb(int UNUSED(fd), short UNUSED(which), void *data)
> sizeof(info->uid)) < (ssize_t)sizeof(info->uid)) {
> printerr(0, "WARNING: %s: failed reading uid from krb5 "
> "upcall pipe: %s\n", __func__, strerror(errno));
> - free(info);
> + free_upcall_info(info);
> return;
> }
>
> if (start_upcall_thread(handle_krb5_upcall, info))
> - free(info);
> + free_upcall_info(info);
> }
>
> static struct clnt_info *
> @@ -501,6 +533,7 @@ gssd_get_clnt(struct topdir *tdi, const char *name)
> clp->name = clp->relpath + strlen(tdi->name) + 1;
> clp->krb5_fd = -1;
> clp->gssd_fd = -1;
> + clp->refcount = 1;
>
> TAILQ_INSERT_HEAD(&tdi->clnt_list, clp, list);
> return clp;
> @@ -651,7 +684,7 @@ gssd_scan_topdir(const char *name)
> if (clp->scanned)
> continue;
>
> - printerr(3, "destroying client %s\n", clp->relpath);
> + printerr(3, "orphaned client %s\n", clp->relpath);
> saveprev = clp->list.tqe_prev;
> TAILQ_REMOVE(&tdi->clnt_list, clp, list);
> gssd_destroy_client(clp);
> diff --git a/utils/gssd/gssd.h b/utils/gssd/gssd.h
> index f4f59754..d33e4dff 100644
> --- a/utils/gssd/gssd.h
> +++ b/utils/gssd/gssd.h
> @@ -63,12 +63,10 @@ extern unsigned int context_timeout;
> extern unsigned int rpc_timeout;
> extern char *preferred_realm;
> extern pthread_mutex_t ple_lock;
> -extern pthread_cond_t pcond;
> -extern pthread_mutex_t pmutex;
> -extern int thread_started;
>
> struct clnt_info {
> TAILQ_ENTRY(clnt_info) list;
> + int refcount;
> int wd;
> bool scanned;
> char *name;
> @@ -94,6 +92,7 @@ struct clnt_upcall_info {
>
> void handle_krb5_upcall(struct clnt_upcall_info *clp);
> void handle_gssd_upcall(struct clnt_upcall_info *clp);
> +void free_upcall_info(struct clnt_upcall_info *info);
>
>
> #endif /* _RPC_GSSD_H_ */
> diff --git a/utils/gssd/gssd_proc.c b/utils/gssd/gssd_proc.c
> index 8fe6605b..05c1da64 100644
> --- a/utils/gssd/gssd_proc.c
> +++ b/utils/gssd/gssd_proc.c
> @@ -730,7 +730,7 @@ handle_krb5_upcall(struct clnt_upcall_info *info)
> printerr(2, "\n%s: uid %d (%s)\n", __func__, info->uid, clp->relpath);
>
> process_krb5_upcall(clp, info->uid, clp->krb5_fd, NULL, NULL, NULL);
> - free(info);
> + free_upcall_info(info);
> }
>
> void
> @@ -830,6 +830,6 @@ handle_gssd_upcall(struct clnt_upcall_info *info)
> out:
> free(upcall_str);
> out_nomem:
> - free(info);
> + free_upcall_info(info);
> return;
> }
> --
> 2.26.2
>
next prev parent reply other threads:[~2020-06-26 21:02 UTC|newest]
Thread overview: 21+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-06-19 21:24 RPC Pipefs: Frequent parsing errors in client database Kraus, Sebastian
2020-06-19 22:04 ` J. Bruce Fields
2020-06-20 11:35 ` Kraus, Sebastian
2020-06-20 17:03 ` J. Bruce Fields
2020-06-20 21:08 ` Kraus, Sebastian
2020-06-22 22:36 ` J. Bruce Fields
2020-06-25 17:43 ` Strange segmentation violations of rpc.gssd in Debian Buster Kraus, Sebastian
2020-06-25 20:14 ` J. Bruce Fields
2020-06-25 21:44 ` Doug Nazar
2020-06-26 12:31 ` Kraus, Sebastian
2020-06-26 17:23 ` Doug Nazar
2020-06-26 19:46 ` J. Bruce Fields
2020-06-26 20:15 ` Doug Nazar
2020-06-26 21:02 ` J. Bruce Fields [this message]
2020-06-26 21:30 ` [PATCH v2] " Doug Nazar
2020-06-26 21:44 ` J. Bruce Fields
2020-06-29 5:39 ` Kraus, Sebastian
2020-06-29 14:09 ` Doug Nazar
2020-07-01 7:39 ` Kraus, Sebastian
2020-07-01 8:13 ` [PATCH v2] " Peter Eriksson
2020-07-01 18:45 ` [PATCH v2] " Doug Nazar
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20200626210243.GD11850@fieldses.org \
--to=bfields@fieldses.org \
--cc=aglo@umich.edu \
--cc=linux-nfs@vger.kernel.org \
--cc=nazard@nazar.ca \
--cc=sebastian.kraus@tu-berlin.de \
--cc=steved@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).