linux-usb.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: John Keeping <john@metanate.com>
To: Udipto Goswami <quic_ugoswami@quicinc.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>,
	linux-usb@vger.kernel.org, Jack Pham <quic_jackp@quicinc.com>,
	Pratham Pratap <quic_ppratap@quicinc.com>,
	Wesley Cheng <quic_wcheng@quicinc.com>
Subject: Re: [v2] usb: gadget: f_fs: Prevent race between functionfs_unbind & ffs_ep0_queue_wait
Date: Tue, 22 Nov 2022 13:07:41 +0000	[thread overview]
Message-ID: <Y3zJnQKQslBE413v@donbot> (raw)
In-Reply-To: <a485330a-489b-7f0d-5291-1931ba2a749a@quicinc.com>

On Tue, Nov 22, 2022 at 05:56:56PM +0530, Udipto Goswami wrote:
> On 11/22/22 5:17 PM, John Keeping wrote:
> > On Mon, Nov 21, 2022 at 09:52:43AM +0530, Udipto Goswami wrote:
> > > Hi John
> > > 
> > > On 11/20/22 11:18 PM, John Keeping wrote:
> > > > On Sun, Nov 20, 2022 at 12:23:50PM +0530, Udipto Goswami wrote:
> > > > > On 11/18/22 9:49 PM, John Keeping wrote:
> > > > > > On Wed, Nov 16, 2022 at 04:49:55PM +0530, Udipto Goswami wrote:
> > > > > > > While performing fast composition switch, there is a possibility that the
> > > > > > > process of ffs_ep0_write/ffs_ep0_read get into a race condition
> > > > > > > due to ep0req being freed up from functionfs_unbind.
> > > > > > > 
> > > > > > > Consider the scenario that the ffs_ep0_write calls the ffs_ep0_queue_wait
> > > > > > > by taking a lock &ffs->ev.waitq.lock. However, the functionfs_unbind isn't
> > > > > > > bounded so it can go ahead and mark the ep0req to NULL, and since there
> > > > > > > is no NULL check in ffs_ep0_queue_wait we will end up in use-after-free.
> > > > > > > 
> > > > > > > Fix this by making a serialized execution between the two functions using
> > > > > > > a mutex_lock(ffs->mutex). Also, dequeue the ep0req to ensure that no
> > > > > > > other function can use it after the free operation.
> > > > > > > 
> > > > > > > Fixes: ddf8abd25994 ("USB: f_fs: the FunctionFS driver")
> > > > > > > Signed-off-by: Udipto Goswami <quic_ugoswami@quicinc.com>
> > > > > > > ---
> > > > > > > v2: Replaces spinlock with mutex & added dequeue operation in unbind.
> > > > > > > 
> > > > > > >     drivers/usb/gadget/function/f_fs.c | 7 +++++++
> > > > > > >     1 file changed, 7 insertions(+)
> > > > > > > 
> > > > > > > diff --git a/drivers/usb/gadget/function/f_fs.c b/drivers/usb/gadget/function/f_fs.c
> > > > > > > index 73dc10a77cde..1439449df39a 100644
> > > > > > > --- a/drivers/usb/gadget/function/f_fs.c
> > > > > > > +++ b/drivers/usb/gadget/function/f_fs.c
> > > > > > > @@ -279,6 +279,9 @@ static int __ffs_ep0_queue_wait(struct ffs_data *ffs, char *data, size_t len)
> > > > > > >     	struct usb_request *req = ffs->ep0req;
> > > > > > >     	int ret;
> > > > > > > +	if (!req)
> > > > > > > +		return -EINVAL;
> > > > > > > +
> > > > > > >     	req->zero     = len < le16_to_cpu(ffs->ev.setup.wLength);
> > > > > > >     	spin_unlock_irq(&ffs->ev.waitq.lock);
> > > > > > > @@ -1892,10 +1895,14 @@ static void functionfs_unbind(struct ffs_data *ffs)
> > > > > > >     	ENTER();
> > > > > > >     	if (!WARN_ON(!ffs->gadget)) {
> > > > > > > +		mutex_lock(&ffs->mutex);
> > > > > > > +		/* dequeue before freeing ep0req */
> > > > > > > +		usb_ep_dequeue(ffs->gadget->ep0, ffs->ep0req);
> > > > > > >     		usb_ep_free_request(ffs->gadget->ep0, ffs->ep0req);
> > > > > > >     		ffs->ep0req = NULL;
> > > > > > >     		ffs->gadget = NULL;
> > > > > > >     		clear_bit(FFS_FL_BOUND, &ffs->flags);
> > > > > > > +		mutex_unlock(&ffs->mutex);
> > > > > > 
> > > > > > There's now a deadlock here if some other thread is waiting in
> > > > > > __ffs_ep0_queue_wait() on ep0req_completion.
> > > > > > 
> > > > > > You need to dequeue before taking the lock.
> > > > > That's a control request right, will it be async?
> > > > > 
> > > > > Anyway I see only 2 possible threads ep0_read/ep0_write who calls
> > > > > ep0_queue_wait and waits for the completion of ep0req and both
> > > > > ep0_read/write are prptected by the mutex lock so i guess execution won't
> > > > > reach there right ?
> > > > > Say functionfs_unbind ran first then ep0_read/write had to wait will the
> > > > > functionfs_unbind is completed so ep_dequeue will ran, will get completed,
> > > > > further free the request, mark in NULL. now ep0_read/write will have ep0req
> > > > > as NULL so bail out.
> > > > > 
> > > > > Is reverse then functionfs_unbind will wait will the ep0_read/write is
> > > > > completed.
> > > > 
> > > > What guarantee is there that the transfer completes?
> > > > 
> > > > If there is such a guarantee, then the request will not be queued, so
> > > > why is usb_ep_dequeue() necessary?
> > > 
> > > I Agree that we cannot say that for sure, but we see that
> > > wait_for_completion in the ep0_queue_wait is also inside mutex which was
> > > acquired in ep0_read/write right?
> > 
> > Correct.
> > 
> > > I Though of maintaining the uniformity for the approaches.
> > 
> > What uniformity?  If one process is blocked waiting for completion and
> > another process wants to cancel the operation, then the cancel
> > (usb_eq_dequeue()) must run concurrently with the wait, otherwise the
> > blocked process will never wake up.
> 
> I get that, we want to rely on the dequeue to get us unblocked.
> But this is also true right that doing dequeue outside might cause this?
> 
> functionfs_unbind
> ep0_dequeue
> 			ffs_ep0_read
> 			mutex_lock()
> giveback		ep0_queue
> 			map request buffer
> unmap buffer
> 
> This can affect the controller's list i.e the pending_list for ep0 or might
> also result on controller accessing a stale memory address isn't it ?
> 
> Or does the mutex would let the ep0_read execute in atomic context? No
> right. Will it not be able to execute parallely? If not then yah we can do
> dequeue outside mutex for sure.

I would expect that if we're in unbind then any attempt to enqueue a new
request will fail, so if the mutex is taken in the case above ep_queue()
should fail with -ESHUTDOWN.

But I can't actually find an point to any code that ensures that is the
case!

This doesn't matter too much though - it's not going to result in any
access to stale memory because either:

	ep0_dequeue
				ffs_ep0_read
				mutex_lock()
				ep0_queue
				... wait for response ...
				read ep0req->status
				mutex_unlock()
	mutex_lock()
	free_ep0_request
	...

or:

				ffs_ep0_read
				mutex_lock()
				ep0_queue
	ep0_dequeue
				wake up
				read ep0req->status
				mutex_unlock()
	mutex_lock()
	free_ep0_request
	...

The first case isn't ideal as we don't want to be waiting for a request
while unbinding, but it's not unsafe.

  reply	other threads:[~2022-11-22 13:08 UTC|newest]

Thread overview: 9+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-11-16 11:19 [v2] usb: gadget: f_fs: Prevent race between functionfs_unbind & ffs_ep0_queue_wait Udipto Goswami
2022-11-18 16:19 ` John Keeping
2022-11-20  6:53   ` Udipto Goswami
2022-11-20 17:48     ` John Keeping
2022-11-21  4:22       ` Udipto Goswami
2022-11-22 11:47         ` John Keeping
2022-11-22 12:26           ` Udipto Goswami
2022-11-22 13:07             ` John Keeping [this message]
2022-11-22 13:40               ` Udipto Goswami

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=Y3zJnQKQslBE413v@donbot \
    --to=john@metanate.com \
    --cc=gregkh@linuxfoundation.org \
    --cc=linux-usb@vger.kernel.org \
    --cc=quic_jackp@quicinc.com \
    --cc=quic_ppratap@quicinc.com \
    --cc=quic_ugoswami@quicinc.com \
    --cc=quic_wcheng@quicinc.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).