linux-usb.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Oliver Neukum <oneukum@suse.de>
To: Greg KH <gregkh@linuxfoundation.org>
Cc: keithp@keithp.com, linux-usb@vger.kernel.org
Subject: Re: [PATCH] USB: chaoskey: fix error case of a timeout
Date: Thu, 07 Nov 2019 18:06:40 +0100	[thread overview]
Message-ID: <1573146400.13325.1.camel@suse.de> (raw)
In-Reply-To: <20191107150140.GA154021@kroah.com>

Am Donnerstag, den 07.11.2019, 16:01 +0100 schrieb Greg KH:
> On Thu, Nov 07, 2019 at 03:28:56PM +0100, Oliver Neukum wrote:
> > In case of a timeout or if a signal aborts a read
> > communication with the device needs to be ended
> > lest we overwrite an active URB the next time we
> > do IO to the device, as the URB may still be active.
> > 
> > Signed-off-by: Oliver Neukum <oneukum@suse.de>
> > ---
> >  drivers/usb/misc/chaoskey.c | 24 +++++++++++++++++++++---
> >  1 file changed, 21 insertions(+), 3 deletions(-)
> > 
> > diff --git a/drivers/usb/misc/chaoskey.c b/drivers/usb/misc/chaoskey.c
> > index 34e6cd6f40d3..87067c3d6109 100644
> > --- a/drivers/usb/misc/chaoskey.c
> > +++ b/drivers/usb/misc/chaoskey.c
> > @@ -384,13 +384,17 @@ static int _chaoskey_fill(struct chaoskey *dev)
> >  		!dev->reading,
> >  		(started ? NAK_TIMEOUT : ALEA_FIRST_TIMEOUT) );
> >  
> > -	if (result < 0)
> > +	if (result < 0) {
> > +		usb_kill_urb(dev->urb);
> >  		goto out;
> > +	}
> >  
> > -	if (result == 0)
> > +	if (result == 0) {
> >  		result = -ETIMEDOUT;
> > -	else
> > +		usb_kill_urb(dev->urb);
> > +	} else {
> >  		result = dev->valid;
> > +	}
> >  out:
> >  	/* Let the device go back to sleep eventually */
> >  	usb_autopm_put_interface(dev->interface);
> > @@ -526,7 +530,21 @@ static int chaoskey_suspend(struct usb_interface *interface,
> >  
> >  static int chaoskey_resume(struct usb_interface *interface)
> >  {
> > +	struct chaoskey *dev;
> > +	struct usb_device *udev = interface_to_usbdev(interface);
> > +
> >  	usb_dbg(interface, "resume");
> > +	dev = usb_get_intfdata(interface);
> > +
> > +	/*
> > +	 * We may have lost power.
> > +	 * In that case the device that needs a long time
> > +	 * for the first requests needs an extended timeout
> > +	 * again
> > +	 */
> > +	if (le16_to_cpu(udev->descriptor.idVendor) == ALEA_VENDOR_ID)
> 
> What is this helping with?

THe quirk is specific for this model. The others do
not need it.

> > +		dev->reads_started = false;
> > +
> >  	return 0;
> >  }
> >  #else
> > -- 
> > 2.16.4
> > 
> 
> Why send this twice?

There must be an issue in my scripts. Investigating.

> And did this pass the syzbot testing?

Inconclusive. The test crashed but in another place.
I cannot claim that fix.

	Regards
		Oliver


      reply	other threads:[~2019-11-07 17:22 UTC|newest]

Thread overview: 4+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2019-11-07 14:28 [PATCH] USB: chaoskey: fix error case of a timeout Oliver Neukum
2019-11-07 14:28 ` Oliver Neukum
2019-11-07 15:01   ` Greg KH
2019-11-07 17:06     ` Oliver Neukum [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=1573146400.13325.1.camel@suse.de \
    --to=oneukum@suse.de \
    --cc=gregkh@linuxfoundation.org \
    --cc=keithp@keithp.com \
    --cc=linux-usb@vger.kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).