From: Scott Branden <firstname.lastname@example.org> To: Luis Chamberlain <email@example.com> Cc: Greg Kroah-Hartman <firstname.lastname@example.org>, Andy Gross <email@example.com>, David Brown <firstname.lastname@example.org>, Alexander Viro <email@example.com>, Shuah Khan <firstname.lastname@example.org>, email@example.com, "Rafael J . Wysocki" <firstname.lastname@example.org>, email@example.com, firstname.lastname@example.org, email@example.com, BCM Kernel Feedback <firstname.lastname@example.org>, Olof Johansson <email@example.com>, Andrew Morton <firstname.lastname@example.org>, Dan Carpenter <email@example.com>, Colin Ian King <firstname.lastname@example.org>, Kees Cook <email@example.com>, Takashi Iwai <firstname.lastname@example.org>, email@example.com Subject: Re: [PATCH 3/3] firmware: add mutex fw_lock_fallback for race condition Date: Tue, 20 Aug 2019 08:54:58 -0700 [thread overview] Message-ID: <firstname.lastname@example.org> (raw) In-Reply-To: <20190820012655.GU16384@42.do-not-panic.com> Hi Luis, I'm glad you are a subject expert in this area. Some more comments inline. On 2019-08-19 6:26 p.m., Luis Chamberlain wrote: > On Mon, Aug 19, 2019 at 09:19:51AM -0700, Scott Branden wrote: >> To be honest, I find the entire firmware code sloppy. > And that is after years of cleanup on my part. Try going back to v4.1 > for instance, check the code out then for an incredible horrific sight :) > >> I don't think the cache/no-cache feature is >> implemented or tested properly nor fallback to begin with. > I'm in total agreement! I *know* there must be holes in that code, and I > acknowledge a few possible gotchas on the commit logs. For instance, I > acknowledged that the firmware cache had a secondary purpose which was > not well documented or understood through commit e44565f62a720 > ("firmware: fix batched requests - wake all waiters"). The firmware > cache allows for batching requests and sharing the same original request > for multiple consecutive requests which *race against each other*. > That's when I started having my doubts about the architecture of the > firmware cache mechanism, it seemed too complex and perhaps overkill > and considered killing it. Great (kill it!). I have no need for cached or batched requests. The would remove a lot of problems. > > As I noted in that commit, the firmware cache is used for: > > 1) Addressing races with file lookups during the suspend/resume cycle by > keeping firmware in memory during the suspend/resume cycle > > 2) Batched requests for the same file rely only on work from the first > file lookup, which keeps the firmware in memory until the last > release_firmware() is called > > Also worth quoting from that commit as well: > > "Batched requests *only* take effect if secondary requests come in > prior to the first user calling release_firmware(). The devres name used > for the internal firmware cache is used as a hint other pending requests > are ongoing, the firmware buffer data is kept in memory until the last > user of the buffer calls release_firmware(), therefore serializing > requests and delaying the release until all requests are done." > > Later we discovered that the firmware cache had a serious security issue > since its inception through commit 422b3db2a503 ("firmware: Fix security > issue with request_firmware_into_buf()"). Granted, exploiting this would > require the ability to load kernel code, so the vector of exploitation > is rather small. > > The cache stuff cannot be removed as it *at least* resolves the fw > suspend stuff, but still, this can likely use a revisit in rachitecture > long term. The second implicit use case for batched requests however > seems complex and not sure if its worth to maintain. I'll note that > at least some drivers *do* their own firmware caching, iwlwifi, is one, > so there is an example there to allow drivers to say "I actually don't > need caching" for the future. > > If you're volunteering to cleaning / testing the cache stuff I highly > welcome that. I would only volunteer to remove it, not test or support it. > That and the fallback stuff has been needing testing for > years. Someoone was working on patches on the test case for cache stuff > a while ago, from Intel, but they disappeared. Again, I would only volunteer to remove the fallback mechanism to remove added race conditions. >> I'm not claiming this patch is the final >> solution and indicated such in the cover letter and the comment above. > I missed that sorry. > >> I hope there is someone more familiar with this code to comment further and >> come up with a proper solution. > Alright, I'll dig in and take a look, and propose an alternative. > >> I have found numerous issues and race conditions with the firmware code (I >> simply added a test). > That is nothing compared to the amount of fixes I have found and > actually fixed too, the code was a nightmare before I took on > maintenance. > >> 1) Try loading the same valid firmware using no-cache once it has already >> been loaded with cache. > :) > >> It won't work, which is why I had to use a different filename in the test >> for request_firmware_into_buf. > Alright, I'll go try to fix this. Thanks for the report. I think it's a minor issue compared to the race conditions present. In reality I don't think anyone will load the same firmware using cache vs. no-cache. It's just something I stumbled upon when adding the test case and then had to avoid. > >> 2) Try removing the "if (opt_flags & FW_OPT_NOCACHE)" in my patch and always >> call the mutex. >> >> The firmware test will lock up during a "no uevent" test. I am not familiar >> with the code to >> >> know why such is true and what issue this exposes in the code. > I hinted in my review of the oops what the issue was. I don't know if it's the same bug for the "no uevent" test case though? The test just hangs and the kernel oops is not present. It might be exposing another underlying issue with the request_firmware code. > >> 3) I have a driver that uses request_firmware_into_buf and have multiple >> instances of the driver > Cool, is the driver upstream? I'm working on cleaning up the driver right now to upstream. First thing is I need the request_firmware_into_buf tests accepted upstream. Then I can add my enhancement to request_firmware_into_buf to partial read the file (previous sent out but needed test case). In order to do so Greg K-H required a test case for this but even the current API had no test. In that patch series I can then add the new driver which requires my enhanced request_firmware_into_buf. > >> loading the same firmware in parallel. Some of the data is not read >> correctly in each instance. > Makes perfect sense considering the lack of testing I noted. > >> I haven't yet to reproduce this issue with the firmware test > That's because of batched firmware request mechanism. Is there a way to not use the batch firmware request mechanism when calling request_firmware_into_buf to see if the problem doesn't happen? > >> but currently >> have a mutex around the entire >> call to request_firmware_into_buf in our driver. > I will take a look at this now. > >> Perhaps it is better at this point to add a mutex in >> request_firmware_into_buf to make is entirely safe? > No, that is not sufficient, although it would also solve the > issue. I don't have another solution with all the other mechanisms in play in the current firmware code. For now I'll leave the mutex in the driver I'm upstreaming so it works reliably. > >> (Perhaps even with every request_firmware functions as none seems to be >> tested properly.) > No, you are incorrect. The other firmware API calls *have* been > elaborately tested. The firmware cache stuff *is a mess* however, > since we *use and support it*, I've done my best to salvage it and > document it. OK, I don't use any of the other mechanisms right now. All I require is request_firmware_into_buf. > > I'll take a look at this and propose an alternative solution. > > Luis
next prev parent reply other threads:[~2019-08-20 15:55 UTC|newest] Thread overview: 16+ messages / expand[flat|nested] mbox.gz Atom feed top 2019-08-16 0:09 [PATCH 0/3] firmware: selftest for request_firmware_into_buf Scott Branden 2019-08-16 0:09 ` [PATCH 1/3] test_firmware: add support " Scott Branden 2019-08-19 5:24 ` Luis Chamberlain 2019-08-19 20:27 ` shuah 2019-08-16 0:09 ` [PATCH 2/3] selftest: firmware: Add request_firmware_into_buf tests Scott Branden 2019-08-19 5:24 ` Luis Chamberlain 2019-08-19 20:27 ` shuah 2019-08-16 0:09 ` [PATCH 3/3] firmware: add mutex fw_lock_fallback for race condition Scott Branden 2019-08-19 5:39 ` Luis Chamberlain 2019-08-19 16:19 ` Scott Branden 2019-08-20 1:26 ` Luis Chamberlain 2019-08-20 15:54 ` Scott Branden [this message] 2019-08-23 10:31 ` Takashi Iwai 2019-08-23 15:43 ` Luis Chamberlain 2019-08-23 19:56 ` Scott Branden 2019-08-23 19:48 ` Scott Branden
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --email@example.com \ --firstname.lastname@example.org \ --email@example.com \ --firstname.lastname@example.org \ --email@example.com \ --firstname.lastname@example.org \ --email@example.com \ --firstname.lastname@example.org \ --email@example.com \ --firstname.lastname@example.org \ --email@example.com \ --firstname.lastname@example.org \ --email@example.com \ --firstname.lastname@example.org \ --email@example.com \ --firstname.lastname@example.org \ --email@example.com \ --firstname.lastname@example.org \ --email@example.com \ --firstname.lastname@example.org \ --email@example.com \ --subject='Re: [PATCH 3/3] firmware: add mutex fw_lock_fallback for race condition' \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: link
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).