From: Alexander Stein <alexander.stein@systec-electronic.com>
To: Ken Goldman <kgold@linux.vnet.ibm.com>
Cc: linux-rt-users@vger.kernel.org, linux-kernel@vger.kernel.org,
tpmdd-devel@lists.sourceforge.net
Subject: Re: [tpmdd-devel] [PATCH v2] tpm_tis: fix stall after iowrite*()s
Date: Thu, 17 Aug 2017 07:57:08 +0200 [thread overview]
Message-ID: <2684601.PnLOm3uNmr@ws-stein> (raw)
In-Reply-To: <13741b28-1b5c-de55-3945-e05911e5a4e2@linux.vnet.ibm.com>
On Wednesday 16 August 2017 17:15:55, Ken Goldman wrote:
> On 8/15/2017 4:13 PM, Haris Okanovic wrote:
> > ioread8() operations to TPM MMIO addresses can stall the cpu when
> > immediately following a sequence of iowrite*()'s to the same region.
> >
> > For example, cyclitest measures ~400us latency spikes when a non-RT
> > usermode application communicates with an SPI-based TPM chip (Intel Atom
> > E3940 system, PREEMPT_RT_FULL kernel). The spikes are caused by a
> > stalling ioread8() operation following a sequence of 30+ iowrite8()s to
> > the same address. I believe this happens because the write sequence is
> > buffered (in cpu or somewhere along the bus), and gets flushed on the
> > first LOAD instruction (ioread*()) that follows.
> >
> > The enclosed change appears to fix this issue: read the TPM chip's
> > access register (status code) after every iowrite*() operation to
> > amortize the cost of flushing data to chip across multiple instructions.
>
> I worry a bit about "appears to fix". It seems odd that the TPM device
> driver would be the first code to uncover this. Can anyone confirm that
> the chipset does indeed have this bug?
No, there was already a similar problem in e1000e where a PCIe read stalled
the CPU, hence no interrupts are serviced. See
https://www.spinics.net/lists/linux-rt-users/msg14077.html
AFAIK there was no outcome though.
> I'd also like an indication of the performance penalty. We're doing a
> lot of work to improve the performance and I worry that "do a read after
> every write" will have a performance impact.
Realtime will always affect performance, but IMHO the latter is much more
important.
Best regards,
Alexander
next prev parent reply other threads:[~2017-08-17 5:57 UTC|newest]
Thread overview: 10+ messages / expand[flat|nested] mbox.gz Atom feed top
[not found] <20170804215651.29247-1-haris.okanovic@ni.com>
2017-08-14 22:53 ` [PATCH] tpm_tis: fix stall after iowrite*()s Haris Okanovic
2017-08-15 6:11 ` Alexander Stein
2017-08-15 20:10 ` Haris Okanovic
[not found] ` <20170804215651.29247-1-haris.okanovic-acOepvfBmUk@public.gmane.org>
2017-08-15 20:13 ` [PATCH v2] " Haris Okanovic
2017-08-16 21:15 ` [tpmdd-devel] " Ken Goldman
2017-08-17 5:57 ` Alexander Stein [this message]
2017-08-17 10:38 ` Sebastian Andrzej Siewior
[not found] ` <20170817103807.ubrbylnud6wxod3s-hfZtesqFncYOwBW4kG4KsQ@public.gmane.org>
2017-08-17 17:17 ` Jason Gunthorpe
[not found] ` <20170817171732.GA22792-ePGOBjL8dl3ta4EC/59zMFaTQe2KTcn/@public.gmane.org>
2017-08-17 20:12 ` Haris Okanovic
2017-08-19 17:03 ` [tpmdd-devel] " Jarkko Sakkinen
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=2684601.PnLOm3uNmr@ws-stein \
--to=alexander.stein@systec-electronic.com \
--cc=kgold@linux.vnet.ibm.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-rt-users@vger.kernel.org \
--cc=tpmdd-devel@lists.sourceforge.net \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).