From: Randy Dunlap <rdunlap@infradead.org>
To: linux-kernel@vger.kernel.org
Cc: Randy Dunlap <rdunlap@infradead.org>,
Jason Gunthorpe <jgg@nvidia.com>,
Leon Romanovsky <leon@kernel.org>,
Dennis Dalessandro <dennis.dalessandro@cornelisnetworks.com>,
linux-rdma@vger.kernel.org, Jeff Dike <jdike@addtoit.com>,
Richard Weinberger <richard@nod.at>,
Anton Ivanov <anton.ivanov@cambridgegreys.com>,
Johannes Berg <johannes@sipsolutions.net>,
linux-um@lists.infradead.org
Subject: [PATCH 2/2] IB/rdmavt: don't use rdmavt for UML
Date: Wed, 30 Nov 2022 12:09:58 -0800 [thread overview]
Message-ID: <20221130200958.25305-1-rdunlap@infradead.org> (raw)
When building rdmavt for ARCH=um, qp.c has a build error on a reference
to the x86-specific cpuinfo field 'x86_cache_size'.
Fix the build errors by making this driver depend on !UML.
Prevents these build errors:
../drivers/infiniband/sw/rdmavt/qp.c: In function ‘rvt_wss_llc_size’:
../drivers/infiniband/sw/rdmavt/qp.c:88:29: error: ‘struct cpuinfo_um’ has no member named ‘x86_cache_size’
88 | return boot_cpu_data.x86_cache_size;
../drivers/infiniband/sw/rdmavt/qp.c: In function ‘cacheless_memcpy’:
../drivers/infiniband/sw/rdmavt/qp.c:100:9: error: implicit declaration of function ‘__copy_user_nocache’; did you mean ‘copy_user_page’? [-Werror=implicit-function-declaration]
100 | __copy_user_nocache(dst, (void __user *)src, n, 0);
../drivers/infiniband/sw/rdmavt/qp.c: In function ‘rvt_wss_llc_size’:
../drivers/infiniband/sw/rdmavt/qp.c:89:1: error: control reaches end of non-void function [-Werror=return-type]
89 | }
Fixes: 68f5d3f3b654 ("um: add PCI over virtio emulation driver")
Signed-off-by: Randy Dunlap <rdunlap@infradead.org>
Cc: Jason Gunthorpe <jgg@nvidia.com>
Cc: Leon Romanovsky <leon@kernel.org>
Cc: Dennis Dalessandro <dennis.dalessandro@cornelisnetworks.com>
Cc: linux-rdma@vger.kernel.org
Cc: Jeff Dike <jdike@addtoit.com>
Cc: Richard Weinberger <richard@nod.at>
Cc: Anton Ivanov <anton.ivanov@cambridgegreys.com>
Cc: Johannes Berg <johannes@sipsolutions.net>
Cc: linux-um@lists.infradead.org
---
v2: rebase & resend
drivers/infiniband/sw/rdmavt/Kconfig | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff -- a/drivers/infiniband/sw/rdmavt/Kconfig b/drivers/infiniband/sw/rdmavt/Kconfig
--- a/drivers/infiniband/sw/rdmavt/Kconfig
+++ b/drivers/infiniband/sw/rdmavt/Kconfig
@@ -2,7 +2,7 @@
config INFINIBAND_RDMAVT
tristate "RDMA verbs transport library"
depends on INFINIBAND_VIRT_DMA
- depends on X86_64
+ depends on X86_64 && !UML
depends on PCI
help
This is a common software verbs provider for RDMA networks.
next reply other threads:[~2022-11-30 20:11 UTC|newest]
Thread overview: 2+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-11-30 20:09 Randy Dunlap [this message]
2022-12-01 9:22 ` [PATCH 2/2] IB/rdmavt: don't use rdmavt for UML Leon Romanovsky
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20221130200958.25305-1-rdunlap@infradead.org \
--to=rdunlap@infradead.org \
--cc=anton.ivanov@cambridgegreys.com \
--cc=dennis.dalessandro@cornelisnetworks.com \
--cc=jdike@addtoit.com \
--cc=jgg@nvidia.com \
--cc=johannes@sipsolutions.net \
--cc=leon@kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-rdma@vger.kernel.org \
--cc=linux-um@lists.infradead.org \
--cc=richard@nod.at \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).