All of lore.kernel.org
 help / color / mirror / Atom feed
From: Dave Hansen <dave@sr71.net>
To: linux-kernel@vger.kernel.org
Cc: linux-mm@kvack.org, x86@kernel.org, Dave Hansen <dave@sr71.net>,
	dave.hansen@linux.intel.com
Subject: [PATCH 13/31] x86, pkeys: fill in pkey field in siginfo
Date: Wed, 06 Jan 2016 16:01:23 -0800	[thread overview]
Message-ID: <20160107000123.DEB64910@viggo.jf.intel.com> (raw)
In-Reply-To: <20160107000104.1A105322@viggo.jf.intel.com>


From: Dave Hansen <dave.hansen@linux.intel.com>

This fills in the new siginfo field: si_pkey to indicate to
userspace which protection key was set on the PTE that we faulted
on.

Note though that *ALL* protection key faults have to be generated
by a valid, present PTE at some point.  But this code does no PTE
lookups which seeds odd.  The reason is that we take advantage of
the way we generate PTEs from VMAs.  All PTEs under a VMA share
some attributes.  For instance, they are _all_ either PROT_READ
*OR* PROT_NONE.  They also always share a protection key, so we
never have to walk the page tables; we just use the VMA.

Note that _pkey is a 64-bit value.  The current hardware only
supports 4-bit protection keys.  We do this because there is
_plenty_ of space in _sigfault and it is possible that future
processors would support more than 4 bits of protection keys.

Signed-off-by: Dave Hansen <dave.hansen@linux.intel.com>
Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
---

 b/arch/x86/include/asm/pgtable_types.h |    5 ++
 b/arch/x86/mm/fault.c                  |   64 ++++++++++++++++++++++++++++++++-
 2 files changed, 68 insertions(+), 1 deletion(-)

diff -puN arch/x86/include/asm/pgtable_types.h~pkeys-09-siginfo-x86 arch/x86/include/asm/pgtable_types.h
--- a/arch/x86/include/asm/pgtable_types.h~pkeys-09-siginfo-x86	2016-01-06 15:50:08.273276052 -0800
+++ b/arch/x86/include/asm/pgtable_types.h	2016-01-06 15:50:08.278276277 -0800
@@ -64,6 +64,11 @@
 #endif
 #define __HAVE_ARCH_PTE_SPECIAL
 
+#define _PAGE_PKEY_MASK (_PAGE_PKEY_BIT0 | \
+			 _PAGE_PKEY_BIT1 | \
+			 _PAGE_PKEY_BIT2 | \
+			 _PAGE_PKEY_BIT3)
+
 #ifdef CONFIG_KMEMCHECK
 #define _PAGE_HIDDEN	(_AT(pteval_t, 1) << _PAGE_BIT_HIDDEN)
 #else
diff -puN arch/x86/mm/fault.c~pkeys-09-siginfo-x86 arch/x86/mm/fault.c
--- a/arch/x86/mm/fault.c~pkeys-09-siginfo-x86	2016-01-06 15:50:08.275276142 -0800
+++ b/arch/x86/mm/fault.c	2016-01-06 15:50:08.279276323 -0800
@@ -15,12 +15,14 @@
 #include <linux/context_tracking.h>	/* exception_enter(), ...	*/
 #include <linux/uaccess.h>		/* faulthandler_disabled()	*/
 
+#include <asm/cpufeature.h>		/* boot_cpu_has, ...		*/
 #include <asm/traps.h>			/* dotraplinkage, ...		*/
 #include <asm/pgalloc.h>		/* pgd_*(), ...			*/
 #include <asm/kmemcheck.h>		/* kmemcheck_*(), ...		*/
 #include <asm/fixmap.h>			/* VSYSCALL_ADDR		*/
 #include <asm/vsyscall.h>		/* emulate_vsyscall		*/
 #include <asm/vm86.h>			/* struct vm86			*/
+#include <asm/mmu_context.h>		/* vma_pkey()			*/
 
 #define CREATE_TRACE_POINTS
 #include <asm/trace/exceptions.h>
@@ -169,6 +171,56 @@ is_prefetch(struct pt_regs *regs, unsign
 	return prefetch;
 }
 
+/*
+ * A protection key fault means that the PKRU value did not allow
+ * access to some PTE.  Userspace can figure out what PKRU was
+ * from the XSAVE state, and this function fills out a field in
+ * siginfo so userspace can discover which protection key was set
+ * on the PTE.
+ *
+ * If we get here, we know that the hardware signaled a PF_PK
+ * fault and that there was a VMA once we got in the fault
+ * handler.  It does *not* guarantee that the VMA we find here
+ * was the one that we faulted on.
+ *
+ * 1. T1   : mprotect_key(foo, PAGE_SIZE, pkey=4);
+ * 2. T1   : set PKRU to deny access to pkey=4, touches page
+ * 3. T1   : faults...
+ * 4.    T2: mprotect_key(foo, PAGE_SIZE, pkey=5);
+ * 5. T1   : enters fault handler, takes mmap_sem, etc...
+ * 6. T1   : reaches here, sees vma_pkey(vma)=5, when we really
+ *	     faulted on a pte with its pkey=4.
+ */
+static void fill_sig_info_pkey(int si_code, siginfo_t *info,
+		struct vm_area_struct *vma)
+{
+	/* This is effectively an #ifdef */
+	if (!boot_cpu_has(X86_FEATURE_OSPKE))
+		return;
+
+	/* Fault not from Protection Keys: nothing to do */
+	if (si_code != SEGV_PKUERR)
+		return;
+	/*
+	 * force_sig_info_fault() is called from a number of
+	 * contexts, some of which have a VMA and some of which
+	 * do not.  The PF_PK handing happens after we have a
+	 * valid VMA, so we should never reach this without a
+	 * valid VMA.
+	 */
+	if (!vma) {
+		WARN_ONCE(1, "PKU fault with no VMA passed in");
+		info->si_pkey = 0;
+		return;
+	}
+	/*
+	 * si_pkey should be thought of as a strong hint, but not
+	 * absolutely guranteed to be 100% accurate because of
+	 * the race explained above.
+	 */
+	info->si_pkey = vma_pkey(vma);
+}
+
 static void
 force_sig_info_fault(int si_signo, int si_code, unsigned long address,
 		     struct task_struct *tsk, struct vm_area_struct *vma,
@@ -187,6 +239,8 @@ force_sig_info_fault(int si_signo, int s
 		lsb = PAGE_SHIFT;
 	info.si_addr_lsb = lsb;
 
+	fill_sig_info_pkey(si_code, &info, vma);
+
 	force_sig_info(si_signo, &info, tsk);
 }
 
@@ -847,7 +901,15 @@ static noinline void
 bad_area_access_error(struct pt_regs *regs, unsigned long error_code,
 		      unsigned long address, struct vm_area_struct *vma)
 {
-	__bad_area(regs, error_code, address, vma, SEGV_ACCERR);
+	/*
+	 * This OSPKE check is not strictly necessary at runtime.
+	 * But, doing it this way allows compiler optimizations
+	 * if pkeys are compiled out.
+	 */
+	if (boot_cpu_has(X86_FEATURE_OSPKE) && (error_code & PF_PK))
+		__bad_area(regs, error_code, address, vma, SEGV_PKUERR);
+	else
+		__bad_area(regs, error_code, address, vma, SEGV_ACCERR);
 }
 
 static void
_

WARNING: multiple messages have this Message-ID (diff)
From: Dave Hansen <dave@sr71.net>
To: linux-kernel@vger.kernel.org
Cc: linux-mm@kvack.org, x86@kernel.org, Dave Hansen <dave@sr71.net>,
	dave.hansen@linux.intel.com
Subject: [PATCH 13/31] x86, pkeys: fill in pkey field in siginfo
Date: Wed, 06 Jan 2016 16:01:23 -0800	[thread overview]
Message-ID: <20160107000123.DEB64910@viggo.jf.intel.com> (raw)
In-Reply-To: <20160107000104.1A105322@viggo.jf.intel.com>


From: Dave Hansen <dave.hansen@linux.intel.com>

This fills in the new siginfo field: si_pkey to indicate to
userspace which protection key was set on the PTE that we faulted
on.

Note though that *ALL* protection key faults have to be generated
by a valid, present PTE at some point.  But this code does no PTE
lookups which seeds odd.  The reason is that we take advantage of
the way we generate PTEs from VMAs.  All PTEs under a VMA share
some attributes.  For instance, they are _all_ either PROT_READ
*OR* PROT_NONE.  They also always share a protection key, so we
never have to walk the page tables; we just use the VMA.

Note that _pkey is a 64-bit value.  The current hardware only
supports 4-bit protection keys.  We do this because there is
_plenty_ of space in _sigfault and it is possible that future
processors would support more than 4 bits of protection keys.

Signed-off-by: Dave Hansen <dave.hansen@linux.intel.com>
Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
---

 b/arch/x86/include/asm/pgtable_types.h |    5 ++
 b/arch/x86/mm/fault.c                  |   64 ++++++++++++++++++++++++++++++++-
 2 files changed, 68 insertions(+), 1 deletion(-)

diff -puN arch/x86/include/asm/pgtable_types.h~pkeys-09-siginfo-x86 arch/x86/include/asm/pgtable_types.h
--- a/arch/x86/include/asm/pgtable_types.h~pkeys-09-siginfo-x86	2016-01-06 15:50:08.273276052 -0800
+++ b/arch/x86/include/asm/pgtable_types.h	2016-01-06 15:50:08.278276277 -0800
@@ -64,6 +64,11 @@
 #endif
 #define __HAVE_ARCH_PTE_SPECIAL
 
+#define _PAGE_PKEY_MASK (_PAGE_PKEY_BIT0 | \
+			 _PAGE_PKEY_BIT1 | \
+			 _PAGE_PKEY_BIT2 | \
+			 _PAGE_PKEY_BIT3)
+
 #ifdef CONFIG_KMEMCHECK
 #define _PAGE_HIDDEN	(_AT(pteval_t, 1) << _PAGE_BIT_HIDDEN)
 #else
diff -puN arch/x86/mm/fault.c~pkeys-09-siginfo-x86 arch/x86/mm/fault.c
--- a/arch/x86/mm/fault.c~pkeys-09-siginfo-x86	2016-01-06 15:50:08.275276142 -0800
+++ b/arch/x86/mm/fault.c	2016-01-06 15:50:08.279276323 -0800
@@ -15,12 +15,14 @@
 #include <linux/context_tracking.h>	/* exception_enter(), ...	*/
 #include <linux/uaccess.h>		/* faulthandler_disabled()	*/
 
+#include <asm/cpufeature.h>		/* boot_cpu_has, ...		*/
 #include <asm/traps.h>			/* dotraplinkage, ...		*/
 #include <asm/pgalloc.h>		/* pgd_*(), ...			*/
 #include <asm/kmemcheck.h>		/* kmemcheck_*(), ...		*/
 #include <asm/fixmap.h>			/* VSYSCALL_ADDR		*/
 #include <asm/vsyscall.h>		/* emulate_vsyscall		*/
 #include <asm/vm86.h>			/* struct vm86			*/
+#include <asm/mmu_context.h>		/* vma_pkey()			*/
 
 #define CREATE_TRACE_POINTS
 #include <asm/trace/exceptions.h>
@@ -169,6 +171,56 @@ is_prefetch(struct pt_regs *regs, unsign
 	return prefetch;
 }
 
+/*
+ * A protection key fault means that the PKRU value did not allow
+ * access to some PTE.  Userspace can figure out what PKRU was
+ * from the XSAVE state, and this function fills out a field in
+ * siginfo so userspace can discover which protection key was set
+ * on the PTE.
+ *
+ * If we get here, we know that the hardware signaled a PF_PK
+ * fault and that there was a VMA once we got in the fault
+ * handler.  It does *not* guarantee that the VMA we find here
+ * was the one that we faulted on.
+ *
+ * 1. T1   : mprotect_key(foo, PAGE_SIZE, pkey=4);
+ * 2. T1   : set PKRU to deny access to pkey=4, touches page
+ * 3. T1   : faults...
+ * 4.    T2: mprotect_key(foo, PAGE_SIZE, pkey=5);
+ * 5. T1   : enters fault handler, takes mmap_sem, etc...
+ * 6. T1   : reaches here, sees vma_pkey(vma)=5, when we really
+ *	     faulted on a pte with its pkey=4.
+ */
+static void fill_sig_info_pkey(int si_code, siginfo_t *info,
+		struct vm_area_struct *vma)
+{
+	/* This is effectively an #ifdef */
+	if (!boot_cpu_has(X86_FEATURE_OSPKE))
+		return;
+
+	/* Fault not from Protection Keys: nothing to do */
+	if (si_code != SEGV_PKUERR)
+		return;
+	/*
+	 * force_sig_info_fault() is called from a number of
+	 * contexts, some of which have a VMA and some of which
+	 * do not.  The PF_PK handing happens after we have a
+	 * valid VMA, so we should never reach this without a
+	 * valid VMA.
+	 */
+	if (!vma) {
+		WARN_ONCE(1, "PKU fault with no VMA passed in");
+		info->si_pkey = 0;
+		return;
+	}
+	/*
+	 * si_pkey should be thought of as a strong hint, but not
+	 * absolutely guranteed to be 100% accurate because of
+	 * the race explained above.
+	 */
+	info->si_pkey = vma_pkey(vma);
+}
+
 static void
 force_sig_info_fault(int si_signo, int si_code, unsigned long address,
 		     struct task_struct *tsk, struct vm_area_struct *vma,
@@ -187,6 +239,8 @@ force_sig_info_fault(int si_signo, int s
 		lsb = PAGE_SHIFT;
 	info.si_addr_lsb = lsb;
 
+	fill_sig_info_pkey(si_code, &info, vma);
+
 	force_sig_info(si_signo, &info, tsk);
 }
 
@@ -847,7 +901,15 @@ static noinline void
 bad_area_access_error(struct pt_regs *regs, unsigned long error_code,
 		      unsigned long address, struct vm_area_struct *vma)
 {
-	__bad_area(regs, error_code, address, vma, SEGV_ACCERR);
+	/*
+	 * This OSPKE check is not strictly necessary at runtime.
+	 * But, doing it this way allows compiler optimizations
+	 * if pkeys are compiled out.
+	 */
+	if (boot_cpu_has(X86_FEATURE_OSPKE) && (error_code & PF_PK))
+		__bad_area(regs, error_code, address, vma, SEGV_PKUERR);
+	else
+		__bad_area(regs, error_code, address, vma, SEGV_ACCERR);
 }
 
 static void
_

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

  parent reply	other threads:[~2016-01-07  0:06 UTC|newest]

Thread overview: 87+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2016-01-07  0:01 [PATCH 00/31] x86: Memory Protection Keys (v8) Dave Hansen
2016-01-07  0:01 ` Dave Hansen
2016-01-07  0:01 ` [PATCH 01/31] mm, gup: introduce concept of "foreign" get_user_pages() Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-13 19:00   ` Vlastimil Babka
2016-01-13 19:00     ` Vlastimil Babka
2016-01-13 19:16     ` Dave Hansen
2016-01-13 19:16       ` Dave Hansen
2016-01-07  0:01 ` [PATCH 02/31] x86, fpu: add placeholder for Processor Trace XSAVE state Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 03/31] x86, pkeys: Add Kconfig option Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 04/31] x86, pkeys: cpuid bit definition Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 05/31] x86, pkeys: define new CR4 bit Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 06/31] x86, pkeys: add PKRU xsave fields and data structure(s) Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 07/31] x86, pkeys: PTE bits for storing protection key Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 08/31] x86, pkeys: new page fault error code bit: PF_PK Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 09/31] x86, pkeys: store protection in high VMA flags Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 10/31] x86, pkeys: arch-specific protection bits Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-08 19:31   ` Thomas Gleixner
2016-01-07  0:01 ` [PATCH 11/31] x86, pkeys: pass VMA down in to fault signal generation code Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 12/31] signals, pkeys: notify userspace about protection key faults Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` Dave Hansen [this message]
2016-01-07  0:01   ` [PATCH 13/31] x86, pkeys: fill in pkey field in siginfo Dave Hansen
2016-01-07  0:01 ` [PATCH 14/31] x86, pkeys: add functions to fetch PKRU Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-08 19:32   ` Thomas Gleixner
2016-01-07  0:01 ` [PATCH 15/31] mm: factor out VMA fault permission checking Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 16/31] x86, mm: simplify get_user_pages() PTE bit handling Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 17/31] x86, pkeys: check VMAs and PTEs for protection keys Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 18/31] mm: add gup flag to indicate "foreign" mm access Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 19/31] x86, pkeys: optimize fault handling in access_error() Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 20/31] x86, pkeys: differentiate instruction fetches Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 21/31] x86, pkeys: dump PKRU with other kernel registers Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 22/31] x86, pkeys: dump pkey from VMA in /proc/pid/smaps Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 23/31] x86, pkeys: add Kconfig prompt to existing config option Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 24/31] x86, pkeys: actually enable Memory Protection Keys in CPU Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 25/31] mm, multi-arch: pass a protection key in to calc_vm_flag_bits() Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 26/31] x86, pkeys: add arch_validate_pkey() Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-08 19:34   ` Thomas Gleixner
2016-01-07  0:01 ` [PATCH 27/31] x86: separate out LDT init from context init Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 28/31] x86, fpu: allow setting of XSAVE state Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07  0:01 ` [PATCH 29/31] x86, pkeys: allow kernel to modify user pkey rights register Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-08 19:40   ` Thomas Gleixner
2016-01-07  0:01 ` [PATCH 30/31] x86, pkeys: create an x86 arch_calc_vm_prot_bits() for VMA flags Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-08 19:40   ` Thomas Gleixner
2016-01-07  0:01 ` [PATCH 31/31] x86, pkeys: execute-only support Dave Hansen
2016-01-07  0:01   ` Dave Hansen
2016-01-07 21:02   ` Kees Cook
2016-01-07 21:02     ` Kees Cook
2016-01-07 22:25     ` Dave Hansen
2016-01-07 22:25       ` Dave Hansen
2016-01-07 21:10   ` Andy Lutomirski
2016-01-07 21:10     ` Andy Lutomirski
2016-01-07 22:13     ` Dave Hansen
2016-01-07 22:13       ` Dave Hansen
2016-01-07 22:44       ` Andy Lutomirski
2016-01-07 22:44         ` Andy Lutomirski
2016-01-08 19:51   ` Thomas Gleixner
2016-01-08 19:51     ` Thomas Gleixner
2016-01-29 18:16 [PATCH 00/31] x86: Memory Protection Keys (v9) Dave Hansen
2016-01-29 18:17 ` [PATCH 13/31] x86, pkeys: fill in pkey field in siginfo Dave Hansen
2016-01-29 18:17   ` Dave Hansen

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20160107000123.DEB64910@viggo.jf.intel.com \
    --to=dave@sr71.net \
    --cc=dave.hansen@linux.intel.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=x86@kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.