All of lore.kernel.org
 help / color / mirror / Atom feed
From: Kees Cook <keescook@chromium.org>
To: linux-kernel@vger.kernel.org
Cc: Kees Cook <keescook@chromium.org>, Rik van Riel <riel@redhat.com>,
	Casey Schaufler <casey@schaufler-ca.com>,
	PaX Team <pageexec@freemail.hu>,
	Brad Spengler <spender@grsecurity.net>,
	Russell King <linux@armlinux.org.uk>,
	Catalin Marinas <catalin.marinas@arm.com>,
	Will Deacon <will.deacon@arm.com>,
	Ard Biesheuvel <ard.biesheuvel@linaro.org>,
	Benjamin Herrenschmidt <benh@kernel.crashing.org>,
	Michael Ellerman <mpe@ellerman.id.au>,
	Tony Luck <tony.luck@intel.com>,
	Fenghua Yu <fenghua.yu@intel.com>,
	"David S. Miller" <davem@davemloft.net>,
	x86@kernel.org, Christoph Lameter <cl@linux.com>,
	Pekka Enberg <penberg@kernel.org>,
	David Rientjes <rientjes@google.com>,
	Joonsoo Kim <iamjoonsoo.kim@lge.com>,
	Andrew Morton <akpm@linux-foundation.org>,
	Andy Lutomirski <luto@kernel.org>, Borislav Petkov <bp@suse.de>,
	Mathias Krause <minipli@googlemail.com>, Jan Kara <jack@suse.cz>,
	Vitaly Wool <vitalywool@gmail.com>,
	Andrea Arcangeli <aarcange@redhat.com>,
	Dmitry Vyukov <dvyukov@google.com>,
	Laura Abbott <labbott@fedoraproject.org>,
	linux-arm-kernel@lists.infradead.org, linux-ia64@vger.kernel.org,
	linuxppc-dev@lists.ozlabs.org, sparclinux@vger.kernel.org,
	linux-arch@vger.kernel.org, linux-mm@kvack.org,
	kernel-hardening@lists.openwall.com
Subject: [PATCH 9/9] mm: SLUB hardened usercopy support
Date: Wed,  6 Jul 2016 15:25:28 -0700	[thread overview]
Message-ID: <1467843928-29351-10-git-send-email-keescook@chromium.org> (raw)
In-Reply-To: <1467843928-29351-1-git-send-email-keescook@chromium.org>

Under CONFIG_HARDENED_USERCOPY, this adds object size checking to the
SLUB allocator to catch any copies that may span objects.

Based on code from PaX and grsecurity.

Signed-off-by: Kees Cook <keescook@chromium.org>
---
 init/Kconfig |  1 +
 mm/slub.c    | 27 +++++++++++++++++++++++++++
 2 files changed, 28 insertions(+)

diff --git a/init/Kconfig b/init/Kconfig
index 798c2020ee7c..1c4711819dfd 100644
--- a/init/Kconfig
+++ b/init/Kconfig
@@ -1765,6 +1765,7 @@ config SLAB
 
 config SLUB
 	bool "SLUB (Unqueued Allocator)"
+	select HAVE_HARDENED_USERCOPY_ALLOCATOR
 	help
 	   SLUB is a slab allocator that minimizes cache line usage
 	   instead of managing queues of cached objects (SLAB approach).
diff --git a/mm/slub.c b/mm/slub.c
index 825ff4505336..0c8ace04f075 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -3614,6 +3614,33 @@ void *__kmalloc_node(size_t size, gfp_t flags, int node)
 EXPORT_SYMBOL(__kmalloc_node);
 #endif
 
+#ifdef CONFIG_HARDENED_USERCOPY
+/*
+ * Rejects objects that are incorrectly sized.
+ *
+ * Returns NULL if check passes, otherwise const char * to name of cache
+ * to indicate an error.
+ */
+const char *__check_heap_object(const void *ptr, unsigned long n,
+				struct page *page)
+{
+	struct kmem_cache *s;
+	unsigned long offset;
+
+	/* Find object. */
+	s = page->slab_cache;
+
+	/* Find offset within object. */
+	offset = (ptr - page_address(page)) % s->size;
+
+	/* Allow address range falling entirely within object size. */
+	if (offset <= s->object_size && n <= s->object_size - offset)
+		return NULL;
+
+	return s->name;
+}
+#endif /* CONFIG_HARDENED_USERCOPY */
+
 static size_t __ksize(const void *object)
 {
 	struct page *page;
-- 
2.7.4

WARNING: multiple messages have this Message-ID (diff)
From: Kees Cook <keescook@chromium.org>
To: linux-kernel@vger.kernel.org
Cc: Kees Cook <keescook@chromium.org>, Rik van Riel <riel@redhat.com>,
	Casey Schaufler <casey@schaufler-ca.com>,
	PaX Team <pageexec@freemail.hu>,
	Brad Spengler <spender@grsecurity.net>,
	Russell King <linux@armlinux.org.uk>,
	Catalin Marinas <catalin.marinas@arm.com>,
	Will Deacon <will.deacon@arm.com>,
	Ard Biesheuvel <ard.biesheuvel@linaro.org>,
	Benjamin Herrenschmidt <benh@kernel.crashing.org>,
	Michael Ellerman <mpe@ellerman.id.au>,
	Tony Luck <tony.luck@intel.com>,
	Fenghua Yu <fenghua.yu@intel.com>,
	"David S. Miller" <davem@davemloft.net>,
	x86@kernel.org, Christoph Lameter <cl@linux.com>,
	Pekka Enberg <penberg@kernel.org>,
	David Rientjes <rientjes@google.com>,
	Joonsoo Kim <iamjoonsoo.kim@lge.com>,
	Andrew Morton <akpm@linux-foundation.org>,
	Andy Lutomirski <luto@kernel.org>, Borislav Petkov <bp@suse.de>,
	Mathias Krause <minipli@googlemail.com>, Jan Kara <jack@suse.cz>
Subject: [PATCH 9/9] mm: SLUB hardened usercopy support
Date: Wed,  6 Jul 2016 15:25:28 -0700	[thread overview]
Message-ID: <1467843928-29351-10-git-send-email-keescook@chromium.org> (raw)
In-Reply-To: <1467843928-29351-1-git-send-email-keescook@chromium.org>

Under CONFIG_HARDENED_USERCOPY, this adds object size checking to the
SLUB allocator to catch any copies that may span objects.

Based on code from PaX and grsecurity.

Signed-off-by: Kees Cook <keescook@chromium.org>
---
 init/Kconfig |  1 +
 mm/slub.c    | 27 +++++++++++++++++++++++++++
 2 files changed, 28 insertions(+)

diff --git a/init/Kconfig b/init/Kconfig
index 798c2020ee7c..1c4711819dfd 100644
--- a/init/Kconfig
+++ b/init/Kconfig
@@ -1765,6 +1765,7 @@ config SLAB
 
 config SLUB
 	bool "SLUB (Unqueued Allocator)"
+	select HAVE_HARDENED_USERCOPY_ALLOCATOR
 	help
 	   SLUB is a slab allocator that minimizes cache line usage
 	   instead of managing queues of cached objects (SLAB approach).
diff --git a/mm/slub.c b/mm/slub.c
index 825ff4505336..0c8ace04f075 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -3614,6 +3614,33 @@ void *__kmalloc_node(size_t size, gfp_t flags, int node)
 EXPORT_SYMBOL(__kmalloc_node);
 #endif
 
+#ifdef CONFIG_HARDENED_USERCOPY
+/*
+ * Rejects objects that are incorrectly sized.
+ *
+ * Returns NULL if check passes, otherwise const char * to name of cache
+ * to indicate an error.
+ */
+const char *__check_heap_object(const void *ptr, unsigned long n,
+				struct page *page)
+{
+	struct kmem_cache *s;
+	unsigned long offset;
+
+	/* Find object. */
+	s = page->slab_cache;
+
+	/* Find offset within object. */
+	offset = (ptr - page_address(page)) % s->size;
+
+	/* Allow address range falling entirely within object size. */
+	if (offset <= s->object_size && n <= s->object_size - offset)
+		return NULL;
+
+	return s->name;
+}
+#endif /* CONFIG_HARDENED_USERCOPY */
+
 static size_t __ksize(const void *object)
 {
 	struct page *page;
-- 
2.7.4

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

WARNING: multiple messages have this Message-ID (diff)
From: Kees Cook <keescook@chromium.org>
To: linux-kernel@vger.kernel.org
Cc: Kees Cook <keescook@chromium.org>, Rik van Riel <riel@redhat.com>,
	Casey Schaufler <casey@schaufler-ca.com>,
	PaX Team <pageexec@freemail.hu>,
	Brad Spengler <spender@grsecurity.net>,
	Russell King <linux@armlinux.org.uk>,
	Catalin Marinas <catalin.marinas@arm.com>,
	Will Deacon <will.deacon@arm.com>,
	Ard Biesheuvel <ard.biesheuvel@linaro.org>,
	Benjamin Herrenschmidt <benh@kernel.crashing.org>,
	Michael Ellerman <mpe@ellerman.id.au>,
	Tony Luck <tony.luck@intel.com>,
	Fenghua Yu <fenghua.yu@intel.com>,
	"David S. Miller" <davem@davemloft.net>,
	x86@kernel.org, Christoph Lameter <cl@linux.com>,
	Pekka Enberg <penberg@kernel.org>,
	David Rientjes <rientjes@google.com>,
	Joonsoo Kim <iamjoonsoo.kim@lge.com>,
	Andrew Morton <akpm@linux-foundation.org>,
	Andy Lutomirski <luto@kernel.org>, Borislav Petkov <bp@suse.de>,
	Mathias Krause <minipli@googlemail.com>, Jan Kara <jack@suse.cz>,
	Vitaly Wool <vitalywool@gmail.com>,
	Andrea Arcangeli <aarcange@redhat.com>,
	Dmitry Vyukov <dvyukov@google.com>,
	Laura Abbott <labbott@fedoraproject.org>,
	linux-arm-kernel@lists.infradead.org, linux-ia64@vger.kernel.org,
	linuxppc-dev@lists.ozlabs.org, sparclinux@vger.kernel.org,
	linux-arch@vger.kernel.org, linux-mm@kvack.org,
	kernel-hardening@lists.openwall.com
Subject: [PATCH 9/9] mm: SLUB hardened usercopy support
Date: Wed, 06 Jul 2016 22:25:28 +0000	[thread overview]
Message-ID: <1467843928-29351-10-git-send-email-keescook@chromium.org> (raw)
In-Reply-To: <1467843928-29351-1-git-send-email-keescook@chromium.org>

Under CONFIG_HARDENED_USERCOPY, this adds object size checking to the
SLUB allocator to catch any copies that may span objects.

Based on code from PaX and grsecurity.

Signed-off-by: Kees Cook <keescook@chromium.org>
---
 init/Kconfig |  1 +
 mm/slub.c    | 27 +++++++++++++++++++++++++++
 2 files changed, 28 insertions(+)

diff --git a/init/Kconfig b/init/Kconfig
index 798c2020ee7c..1c4711819dfd 100644
--- a/init/Kconfig
+++ b/init/Kconfig
@@ -1765,6 +1765,7 @@ config SLAB
 
 config SLUB
 	bool "SLUB (Unqueued Allocator)"
+	select HAVE_HARDENED_USERCOPY_ALLOCATOR
 	help
 	   SLUB is a slab allocator that minimizes cache line usage
 	   instead of managing queues of cached objects (SLAB approach).
diff --git a/mm/slub.c b/mm/slub.c
index 825ff4505336..0c8ace04f075 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -3614,6 +3614,33 @@ void *__kmalloc_node(size_t size, gfp_t flags, int node)
 EXPORT_SYMBOL(__kmalloc_node);
 #endif
 
+#ifdef CONFIG_HARDENED_USERCOPY
+/*
+ * Rejects objects that are incorrectly sized.
+ *
+ * Returns NULL if check passes, otherwise const char * to name of cache
+ * to indicate an error.
+ */
+const char *__check_heap_object(const void *ptr, unsigned long n,
+				struct page *page)
+{
+	struct kmem_cache *s;
+	unsigned long offset;
+
+	/* Find object. */
+	s = page->slab_cache;
+
+	/* Find offset within object. */
+	offset = (ptr - page_address(page)) % s->size;
+
+	/* Allow address range falling entirely within object size. */
+	if (offset <= s->object_size && n <= s->object_size - offset)
+		return NULL;
+
+	return s->name;
+}
+#endif /* CONFIG_HARDENED_USERCOPY */
+
 static size_t __ksize(const void *object)
 {
 	struct page *page;
-- 
2.7.4


WARNING: multiple messages have this Message-ID (diff)
From: Kees Cook <keescook@chromium.org>
To: linux-kernel@vger.kernel.org
Cc: Kees Cook <keescook@chromium.org>, Rik van Riel <riel@redhat.com>,
	Casey Schaufler <casey@schaufler-ca.com>,
	PaX Team <pageexec@freemail.hu>,
	Brad Spengler <spender@grsecurity.net>,
	Russell King <linux@armlinux.org.uk>,
	Catalin Marinas <catalin.marinas@arm.com>,
	Will Deacon <will.deacon@arm.com>,
	Ard Biesheuvel <ard.biesheuvel@linaro.org>,
	Benjamin Herrenschmidt <benh@kernel.crashing.org>,
	Michael Ellerman <mpe@ellerman.id.au>,
	Tony Luck <tony.luck@intel.com>,
	Fenghua Yu <fenghua.yu@intel.com>,
	"David S. Miller" <davem@davemloft.net>,
	x86@kernel.org, Christoph Lameter <cl@linux.com>,
	Pekka Enberg <penberg@kernel.org>,
	David Rientjes <rientjes@google.com>,
	Joonsoo Kim <iamjoonsoo.kim@lge.com>,
	Andrew Morton <akpm@linux-foundation.org>,
	Andy Lutomirski <luto@kernel.org>, Borislav Petkov <bp@suse.de>,
	Mathias Krause <minipli@googlemail.com>, Jan Kara <jack@suse.cz>,
	Vitaly Wool <vitalywool@gmail.com>,
	Andrea Arcangeli <aarcange@redhat.com>,
	Dmitry Vyukov <dvyukov@google.com>,
	Laura Abbott <labbott@fedoraproject.org>,
	linux-arm-kernel@lists.infradead.org, linux-ia64@vger.kernel.org,
	linuxppc-dev@lists.ozlabs.org, sparclinux@vger.kernel.org,
	linux-arch@vger.kernel.org, linux-mm@kvack.org,
	kernel-hardening@lists.openwall.com
Subject: [PATCH 9/9] mm: SLUB hardened usercopy support
Date: Wed,  6 Jul 2016 15:25:28 -0700	[thread overview]
Message-ID: <1467843928-29351-10-git-send-email-keescook@chromium.org> (raw)
In-Reply-To: <1467843928-29351-1-git-send-email-keescook@chromium.org>

Under CONFIG_HARDENED_USERCOPY, this adds object size checking to the
SLUB allocator to catch any copies that may span objects.

Based on code from PaX and grsecurity.

Signed-off-by: Kees Cook <keescook@chromium.org>
---
 init/Kconfig |  1 +
 mm/slub.c    | 27 +++++++++++++++++++++++++++
 2 files changed, 28 insertions(+)

diff --git a/init/Kconfig b/init/Kconfig
index 798c2020ee7c..1c4711819dfd 100644
--- a/init/Kconfig
+++ b/init/Kconfig
@@ -1765,6 +1765,7 @@ config SLAB
 
 config SLUB
 	bool "SLUB (Unqueued Allocator)"
+	select HAVE_HARDENED_USERCOPY_ALLOCATOR
 	help
 	   SLUB is a slab allocator that minimizes cache line usage
 	   instead of managing queues of cached objects (SLAB approach).
diff --git a/mm/slub.c b/mm/slub.c
index 825ff4505336..0c8ace04f075 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -3614,6 +3614,33 @@ void *__kmalloc_node(size_t size, gfp_t flags, int node)
 EXPORT_SYMBOL(__kmalloc_node);
 #endif
 
+#ifdef CONFIG_HARDENED_USERCOPY
+/*
+ * Rejects objects that are incorrectly sized.
+ *
+ * Returns NULL if check passes, otherwise const char * to name of cache
+ * to indicate an error.
+ */
+const char *__check_heap_object(const void *ptr, unsigned long n,
+				struct page *page)
+{
+	struct kmem_cache *s;
+	unsigned long offset;
+
+	/* Find object. */
+	s = page->slab_cache;
+
+	/* Find offset within object. */
+	offset = (ptr - page_address(page)) % s->size;
+
+	/* Allow address range falling entirely within object size. */
+	if (offset <= s->object_size && n <= s->object_size - offset)
+		return NULL;
+
+	return s->name;
+}
+#endif /* CONFIG_HARDENED_USERCOPY */
+
 static size_t __ksize(const void *object)
 {
 	struct page *page;
-- 
2.7.4

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

WARNING: multiple messages have this Message-ID (diff)
From: keescook@chromium.org (Kees Cook)
To: linux-arm-kernel@lists.infradead.org
Subject: [PATCH 9/9] mm: SLUB hardened usercopy support
Date: Wed,  6 Jul 2016 15:25:28 -0700	[thread overview]
Message-ID: <1467843928-29351-10-git-send-email-keescook@chromium.org> (raw)
In-Reply-To: <1467843928-29351-1-git-send-email-keescook@chromium.org>

Under CONFIG_HARDENED_USERCOPY, this adds object size checking to the
SLUB allocator to catch any copies that may span objects.

Based on code from PaX and grsecurity.

Signed-off-by: Kees Cook <keescook@chromium.org>
---
 init/Kconfig |  1 +
 mm/slub.c    | 27 +++++++++++++++++++++++++++
 2 files changed, 28 insertions(+)

diff --git a/init/Kconfig b/init/Kconfig
index 798c2020ee7c..1c4711819dfd 100644
--- a/init/Kconfig
+++ b/init/Kconfig
@@ -1765,6 +1765,7 @@ config SLAB
 
 config SLUB
 	bool "SLUB (Unqueued Allocator)"
+	select HAVE_HARDENED_USERCOPY_ALLOCATOR
 	help
 	   SLUB is a slab allocator that minimizes cache line usage
 	   instead of managing queues of cached objects (SLAB approach).
diff --git a/mm/slub.c b/mm/slub.c
index 825ff4505336..0c8ace04f075 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -3614,6 +3614,33 @@ void *__kmalloc_node(size_t size, gfp_t flags, int node)
 EXPORT_SYMBOL(__kmalloc_node);
 #endif
 
+#ifdef CONFIG_HARDENED_USERCOPY
+/*
+ * Rejects objects that are incorrectly sized.
+ *
+ * Returns NULL if check passes, otherwise const char * to name of cache
+ * to indicate an error.
+ */
+const char *__check_heap_object(const void *ptr, unsigned long n,
+				struct page *page)
+{
+	struct kmem_cache *s;
+	unsigned long offset;
+
+	/* Find object. */
+	s = page->slab_cache;
+
+	/* Find offset within object. */
+	offset = (ptr - page_address(page)) % s->size;
+
+	/* Allow address range falling entirely within object size. */
+	if (offset <= s->object_size && n <= s->object_size - offset)
+		return NULL;
+
+	return s->name;
+}
+#endif /* CONFIG_HARDENED_USERCOPY */
+
 static size_t __ksize(const void *object)
 {
 	struct page *page;
-- 
2.7.4

WARNING: multiple messages have this Message-ID (diff)
From: Kees Cook <keescook@chromium.org>
To: linux-kernel@vger.kernel.org
Cc: Kees Cook <keescook@chromium.org>, Rik van Riel <riel@redhat.com>,
	Casey Schaufler <casey@schaufler-ca.com>,
	PaX Team <pageexec@freemail.hu>,
	Brad Spengler <spender@grsecurity.net>,
	Russell King <linux@armlinux.org.uk>,
	Catalin Marinas <catalin.marinas@arm.com>,
	Will Deacon <will.deacon@arm.com>,
	Ard Biesheuvel <ard.biesheuvel@linaro.org>,
	Benjamin Herrenschmidt <benh@kernel.crashing.org>,
	Michael Ellerman <mpe@ellerman.id.au>,
	Tony Luck <tony.luck@intel.com>,
	Fenghua Yu <fenghua.yu@intel.com>,
	"David S. Miller" <davem@davemloft.net>,
	x86@kernel.org, Christoph Lameter <cl@linux.com>,
	Pekka Enberg <penberg@kernel.org>,
	David Rientjes <rientjes@google.com>,
	Joonsoo Kim <iamjoonsoo.kim@lge.com>,
	Andrew Morton <akpm@linux-foundation.org>,
	Andy Lutomirski <luto@kernel.org>, Borislav Petkov <bp@suse.de>,
	Mathias Krause <minipli@googlemail.com>, Jan Kara <jack@suse.cz>,
	Vitaly Wool <vitalywool@gmail.com>,
	Andrea Arcangeli <aarcange@redhat.com>,
	Dmitry Vyukov <dvyukov@google.com>,
	Laura Abbott <labbott@fedoraproject.org>,
	linux-arm-kernel@lists.infradead.org, linux-ia64@vger.kernel.org,
	linuxppc-dev@lists.ozlabs.org, sparclinux@vger.kernel.org,
	linux-arch@vger.kernel.org, linux-mm@kvack.org,
	kernel-hardening@lists.openwall.com
Subject: [kernel-hardening] [PATCH 9/9] mm: SLUB hardened usercopy support
Date: Wed,  6 Jul 2016 15:25:28 -0700	[thread overview]
Message-ID: <1467843928-29351-10-git-send-email-keescook@chromium.org> (raw)
In-Reply-To: <1467843928-29351-1-git-send-email-keescook@chromium.org>

Under CONFIG_HARDENED_USERCOPY, this adds object size checking to the
SLUB allocator to catch any copies that may span objects.

Based on code from PaX and grsecurity.

Signed-off-by: Kees Cook <keescook@chromium.org>
---
 init/Kconfig |  1 +
 mm/slub.c    | 27 +++++++++++++++++++++++++++
 2 files changed, 28 insertions(+)

diff --git a/init/Kconfig b/init/Kconfig
index 798c2020ee7c..1c4711819dfd 100644
--- a/init/Kconfig
+++ b/init/Kconfig
@@ -1765,6 +1765,7 @@ config SLAB
 
 config SLUB
 	bool "SLUB (Unqueued Allocator)"
+	select HAVE_HARDENED_USERCOPY_ALLOCATOR
 	help
 	   SLUB is a slab allocator that minimizes cache line usage
 	   instead of managing queues of cached objects (SLAB approach).
diff --git a/mm/slub.c b/mm/slub.c
index 825ff4505336..0c8ace04f075 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -3614,6 +3614,33 @@ void *__kmalloc_node(size_t size, gfp_t flags, int node)
 EXPORT_SYMBOL(__kmalloc_node);
 #endif
 
+#ifdef CONFIG_HARDENED_USERCOPY
+/*
+ * Rejects objects that are incorrectly sized.
+ *
+ * Returns NULL if check passes, otherwise const char * to name of cache
+ * to indicate an error.
+ */
+const char *__check_heap_object(const void *ptr, unsigned long n,
+				struct page *page)
+{
+	struct kmem_cache *s;
+	unsigned long offset;
+
+	/* Find object. */
+	s = page->slab_cache;
+
+	/* Find offset within object. */
+	offset = (ptr - page_address(page)) % s->size;
+
+	/* Allow address range falling entirely within object size. */
+	if (offset <= s->object_size && n <= s->object_size - offset)
+		return NULL;
+
+	return s->name;
+}
+#endif /* CONFIG_HARDENED_USERCOPY */
+
 static size_t __ksize(const void *object)
 {
 	struct page *page;
-- 
2.7.4

  parent reply	other threads:[~2016-07-06 22:26 UTC|newest]

Thread overview: 366+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2016-07-06 22:25 [PATCH 0/9] mm: Hardened usercopy Kees Cook
2016-07-06 22:25 ` [kernel-hardening] " Kees Cook
2016-07-06 22:25 ` Kees Cook
2016-07-06 22:25 ` Kees Cook
2016-07-06 22:25 ` Kees Cook
2016-07-06 22:25 ` Kees Cook
2016-07-06 22:25 ` [PATCH 1/9] " Kees Cook
2016-07-06 22:25   ` [kernel-hardening] " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-07  5:37   ` Baruch Siach
2016-07-07  5:37     ` [kernel-hardening] " Baruch Siach
2016-07-07  5:37     ` Baruch Siach
2016-07-07  5:37     ` Baruch Siach
2016-07-07  5:37     ` Baruch Siach
2016-07-07  5:37     ` Baruch Siach
2016-07-07 17:25     ` Kees Cook
2016-07-07 17:25       ` [kernel-hardening] " Kees Cook
2016-07-07 17:25       ` Kees Cook
2016-07-07 17:25       ` Kees Cook
2016-07-07 17:25       ` Kees Cook
2016-07-07 17:25       ` Kees Cook
2016-07-07 17:25       ` Kees Cook
2016-07-07 18:35       ` Baruch Siach
2016-07-07 18:35         ` [kernel-hardening] " Baruch Siach
2016-07-07 18:35         ` Baruch Siach
2016-07-07 18:35         ` Baruch Siach
2016-07-07 18:35         ` Baruch Siach
2016-07-07 18:35         ` Baruch Siach
2016-07-07 18:35         ` Baruch Siach
2016-07-07  7:42   ` Thomas Gleixner
2016-07-07  7:42     ` [kernel-hardening] " Thomas Gleixner
2016-07-07  7:42     ` Thomas Gleixner
2016-07-07  7:42     ` Thomas Gleixner
2016-07-07  7:42     ` Thomas Gleixner
2016-07-07  7:42     ` Thomas Gleixner
2016-07-07 17:29     ` Kees Cook
2016-07-07 17:29       ` [kernel-hardening] " Kees Cook
2016-07-07 17:29       ` Kees Cook
2016-07-07 17:29       ` Kees Cook
2016-07-07 17:29       ` Kees Cook
2016-07-07 17:29       ` Kees Cook
2016-07-07 17:29       ` Kees Cook
2016-07-07 19:34       ` Thomas Gleixner
2016-07-07 19:34         ` [kernel-hardening] " Thomas Gleixner
2016-07-07 19:34         ` Thomas Gleixner
2016-07-07 19:34         ` Thomas Gleixner
2016-07-07 19:34         ` Thomas Gleixner
2016-07-07 19:34         ` Thomas Gleixner
2016-07-07 19:34         ` Thomas Gleixner
2016-07-07  8:01   ` Arnd Bergmann
2016-07-07  8:01     ` [kernel-hardening] " Arnd Bergmann
2016-07-07  8:01     ` Arnd Bergmann
2016-07-07  8:01     ` Arnd Bergmann
2016-07-07  8:01     ` Arnd Bergmann
2016-07-07  8:01     ` Arnd Bergmann
2016-07-07 17:37     ` Kees Cook
2016-07-07 17:37       ` [kernel-hardening] " Kees Cook
2016-07-07 17:37       ` Kees Cook
2016-07-07 17:37       ` Kees Cook
2016-07-07 17:37       ` Kees Cook
2016-07-07 17:37       ` Kees Cook
2016-07-07 17:37       ` Kees Cook
2016-07-08  5:34       ` Michael Ellerman
2016-07-08  5:34       ` Michael Ellerman
2016-07-08  5:34         ` [kernel-hardening] " Michael Ellerman
2016-07-08  5:34       ` Michael Ellerman
2016-07-08  5:34       ` Michael Ellerman
2016-07-08  5:34         ` Michael Ellerman
2016-07-08  5:34       ` Michael Ellerman
2016-07-08  9:22       ` Arnd Bergmann
2016-07-08  9:22         ` [kernel-hardening] " Arnd Bergmann
2016-07-08  9:22         ` Arnd Bergmann
2016-07-08  9:22         ` Arnd Bergmann
2016-07-08  9:22         ` Arnd Bergmann
2016-07-08  9:22         ` Arnd Bergmann
2016-07-08  9:22         ` Arnd Bergmann
2016-07-07 16:19   ` Rik van Riel
2016-07-07 16:19     ` [kernel-hardening] " Rik van Riel
2016-07-07 16:19     ` Rik van Riel
2016-07-07 16:19     ` Rik van Riel
2016-07-07 16:19     ` Rik van Riel
2016-07-07 16:35   ` Rik van Riel
2016-07-07 16:35     ` [kernel-hardening] " Rik van Riel
2016-07-07 16:35     ` Rik van Riel
2016-07-07 16:35     ` Rik van Riel
2016-07-07 16:35     ` Rik van Riel
2016-07-07 17:41     ` Kees Cook
2016-07-07 17:41       ` [kernel-hardening] " Kees Cook
2016-07-07 17:41       ` Kees Cook
2016-07-07 17:41       ` Kees Cook
2016-07-07 17:41       ` Kees Cook
2016-07-07 17:41       ` Kees Cook
2016-07-07 17:41       ` Kees Cook
2016-07-06 22:25 ` [PATCH 2/9] x86/uaccess: Enable hardened usercopy Kees Cook
2016-07-06 22:25   ` [kernel-hardening] " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25 ` [PATCH 3/9] ARM: uaccess: " Kees Cook
2016-07-06 22:25   ` [kernel-hardening] " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25 ` [PATCH 4/9] arm64/uaccess: " Kees Cook
2016-07-06 22:25   ` [kernel-hardening] " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-07 10:07   ` Mark Rutland
2016-07-07 10:07     ` [kernel-hardening] " Mark Rutland
2016-07-07 10:07     ` Mark Rutland
2016-07-07 10:07     ` Mark Rutland
2016-07-07 10:07     ` Mark Rutland
2016-07-07 10:07     ` Mark Rutland
2016-07-07 17:19     ` Kees Cook
2016-07-07 17:19       ` [kernel-hardening] " Kees Cook
2016-07-07 17:19       ` Kees Cook
2016-07-07 17:19       ` Kees Cook
2016-07-07 17:19       ` Kees Cook
2016-07-07 17:19       ` Kees Cook
2016-07-07 17:19       ` Kees Cook
2016-07-06 22:25 ` [PATCH 5/9] ia64/uaccess: " Kees Cook
2016-07-06 22:25   ` [kernel-hardening] " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25 ` [PATCH 6/9] powerpc/uaccess: " Kees Cook
2016-07-06 22:25   ` [kernel-hardening] " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25 ` [PATCH 7/9] sparc/uaccess: " Kees Cook
2016-07-06 22:25   ` [kernel-hardening] " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25 ` [PATCH 8/9] mm: SLAB hardened usercopy support Kees Cook
2016-07-06 22:25   ` [kernel-hardening] " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25 ` Kees Cook [this message]
2016-07-06 22:25   ` [kernel-hardening] [PATCH 9/9] mm: SLUB " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-07  4:35   ` Michael Ellerman
2016-07-07  4:35     ` [kernel-hardening] " Michael Ellerman
2016-07-07  4:35   ` Michael Ellerman
2016-07-07  4:35   ` Michael Ellerman
2016-07-07  4:35     ` Michael Ellerman
2016-07-07  4:35   ` Michael Ellerman
2016-07-07  4:35   ` Michael Ellerman
     [not found]   ` <577ddc18.d351190a.1fa54.ffffbe79SMTPIN_ADDED_BROKEN@mx.google.com>
2016-07-07 18:56     ` [kernel-hardening] " Kees Cook
2016-07-07 18:56       ` Kees Cook
2016-07-07 18:56       ` Kees Cook
2016-07-07 18:56       ` Kees Cook
2016-07-07 18:56       ` Kees Cook
2016-07-08 10:19       ` Michael Ellerman
2016-07-08 13:45         ` Christoph Lameter
2016-07-08 13:45           ` Christoph Lameter
2016-07-08 13:45           ` Christoph Lameter
2016-07-08 13:45           ` Christoph Lameter
2016-07-08 13:45           ` Christoph Lameter
2016-07-08 16:07           ` Kees Cook
2016-07-08 16:07             ` Kees Cook
2016-07-08 16:07             ` Kees Cook
2016-07-08 16:07             ` Kees Cook
2016-07-08 16:07             ` Kees Cook
2016-07-08 16:20             ` Christoph Lameter
2016-07-08 16:20               ` Christoph Lameter
2016-07-08 16:20               ` Christoph Lameter
2016-07-08 16:20               ` Christoph Lameter
2016-07-08 16:20               ` Christoph Lameter
2016-07-08 17:41               ` [kernel-hardening] " Kees Cook
2016-07-08 17:41                 ` Kees Cook
2016-07-08 17:41                 ` Kees Cook
2016-07-08 17:41                 ` Kees Cook
2016-07-08 17:41                 ` Kees Cook
2016-07-08 17:41                 ` Kees Cook
2016-07-08 20:48                 ` Kees Cook
2016-07-08 20:48                   ` Kees Cook
2016-07-08 20:48                   ` Kees Cook
2016-07-08 20:48                   ` Kees Cook
2016-07-08 20:48                   ` Kees Cook
2016-07-08 20:48                   ` Kees Cook
2016-07-09  5:58                   ` Michael Ellerman
2016-07-09  5:58                   ` [kernel-hardening] " Michael Ellerman
2016-07-09  5:58                     ` Michael Ellerman
2016-07-09  5:58                   ` Michael Ellerman
2016-07-09  5:58                   ` Michael Ellerman
2016-07-09  5:58                   ` Michael Ellerman
2016-07-09  5:58                   ` Michael Ellerman
2016-07-09  5:58                     ` Michael Ellerman
2016-07-09  5:58                     ` Michael Ellerman
2016-07-09  6:07                     ` Michael Ellerman
2016-07-09  6:07                     ` Michael Ellerman
2016-07-09  6:07                     ` Michael Ellerman
2016-07-09  6:07                       ` Michael Ellerman
2016-07-09  6:07                     ` Michael Ellerman
2016-07-09  6:07                       ` Michael Ellerman
2016-07-09  6:07                       ` Michael Ellerman
2016-07-09  6:07                     ` Michael Ellerman
2016-07-09  6:07                     ` Michael Ellerman
     [not found]                   ` <57809299.84b3370a.5390c.ffff9e58SMTPIN_ADDED_BROKEN@mx.google.com>
2016-07-09  6:17                     ` [kernel-hardening] " Valdis.Kletnieks
2016-07-09  6:17                       ` Valdis.Kletnieks at vt.edu
2016-07-09  6:17                       ` Valdis.Kletnieks
2016-07-09  6:17                       ` Valdis.Kletnieks
2016-07-09  6:17                       ` Valdis.Kletnieks
2016-07-09  6:17                       ` Valdis.Kletnieks
2016-07-09 17:07                       ` Kees Cook
2016-07-09 17:07                         ` Kees Cook
2016-07-09 17:07                         ` Kees Cook
2016-07-09 17:07                         ` Kees Cook
2016-07-09 17:07                         ` Kees Cook
2016-07-09 17:07                         ` Kees Cook
2016-07-11  6:08                   ` Joonsoo Kim
2016-07-11  6:08                     ` Joonsoo Kim
2016-07-11  6:08                     ` Joonsoo Kim
2016-07-11  6:08                     ` Joonsoo Kim
2016-07-11  6:08                     ` Joonsoo Kim
2016-07-11  6:08                     ` Joonsoo Kim
2016-07-08 10:19       ` Michael Ellerman
2016-07-08 10:19       ` Michael Ellerman
2016-07-08 10:19       ` Michael Ellerman
2016-07-08 10:19       ` Michael Ellerman
2016-07-08 10:19       ` [kernel-hardening] " Michael Ellerman
2016-07-08 10:19         ` Michael Ellerman
2016-07-07  7:30 ` [PATCH 0/9] mm: Hardened usercopy Christian Borntraeger
2016-07-07  7:30   ` [kernel-hardening] " Christian Borntraeger
2016-07-07  7:30   ` Christian Borntraeger
2016-07-07  7:30   ` Christian Borntraeger
2016-07-07  7:30   ` Christian Borntraeger
2016-07-07  7:30   ` Christian Borntraeger
2016-07-07 17:27   ` Kees Cook
2016-07-07 17:27     ` [kernel-hardening] " Kees Cook
2016-07-07 17:27     ` Kees Cook
2016-07-07 17:27     ` Kees Cook
2016-07-07 17:27     ` Kees Cook
2016-07-07 17:27     ` Kees Cook
2016-07-07 17:27     ` Kees Cook
2016-07-08  8:46 ` Ingo Molnar
2016-07-08  8:46   ` [kernel-hardening] " Ingo Molnar
2016-07-08  8:46   ` Ingo Molnar
2016-07-08  8:46   ` Ingo Molnar
2016-07-08  8:46   ` Ingo Molnar
2016-07-08  8:46   ` Ingo Molnar
2016-07-08 16:19   ` Linus Torvalds
2016-07-08 16:19     ` [kernel-hardening] " Linus Torvalds
2016-07-08 16:19     ` Linus Torvalds
2016-07-08 16:19     ` Linus Torvalds
2016-07-08 16:19     ` Linus Torvalds
2016-07-08 16:19     ` Linus Torvalds
2016-07-08 16:19     ` Linus Torvalds
2016-07-08 18:23     ` Ingo Molnar
2016-07-08 18:23       ` [kernel-hardening] " Ingo Molnar
2016-07-08 18:23       ` Ingo Molnar
2016-07-08 18:23       ` Ingo Molnar
2016-07-08 18:23       ` Ingo Molnar
2016-07-08 18:23       ` Ingo Molnar
2016-07-08 18:23       ` Ingo Molnar
2016-07-09  2:22 ` Laura Abbott
2016-07-09  2:22   ` [kernel-hardening] " Laura Abbott
2016-07-09  2:22   ` Laura Abbott
2016-07-09  2:22   ` Laura Abbott
2016-07-09  2:44   ` Rik van Riel
2016-07-09  2:44     ` [kernel-hardening] " Rik van Riel
2016-07-09  2:44     ` Rik van Riel
2016-07-09  2:44     ` Rik van Riel
2016-07-09  2:44     ` Rik van Riel
2016-07-09  7:55     ` Ingo Molnar
2016-07-09  7:55       ` [kernel-hardening] " Ingo Molnar
2016-07-09  7:55       ` Ingo Molnar
2016-07-09  7:55       ` Ingo Molnar
2016-07-09  7:55       ` Ingo Molnar
2016-07-09  7:55       ` Ingo Molnar
2016-07-09  8:25   ` Ard Biesheuvel
2016-07-09  8:25     ` Ard Biesheuvel
2016-07-09  8:25     ` [kernel-hardening] " Ard Biesheuvel
2016-07-09  8:25     ` Ard Biesheuvel
2016-07-09  8:25     ` Ard Biesheuvel
2016-07-09  8:25     ` Ard Biesheuvel
2016-07-09  8:25     ` Ard Biesheuvel
2016-07-09  8:25     ` Ard Biesheuvel
2016-07-09 12:58     ` Laura Abbott
2016-07-09 12:58       ` [kernel-hardening] " Laura Abbott
2016-07-09 12:58       ` Laura Abbott
2016-07-09 17:03     ` Kees Cook
2016-07-09 17:03       ` Kees Cook
2016-07-09 17:03       ` [kernel-hardening] " Kees Cook
2016-07-09 17:03       ` Kees Cook
2016-07-09 17:03       ` Kees Cook
2016-07-09 17:03       ` Kees Cook
2016-07-09 17:03       ` Kees Cook
2016-07-09 17:03       ` Kees Cook
2016-07-09 17:01   ` Kees Cook
2016-07-09 17:01     ` [kernel-hardening] " Kees Cook
2016-07-09 17:01     ` Kees Cook
2016-07-09 17:01     ` Kees Cook
2016-07-09 17:01     ` Kees Cook
2016-07-09 17:01     ` Kees Cook
2016-07-09 17:01     ` Kees Cook
2016-07-09 21:27 ` Andy Lutomirski
2016-07-09 21:27   ` [kernel-hardening] " Andy Lutomirski
2016-07-09 21:27   ` Andy Lutomirski
2016-07-09 21:27   ` Andy Lutomirski
2016-07-09 21:27   ` Andy Lutomirski
2016-07-09 21:27   ` Andy Lutomirski
2016-07-09 21:27   ` Andy Lutomirski
2016-07-09 23:16   ` PaX Team
2016-07-09 23:16     ` [kernel-hardening] " PaX Team
2016-07-09 23:16     ` PaX Team
2016-07-09 23:16     ` PaX Team
2016-07-09 23:16     ` PaX Team
2016-07-09 23:16     ` PaX Team
2016-07-09 23:16     ` PaX Team
2016-07-09 23:16     ` PaX Team
2016-07-10  9:16     ` Ingo Molnar
2016-07-10  9:16       ` [kernel-hardening] " Ingo Molnar
2016-07-10  9:16       ` Ingo Molnar
2016-07-10  9:16       ` Ingo Molnar
2016-07-10  9:16       ` Ingo Molnar
2016-07-10  9:16       ` Ingo Molnar
2016-07-10  9:16       ` Ingo Molnar
2016-07-10 12:03       ` PaX Team
2016-07-10 12:03         ` [kernel-hardening] " PaX Team
2016-07-10 12:03         ` PaX Team
2016-07-10 12:03         ` PaX Team
2016-07-10 12:03         ` PaX Team
2016-07-10 12:03         ` PaX Team
2016-07-10 12:03         ` PaX Team
2016-07-10 12:03         ` PaX Team
2016-07-10 12:38         ` Andy Lutomirski
2016-07-10 12:38           ` [kernel-hardening] " Andy Lutomirski
2016-07-10 12:38           ` Andy Lutomirski
2016-07-10 12:38           ` Andy Lutomirski
2016-07-10 12:38           ` Andy Lutomirski
2016-07-10 12:38           ` Andy Lutomirski
2016-07-10 12:38           ` Andy Lutomirski
2016-07-11 18:40           ` Kees Cook
2016-07-11 18:40             ` [kernel-hardening] " Kees Cook
2016-07-11 18:40             ` Kees Cook
2016-07-11 18:40             ` Kees Cook
2016-07-11 18:40             ` Kees Cook
2016-07-11 18:40             ` Kees Cook
2016-07-11 18:40             ` Kees Cook
2016-07-11 18:34         ` Kees Cook
2016-07-11 18:34           ` [kernel-hardening] " Kees Cook
2016-07-11 18:34           ` Kees Cook
2016-07-11 18:34           ` Kees Cook
2016-07-11 18:34           ` Kees Cook
2016-07-11 18:34           ` Kees Cook
2016-07-11 18:34           ` Kees Cook
2016-07-12 18:26 ` [kernel-hardening] " Valdis.Kletnieks
2016-07-12 18:44   ` Kees Cook

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=1467843928-29351-10-git-send-email-keescook@chromium.org \
    --to=keescook@chromium.org \
    --cc=aarcange@redhat.com \
    --cc=akpm@linux-foundation.org \
    --cc=ard.biesheuvel@linaro.org \
    --cc=benh@kernel.crashing.org \
    --cc=bp@suse.de \
    --cc=casey@schaufler-ca.com \
    --cc=catalin.marinas@arm.com \
    --cc=cl@linux.com \
    --cc=davem@davemloft.net \
    --cc=dvyukov@google.com \
    --cc=fenghua.yu@intel.com \
    --cc=iamjoonsoo.kim@lge.com \
    --cc=jack@suse.cz \
    --cc=kernel-hardening@lists.openwall.com \
    --cc=labbott@fedoraproject.org \
    --cc=linux-arch@vger.kernel.org \
    --cc=linux-arm-kernel@lists.infradead.org \
    --cc=linux-ia64@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=linux@armlinux.org.uk \
    --cc=linuxppc-dev@lists.ozlabs.org \
    --cc=luto@kernel.org \
    --cc=minipli@googlemail.com \
    --cc=mpe@ellerman.id.au \
    --cc=pageexec@freemail.hu \
    --cc=penberg@kernel.org \
    --cc=riel@redhat.com \
    --cc=rientjes@google.com \
    --cc=sparclinux@vger.kernel.org \
    --cc=spender@grsecurity.net \
    --cc=tony.luck@intel.com \
    --cc=vitalywool@gmail.com \
    --cc=will.deacon@arm.com \
    --cc=x86@kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.