All of lore.kernel.org
 help / color / mirror / Atom feed
From: Kees Cook <keescook@chromium.org>
To: Arnd Bergmann <arnd@arndb.de>
Cc: "linuxppc-dev@lists.ozlabs.org" <linuxppc-dev@lists.ozlabs.org>,
	LKML <linux-kernel@vger.kernel.org>, Jan Kara <jack@suse.cz>,
	"kernel-hardening@lists.openwall.com" 
	<kernel-hardening@lists.openwall.com>,
	Catalin Marinas <catalin.marinas@arm.com>,
	Will Deacon <will.deacon@arm.com>, Linux-MM <linux-mm@kvack.org>,
	sparclinux <sparclinux@vger.kernel.org>,
	linux-ia64@vger.kernel.org, Christoph Lameter <cl@linux.com>,
	Andrea Arcangeli <aarcange@redhat.com>,
	linux-arch <linux-arch@vger.kernel.org>,
	"x86@kernel.org" <x86@kernel.org>,
	Russell King <linux@armlinux.org.uk>,
	"linux-arm-kernel@lists.infradead.org" 
	<linux-arm-kernel@lists.infradead.org>,
	PaX Team <pageexec@freemail.hu>,
	Mathias Krause <minipli@googlemail.com>,
	Fenghua Yu <fenghua.yu@intel.com>, Rik van Riel <riel@redhat.com>,
	David Rientjes <rientjes@google.com>,
	Tony Luck <tony.luck@intel.com>,
	Andy Lutomirski <luto@kernel.org>,
	Joonsoo Kim <iamjoonsoo.kim@lge.com>,
	Dmitry Vyukov <dvyukov@google.com>,
	Laura Abbott <labbott@fedoraproject.org>,
	Brad Spengler <spender@grsecurity.net>,
	Ard Biesheuvel <ard.biesheuvel@linaro.org>,
	Pekka Enberg <penberg@kernel.org>,
	Casey Schaufler <casey@schaufler-ca.com>,
	Andrew Morton <akpm@linux-foundation.org>,
	"David S. Miller" <davem@davemloft.net>
Subject: Re: [PATCH 1/9] mm: Hardened usercopy
Date: Thu, 7 Jul 2016 13:37:43 -0400	[thread overview]
Message-ID: <CAGXu5jLyBfqXJKxohHiZgztRVrFyqwbta1W_Dw6KyyGM3LzshQ@mail.gmail.com> (raw)
In-Reply-To: <3418914.byvl8Wuxlf@wuerfel>

On Thu, Jul 7, 2016 at 4:01 AM, Arnd Bergmann <arnd@arndb.de> wrote:
> On Wednesday, July 6, 2016 3:25:20 PM CEST Kees Cook wrote:
>> This is the start of porting PAX_USERCOPY into the mainline kernel. This
>> is the first set of features, controlled by CONFIG_HARDENED_USERCOPY. The
>> work is based on code by PaX Team and Brad Spengler, and an earlier port
>> from Casey Schaufler. Additional non-slab page tests are from Rik van Riel.
>>
>> This patch contains the logic for validating several conditions when
>> performing copy_to_user() and copy_from_user() on the kernel object
>> being copied to/from:
>> - address range doesn't wrap around
>> - address range isn't NULL or zero-allocated (with a non-zero copy size)
>> - if on the slab allocator:
>>   - object size must be less than or equal to copy size (when check is
>>     implemented in the allocator, which appear in subsequent patches)
>> - otherwise, object must not span page allocations
>> - if on the stack
>>   - object must not extend before/after the current process task
>>   - object must be contained by the current stack frame (when there is
>>     arch/build support for identifying stack frames)
>> - object must not overlap with kernel text
>>
>> Signed-off-by: Kees Cook <keescook@chromium.org>
>
> Nice!
>
> I have a few further thoughts, most of which have probably been
> considered before:
>
>> +static inline const char *check_bogus_address(const void *ptr, unsigned long n)
>> +{
>> +     /* Reject if object wraps past end of memory. */
>> +     if (ptr + n < ptr)
>> +             return "<wrapped address>";
>> +
>> +     /* Reject if NULL or ZERO-allocation. */
>> +     if (ZERO_OR_NULL_PTR(ptr))
>> +             return "<null>";
>> +
>> +     return NULL;
>> +}
>
> This checks against address (void*)16, but I guess on most architectures the
> lowest possible kernel address is much higher. While there may not be much
> that to exploit if the expected kernel address points to userland, forbidding
> any obviously incorrect address that is outside of the kernel may be easier.
>
> Even on architectures like s390 that start the kernel memory at (void *)0x0,
> the lowest address to which we may want to do a copy_to_user would be much
> higher than (void*)0x16.

Yeah, that's worth exploring, but given the shenanigans around
set_fs(), I'd like to leave this as-is, and we can add to these checks
as we remove as much of the insane usage of set_fs().

>> +
>> +     /* Allow kernel rodata region (if not marked as Reserved). */
>> +     if (ptr >= (const void *)__start_rodata &&
>> +         end <= (const void *)__end_rodata)
>> +             return NULL;
>
> Should we explicitly forbid writing to rodata, or is it enough to
> rely on page protection here?

Hm, interesting. That's a very small check to add. My knee-jerk is to
just leave it up to page protection. I'm on the fence. :)

>
>> +     /* Allow kernel bss region (if not marked as Reserved). */
>> +     if (ptr >= (const void *)__bss_start &&
>> +         end <= (const void *)__bss_stop)
>> +             return NULL;
>
> accesses to .data/.rodata/.bss are probably not performance critical,
> so we could go further here and check the kallsyms table to ensure
> that we are not spanning multiple symbols here.

Oh, interesting! Yeah, would you be willing to put together that patch
and test it? I wonder if there are any cases where there are
legitimate usercopys across multiple symbols.

> For stuff that is performance critical, should there be a way to
> opt out of the checks, or do we assume it already uses functions
> that avoid the checks? I looked at the file and network I/O path
> briefly and they seem to use kmap_atomic() to get to the user pages
> at least in some of the common cases (but I may well be missing
> important ones).

I don't want to start with an exemption here, so until such a case is
found, I'd rather leave this as-is. That said, the primary protection
here tends to be buggy lengths (which is why put/get_user() is
untouched). For constant-sized copies, some checks could be skipped.
In the second part of this protection (what I named
CONFIG_HARDENED_USERCOPY_WHITELIST in the RFC version of this series),
there are cases where we want to skip the whitelist checking since it
is for a constant-sized copy the code understands is okay to pull out
of an otherwise disallowed allocator object.

-Kees

-- 
Kees Cook
Chrome OS & Brillo Security

WARNING: multiple messages have this Message-ID (diff)
From: Kees Cook <keescook@chromium.org>
To: Arnd Bergmann <arnd@arndb.de>
Cc: "linuxppc-dev@lists.ozlabs.org" <linuxppc-dev@lists.ozlabs.org>,
	LKML <linux-kernel@vger.kernel.org>, Jan Kara <jack@suse.cz>,
	"kernel-hardening@lists.openwall.com"
	<kernel-hardening@lists.openwall.com>,
	Catalin Marinas <catalin.marinas@arm.com>,
	Will Deacon <will.deacon@arm.com>, Linux-MM <linux-mm@kvack.org>,
	sparclinux <sparclinux@vger.kernel.org>,
	linux-ia64@vger.kernel.org, Christoph Lameter <cl@linux.com>,
	Andrea Arcangeli <aarcange@redhat.com>,
	linux-arch <linux-arch@vger.kernel.org>,
	"x86@kernel.org" <x86@kernel.org>,
	Russell King <linux@armlinux.org.uk>,
	"linux-arm-kernel@lists.infradead.org"
	<linux-arm-kernel@lists.infradead.org>,
	PaX Team <pageexec@freemail.hu>,
	Mathias Krause <minipli@googlemail.com>,
	Fenghua Yu <fenghua.yu@intel.com>, Rik van Riel <riel@redhat.com>,
	David Rientjes <rientjes@google.com>,
	Tony Luck <tony.luck@intel.com>,
	Andy
Subject: Re: [PATCH 1/9] mm: Hardened usercopy
Date: Thu, 7 Jul 2016 13:37:43 -0400	[thread overview]
Message-ID: <CAGXu5jLyBfqXJKxohHiZgztRVrFyqwbta1W_Dw6KyyGM3LzshQ@mail.gmail.com> (raw)
In-Reply-To: <3418914.byvl8Wuxlf@wuerfel>

On Thu, Jul 7, 2016 at 4:01 AM, Arnd Bergmann <arnd@arndb.de> wrote:
> On Wednesday, July 6, 2016 3:25:20 PM CEST Kees Cook wrote:
>> This is the start of porting PAX_USERCOPY into the mainline kernel. This
>> is the first set of features, controlled by CONFIG_HARDENED_USERCOPY. The
>> work is based on code by PaX Team and Brad Spengler, and an earlier port
>> from Casey Schaufler. Additional non-slab page tests are from Rik van Riel.
>>
>> This patch contains the logic for validating several conditions when
>> performing copy_to_user() and copy_from_user() on the kernel object
>> being copied to/from:
>> - address range doesn't wrap around
>> - address range isn't NULL or zero-allocated (with a non-zero copy size)
>> - if on the slab allocator:
>>   - object size must be less than or equal to copy size (when check is
>>     implemented in the allocator, which appear in subsequent patches)
>> - otherwise, object must not span page allocations
>> - if on the stack
>>   - object must not extend before/after the current process task
>>   - object must be contained by the current stack frame (when there is
>>     arch/build support for identifying stack frames)
>> - object must not overlap with kernel text
>>
>> Signed-off-by: Kees Cook <keescook@chromium.org>
>
> Nice!
>
> I have a few further thoughts, most of which have probably been
> considered before:
>
>> +static inline const char *check_bogus_address(const void *ptr, unsigned long n)
>> +{
>> +     /* Reject if object wraps past end of memory. */
>> +     if (ptr + n < ptr)
>> +             return "<wrapped address>";
>> +
>> +     /* Reject if NULL or ZERO-allocation. */
>> +     if (ZERO_OR_NULL_PTR(ptr))
>> +             return "<null>";
>> +
>> +     return NULL;
>> +}
>
> This checks against address (void*)16, but I guess on most architectures the
> lowest possible kernel address is much higher. While there may not be much
> that to exploit if the expected kernel address points to userland, forbidding
> any obviously incorrect address that is outside of the kernel may be easier.
>
> Even on architectures like s390 that start the kernel memory at (void *)0x0,
> the lowest address to which we may want to do a copy_to_user would be much
> higher than (void*)0x16.

Yeah, that's worth exploring, but given the shenanigans around
set_fs(), I'd like to leave this as-is, and we can add to these checks
as we remove as much of the insane usage of set_fs().

>> +
>> +     /* Allow kernel rodata region (if not marked as Reserved). */
>> +     if (ptr >= (const void *)__start_rodata &&
>> +         end <= (const void *)__end_rodata)
>> +             return NULL;
>
> Should we explicitly forbid writing to rodata, or is it enough to
> rely on page protection here?

Hm, interesting. That's a very small check to add. My knee-jerk is to
just leave it up to page protection. I'm on the fence. :)

>
>> +     /* Allow kernel bss region (if not marked as Reserved). */
>> +     if (ptr >= (const void *)__bss_start &&
>> +         end <= (const void *)__bss_stop)
>> +             return NULL;
>
> accesses to .data/.rodata/.bss are probably not performance critical,
> so we could go further here and check the kallsyms table to ensure
> that we are not spanning multiple symbols here.

Oh, interesting! Yeah, would you be willing to put together that patch
and test it? I wonder if there are any cases where there are
legitimate usercopys across multiple symbols.

> For stuff that is performance critical, should there be a way to
> opt out of the checks, or do we assume it already uses functions
> that avoid the checks? I looked at the file and network I/O path
> briefly and they seem to use kmap_atomic() to get to the user pages
> at least in some of the common cases (but I may well be missing
> important ones).

I don't want to start with an exemption here, so until such a case is
found, I'd rather leave this as-is. That said, the primary protection
here tends to be buggy lengths (which is why put/get_user() is
untouched). For constant-sized copies, some checks could be skipped.
In the second part of this protection (what I named
CONFIG_HARDENED_USERCOPY_WHITELIST in the RFC version of this series),
there are cases where we want to skip the whitelist checking since it
is for a constant-sized copy the code understands is okay to pull out
of an otherwise disallowed allocator object.

-Kees

-- 
Kees Cook
Chrome OS & Brillo Security

WARNING: multiple messages have this Message-ID (diff)
From: Kees Cook <keescook@chromium.org>
To: Arnd Bergmann <arnd@arndb.de>
Cc: "linuxppc-dev@lists.ozlabs.org" <linuxppc-dev@lists.ozlabs.org>,
	LKML <linux-kernel@vger.kernel.org>, Jan Kara <jack@suse.cz>,
	"kernel-hardening@lists.openwall.com"
	<kernel-hardening@lists.openwall.com>,
	Catalin Marinas <catalin.marinas@arm.com>,
	Will Deacon <will.deacon@arm.com>, Linux-MM <linux-mm@kvack.org>,
	sparclinux <sparclinux@vger.kernel.org>,
	linux-ia64@vger.kernel.org, Christoph Lameter <cl@linux.com>,
	Andrea Arcangeli <aarcange@redhat.com>,
	linux-arch <linux-arch@vger.kernel.org>,
	"x86@kernel.org" <x86@kernel.org>,
	Russell King <linux@armlinux.org.uk>,
	"linux-arm-kernel@lists.infradead.org"
	<linux-arm-kernel@lists.infradead.org>,
	PaX Team <pageexec@freemail.hu>,
	Mathias Krause <minipli@googlemail.com>,
	Fenghua Yu <fenghua.yu@intel.com>, Rik van Riel <riel@redhat.com>,
	David Rientjes <rientjes@google.com>,
	Tony Luck <tony.luck@intel.com>,
	Andy Lutomirski <luto@kernel.org>,
	Joonsoo Kim <iamjoonsoo.kim@lge.com>,
	Dmitry Vyukov <dvyukov@google.com>,
	Laura Abbott <labbott@fedoraproject.org>,
	Brad Spengler <spender@grsecurity.net>,
	Ard Biesheuvel <ard.biesheuvel@linaro.org>,
	Pekka Enberg <penberg@kernel.org>,
	Casey Schaufler <casey@schaufler-ca.com>,
	Andrew Morton <akpm@linux-foundation.org>,
	"David S. Miller" <davem@davemloft.net>
Subject: Re: [PATCH 1/9] mm: Hardened usercopy
Date: Thu, 07 Jul 2016 17:37:43 +0000	[thread overview]
Message-ID: <CAGXu5jLyBfqXJKxohHiZgztRVrFyqwbta1W_Dw6KyyGM3LzshQ@mail.gmail.com> (raw)
In-Reply-To: <3418914.byvl8Wuxlf@wuerfel>

On Thu, Jul 7, 2016 at 4:01 AM, Arnd Bergmann <arnd@arndb.de> wrote:
> On Wednesday, July 6, 2016 3:25:20 PM CEST Kees Cook wrote:
>> This is the start of porting PAX_USERCOPY into the mainline kernel. This
>> is the first set of features, controlled by CONFIG_HARDENED_USERCOPY. The
>> work is based on code by PaX Team and Brad Spengler, and an earlier port
>> from Casey Schaufler. Additional non-slab page tests are from Rik van Riel.
>>
>> This patch contains the logic for validating several conditions when
>> performing copy_to_user() and copy_from_user() on the kernel object
>> being copied to/from:
>> - address range doesn't wrap around
>> - address range isn't NULL or zero-allocated (with a non-zero copy size)
>> - if on the slab allocator:
>>   - object size must be less than or equal to copy size (when check is
>>     implemented in the allocator, which appear in subsequent patches)
>> - otherwise, object must not span page allocations
>> - if on the stack
>>   - object must not extend before/after the current process task
>>   - object must be contained by the current stack frame (when there is
>>     arch/build support for identifying stack frames)
>> - object must not overlap with kernel text
>>
>> Signed-off-by: Kees Cook <keescook@chromium.org>
>
> Nice!
>
> I have a few further thoughts, most of which have probably been
> considered before:
>
>> +static inline const char *check_bogus_address(const void *ptr, unsigned long n)
>> +{
>> +     /* Reject if object wraps past end of memory. */
>> +     if (ptr + n < ptr)
>> +             return "<wrapped address>";
>> +
>> +     /* Reject if NULL or ZERO-allocation. */
>> +     if (ZERO_OR_NULL_PTR(ptr))
>> +             return "<null>";
>> +
>> +     return NULL;
>> +}
>
> This checks against address (void*)16, but I guess on most architectures the
> lowest possible kernel address is much higher. While there may not be much
> that to exploit if the expected kernel address points to userland, forbidding
> any obviously incorrect address that is outside of the kernel may be easier.
>
> Even on architectures like s390 that start the kernel memory at (void *)0x0,
> the lowest address to which we may want to do a copy_to_user would be much
> higher than (void*)0x16.

Yeah, that's worth exploring, but given the shenanigans around
set_fs(), I'd like to leave this as-is, and we can add to these checks
as we remove as much of the insane usage of set_fs().

>> +
>> +     /* Allow kernel rodata region (if not marked as Reserved). */
>> +     if (ptr >= (const void *)__start_rodata &&
>> +         end <= (const void *)__end_rodata)
>> +             return NULL;
>
> Should we explicitly forbid writing to rodata, or is it enough to
> rely on page protection here?

Hm, interesting. That's a very small check to add. My knee-jerk is to
just leave it up to page protection. I'm on the fence. :)

>
>> +     /* Allow kernel bss region (if not marked as Reserved). */
>> +     if (ptr >= (const void *)__bss_start &&
>> +         end <= (const void *)__bss_stop)
>> +             return NULL;
>
> accesses to .data/.rodata/.bss are probably not performance critical,
> so we could go further here and check the kallsyms table to ensure
> that we are not spanning multiple symbols here.

Oh, interesting! Yeah, would you be willing to put together that patch
and test it? I wonder if there are any cases where there are
legitimate usercopys across multiple symbols.

> For stuff that is performance critical, should there be a way to
> opt out of the checks, or do we assume it already uses functions
> that avoid the checks? I looked at the file and network I/O path
> briefly and they seem to use kmap_atomic() to get to the user pages
> at least in some of the common cases (but I may well be missing
> important ones).

I don't want to start with an exemption here, so until such a case is
found, I'd rather leave this as-is. That said, the primary protection
here tends to be buggy lengths (which is why put/get_user() is
untouched). For constant-sized copies, some checks could be skipped.
In the second part of this protection (what I named
CONFIG_HARDENED_USERCOPY_WHITELIST in the RFC version of this series),
there are cases where we want to skip the whitelist checking since it
is for a constant-sized copy the code understands is okay to pull out
of an otherwise disallowed allocator object.

-Kees

-- 
Kees Cook
Chrome OS & Brillo Security

WARNING: multiple messages have this Message-ID (diff)
From: Kees Cook <keescook@chromium.org>
To: Arnd Bergmann <arnd@arndb.de>
Cc: "linuxppc-dev@lists.ozlabs.org" <linuxppc-dev@lists.ozlabs.org>,
	LKML <linux-kernel@vger.kernel.org>, Jan Kara <jack@suse.cz>,
	"kernel-hardening@lists.openwall.com"
	<kernel-hardening@lists.openwall.com>,
	Catalin Marinas <catalin.marinas@arm.com>,
	Will Deacon <will.deacon@arm.com>, Linux-MM <linux-mm@kvack.org>,
	sparclinux <sparclinux@vger.kernel.org>,
	linux-ia64@vger.kernel.org, Christoph Lameter <cl@linux.com>,
	Andrea Arcangeli <aarcange@redhat.com>,
	linux-arch <linux-arch@vger.kernel.org>,
	"x86@kernel.org" <x86@kernel.org>,
	Russell King <linux@armlinux.org.uk>,
	"linux-arm-kernel@lists.infradead.org"
	<linux-arm-kernel@lists.infradead.org>,
	PaX Team <pageexec@freemail.hu>,
	Mathias Krause <minipli@googlemail.com>,
	Fenghua Yu <fenghua.yu@intel.com>, Rik van Riel <riel@redhat.com>,
	David Rientjes <rientjes@google.com>,
	Tony Luck <tony.luck@intel.com>,
	Andy Lutomirski <luto@kernel.org>,
	Joonsoo Kim <iamjoonsoo.kim@lge.com>,
	Dmitry Vyukov <dvyukov@google.com>,
	Laura Abbott <labbott@fedoraproject.org>,
	Brad Spengler <spender@grsecurity.net>,
	Ard Biesheuvel <ard.biesheuvel@linaro.org>,
	Pekka Enberg <penberg@kernel.org>,
	Casey Schaufler <casey@schaufler-ca.com>,
	Andrew Morton <akpm@linux-foundation.org>,
	"David S. Miller" <davem@davemloft.net>
Subject: Re: [PATCH 1/9] mm: Hardened usercopy
Date: Thu, 7 Jul 2016 13:37:43 -0400	[thread overview]
Message-ID: <CAGXu5jLyBfqXJKxohHiZgztRVrFyqwbta1W_Dw6KyyGM3LzshQ@mail.gmail.com> (raw)
In-Reply-To: <3418914.byvl8Wuxlf@wuerfel>

On Thu, Jul 7, 2016 at 4:01 AM, Arnd Bergmann <arnd@arndb.de> wrote:
> On Wednesday, July 6, 2016 3:25:20 PM CEST Kees Cook wrote:
>> This is the start of porting PAX_USERCOPY into the mainline kernel. This
>> is the first set of features, controlled by CONFIG_HARDENED_USERCOPY. The
>> work is based on code by PaX Team and Brad Spengler, and an earlier port
>> from Casey Schaufler. Additional non-slab page tests are from Rik van Riel.
>>
>> This patch contains the logic for validating several conditions when
>> performing copy_to_user() and copy_from_user() on the kernel object
>> being copied to/from:
>> - address range doesn't wrap around
>> - address range isn't NULL or zero-allocated (with a non-zero copy size)
>> - if on the slab allocator:
>>   - object size must be less than or equal to copy size (when check is
>>     implemented in the allocator, which appear in subsequent patches)
>> - otherwise, object must not span page allocations
>> - if on the stack
>>   - object must not extend before/after the current process task
>>   - object must be contained by the current stack frame (when there is
>>     arch/build support for identifying stack frames)
>> - object must not overlap with kernel text
>>
>> Signed-off-by: Kees Cook <keescook@chromium.org>
>
> Nice!
>
> I have a few further thoughts, most of which have probably been
> considered before:
>
>> +static inline const char *check_bogus_address(const void *ptr, unsigned long n)
>> +{
>> +     /* Reject if object wraps past end of memory. */
>> +     if (ptr + n < ptr)
>> +             return "<wrapped address>";
>> +
>> +     /* Reject if NULL or ZERO-allocation. */
>> +     if (ZERO_OR_NULL_PTR(ptr))
>> +             return "<null>";
>> +
>> +     return NULL;
>> +}
>
> This checks against address (void*)16, but I guess on most architectures the
> lowest possible kernel address is much higher. While there may not be much
> that to exploit if the expected kernel address points to userland, forbidding
> any obviously incorrect address that is outside of the kernel may be easier.
>
> Even on architectures like s390 that start the kernel memory at (void *)0x0,
> the lowest address to which we may want to do a copy_to_user would be much
> higher than (void*)0x16.

Yeah, that's worth exploring, but given the shenanigans around
set_fs(), I'd like to leave this as-is, and we can add to these checks
as we remove as much of the insane usage of set_fs().

>> +
>> +     /* Allow kernel rodata region (if not marked as Reserved). */
>> +     if (ptr >= (const void *)__start_rodata &&
>> +         end <= (const void *)__end_rodata)
>> +             return NULL;
>
> Should we explicitly forbid writing to rodata, or is it enough to
> rely on page protection here?

Hm, interesting. That's a very small check to add. My knee-jerk is to
just leave it up to page protection. I'm on the fence. :)

>
>> +     /* Allow kernel bss region (if not marked as Reserved). */
>> +     if (ptr >= (const void *)__bss_start &&
>> +         end <= (const void *)__bss_stop)
>> +             return NULL;
>
> accesses to .data/.rodata/.bss are probably not performance critical,
> so we could go further here and check the kallsyms table to ensure
> that we are not spanning multiple symbols here.

Oh, interesting! Yeah, would you be willing to put together that patch
and test it? I wonder if there are any cases where there are
legitimate usercopys across multiple symbols.

> For stuff that is performance critical, should there be a way to
> opt out of the checks, or do we assume it already uses functions
> that avoid the checks? I looked at the file and network I/O path
> briefly and they seem to use kmap_atomic() to get to the user pages
> at least in some of the common cases (but I may well be missing
> important ones).

I don't want to start with an exemption here, so until such a case is
found, I'd rather leave this as-is. That said, the primary protection
here tends to be buggy lengths (which is why put/get_user() is
untouched). For constant-sized copies, some checks could be skipped.
In the second part of this protection (what I named
CONFIG_HARDENED_USERCOPY_WHITELIST in the RFC version of this series),
there are cases where we want to skip the whitelist checking since it
is for a constant-sized copy the code understands is okay to pull out
of an otherwise disallowed allocator object.

-Kees

-- 
Kees Cook
Chrome OS & Brillo Security

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

WARNING: multiple messages have this Message-ID (diff)
From: keescook@chromium.org (Kees Cook)
To: linux-arm-kernel@lists.infradead.org
Subject: [PATCH 1/9] mm: Hardened usercopy
Date: Thu, 7 Jul 2016 13:37:43 -0400	[thread overview]
Message-ID: <CAGXu5jLyBfqXJKxohHiZgztRVrFyqwbta1W_Dw6KyyGM3LzshQ@mail.gmail.com> (raw)
In-Reply-To: <3418914.byvl8Wuxlf@wuerfel>

On Thu, Jul 7, 2016 at 4:01 AM, Arnd Bergmann <arnd@arndb.de> wrote:
> On Wednesday, July 6, 2016 3:25:20 PM CEST Kees Cook wrote:
>> This is the start of porting PAX_USERCOPY into the mainline kernel. This
>> is the first set of features, controlled by CONFIG_HARDENED_USERCOPY. The
>> work is based on code by PaX Team and Brad Spengler, and an earlier port
>> from Casey Schaufler. Additional non-slab page tests are from Rik van Riel.
>>
>> This patch contains the logic for validating several conditions when
>> performing copy_to_user() and copy_from_user() on the kernel object
>> being copied to/from:
>> - address range doesn't wrap around
>> - address range isn't NULL or zero-allocated (with a non-zero copy size)
>> - if on the slab allocator:
>>   - object size must be less than or equal to copy size (when check is
>>     implemented in the allocator, which appear in subsequent patches)
>> - otherwise, object must not span page allocations
>> - if on the stack
>>   - object must not extend before/after the current process task
>>   - object must be contained by the current stack frame (when there is
>>     arch/build support for identifying stack frames)
>> - object must not overlap with kernel text
>>
>> Signed-off-by: Kees Cook <keescook@chromium.org>
>
> Nice!
>
> I have a few further thoughts, most of which have probably been
> considered before:
>
>> +static inline const char *check_bogus_address(const void *ptr, unsigned long n)
>> +{
>> +     /* Reject if object wraps past end of memory. */
>> +     if (ptr + n < ptr)
>> +             return "<wrapped address>";
>> +
>> +     /* Reject if NULL or ZERO-allocation. */
>> +     if (ZERO_OR_NULL_PTR(ptr))
>> +             return "<null>";
>> +
>> +     return NULL;
>> +}
>
> This checks against address (void*)16, but I guess on most architectures the
> lowest possible kernel address is much higher. While there may not be much
> that to exploit if the expected kernel address points to userland, forbidding
> any obviously incorrect address that is outside of the kernel may be easier.
>
> Even on architectures like s390 that start the kernel memory at (void *)0x0,
> the lowest address to which we may want to do a copy_to_user would be much
> higher than (void*)0x16.

Yeah, that's worth exploring, but given the shenanigans around
set_fs(), I'd like to leave this as-is, and we can add to these checks
as we remove as much of the insane usage of set_fs().

>> +
>> +     /* Allow kernel rodata region (if not marked as Reserved). */
>> +     if (ptr >= (const void *)__start_rodata &&
>> +         end <= (const void *)__end_rodata)
>> +             return NULL;
>
> Should we explicitly forbid writing to rodata, or is it enough to
> rely on page protection here?

Hm, interesting. That's a very small check to add. My knee-jerk is to
just leave it up to page protection. I'm on the fence. :)

>
>> +     /* Allow kernel bss region (if not marked as Reserved). */
>> +     if (ptr >= (const void *)__bss_start &&
>> +         end <= (const void *)__bss_stop)
>> +             return NULL;
>
> accesses to .data/.rodata/.bss are probably not performance critical,
> so we could go further here and check the kallsyms table to ensure
> that we are not spanning multiple symbols here.

Oh, interesting! Yeah, would you be willing to put together that patch
and test it? I wonder if there are any cases where there are
legitimate usercopys across multiple symbols.

> For stuff that is performance critical, should there be a way to
> opt out of the checks, or do we assume it already uses functions
> that avoid the checks? I looked at the file and network I/O path
> briefly and they seem to use kmap_atomic() to get to the user pages
> at least in some of the common cases (but I may well be missing
> important ones).

I don't want to start with an exemption here, so until such a case is
found, I'd rather leave this as-is. That said, the primary protection
here tends to be buggy lengths (which is why put/get_user() is
untouched). For constant-sized copies, some checks could be skipped.
In the second part of this protection (what I named
CONFIG_HARDENED_USERCOPY_WHITELIST in the RFC version of this series),
there are cases where we want to skip the whitelist checking since it
is for a constant-sized copy the code understands is okay to pull out
of an otherwise disallowed allocator object.

-Kees

-- 
Kees Cook
Chrome OS & Brillo Security

WARNING: multiple messages have this Message-ID (diff)
From: Kees Cook <keescook@chromium.org>
To: Arnd Bergmann <arnd@arndb.de>
Cc: "linuxppc-dev@lists.ozlabs.org" <linuxppc-dev@lists.ozlabs.org>,
	LKML <linux-kernel@vger.kernel.org>, Jan Kara <jack@suse.cz>,
	"kernel-hardening@lists.openwall.com"
	<kernel-hardening@lists.openwall.com>,
	Catalin Marinas <catalin.marinas@arm.com>,
	Will Deacon <will.deacon@arm.com>, Linux-MM <linux-mm@kvack.org>,
	sparclinux <sparclinux@vger.kernel.org>,
	linux-ia64@vger.kernel.org, Christoph Lameter <cl@linux.com>,
	Andrea Arcangeli <aarcange@redhat.com>,
	linux-arch <linux-arch@vger.kernel.org>,
	"x86@kernel.org" <x86@kernel.org>,
	Russell King <linux@armlinux.org.uk>,
	"linux-arm-kernel@lists.infradead.org"
	<linux-arm-kernel@lists.infradead.org>,
	PaX Team <pageexec@freemail.hu>,
	Mathias Krause <minipli@googlemail.com>,
	Fenghua Yu <fenghua.yu@intel.com>, Rik van Riel <riel@redhat.com>,
	David Rientjes <rientjes@google.com>,
	Tony Luck <tony.luck@intel.com>,
	Andy Lutomirski <luto@kernel.org>,
	Joonsoo Kim <iamjoonsoo.kim@lge.com>,
	Dmitry Vyukov <dvyukov@google.com>,
	Laura Abbott <labbott@fedoraproject.org>,
	Brad Spengler <spender@grsecurity.net>,
	Ard Biesheuvel <ard.biesheuvel@linaro.org>,
	Pekka Enberg <penberg@kernel.org>,
	Casey Schaufler <casey@schaufler-ca.com>,
	Andrew Morton <akpm@linux-foundation.org>,
	"David S. Miller" <davem@davemloft.net>
Subject: [kernel-hardening] Re: [PATCH 1/9] mm: Hardened usercopy
Date: Thu, 7 Jul 2016 13:37:43 -0400	[thread overview]
Message-ID: <CAGXu5jLyBfqXJKxohHiZgztRVrFyqwbta1W_Dw6KyyGM3LzshQ@mail.gmail.com> (raw)
In-Reply-To: <3418914.byvl8Wuxlf@wuerfel>

On Thu, Jul 7, 2016 at 4:01 AM, Arnd Bergmann <arnd@arndb.de> wrote:
> On Wednesday, July 6, 2016 3:25:20 PM CEST Kees Cook wrote:
>> This is the start of porting PAX_USERCOPY into the mainline kernel. This
>> is the first set of features, controlled by CONFIG_HARDENED_USERCOPY. The
>> work is based on code by PaX Team and Brad Spengler, and an earlier port
>> from Casey Schaufler. Additional non-slab page tests are from Rik van Riel.
>>
>> This patch contains the logic for validating several conditions when
>> performing copy_to_user() and copy_from_user() on the kernel object
>> being copied to/from:
>> - address range doesn't wrap around
>> - address range isn't NULL or zero-allocated (with a non-zero copy size)
>> - if on the slab allocator:
>>   - object size must be less than or equal to copy size (when check is
>>     implemented in the allocator, which appear in subsequent patches)
>> - otherwise, object must not span page allocations
>> - if on the stack
>>   - object must not extend before/after the current process task
>>   - object must be contained by the current stack frame (when there is
>>     arch/build support for identifying stack frames)
>> - object must not overlap with kernel text
>>
>> Signed-off-by: Kees Cook <keescook@chromium.org>
>
> Nice!
>
> I have a few further thoughts, most of which have probably been
> considered before:
>
>> +static inline const char *check_bogus_address(const void *ptr, unsigned long n)
>> +{
>> +     /* Reject if object wraps past end of memory. */
>> +     if (ptr + n < ptr)
>> +             return "<wrapped address>";
>> +
>> +     /* Reject if NULL or ZERO-allocation. */
>> +     if (ZERO_OR_NULL_PTR(ptr))
>> +             return "<null>";
>> +
>> +     return NULL;
>> +}
>
> This checks against address (void*)16, but I guess on most architectures the
> lowest possible kernel address is much higher. While there may not be much
> that to exploit if the expected kernel address points to userland, forbidding
> any obviously incorrect address that is outside of the kernel may be easier.
>
> Even on architectures like s390 that start the kernel memory at (void *)0x0,
> the lowest address to which we may want to do a copy_to_user would be much
> higher than (void*)0x16.

Yeah, that's worth exploring, but given the shenanigans around
set_fs(), I'd like to leave this as-is, and we can add to these checks
as we remove as much of the insane usage of set_fs().

>> +
>> +     /* Allow kernel rodata region (if not marked as Reserved). */
>> +     if (ptr >= (const void *)__start_rodata &&
>> +         end <= (const void *)__end_rodata)
>> +             return NULL;
>
> Should we explicitly forbid writing to rodata, or is it enough to
> rely on page protection here?

Hm, interesting. That's a very small check to add. My knee-jerk is to
just leave it up to page protection. I'm on the fence. :)

>
>> +     /* Allow kernel bss region (if not marked as Reserved). */
>> +     if (ptr >= (const void *)__bss_start &&
>> +         end <= (const void *)__bss_stop)
>> +             return NULL;
>
> accesses to .data/.rodata/.bss are probably not performance critical,
> so we could go further here and check the kallsyms table to ensure
> that we are not spanning multiple symbols here.

Oh, interesting! Yeah, would you be willing to put together that patch
and test it? I wonder if there are any cases where there are
legitimate usercopys across multiple symbols.

> For stuff that is performance critical, should there be a way to
> opt out of the checks, or do we assume it already uses functions
> that avoid the checks? I looked at the file and network I/O path
> briefly and they seem to use kmap_atomic() to get to the user pages
> at least in some of the common cases (but I may well be missing
> important ones).

I don't want to start with an exemption here, so until such a case is
found, I'd rather leave this as-is. That said, the primary protection
here tends to be buggy lengths (which is why put/get_user() is
untouched). For constant-sized copies, some checks could be skipped.
In the second part of this protection (what I named
CONFIG_HARDENED_USERCOPY_WHITELIST in the RFC version of this series),
there are cases where we want to skip the whitelist checking since it
is for a constant-sized copy the code understands is okay to pull out
of an otherwise disallowed allocator object.

-Kees

-- 
Kees Cook
Chrome OS & Brillo Security

  reply	other threads:[~2016-07-07 17:37 UTC|newest]

Thread overview: 366+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2016-07-06 22:25 [PATCH 0/9] mm: Hardened usercopy Kees Cook
2016-07-06 22:25 ` [kernel-hardening] " Kees Cook
2016-07-06 22:25 ` Kees Cook
2016-07-06 22:25 ` Kees Cook
2016-07-06 22:25 ` Kees Cook
2016-07-06 22:25 ` Kees Cook
2016-07-06 22:25 ` [PATCH 1/9] " Kees Cook
2016-07-06 22:25   ` [kernel-hardening] " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-07  5:37   ` Baruch Siach
2016-07-07  5:37     ` [kernel-hardening] " Baruch Siach
2016-07-07  5:37     ` Baruch Siach
2016-07-07  5:37     ` Baruch Siach
2016-07-07  5:37     ` Baruch Siach
2016-07-07  5:37     ` Baruch Siach
2016-07-07 17:25     ` Kees Cook
2016-07-07 17:25       ` [kernel-hardening] " Kees Cook
2016-07-07 17:25       ` Kees Cook
2016-07-07 17:25       ` Kees Cook
2016-07-07 17:25       ` Kees Cook
2016-07-07 17:25       ` Kees Cook
2016-07-07 17:25       ` Kees Cook
2016-07-07 18:35       ` Baruch Siach
2016-07-07 18:35         ` [kernel-hardening] " Baruch Siach
2016-07-07 18:35         ` Baruch Siach
2016-07-07 18:35         ` Baruch Siach
2016-07-07 18:35         ` Baruch Siach
2016-07-07 18:35         ` Baruch Siach
2016-07-07 18:35         ` Baruch Siach
2016-07-07  7:42   ` Thomas Gleixner
2016-07-07  7:42     ` [kernel-hardening] " Thomas Gleixner
2016-07-07  7:42     ` Thomas Gleixner
2016-07-07  7:42     ` Thomas Gleixner
2016-07-07  7:42     ` Thomas Gleixner
2016-07-07  7:42     ` Thomas Gleixner
2016-07-07 17:29     ` Kees Cook
2016-07-07 17:29       ` [kernel-hardening] " Kees Cook
2016-07-07 17:29       ` Kees Cook
2016-07-07 17:29       ` Kees Cook
2016-07-07 17:29       ` Kees Cook
2016-07-07 17:29       ` Kees Cook
2016-07-07 17:29       ` Kees Cook
2016-07-07 19:34       ` Thomas Gleixner
2016-07-07 19:34         ` [kernel-hardening] " Thomas Gleixner
2016-07-07 19:34         ` Thomas Gleixner
2016-07-07 19:34         ` Thomas Gleixner
2016-07-07 19:34         ` Thomas Gleixner
2016-07-07 19:34         ` Thomas Gleixner
2016-07-07 19:34         ` Thomas Gleixner
2016-07-07  8:01   ` Arnd Bergmann
2016-07-07  8:01     ` [kernel-hardening] " Arnd Bergmann
2016-07-07  8:01     ` Arnd Bergmann
2016-07-07  8:01     ` Arnd Bergmann
2016-07-07  8:01     ` Arnd Bergmann
2016-07-07  8:01     ` Arnd Bergmann
2016-07-07 17:37     ` Kees Cook [this message]
2016-07-07 17:37       ` [kernel-hardening] " Kees Cook
2016-07-07 17:37       ` Kees Cook
2016-07-07 17:37       ` Kees Cook
2016-07-07 17:37       ` Kees Cook
2016-07-07 17:37       ` Kees Cook
2016-07-07 17:37       ` Kees Cook
2016-07-08  5:34       ` Michael Ellerman
2016-07-08  5:34       ` Michael Ellerman
2016-07-08  5:34         ` [kernel-hardening] " Michael Ellerman
2016-07-08  5:34       ` Michael Ellerman
2016-07-08  5:34       ` Michael Ellerman
2016-07-08  5:34         ` Michael Ellerman
2016-07-08  5:34       ` Michael Ellerman
2016-07-08  9:22       ` Arnd Bergmann
2016-07-08  9:22         ` [kernel-hardening] " Arnd Bergmann
2016-07-08  9:22         ` Arnd Bergmann
2016-07-08  9:22         ` Arnd Bergmann
2016-07-08  9:22         ` Arnd Bergmann
2016-07-08  9:22         ` Arnd Bergmann
2016-07-08  9:22         ` Arnd Bergmann
2016-07-07 16:19   ` Rik van Riel
2016-07-07 16:19     ` [kernel-hardening] " Rik van Riel
2016-07-07 16:19     ` Rik van Riel
2016-07-07 16:19     ` Rik van Riel
2016-07-07 16:19     ` Rik van Riel
2016-07-07 16:35   ` Rik van Riel
2016-07-07 16:35     ` [kernel-hardening] " Rik van Riel
2016-07-07 16:35     ` Rik van Riel
2016-07-07 16:35     ` Rik van Riel
2016-07-07 16:35     ` Rik van Riel
2016-07-07 17:41     ` Kees Cook
2016-07-07 17:41       ` [kernel-hardening] " Kees Cook
2016-07-07 17:41       ` Kees Cook
2016-07-07 17:41       ` Kees Cook
2016-07-07 17:41       ` Kees Cook
2016-07-07 17:41       ` Kees Cook
2016-07-07 17:41       ` Kees Cook
2016-07-06 22:25 ` [PATCH 2/9] x86/uaccess: Enable hardened usercopy Kees Cook
2016-07-06 22:25   ` [kernel-hardening] " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25 ` [PATCH 3/9] ARM: uaccess: " Kees Cook
2016-07-06 22:25   ` [kernel-hardening] " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25 ` [PATCH 4/9] arm64/uaccess: " Kees Cook
2016-07-06 22:25   ` [kernel-hardening] " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-07 10:07   ` Mark Rutland
2016-07-07 10:07     ` [kernel-hardening] " Mark Rutland
2016-07-07 10:07     ` Mark Rutland
2016-07-07 10:07     ` Mark Rutland
2016-07-07 10:07     ` Mark Rutland
2016-07-07 10:07     ` Mark Rutland
2016-07-07 17:19     ` Kees Cook
2016-07-07 17:19       ` [kernel-hardening] " Kees Cook
2016-07-07 17:19       ` Kees Cook
2016-07-07 17:19       ` Kees Cook
2016-07-07 17:19       ` Kees Cook
2016-07-07 17:19       ` Kees Cook
2016-07-07 17:19       ` Kees Cook
2016-07-06 22:25 ` [PATCH 5/9] ia64/uaccess: " Kees Cook
2016-07-06 22:25   ` [kernel-hardening] " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25 ` [PATCH 6/9] powerpc/uaccess: " Kees Cook
2016-07-06 22:25   ` [kernel-hardening] " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25 ` [PATCH 7/9] sparc/uaccess: " Kees Cook
2016-07-06 22:25   ` [kernel-hardening] " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25 ` [PATCH 8/9] mm: SLAB hardened usercopy support Kees Cook
2016-07-06 22:25   ` [kernel-hardening] " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25 ` [PATCH 9/9] mm: SLUB " Kees Cook
2016-07-06 22:25   ` [kernel-hardening] " Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-06 22:25   ` Kees Cook
2016-07-07  4:35   ` Michael Ellerman
2016-07-07  4:35     ` [kernel-hardening] " Michael Ellerman
2016-07-07  4:35   ` Michael Ellerman
2016-07-07  4:35   ` Michael Ellerman
2016-07-07  4:35     ` Michael Ellerman
2016-07-07  4:35   ` Michael Ellerman
2016-07-07  4:35   ` Michael Ellerman
     [not found]   ` <577ddc18.d351190a.1fa54.ffffbe79SMTPIN_ADDED_BROKEN@mx.google.com>
2016-07-07 18:56     ` [kernel-hardening] " Kees Cook
2016-07-07 18:56       ` Kees Cook
2016-07-07 18:56       ` Kees Cook
2016-07-07 18:56       ` Kees Cook
2016-07-07 18:56       ` Kees Cook
2016-07-08 10:19       ` Michael Ellerman
2016-07-08 13:45         ` Christoph Lameter
2016-07-08 13:45           ` Christoph Lameter
2016-07-08 13:45           ` Christoph Lameter
2016-07-08 13:45           ` Christoph Lameter
2016-07-08 13:45           ` Christoph Lameter
2016-07-08 16:07           ` Kees Cook
2016-07-08 16:07             ` Kees Cook
2016-07-08 16:07             ` Kees Cook
2016-07-08 16:07             ` Kees Cook
2016-07-08 16:07             ` Kees Cook
2016-07-08 16:20             ` Christoph Lameter
2016-07-08 16:20               ` Christoph Lameter
2016-07-08 16:20               ` Christoph Lameter
2016-07-08 16:20               ` Christoph Lameter
2016-07-08 16:20               ` Christoph Lameter
2016-07-08 17:41               ` [kernel-hardening] " Kees Cook
2016-07-08 17:41                 ` Kees Cook
2016-07-08 17:41                 ` Kees Cook
2016-07-08 17:41                 ` Kees Cook
2016-07-08 17:41                 ` Kees Cook
2016-07-08 17:41                 ` Kees Cook
2016-07-08 20:48                 ` Kees Cook
2016-07-08 20:48                   ` Kees Cook
2016-07-08 20:48                   ` Kees Cook
2016-07-08 20:48                   ` Kees Cook
2016-07-08 20:48                   ` Kees Cook
2016-07-08 20:48                   ` Kees Cook
2016-07-09  5:58                   ` Michael Ellerman
2016-07-09  5:58                   ` [kernel-hardening] " Michael Ellerman
2016-07-09  5:58                     ` Michael Ellerman
2016-07-09  5:58                   ` Michael Ellerman
2016-07-09  5:58                   ` Michael Ellerman
2016-07-09  5:58                   ` Michael Ellerman
2016-07-09  5:58                   ` Michael Ellerman
2016-07-09  5:58                     ` Michael Ellerman
2016-07-09  5:58                     ` Michael Ellerman
2016-07-09  6:07                     ` Michael Ellerman
2016-07-09  6:07                     ` Michael Ellerman
2016-07-09  6:07                     ` Michael Ellerman
2016-07-09  6:07                       ` Michael Ellerman
2016-07-09  6:07                     ` Michael Ellerman
2016-07-09  6:07                       ` Michael Ellerman
2016-07-09  6:07                       ` Michael Ellerman
2016-07-09  6:07                     ` Michael Ellerman
2016-07-09  6:07                     ` Michael Ellerman
     [not found]                   ` <57809299.84b3370a.5390c.ffff9e58SMTPIN_ADDED_BROKEN@mx.google.com>
2016-07-09  6:17                     ` [kernel-hardening] " Valdis.Kletnieks
2016-07-09  6:17                       ` Valdis.Kletnieks at vt.edu
2016-07-09  6:17                       ` Valdis.Kletnieks
2016-07-09  6:17                       ` Valdis.Kletnieks
2016-07-09  6:17                       ` Valdis.Kletnieks
2016-07-09  6:17                       ` Valdis.Kletnieks
2016-07-09 17:07                       ` Kees Cook
2016-07-09 17:07                         ` Kees Cook
2016-07-09 17:07                         ` Kees Cook
2016-07-09 17:07                         ` Kees Cook
2016-07-09 17:07                         ` Kees Cook
2016-07-09 17:07                         ` Kees Cook
2016-07-11  6:08                   ` Joonsoo Kim
2016-07-11  6:08                     ` Joonsoo Kim
2016-07-11  6:08                     ` Joonsoo Kim
2016-07-11  6:08                     ` Joonsoo Kim
2016-07-11  6:08                     ` Joonsoo Kim
2016-07-11  6:08                     ` Joonsoo Kim
2016-07-08 10:19       ` Michael Ellerman
2016-07-08 10:19       ` Michael Ellerman
2016-07-08 10:19       ` Michael Ellerman
2016-07-08 10:19       ` Michael Ellerman
2016-07-08 10:19       ` [kernel-hardening] " Michael Ellerman
2016-07-08 10:19         ` Michael Ellerman
2016-07-07  7:30 ` [PATCH 0/9] mm: Hardened usercopy Christian Borntraeger
2016-07-07  7:30   ` [kernel-hardening] " Christian Borntraeger
2016-07-07  7:30   ` Christian Borntraeger
2016-07-07  7:30   ` Christian Borntraeger
2016-07-07  7:30   ` Christian Borntraeger
2016-07-07  7:30   ` Christian Borntraeger
2016-07-07 17:27   ` Kees Cook
2016-07-07 17:27     ` [kernel-hardening] " Kees Cook
2016-07-07 17:27     ` Kees Cook
2016-07-07 17:27     ` Kees Cook
2016-07-07 17:27     ` Kees Cook
2016-07-07 17:27     ` Kees Cook
2016-07-07 17:27     ` Kees Cook
2016-07-08  8:46 ` Ingo Molnar
2016-07-08  8:46   ` [kernel-hardening] " Ingo Molnar
2016-07-08  8:46   ` Ingo Molnar
2016-07-08  8:46   ` Ingo Molnar
2016-07-08  8:46   ` Ingo Molnar
2016-07-08  8:46   ` Ingo Molnar
2016-07-08 16:19   ` Linus Torvalds
2016-07-08 16:19     ` [kernel-hardening] " Linus Torvalds
2016-07-08 16:19     ` Linus Torvalds
2016-07-08 16:19     ` Linus Torvalds
2016-07-08 16:19     ` Linus Torvalds
2016-07-08 16:19     ` Linus Torvalds
2016-07-08 16:19     ` Linus Torvalds
2016-07-08 18:23     ` Ingo Molnar
2016-07-08 18:23       ` [kernel-hardening] " Ingo Molnar
2016-07-08 18:23       ` Ingo Molnar
2016-07-08 18:23       ` Ingo Molnar
2016-07-08 18:23       ` Ingo Molnar
2016-07-08 18:23       ` Ingo Molnar
2016-07-08 18:23       ` Ingo Molnar
2016-07-09  2:22 ` Laura Abbott
2016-07-09  2:22   ` [kernel-hardening] " Laura Abbott
2016-07-09  2:22   ` Laura Abbott
2016-07-09  2:22   ` Laura Abbott
2016-07-09  2:44   ` Rik van Riel
2016-07-09  2:44     ` [kernel-hardening] " Rik van Riel
2016-07-09  2:44     ` Rik van Riel
2016-07-09  2:44     ` Rik van Riel
2016-07-09  2:44     ` Rik van Riel
2016-07-09  7:55     ` Ingo Molnar
2016-07-09  7:55       ` [kernel-hardening] " Ingo Molnar
2016-07-09  7:55       ` Ingo Molnar
2016-07-09  7:55       ` Ingo Molnar
2016-07-09  7:55       ` Ingo Molnar
2016-07-09  7:55       ` Ingo Molnar
2016-07-09  8:25   ` Ard Biesheuvel
2016-07-09  8:25     ` Ard Biesheuvel
2016-07-09  8:25     ` [kernel-hardening] " Ard Biesheuvel
2016-07-09  8:25     ` Ard Biesheuvel
2016-07-09  8:25     ` Ard Biesheuvel
2016-07-09  8:25     ` Ard Biesheuvel
2016-07-09  8:25     ` Ard Biesheuvel
2016-07-09  8:25     ` Ard Biesheuvel
2016-07-09 12:58     ` Laura Abbott
2016-07-09 12:58       ` [kernel-hardening] " Laura Abbott
2016-07-09 12:58       ` Laura Abbott
2016-07-09 17:03     ` Kees Cook
2016-07-09 17:03       ` Kees Cook
2016-07-09 17:03       ` [kernel-hardening] " Kees Cook
2016-07-09 17:03       ` Kees Cook
2016-07-09 17:03       ` Kees Cook
2016-07-09 17:03       ` Kees Cook
2016-07-09 17:03       ` Kees Cook
2016-07-09 17:03       ` Kees Cook
2016-07-09 17:01   ` Kees Cook
2016-07-09 17:01     ` [kernel-hardening] " Kees Cook
2016-07-09 17:01     ` Kees Cook
2016-07-09 17:01     ` Kees Cook
2016-07-09 17:01     ` Kees Cook
2016-07-09 17:01     ` Kees Cook
2016-07-09 17:01     ` Kees Cook
2016-07-09 21:27 ` Andy Lutomirski
2016-07-09 21:27   ` [kernel-hardening] " Andy Lutomirski
2016-07-09 21:27   ` Andy Lutomirski
2016-07-09 21:27   ` Andy Lutomirski
2016-07-09 21:27   ` Andy Lutomirski
2016-07-09 21:27   ` Andy Lutomirski
2016-07-09 21:27   ` Andy Lutomirski
2016-07-09 23:16   ` PaX Team
2016-07-09 23:16     ` [kernel-hardening] " PaX Team
2016-07-09 23:16     ` PaX Team
2016-07-09 23:16     ` PaX Team
2016-07-09 23:16     ` PaX Team
2016-07-09 23:16     ` PaX Team
2016-07-09 23:16     ` PaX Team
2016-07-09 23:16     ` PaX Team
2016-07-10  9:16     ` Ingo Molnar
2016-07-10  9:16       ` [kernel-hardening] " Ingo Molnar
2016-07-10  9:16       ` Ingo Molnar
2016-07-10  9:16       ` Ingo Molnar
2016-07-10  9:16       ` Ingo Molnar
2016-07-10  9:16       ` Ingo Molnar
2016-07-10  9:16       ` Ingo Molnar
2016-07-10 12:03       ` PaX Team
2016-07-10 12:03         ` [kernel-hardening] " PaX Team
2016-07-10 12:03         ` PaX Team
2016-07-10 12:03         ` PaX Team
2016-07-10 12:03         ` PaX Team
2016-07-10 12:03         ` PaX Team
2016-07-10 12:03         ` PaX Team
2016-07-10 12:03         ` PaX Team
2016-07-10 12:38         ` Andy Lutomirski
2016-07-10 12:38           ` [kernel-hardening] " Andy Lutomirski
2016-07-10 12:38           ` Andy Lutomirski
2016-07-10 12:38           ` Andy Lutomirski
2016-07-10 12:38           ` Andy Lutomirski
2016-07-10 12:38           ` Andy Lutomirski
2016-07-10 12:38           ` Andy Lutomirski
2016-07-11 18:40           ` Kees Cook
2016-07-11 18:40             ` [kernel-hardening] " Kees Cook
2016-07-11 18:40             ` Kees Cook
2016-07-11 18:40             ` Kees Cook
2016-07-11 18:40             ` Kees Cook
2016-07-11 18:40             ` Kees Cook
2016-07-11 18:40             ` Kees Cook
2016-07-11 18:34         ` Kees Cook
2016-07-11 18:34           ` [kernel-hardening] " Kees Cook
2016-07-11 18:34           ` Kees Cook
2016-07-11 18:34           ` Kees Cook
2016-07-11 18:34           ` Kees Cook
2016-07-11 18:34           ` Kees Cook
2016-07-11 18:34           ` Kees Cook
2016-07-12 18:26 ` [kernel-hardening] " Valdis.Kletnieks
2016-07-12 18:44   ` Kees Cook

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=CAGXu5jLyBfqXJKxohHiZgztRVrFyqwbta1W_Dw6KyyGM3LzshQ@mail.gmail.com \
    --to=keescook@chromium.org \
    --cc=aarcange@redhat.com \
    --cc=akpm@linux-foundation.org \
    --cc=ard.biesheuvel@linaro.org \
    --cc=arnd@arndb.de \
    --cc=casey@schaufler-ca.com \
    --cc=catalin.marinas@arm.com \
    --cc=cl@linux.com \
    --cc=davem@davemloft.net \
    --cc=dvyukov@google.com \
    --cc=fenghua.yu@intel.com \
    --cc=iamjoonsoo.kim@lge.com \
    --cc=jack@suse.cz \
    --cc=kernel-hardening@lists.openwall.com \
    --cc=labbott@fedoraproject.org \
    --cc=linux-arch@vger.kernel.org \
    --cc=linux-arm-kernel@lists.infradead.org \
    --cc=linux-ia64@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=linux@armlinux.org.uk \
    --cc=linuxppc-dev@lists.ozlabs.org \
    --cc=luto@kernel.org \
    --cc=minipli@googlemail.com \
    --cc=pageexec@freemail.hu \
    --cc=penberg@kernel.org \
    --cc=riel@redhat.com \
    --cc=rientjes@google.com \
    --cc=sparclinux@vger.kernel.org \
    --cc=spender@grsecurity.net \
    --cc=tony.luck@intel.com \
    --cc=will.deacon@arm.com \
    --cc=x86@kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.