From mboxrd@z Thu Jan 1 00:00:00 1970 From: Andy Lutomirski Subject: Re: [RFC PATCH 1/9] kernel: add support for patchable function pointers Date: Fri, 5 Oct 2018 10:28:55 -0700 Message-ID: References: <20181005081333.15018-1-ard.biesheuvel@linaro.org> <20181005081333.15018-2-ard.biesheuvel@linaro.org> <20181005141433.GS19272@hirez.programming.kicks-ass.net> <9E0E08C8-0DFC-4E50-A4FA-73208835EF9E@amacapital.net> Mime-Version: 1.0 Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: quoted-printable Cc: Andrew Lutomirski , Peter Zijlstra , LKML , "Jason A. Donenfeld" , Eric Biggers , Samuel Neves , Arnd Bergmann , Herbert Xu , "David S. Miller" , Catalin Marinas , Will Deacon , Benjamin Herrenschmidt , Paul Mackerras , Michael Ellerman , Thomas Gleixner , Ingo Molnar , Kees Cook , "Martin K. Petersen" , Greg KH , Andrew Morton Return-path: In-Reply-To: Sender: linux-kernel-owner@vger.kernel.org List-Id: linux-crypto.vger.kernel.org On Fri, Oct 5, 2018 at 10:23 AM Ard Biesheuvel wrote: > > On 5 October 2018 at 19:20, Andy Lutomirski wrote: > > On Fri, Oct 5, 2018 at 10:11 AM Ard Biesheuvel > > wrote: > >> > >> On 5 October 2018 at 18:58, Andy Lutomirski wrote: > >> > On Fri, Oct 5, 2018 at 8:24 AM Ard Biesheuvel wrote: > >> >> > >> >> On 5 October 2018 at 17:08, Andy Lutomirski w= rote: > >> >> > > >> >> > > >> >> >> On Oct 5, 2018, at 7:14 AM, Peter Zijlstra wrote: > >> >> >> > >> >> >>> On Fri, Oct 05, 2018 at 10:13:25AM +0200, Ard Biesheuvel wrote: > >> >> >>> diff --git a/include/linux/ffp.h b/include/linux/ffp.h > >> >> >>> new file mode 100644 > >> >> >>> index 000000000000..8fc3b4c9b38f > >> >> >>> --- /dev/null > >> >> >>> +++ b/include/linux/ffp.h > >> >> >>> @@ -0,0 +1,43 @@ > >> >> >>> +/* SPDX-License-Identifier: GPL-2.0 */ > >> >> >>> + > >> >> >>> +#ifndef __LINUX_FFP_H > >> >> >>> +#define __LINUX_FFP_H > >> >> >>> + > >> >> >>> +#include > >> >> >>> +#include > >> >> >>> + > >> >> >>> +#ifdef CONFIG_HAVE_ARCH_FFP > >> >> >>> +#include > >> >> >>> +#else > >> >> >>> + > >> >> >>> +struct ffp { > >> >> >>> + void (**fn)(void); > >> >> >>> + void (*default_fn)(void); > >> >> >>> +}; > >> >> >>> + > >> >> >>> +#define DECLARE_FFP(_fn, _def) \ > >> >> >>> + extern typeof(_def) *_fn; \ > >> >> >>> + extern struct ffp const __ffp_ ## _fn > >> >> >>> + > >> >> >>> +#define DEFINE_FFP(_fn, _def) \ > >> >> >>> + typeof(_def) *_fn =3D &_def; \ > >> >> >>> + struct ffp const __ffp_ ## _fn \ > >> >> >>> + =3D { (void(**)(void))&_fn, (void(*)(void))&_def }; = \ > >> >> >>> + EXPORT_SYMBOL(__ffp_ ## _fn) > >> >> >>> + > >> >> >>> +static inline void ffp_set_target(const struct ffp *m, void *n= ew_fn) > >> >> >>> +{ > >> >> >>> + WRITE_ONCE(*m->fn, new_fn); > >> >> >>> +} > >> >> >>> + > >> >> >>> +static inline void ffp_reset_target(const struct ffp *m) > >> >> >>> +{ > >> >> >>> + WRITE_ONCE(*m->fn, m->default_fn); > >> >> >>> +} > >> >> >>> + > >> >> >>> +#endif > >> >> >>> + > >> >> >>> +#define SET_FFP(_fn, _new) ffp_set_target(&__ffp_ ## _fn, _= new) > >> >> >>> +#define RESET_FFP(_fn) ffp_reset_target(&__ffp_ ## _fn) > >> >> >>> + > >> >> >>> +#endif > >> >> >> > >> >> >> I don't understand this interface. There is no wrapper for the c= all > >> >> >> site, so how are we going to patch all call-sites when you updat= e the > >> >> >> target? > >> >> > > >> >> > I=E2=80=99m also confused. > >> >> > > >> >> > Anyway, we have patchable functions on x86. They=E2=80=99re calle= d PVOPs, and they=E2=80=99re way overcomplicated. > >> >> > > >> >> > I=E2=80=99ve proposed a better way that should generate better co= de, be more portable, and be more maintainable. It goes like this. > >> >> > > >> >> > To call the function, you literally just call the default implem= entation. It *might* be necessary to call a nonexistent wrapper to avoid a= nnoying optimizations. At build time, the kernel is built with relocations,= so the object files contain relocation entries for the call. We collect th= ese entries into a table. If we=E2=80=99re using the =E2=80=9Cnonexistent w= rapper=E2=80=9D approach, we can link in a .S or linker script to alias the= m to the default implementation. > >> >> > > >> >> > To patch them, we just patch them. It can=E2=80=99t necessarily b= e done concurrently because nothing forces the right alignment. But we can = do it at boot time and module load time. (Maybe we can patch at runtime on = architectures with appropriate instruction alignment. Or we ask gcc for an= extension to align calls to a function.) > >> >> > > >> >> > Most of the machinery already exists: this is roughly how the mod= ule loader resolves calls outside of a module. > >> >> > >> >> Yeah nothing is ever simple on x86 :-( > >> >> > >> >> So are you saying the approach i use in patch #2 (which would > >> >> translate to emitting a jmpq instruction pointing to the default > >> >> implementation, and patching it at runtime to point elsewhere) woul= d > >> >> not fly on x86? > >> > > >> > After getting some more sleep, I'm obviously wrong. The > >> > text_poke_bp() mechanism will work. It's just really slow. > >> > > >> > >> OK > >> > >> > Let me try to summarize some of the issues. First, when emitting > >> > jumps and calls from inline asm on x86, there are a few consideratio= ns > >> > that are annoying: > >> > > >> > 1. Following the x86_64 ABI calling conventions is basically > >> > impossible. x86_64 requires a 128-byte redzone and 16-byte stack > >> > alignment. After much discussion a while back, we decided that it w= as > >> > flat-out impossible on current gcc to get the stack pointer aligned = in > >> > a known manner in an inline asm statement. Instead, if we actually > >> > need alignment, we need to align manually. Fortunately, the kernel = is > >> > built with an override that forces only 8-byte alignment (on *most* > >> > GCC versions). But for crypto in particular, it sucks extra, since > >> > the crypto code is basically the only thing in the kernel that > >> > actually wants 16-byte alignment. I don't think this is a huge > >> > problem in practice, but it's annoying. And the kernel is built > >> > without a redzone. > >> > > >> > 2. On x86_64, depending on config, we either need frame pointers or > >> > ORC. ORC is no big deal -- it Just Works (tm). Frame pointers need > >> > extra asm hackery. It's doable, but it's still annoying. > >> > > >> > 3. Actually getting the asm constraints right to do what a C > >> > programmer expects is distinctly nontrivial. I just fixed an > >> > extremely longstanding bug in the vDSO code in which the asm > >> > constraints for the syscall fallback were wrong in such a way that G= CC > >> > didn't notice that the fallback wrote to its output parameter. > >> > Whoops. > >> > > >> > >> OK, so the thing I am missing is why this all matters. > >> > >> Note that the compiler should take care of all of this. It emits a > >> call a function with external linkage having prototype X, and all the > >> inline asm does is emit a jmp to some function having that same > >> prototype, either the default one or the one we patched in. > >> > >> Apologies if I am missing something obvious here: as you know, x86 is > >> not my focus in general. > > > > The big issue that bothers me isn't the x86-ism so much as the nasty > > interactions with the optimizer. On x86, we have all of this working. > > It's in arch/x86/include/asm/paravirt_types.h, and it looks roughly > > like: > > > > asm volatile(pre = \ > > paravirt_alt(PARAVIRT_CALL) = \ > > post = \ > > : call_clbr, ASM_CALL_CONSTRAINT = \ > > : paravirt_type(op), = \ > > paravirt_clobber(clbr), = \ > > ##__VA_ARGS__ = \ > > : "memory", "cc" extra_clbr); = \ > > > > With some extra magic for the constraints. And I don't even think > > this is strictly correct -- from very recent experience, telling the > > compiler that "memory" is clobbered and that a bunch of arguments are > > used as numeric inputs may not actually imply that the asm modifies > > the target of pointer arguments. Checks this lovely bug out: > > > > https://git.kernel.org/pub/scm/linux/kernel/git/luto/linux.git/commit/?= h=3Dx86/vdso-tglx&id=3D715bd9d12f84d8f5cc8ad21d888f9bc304a8eb0b > > > > As far as I can tell, the whole PVOP infrastructure has the same bug. > > And I don't see how to avoid it generically on x86 or any other > > architecture. (PeterZ, am I wrong? Are we really just getting lucky > > that x86 pvop calls actually work? Or do we not have enough of them > > that take pointers as arguments for this to matter?) > > > > Plus, asm volatile ( ..., "memory" ) is a barrier and makes code > > generation suck. > > > > Whereas, if we use my suggestion the semantics are precisely those of > > any other C function call because, as far as GCC is concerned, it *is* > > a C function call. So the generated code *is* a function call. > > > > But it is the *compiler* that actually emits the call. > > Only, the target of that call is a jmpq to another location where some > version of the routine lives, and all have the same prototype. > > How is that any different from PLTs in shared libraries? Ah, I see, I misunderstood your code. See other email. I think that, if you rework your series a bit to have a generic version that works on all architectures, then do it like you did on ARM, and make sure you leave the door open for the inline patching approach, then it looks pretty good. (None of this is to say that I disagree with Jason, though -- I'm not entirely convinced that this makes sense for Zinc. But maybe it can be done in a way that makes everyone happy.) From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-4.1 required=3.0 tests=DKIMWL_WL_HIGH,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_PASS, URIBL_BLOCKED autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 692D6C65BA7 for ; Fri, 5 Oct 2018 17:29:12 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id EED6121473 for ; Fri, 5 Oct 2018 17:29:11 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (1024-bit key) header.d=kernel.org header.i=@kernel.org header.b="X+0UpW2U" DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org EED6121473 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=kernel.org Authentication-Results: mail.kernel.org; spf=none smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728900AbeJFA2v (ORCPT ); Fri, 5 Oct 2018 20:28:51 -0400 Received: from mail.kernel.org ([198.145.29.99]:50222 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1727958AbeJFA2v (ORCPT ); Fri, 5 Oct 2018 20:28:51 -0400 Received: from mail-wr1-f49.google.com (mail-wr1-f49.google.com [209.85.221.49]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPSA id 6395221477 for ; Fri, 5 Oct 2018 17:29:08 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=default; t=1538760548; bh=MkvjJ/fQC95RoHlI0Pp5CoWyQvuhqmSPBPMV4OpYxEQ=; h=References:In-Reply-To:From:Date:Subject:To:Cc:From; b=X+0UpW2U+pHqqAphMYRCo//3SQj4NuS7xc9LTPyJFHWfmjxJAVS5pwm43xKWuHkTZ 65A966K6u2TW9u3kH2ngZV7igjEBFYSzn1vycME/lah8xLu/rFg1O0U9VWNBxqC+QV n15tPtzx0rK+RtpjlrvYq7Bpd3VmNPDZ/U1sq1bM= Received: by mail-wr1-f49.google.com with SMTP id z4-v6so14347652wrb.1 for ; Fri, 05 Oct 2018 10:29:08 -0700 (PDT) X-Gm-Message-State: ABuFfog1IyZvGsGWFQPKAjVRrO8CeL7urqSMQ+F7KEu3Cjy7mh37zaOW 7Alb90bmNyI1hcE+41BWpq2LiU+h42ZV7iW6dQ/qJg== X-Google-Smtp-Source: ACcGV63xLl2YYe896osC7EpaN0FZRIySNw24yY8hHovxpnPbqKbvMY9lxrPhKi28X6Zkug7nAPZv6ko5pP5wazwcL7M= X-Received: by 2002:adf:9792:: with SMTP id s18-v6mr9365531wrb.283.1538760546759; Fri, 05 Oct 2018 10:29:06 -0700 (PDT) MIME-Version: 1.0 References: <20181005081333.15018-1-ard.biesheuvel@linaro.org> <20181005081333.15018-2-ard.biesheuvel@linaro.org> <20181005141433.GS19272@hirez.programming.kicks-ass.net> <9E0E08C8-0DFC-4E50-A4FA-73208835EF9E@amacapital.net> In-Reply-To: From: Andy Lutomirski Date: Fri, 5 Oct 2018 10:28:55 -0700 X-Gmail-Original-Message-ID: Message-ID: Subject: Re: [RFC PATCH 1/9] kernel: add support for patchable function pointers To: Ard Biesheuvel Cc: Andrew Lutomirski , Peter Zijlstra , LKML , "Jason A. Donenfeld" , Eric Biggers , Samuel Neves , Arnd Bergmann , Herbert Xu , "David S. Miller" , Catalin Marinas , Will Deacon , Benjamin Herrenschmidt , Paul Mackerras , Michael Ellerman , Thomas Gleixner , Ingo Molnar , Kees Cook , "Martin K. Petersen" , Greg KH , Andrew Morton , Richard Weinberger , Linux Crypto Mailing List , linux-arm-kernel , linuxppc-dev Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: quoted-printable Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Oct 5, 2018 at 10:23 AM Ard Biesheuvel wrote: > > On 5 October 2018 at 19:20, Andy Lutomirski wrote: > > On Fri, Oct 5, 2018 at 10:11 AM Ard Biesheuvel > > wrote: > >> > >> On 5 October 2018 at 18:58, Andy Lutomirski wrote: > >> > On Fri, Oct 5, 2018 at 8:24 AM Ard Biesheuvel wrote: > >> >> > >> >> On 5 October 2018 at 17:08, Andy Lutomirski w= rote: > >> >> > > >> >> > > >> >> >> On Oct 5, 2018, at 7:14 AM, Peter Zijlstra wrote: > >> >> >> > >> >> >>> On Fri, Oct 05, 2018 at 10:13:25AM +0200, Ard Biesheuvel wrote: > >> >> >>> diff --git a/include/linux/ffp.h b/include/linux/ffp.h > >> >> >>> new file mode 100644 > >> >> >>> index 000000000000..8fc3b4c9b38f > >> >> >>> --- /dev/null > >> >> >>> +++ b/include/linux/ffp.h > >> >> >>> @@ -0,0 +1,43 @@ > >> >> >>> +/* SPDX-License-Identifier: GPL-2.0 */ > >> >> >>> + > >> >> >>> +#ifndef __LINUX_FFP_H > >> >> >>> +#define __LINUX_FFP_H > >> >> >>> + > >> >> >>> +#include > >> >> >>> +#include > >> >> >>> + > >> >> >>> +#ifdef CONFIG_HAVE_ARCH_FFP > >> >> >>> +#include > >> >> >>> +#else > >> >> >>> + > >> >> >>> +struct ffp { > >> >> >>> + void (**fn)(void); > >> >> >>> + void (*default_fn)(void); > >> >> >>> +}; > >> >> >>> + > >> >> >>> +#define DECLARE_FFP(_fn, _def) \ > >> >> >>> + extern typeof(_def) *_fn; \ > >> >> >>> + extern struct ffp const __ffp_ ## _fn > >> >> >>> + > >> >> >>> +#define DEFINE_FFP(_fn, _def) \ > >> >> >>> + typeof(_def) *_fn =3D &_def; \ > >> >> >>> + struct ffp const __ffp_ ## _fn \ > >> >> >>> + =3D { (void(**)(void))&_fn, (void(*)(void))&_def }; = \ > >> >> >>> + EXPORT_SYMBOL(__ffp_ ## _fn) > >> >> >>> + > >> >> >>> +static inline void ffp_set_target(const struct ffp *m, void *n= ew_fn) > >> >> >>> +{ > >> >> >>> + WRITE_ONCE(*m->fn, new_fn); > >> >> >>> +} > >> >> >>> + > >> >> >>> +static inline void ffp_reset_target(const struct ffp *m) > >> >> >>> +{ > >> >> >>> + WRITE_ONCE(*m->fn, m->default_fn); > >> >> >>> +} > >> >> >>> + > >> >> >>> +#endif > >> >> >>> + > >> >> >>> +#define SET_FFP(_fn, _new) ffp_set_target(&__ffp_ ## _fn, _= new) > >> >> >>> +#define RESET_FFP(_fn) ffp_reset_target(&__ffp_ ## _fn) > >> >> >>> + > >> >> >>> +#endif > >> >> >> > >> >> >> I don't understand this interface. There is no wrapper for the c= all > >> >> >> site, so how are we going to patch all call-sites when you updat= e the > >> >> >> target? > >> >> > > >> >> > I=E2=80=99m also confused. > >> >> > > >> >> > Anyway, we have patchable functions on x86. They=E2=80=99re calle= d PVOPs, and they=E2=80=99re way overcomplicated. > >> >> > > >> >> > I=E2=80=99ve proposed a better way that should generate better co= de, be more portable, and be more maintainable. It goes like this. > >> >> > > >> >> > To call the function, you literally just call the default implem= entation. It *might* be necessary to call a nonexistent wrapper to avoid a= nnoying optimizations. At build time, the kernel is built with relocations,= so the object files contain relocation entries for the call. We collect th= ese entries into a table. If we=E2=80=99re using the =E2=80=9Cnonexistent w= rapper=E2=80=9D approach, we can link in a .S or linker script to alias the= m to the default implementation. > >> >> > > >> >> > To patch them, we just patch them. It can=E2=80=99t necessarily b= e done concurrently because nothing forces the right alignment. But we can = do it at boot time and module load time. (Maybe we can patch at runtime on = architectures with appropriate instruction alignment. Or we ask gcc for an= extension to align calls to a function.) > >> >> > > >> >> > Most of the machinery already exists: this is roughly how the mod= ule loader resolves calls outside of a module. > >> >> > >> >> Yeah nothing is ever simple on x86 :-( > >> >> > >> >> So are you saying the approach i use in patch #2 (which would > >> >> translate to emitting a jmpq instruction pointing to the default > >> >> implementation, and patching it at runtime to point elsewhere) woul= d > >> >> not fly on x86? > >> > > >> > After getting some more sleep, I'm obviously wrong. The > >> > text_poke_bp() mechanism will work. It's just really slow. > >> > > >> > >> OK > >> > >> > Let me try to summarize some of the issues. First, when emitting > >> > jumps and calls from inline asm on x86, there are a few consideratio= ns > >> > that are annoying: > >> > > >> > 1. Following the x86_64 ABI calling conventions is basically > >> > impossible. x86_64 requires a 128-byte redzone and 16-byte stack > >> > alignment. After much discussion a while back, we decided that it w= as > >> > flat-out impossible on current gcc to get the stack pointer aligned = in > >> > a known manner in an inline asm statement. Instead, if we actually > >> > need alignment, we need to align manually. Fortunately, the kernel = is > >> > built with an override that forces only 8-byte alignment (on *most* > >> > GCC versions). But for crypto in particular, it sucks extra, since > >> > the crypto code is basically the only thing in the kernel that > >> > actually wants 16-byte alignment. I don't think this is a huge > >> > problem in practice, but it's annoying. And the kernel is built > >> > without a redzone. > >> > > >> > 2. On x86_64, depending on config, we either need frame pointers or > >> > ORC. ORC is no big deal -- it Just Works (tm). Frame pointers need > >> > extra asm hackery. It's doable, but it's still annoying. > >> > > >> > 3. Actually getting the asm constraints right to do what a C > >> > programmer expects is distinctly nontrivial. I just fixed an > >> > extremely longstanding bug in the vDSO code in which the asm > >> > constraints for the syscall fallback were wrong in such a way that G= CC > >> > didn't notice that the fallback wrote to its output parameter. > >> > Whoops. > >> > > >> > >> OK, so the thing I am missing is why this all matters. > >> > >> Note that the compiler should take care of all of this. It emits a > >> call a function with external linkage having prototype X, and all the > >> inline asm does is emit a jmp to some function having that same > >> prototype, either the default one or the one we patched in. > >> > >> Apologies if I am missing something obvious here: as you know, x86 is > >> not my focus in general. > > > > The big issue that bothers me isn't the x86-ism so much as the nasty > > interactions with the optimizer. On x86, we have all of this working. > > It's in arch/x86/include/asm/paravirt_types.h, and it looks roughly > > like: > > > > asm volatile(pre = \ > > paravirt_alt(PARAVIRT_CALL) = \ > > post = \ > > : call_clbr, ASM_CALL_CONSTRAINT = \ > > : paravirt_type(op), = \ > > paravirt_clobber(clbr), = \ > > ##__VA_ARGS__ = \ > > : "memory", "cc" extra_clbr); = \ > > > > With some extra magic for the constraints. And I don't even think > > this is strictly correct -- from very recent experience, telling the > > compiler that "memory" is clobbered and that a bunch of arguments are > > used as numeric inputs may not actually imply that the asm modifies > > the target of pointer arguments. Checks this lovely bug out: > > > > https://git.kernel.org/pub/scm/linux/kernel/git/luto/linux.git/commit/?= h=3Dx86/vdso-tglx&id=3D715bd9d12f84d8f5cc8ad21d888f9bc304a8eb0b > > > > As far as I can tell, the whole PVOP infrastructure has the same bug. > > And I don't see how to avoid it generically on x86 or any other > > architecture. (PeterZ, am I wrong? Are we really just getting lucky > > that x86 pvop calls actually work? Or do we not have enough of them > > that take pointers as arguments for this to matter?) > > > > Plus, asm volatile ( ..., "memory" ) is a barrier and makes code > > generation suck. > > > > Whereas, if we use my suggestion the semantics are precisely those of > > any other C function call because, as far as GCC is concerned, it *is* > > a C function call. So the generated code *is* a function call. > > > > But it is the *compiler* that actually emits the call. > > Only, the target of that call is a jmpq to another location where some > version of the routine lives, and all have the same prototype. > > How is that any different from PLTs in shared libraries? Ah, I see, I misunderstood your code. See other email. I think that, if you rework your series a bit to have a generic version that works on all architectures, then do it like you did on ARM, and make sure you leave the door open for the inline patching approach, then it looks pretty good. (None of this is to say that I disagree with Jason, though -- I'm not entirely convinced that this makes sense for Zinc. But maybe it can be done in a way that makes everyone happy.) From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-3.8 required=3.0 tests=DKIM_INVALID,DKIM_SIGNED, INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_PASS,URIBL_BLOCKED autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id C736CC00449 for ; Fri, 5 Oct 2018 18:35:21 +0000 (UTC) Received: from lists.ozlabs.org (lists.ozlabs.org [203.11.71.2]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 3F6D821477 for ; Fri, 5 Oct 2018 18:35:21 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=fail reason="signature verification failed" (1024-bit key) header.d=kernel.org header.i=@kernel.org header.b="cGjgBpIW" DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 3F6D821477 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=kernel.org Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=linuxppc-dev-bounces+linuxppc-dev=archiver.kernel.org@lists.ozlabs.org Received: from lists.ozlabs.org (lists.ozlabs.org [IPv6:2401:3900:2:1::3]) by lists.ozlabs.org (Postfix) with ESMTP id 42RdjL73rFzF3Kn for ; Sat, 6 Oct 2018 04:35:18 +1000 (AEST) Authentication-Results: lists.ozlabs.org; dmarc=pass (p=none dis=none) header.from=kernel.org Authentication-Results: lists.ozlabs.org; dkim=fail reason="signature verification failed" (1024-bit key; unprotected) header.d=kernel.org header.i=@kernel.org header.b="cGjgBpIW"; dkim-atps=neutral Authentication-Results: lists.ozlabs.org; spf=pass (mailfrom) smtp.mailfrom=kernel.org (client-ip=198.145.29.99; helo=mail.kernel.org; envelope-from=luto@kernel.org; receiver=) Authentication-Results: lists.ozlabs.org; dmarc=pass (p=none dis=none) header.from=kernel.org Authentication-Results: lists.ozlabs.org; dkim=pass (1024-bit key; unprotected) header.d=kernel.org header.i=@kernel.org header.b="cGjgBpIW"; dkim-atps=neutral Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by lists.ozlabs.org (Postfix) with ESMTPS id 42RcF45F2pzF3cM for ; Sat, 6 Oct 2018 03:29:12 +1000 (AEST) Received: from mail-wr1-f49.google.com (mail-wr1-f49.google.com [209.85.221.49]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPSA id 929BD21486 for ; Fri, 5 Oct 2018 17:29:10 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=default; t=1538760551; bh=MkvjJ/fQC95RoHlI0Pp5CoWyQvuhqmSPBPMV4OpYxEQ=; h=References:In-Reply-To:From:Date:Subject:To:Cc:From; b=cGjgBpIW/of4aV60wUvf9WMoYgQVrTwLlQ6WyRBBC1cReErZKqGEDiDQfJsyWl9wZ 5tc8ZJnZ8TwTlyXKMH+uQehRY4CBWhzQKkTO9XOSRZPJywvghRL326afWoTbsXT+nV k/2YDWdWVLYrx+Nq/eDVGR3dI7yMLqMmJiHOlkWE= Received: by mail-wr1-f49.google.com with SMTP id n1-v6so14333107wrt.10 for ; Fri, 05 Oct 2018 10:29:10 -0700 (PDT) X-Gm-Message-State: ABuFfoiri0ivjzQnN7OyRKDSPm8JYWCp7S5w5B3tFF85DZ2MCDL7QJ32 MEa3rzdNGbsnT4rIkhwt/kXNwLajknZZ8qJHcuyISw== X-Google-Smtp-Source: ACcGV63xLl2YYe896osC7EpaN0FZRIySNw24yY8hHovxpnPbqKbvMY9lxrPhKi28X6Zkug7nAPZv6ko5pP5wazwcL7M= X-Received: by 2002:adf:9792:: with SMTP id s18-v6mr9365531wrb.283.1538760546759; Fri, 05 Oct 2018 10:29:06 -0700 (PDT) MIME-Version: 1.0 References: <20181005081333.15018-1-ard.biesheuvel@linaro.org> <20181005081333.15018-2-ard.biesheuvel@linaro.org> <20181005141433.GS19272@hirez.programming.kicks-ass.net> <9E0E08C8-0DFC-4E50-A4FA-73208835EF9E@amacapital.net> In-Reply-To: From: Andy Lutomirski Date: Fri, 5 Oct 2018 10:28:55 -0700 X-Gmail-Original-Message-ID: Message-ID: Subject: Re: [RFC PATCH 1/9] kernel: add support for patchable function pointers To: Ard Biesheuvel Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: quoted-printable X-BeenThere: linuxppc-dev@lists.ozlabs.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: Linux on PowerPC Developers Mail List List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: "Jason A. Donenfeld" , Peter Zijlstra , Catalin Marinas , Will Deacon , Samuel Neves , Paul Mackerras , Herbert Xu , Richard Weinberger , Eric Biggers , Ingo Molnar , Kees Cook , Arnd Bergmann , Andrew Lutomirski , Thomas Gleixner , linux-arm-kernel , "Martin K. Petersen" , Greg KH , LKML , Linux Crypto Mailing List , Andrew Morton , linuxppc-dev , "David S. Miller" Errors-To: linuxppc-dev-bounces+linuxppc-dev=archiver.kernel.org@lists.ozlabs.org Sender: "Linuxppc-dev" On Fri, Oct 5, 2018 at 10:23 AM Ard Biesheuvel wrote: > > On 5 October 2018 at 19:20, Andy Lutomirski wrote: > > On Fri, Oct 5, 2018 at 10:11 AM Ard Biesheuvel > > wrote: > >> > >> On 5 October 2018 at 18:58, Andy Lutomirski wrote: > >> > On Fri, Oct 5, 2018 at 8:24 AM Ard Biesheuvel wrote: > >> >> > >> >> On 5 October 2018 at 17:08, Andy Lutomirski w= rote: > >> >> > > >> >> > > >> >> >> On Oct 5, 2018, at 7:14 AM, Peter Zijlstra wrote: > >> >> >> > >> >> >>> On Fri, Oct 05, 2018 at 10:13:25AM +0200, Ard Biesheuvel wrote: > >> >> >>> diff --git a/include/linux/ffp.h b/include/linux/ffp.h > >> >> >>> new file mode 100644 > >> >> >>> index 000000000000..8fc3b4c9b38f > >> >> >>> --- /dev/null > >> >> >>> +++ b/include/linux/ffp.h > >> >> >>> @@ -0,0 +1,43 @@ > >> >> >>> +/* SPDX-License-Identifier: GPL-2.0 */ > >> >> >>> + > >> >> >>> +#ifndef __LINUX_FFP_H > >> >> >>> +#define __LINUX_FFP_H > >> >> >>> + > >> >> >>> +#include > >> >> >>> +#include > >> >> >>> + > >> >> >>> +#ifdef CONFIG_HAVE_ARCH_FFP > >> >> >>> +#include > >> >> >>> +#else > >> >> >>> + > >> >> >>> +struct ffp { > >> >> >>> + void (**fn)(void); > >> >> >>> + void (*default_fn)(void); > >> >> >>> +}; > >> >> >>> + > >> >> >>> +#define DECLARE_FFP(_fn, _def) \ > >> >> >>> + extern typeof(_def) *_fn; \ > >> >> >>> + extern struct ffp const __ffp_ ## _fn > >> >> >>> + > >> >> >>> +#define DEFINE_FFP(_fn, _def) \ > >> >> >>> + typeof(_def) *_fn =3D &_def; \ > >> >> >>> + struct ffp const __ffp_ ## _fn \ > >> >> >>> + =3D { (void(**)(void))&_fn, (void(*)(void))&_def }; = \ > >> >> >>> + EXPORT_SYMBOL(__ffp_ ## _fn) > >> >> >>> + > >> >> >>> +static inline void ffp_set_target(const struct ffp *m, void *n= ew_fn) > >> >> >>> +{ > >> >> >>> + WRITE_ONCE(*m->fn, new_fn); > >> >> >>> +} > >> >> >>> + > >> >> >>> +static inline void ffp_reset_target(const struct ffp *m) > >> >> >>> +{ > >> >> >>> + WRITE_ONCE(*m->fn, m->default_fn); > >> >> >>> +} > >> >> >>> + > >> >> >>> +#endif > >> >> >>> + > >> >> >>> +#define SET_FFP(_fn, _new) ffp_set_target(&__ffp_ ## _fn, _= new) > >> >> >>> +#define RESET_FFP(_fn) ffp_reset_target(&__ffp_ ## _fn) > >> >> >>> + > >> >> >>> +#endif > >> >> >> > >> >> >> I don't understand this interface. There is no wrapper for the c= all > >> >> >> site, so how are we going to patch all call-sites when you updat= e the > >> >> >> target? > >> >> > > >> >> > I=E2=80=99m also confused. > >> >> > > >> >> > Anyway, we have patchable functions on x86. They=E2=80=99re calle= d PVOPs, and they=E2=80=99re way overcomplicated. > >> >> > > >> >> > I=E2=80=99ve proposed a better way that should generate better co= de, be more portable, and be more maintainable. It goes like this. > >> >> > > >> >> > To call the function, you literally just call the default implem= entation. It *might* be necessary to call a nonexistent wrapper to avoid a= nnoying optimizations. At build time, the kernel is built with relocations,= so the object files contain relocation entries for the call. We collect th= ese entries into a table. If we=E2=80=99re using the =E2=80=9Cnonexistent w= rapper=E2=80=9D approach, we can link in a .S or linker script to alias the= m to the default implementation. > >> >> > > >> >> > To patch them, we just patch them. It can=E2=80=99t necessarily b= e done concurrently because nothing forces the right alignment. But we can = do it at boot time and module load time. (Maybe we can patch at runtime on = architectures with appropriate instruction alignment. Or we ask gcc for an= extension to align calls to a function.) > >> >> > > >> >> > Most of the machinery already exists: this is roughly how the mod= ule loader resolves calls outside of a module. > >> >> > >> >> Yeah nothing is ever simple on x86 :-( > >> >> > >> >> So are you saying the approach i use in patch #2 (which would > >> >> translate to emitting a jmpq instruction pointing to the default > >> >> implementation, and patching it at runtime to point elsewhere) woul= d > >> >> not fly on x86? > >> > > >> > After getting some more sleep, I'm obviously wrong. The > >> > text_poke_bp() mechanism will work. It's just really slow. > >> > > >> > >> OK > >> > >> > Let me try to summarize some of the issues. First, when emitting > >> > jumps and calls from inline asm on x86, there are a few consideratio= ns > >> > that are annoying: > >> > > >> > 1. Following the x86_64 ABI calling conventions is basically > >> > impossible. x86_64 requires a 128-byte redzone and 16-byte stack > >> > alignment. After much discussion a while back, we decided that it w= as > >> > flat-out impossible on current gcc to get the stack pointer aligned = in > >> > a known manner in an inline asm statement. Instead, if we actually > >> > need alignment, we need to align manually. Fortunately, the kernel = is > >> > built with an override that forces only 8-byte alignment (on *most* > >> > GCC versions). But for crypto in particular, it sucks extra, since > >> > the crypto code is basically the only thing in the kernel that > >> > actually wants 16-byte alignment. I don't think this is a huge > >> > problem in practice, but it's annoying. And the kernel is built > >> > without a redzone. > >> > > >> > 2. On x86_64, depending on config, we either need frame pointers or > >> > ORC. ORC is no big deal -- it Just Works (tm). Frame pointers need > >> > extra asm hackery. It's doable, but it's still annoying. > >> > > >> > 3. Actually getting the asm constraints right to do what a C > >> > programmer expects is distinctly nontrivial. I just fixed an > >> > extremely longstanding bug in the vDSO code in which the asm > >> > constraints for the syscall fallback were wrong in such a way that G= CC > >> > didn't notice that the fallback wrote to its output parameter. > >> > Whoops. > >> > > >> > >> OK, so the thing I am missing is why this all matters. > >> > >> Note that the compiler should take care of all of this. It emits a > >> call a function with external linkage having prototype X, and all the > >> inline asm does is emit a jmp to some function having that same > >> prototype, either the default one or the one we patched in. > >> > >> Apologies if I am missing something obvious here: as you know, x86 is > >> not my focus in general. > > > > The big issue that bothers me isn't the x86-ism so much as the nasty > > interactions with the optimizer. On x86, we have all of this working. > > It's in arch/x86/include/asm/paravirt_types.h, and it looks roughly > > like: > > > > asm volatile(pre = \ > > paravirt_alt(PARAVIRT_CALL) = \ > > post = \ > > : call_clbr, ASM_CALL_CONSTRAINT = \ > > : paravirt_type(op), = \ > > paravirt_clobber(clbr), = \ > > ##__VA_ARGS__ = \ > > : "memory", "cc" extra_clbr); = \ > > > > With some extra magic for the constraints. And I don't even think > > this is strictly correct -- from very recent experience, telling the > > compiler that "memory" is clobbered and that a bunch of arguments are > > used as numeric inputs may not actually imply that the asm modifies > > the target of pointer arguments. Checks this lovely bug out: > > > > https://git.kernel.org/pub/scm/linux/kernel/git/luto/linux.git/commit/?= h=3Dx86/vdso-tglx&id=3D715bd9d12f84d8f5cc8ad21d888f9bc304a8eb0b > > > > As far as I can tell, the whole PVOP infrastructure has the same bug. > > And I don't see how to avoid it generically on x86 or any other > > architecture. (PeterZ, am I wrong? Are we really just getting lucky > > that x86 pvop calls actually work? Or do we not have enough of them > > that take pointers as arguments for this to matter?) > > > > Plus, asm volatile ( ..., "memory" ) is a barrier and makes code > > generation suck. > > > > Whereas, if we use my suggestion the semantics are precisely those of > > any other C function call because, as far as GCC is concerned, it *is* > > a C function call. So the generated code *is* a function call. > > > > But it is the *compiler* that actually emits the call. > > Only, the target of that call is a jmpq to another location where some > version of the routine lives, and all have the same prototype. > > How is that any different from PLTs in shared libraries? Ah, I see, I misunderstood your code. See other email. I think that, if you rework your series a bit to have a generic version that works on all architectures, then do it like you did on ARM, and make sure you leave the door open for the inline patching approach, then it looks pretty good. (None of this is to say that I disagree with Jason, though -- I'm not entirely convinced that this makes sense for Zinc. But maybe it can be done in a way that makes everyone happy.) From mboxrd@z Thu Jan 1 00:00:00 1970 From: luto@kernel.org (Andy Lutomirski) Date: Fri, 5 Oct 2018 10:28:55 -0700 Subject: [RFC PATCH 1/9] kernel: add support for patchable function pointers In-Reply-To: References: <20181005081333.15018-1-ard.biesheuvel@linaro.org> <20181005081333.15018-2-ard.biesheuvel@linaro.org> <20181005141433.GS19272@hirez.programming.kicks-ass.net> <9E0E08C8-0DFC-4E50-A4FA-73208835EF9E@amacapital.net> Message-ID: To: linux-arm-kernel@lists.infradead.org List-Id: linux-arm-kernel.lists.infradead.org On Fri, Oct 5, 2018 at 10:23 AM Ard Biesheuvel wrote: > > On 5 October 2018 at 19:20, Andy Lutomirski wrote: > > On Fri, Oct 5, 2018 at 10:11 AM Ard Biesheuvel > > wrote: > >> > >> On 5 October 2018 at 18:58, Andy Lutomirski wrote: > >> > On Fri, Oct 5, 2018 at 8:24 AM Ard Biesheuvel wrote: > >> >> > >> >> On 5 October 2018 at 17:08, Andy Lutomirski wrote: > >> >> > > >> >> > > >> >> >> On Oct 5, 2018, at 7:14 AM, Peter Zijlstra wrote: > >> >> >> > >> >> >>> On Fri, Oct 05, 2018 at 10:13:25AM +0200, Ard Biesheuvel wrote: > >> >> >>> diff --git a/include/linux/ffp.h b/include/linux/ffp.h > >> >> >>> new file mode 100644 > >> >> >>> index 000000000000..8fc3b4c9b38f > >> >> >>> --- /dev/null > >> >> >>> +++ b/include/linux/ffp.h > >> >> >>> @@ -0,0 +1,43 @@ > >> >> >>> +/* SPDX-License-Identifier: GPL-2.0 */ > >> >> >>> + > >> >> >>> +#ifndef __LINUX_FFP_H > >> >> >>> +#define __LINUX_FFP_H > >> >> >>> + > >> >> >>> +#include > >> >> >>> +#include > >> >> >>> + > >> >> >>> +#ifdef CONFIG_HAVE_ARCH_FFP > >> >> >>> +#include > >> >> >>> +#else > >> >> >>> + > >> >> >>> +struct ffp { > >> >> >>> + void (**fn)(void); > >> >> >>> + void (*default_fn)(void); > >> >> >>> +}; > >> >> >>> + > >> >> >>> +#define DECLARE_FFP(_fn, _def) \ > >> >> >>> + extern typeof(_def) *_fn; \ > >> >> >>> + extern struct ffp const __ffp_ ## _fn > >> >> >>> + > >> >> >>> +#define DEFINE_FFP(_fn, _def) \ > >> >> >>> + typeof(_def) *_fn = &_def; \ > >> >> >>> + struct ffp const __ffp_ ## _fn \ > >> >> >>> + = { (void(**)(void))&_fn, (void(*)(void))&_def }; \ > >> >> >>> + EXPORT_SYMBOL(__ffp_ ## _fn) > >> >> >>> + > >> >> >>> +static inline void ffp_set_target(const struct ffp *m, void *new_fn) > >> >> >>> +{ > >> >> >>> + WRITE_ONCE(*m->fn, new_fn); > >> >> >>> +} > >> >> >>> + > >> >> >>> +static inline void ffp_reset_target(const struct ffp *m) > >> >> >>> +{ > >> >> >>> + WRITE_ONCE(*m->fn, m->default_fn); > >> >> >>> +} > >> >> >>> + > >> >> >>> +#endif > >> >> >>> + > >> >> >>> +#define SET_FFP(_fn, _new) ffp_set_target(&__ffp_ ## _fn, _new) > >> >> >>> +#define RESET_FFP(_fn) ffp_reset_target(&__ffp_ ## _fn) > >> >> >>> + > >> >> >>> +#endif > >> >> >> > >> >> >> I don't understand this interface. There is no wrapper for the call > >> >> >> site, so how are we going to patch all call-sites when you update the > >> >> >> target? > >> >> > > >> >> > I?m also confused. > >> >> > > >> >> > Anyway, we have patchable functions on x86. They?re called PVOPs, and they?re way overcomplicated. > >> >> > > >> >> > I?ve proposed a better way that should generate better code, be more portable, and be more maintainable. It goes like this. > >> >> > > >> >> > To call the function, you literally just call the default implementation. It *might* be necessary to call a nonexistent wrapper to avoid annoying optimizations. At build time, the kernel is built with relocations, so the object files contain relocation entries for the call. We collect these entries into a table. If we?re using the ?nonexistent wrapper? approach, we can link in a .S or linker script to alias them to the default implementation. > >> >> > > >> >> > To patch them, we just patch them. It can?t necessarily be done concurrently because nothing forces the right alignment. But we can do it at boot time and module load time. (Maybe we can patch at runtime on architectures with appropriate instruction alignment. Or we ask gcc for an extension to align calls to a function.) > >> >> > > >> >> > Most of the machinery already exists: this is roughly how the module loader resolves calls outside of a module. > >> >> > >> >> Yeah nothing is ever simple on x86 :-( > >> >> > >> >> So are you saying the approach i use in patch #2 (which would > >> >> translate to emitting a jmpq instruction pointing to the default > >> >> implementation, and patching it at runtime to point elsewhere) would > >> >> not fly on x86? > >> > > >> > After getting some more sleep, I'm obviously wrong. The > >> > text_poke_bp() mechanism will work. It's just really slow. > >> > > >> > >> OK > >> > >> > Let me try to summarize some of the issues. First, when emitting > >> > jumps and calls from inline asm on x86, there are a few considerations > >> > that are annoying: > >> > > >> > 1. Following the x86_64 ABI calling conventions is basically > >> > impossible. x86_64 requires a 128-byte redzone and 16-byte stack > >> > alignment. After much discussion a while back, we decided that it was > >> > flat-out impossible on current gcc to get the stack pointer aligned in > >> > a known manner in an inline asm statement. Instead, if we actually > >> > need alignment, we need to align manually. Fortunately, the kernel is > >> > built with an override that forces only 8-byte alignment (on *most* > >> > GCC versions). But for crypto in particular, it sucks extra, since > >> > the crypto code is basically the only thing in the kernel that > >> > actually wants 16-byte alignment. I don't think this is a huge > >> > problem in practice, but it's annoying. And the kernel is built > >> > without a redzone. > >> > > >> > 2. On x86_64, depending on config, we either need frame pointers or > >> > ORC. ORC is no big deal -- it Just Works (tm). Frame pointers need > >> > extra asm hackery. It's doable, but it's still annoying. > >> > > >> > 3. Actually getting the asm constraints right to do what a C > >> > programmer expects is distinctly nontrivial. I just fixed an > >> > extremely longstanding bug in the vDSO code in which the asm > >> > constraints for the syscall fallback were wrong in such a way that GCC > >> > didn't notice that the fallback wrote to its output parameter. > >> > Whoops. > >> > > >> > >> OK, so the thing I am missing is why this all matters. > >> > >> Note that the compiler should take care of all of this. It emits a > >> call a function with external linkage having prototype X, and all the > >> inline asm does is emit a jmp to some function having that same > >> prototype, either the default one or the one we patched in. > >> > >> Apologies if I am missing something obvious here: as you know, x86 is > >> not my focus in general. > > > > The big issue that bothers me isn't the x86-ism so much as the nasty > > interactions with the optimizer. On x86, we have all of this working. > > It's in arch/x86/include/asm/paravirt_types.h, and it looks roughly > > like: > > > > asm volatile(pre \ > > paravirt_alt(PARAVIRT_CALL) \ > > post \ > > : call_clbr, ASM_CALL_CONSTRAINT \ > > : paravirt_type(op), \ > > paravirt_clobber(clbr), \ > > ##__VA_ARGS__ \ > > : "memory", "cc" extra_clbr); \ > > > > With some extra magic for the constraints. And I don't even think > > this is strictly correct -- from very recent experience, telling the > > compiler that "memory" is clobbered and that a bunch of arguments are > > used as numeric inputs may not actually imply that the asm modifies > > the target of pointer arguments. Checks this lovely bug out: > > > > https://git.kernel.org/pub/scm/linux/kernel/git/luto/linux.git/commit/?h=x86/vdso-tglx&id=715bd9d12f84d8f5cc8ad21d888f9bc304a8eb0b > > > > As far as I can tell, the whole PVOP infrastructure has the same bug. > > And I don't see how to avoid it generically on x86 or any other > > architecture. (PeterZ, am I wrong? Are we really just getting lucky > > that x86 pvop calls actually work? Or do we not have enough of them > > that take pointers as arguments for this to matter?) > > > > Plus, asm volatile ( ..., "memory" ) is a barrier and makes code > > generation suck. > > > > Whereas, if we use my suggestion the semantics are precisely those of > > any other C function call because, as far as GCC is concerned, it *is* > > a C function call. So the generated code *is* a function call. > > > > But it is the *compiler* that actually emits the call. > > Only, the target of that call is a jmpq to another location where some > version of the routine lives, and all have the same prototype. > > How is that any different from PLTs in shared libraries? Ah, I see, I misunderstood your code. See other email. I think that, if you rework your series a bit to have a generic version that works on all architectures, then do it like you did on ARM, and make sure you leave the door open for the inline patching approach, then it looks pretty good. (None of this is to say that I disagree with Jason, though -- I'm not entirely convinced that this makes sense for Zinc. But maybe it can be done in a way that makes everyone happy.)