From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-11.8 required=3.0 tests=HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_PATCH,MAILING_LIST_MULTI, MENTIONS_GIT_HOSTING,SIGNED_OFF_BY,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 37327CA9EA0 for ; Fri, 18 Oct 2019 16:31:36 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id 105A120659 for ; Fri, 18 Oct 2019 16:31:36 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S2443076AbfJRQbe (ORCPT ); Fri, 18 Oct 2019 12:31:34 -0400 Received: from Galois.linutronix.de ([193.142.43.55]:57821 "EHLO Galois.linutronix.de" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S2442947AbfJRQbK (ORCPT ); Fri, 18 Oct 2019 12:31:10 -0400 Received: from [5.158.153.53] (helo=tip-bot2.lab.linutronix.de) by Galois.linutronix.de with esmtpsa (TLS1.2:DHE_RSA_AES_256_CBC_SHA256:256) (Exim 4.80) (envelope-from ) id 1iLV90-0002j8-Qm; Fri, 18 Oct 2019 18:30:27 +0200 Received: from [127.0.1.1] (localhost [IPv6:::1]) by tip-bot2.lab.linutronix.de (Postfix) with ESMTP id 54A741C048C; Fri, 18 Oct 2019 18:30:26 +0200 (CEST) Date: Fri, 18 Oct 2019 16:30:26 -0000 From: "tip-bot2 for Jiri Slaby" Reply-to: linux-kernel@vger.kernel.org To: linux-tip-commits@vger.kernel.org Subject: [tip: x86/asm] x86/asm/32: Add ENDs to some functions and relabel with SYM_CODE_* Cc: Jiri Slaby , Borislav Petkov , Boris Ostrovsky , "Rafael J. Wysocki" , Andy Lutomirski , "H. Peter Anvin" , Ingo Molnar , Josh Poimboeuf , Juergen Gross , Len Brown , linux-arch@vger.kernel.org, linux-pm@vger.kernel.org, Pavel Machek , Peter Zijlstra , Pingfan Liu , Stefano Stabellini , "Steven Rostedt (VMware)" , Thomas Gleixner , "x86-ml" , xen-devel@lists.xenproject.org, Ingo Molnar , Borislav Petkov , linux-kernel@vger.kernel.org In-Reply-To: <20191011115108.12392-26-jslaby@suse.cz> References: <20191011115108.12392-26-jslaby@suse.cz> MIME-Version: 1.0 Message-ID: <157141622620.29376.2511020268312198257.tip-bot2@tip-bot2> X-Mailer: tip-git-log-daemon Robot-ID: Robot-Unsubscribe: Contact to get blacklisted from these emails Content-Type: text/plain; charset="utf-8" Content-Transfer-Encoding: 7bit X-Linutronix-Spam-Score: -1.0 X-Linutronix-Spam-Level: - X-Linutronix-Spam-Status: No , -1.0 points, 5.0 required, ALL_TRUSTED=-1,SHORTCIRCUIT=-0.0001 Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org The following commit has been merged into the x86/asm branch of tip: Commit-ID: 78762b0e79bc1dd01347be061abdf505202152c9 Gitweb: https://git.kernel.org/tip/78762b0e79bc1dd01347be061abdf505202152c9 Author: Jiri Slaby AuthorDate: Fri, 11 Oct 2019 13:51:05 +02:00 Committer: Borislav Petkov CommitterDate: Fri, 18 Oct 2019 11:58:33 +02:00 x86/asm/32: Add ENDs to some functions and relabel with SYM_CODE_* All these are functions which are invoked from elsewhere but they are not typical C functions. So annotate them using the new SYM_CODE_START. All these were not balanced with any END, so mark their ends by SYM_CODE_END, appropriately. Signed-off-by: Jiri Slaby Signed-off-by: Borislav Petkov Reviewed-by: Boris Ostrovsky [xen bits] Reviewed-by: Rafael J. Wysocki [hibernate] Cc: Andy Lutomirski Cc: "H. Peter Anvin" Cc: Ingo Molnar Cc: Josh Poimboeuf Cc: Juergen Gross Cc: Len Brown Cc: linux-arch@vger.kernel.org Cc: linux-pm@vger.kernel.org Cc: Pavel Machek Cc: Peter Zijlstra Cc: Pingfan Liu Cc: Stefano Stabellini Cc: "Steven Rostedt (VMware)" Cc: Thomas Gleixner Cc: x86-ml Cc: xen-devel@lists.xenproject.org Link: https://lkml.kernel.org/r/20191011115108.12392-26-jslaby@suse.cz --- arch/x86/entry/entry_32.S | 3 ++- arch/x86/kernel/acpi/wakeup_32.S | 7 ++++--- arch/x86/kernel/ftrace_32.S | 3 ++- arch/x86/kernel/head_32.S | 3 ++- arch/x86/power/hibernate_asm_32.S | 6 ++++-- arch/x86/realmode/rm/trampoline_32.S | 6 ++++-- arch/x86/xen/xen-asm_32.S | 7 ++++--- 7 files changed, 22 insertions(+), 13 deletions(-) diff --git a/arch/x86/entry/entry_32.S b/arch/x86/entry/entry_32.S index 4900a6a..64fe7aa 100644 --- a/arch/x86/entry/entry_32.S +++ b/arch/x86/entry/entry_32.S @@ -847,9 +847,10 @@ SYM_ENTRY(__begin_SYSENTER_singlestep_region, SYM_L_GLOBAL, SYM_A_NONE) * Xen doesn't set %esp to be precisely what the normal SYSENTER * entry point expects, so fix it up before using the normal path. */ -ENTRY(xen_sysenter_target) +SYM_CODE_START(xen_sysenter_target) addl $5*4, %esp /* remove xen-provided frame */ jmp .Lsysenter_past_esp +SYM_CODE_END(xen_sysenter_target) #endif /* diff --git a/arch/x86/kernel/acpi/wakeup_32.S b/arch/x86/kernel/acpi/wakeup_32.S index 4272492..daf88f8 100644 --- a/arch/x86/kernel/acpi/wakeup_32.S +++ b/arch/x86/kernel/acpi/wakeup_32.S @@ -9,8 +9,7 @@ .code32 ALIGN -ENTRY(wakeup_pmode_return) -wakeup_pmode_return: +SYM_CODE_START(wakeup_pmode_return) movw $__KERNEL_DS, %ax movw %ax, %ss movw %ax, %fs @@ -39,6 +38,7 @@ wakeup_pmode_return: # jump to place where we left off movl saved_eip, %eax jmp *%eax +SYM_CODE_END(wakeup_pmode_return) bogus_magic: jmp bogus_magic @@ -72,7 +72,7 @@ restore_registers: popfl ret -ENTRY(do_suspend_lowlevel) +SYM_CODE_START(do_suspend_lowlevel) call save_processor_state call save_registers pushl $3 @@ -87,6 +87,7 @@ ret_point: call restore_registers call restore_processor_state ret +SYM_CODE_END(do_suspend_lowlevel) .data ALIGN diff --git a/arch/x86/kernel/ftrace_32.S b/arch/x86/kernel/ftrace_32.S index 219be13..a43ed4c 100644 --- a/arch/x86/kernel/ftrace_32.S +++ b/arch/x86/kernel/ftrace_32.S @@ -89,7 +89,7 @@ WEAK(ftrace_stub) ret END(ftrace_caller) -ENTRY(ftrace_regs_caller) +SYM_CODE_START(ftrace_regs_caller) /* * We're here from an mcount/fentry CALL, and the stack frame looks like: * @@ -163,6 +163,7 @@ SYM_INNER_LABEL(ftrace_regs_call, SYM_L_GLOBAL) popl %eax jmp .Lftrace_ret +SYM_CODE_END(ftrace_regs_caller) #ifdef CONFIG_FUNCTION_GRAPH_TRACER ENTRY(ftrace_graph_caller) diff --git a/arch/x86/kernel/head_32.S b/arch/x86/kernel/head_32.S index e2b3e6c..7029bba 100644 --- a/arch/x86/kernel/head_32.S +++ b/arch/x86/kernel/head_32.S @@ -64,7 +64,7 @@ RESERVE_BRK(pagetables, INIT_MAP_SIZE) * can. */ __HEAD -ENTRY(startup_32) +SYM_CODE_START(startup_32) movl pa(initial_stack),%ecx /* test KEEP_SEGMENTS flag to see if the bootloader is asking @@ -172,6 +172,7 @@ num_subarch_entries = (. - subarch_entries) / 4 #else jmp .Ldefault_entry #endif /* CONFIG_PARAVIRT */ +SYM_CODE_END(startup_32) #ifdef CONFIG_HOTPLUG_CPU /* diff --git a/arch/x86/power/hibernate_asm_32.S b/arch/x86/power/hibernate_asm_32.S index 6fe3830..a19ed3d 100644 --- a/arch/x86/power/hibernate_asm_32.S +++ b/arch/x86/power/hibernate_asm_32.S @@ -35,7 +35,7 @@ ENTRY(swsusp_arch_suspend) ret ENDPROC(swsusp_arch_suspend) -ENTRY(restore_image) +SYM_CODE_START(restore_image) /* prepare to jump to the image kernel */ movl restore_jump_address, %ebx movl restore_cr3, %ebp @@ -45,9 +45,10 @@ ENTRY(restore_image) /* jump to relocated restore code */ movl relocated_restore_code, %eax jmpl *%eax +SYM_CODE_END(restore_image) /* code below has been relocated to a safe page */ -ENTRY(core_restore_code) +SYM_CODE_START(core_restore_code) movl temp_pgt, %eax movl %eax, %cr3 @@ -77,6 +78,7 @@ copy_loop: done: jmpl *%ebx +SYM_CODE_END(core_restore_code) /* code below belongs to the image kernel */ .align PAGE_SIZE diff --git a/arch/x86/realmode/rm/trampoline_32.S b/arch/x86/realmode/rm/trampoline_32.S index ff00594..3fad907 100644 --- a/arch/x86/realmode/rm/trampoline_32.S +++ b/arch/x86/realmode/rm/trampoline_32.S @@ -29,7 +29,7 @@ .code16 .balign PAGE_SIZE -ENTRY(trampoline_start) +SYM_CODE_START(trampoline_start) wbinvd # Needed for NUMA-Q should be harmless for others LJMPW_RM(1f) @@ -54,11 +54,13 @@ ENTRY(trampoline_start) lmsw %dx # into protected mode ljmpl $__BOOT_CS, $pa_startup_32 +SYM_CODE_END(trampoline_start) .section ".text32","ax" .code32 -ENTRY(startup_32) # note: also used from wakeup_asm.S +SYM_CODE_START(startup_32) # note: also used from wakeup_asm.S jmp *%eax +SYM_CODE_END(startup_32) .bss .balign 8 diff --git a/arch/x86/xen/xen-asm_32.S b/arch/x86/xen/xen-asm_32.S index c15db06..8b8f835 100644 --- a/arch/x86/xen/xen-asm_32.S +++ b/arch/x86/xen/xen-asm_32.S @@ -56,7 +56,7 @@ _ASM_EXTABLE(1b,2b) .endm -ENTRY(xen_iret) +SYM_CODE_START(xen_iret) /* test eflags for special cases */ testl $(X86_EFLAGS_VM | XEN_EFLAGS_NMI), 8(%esp) jnz hyper_iret @@ -122,6 +122,7 @@ xen_iret_end_crit: hyper_iret: /* put this out of line since its very rarely used */ jmp hypercall_page + __HYPERVISOR_iret * 32 +SYM_CODE_END(xen_iret) .globl xen_iret_start_crit, xen_iret_end_crit @@ -165,7 +166,7 @@ hyper_iret: * SAVE_ALL state before going on, since it's usermode state which we * eventually need to restore. */ -ENTRY(xen_iret_crit_fixup) +SYM_CODE_START(xen_iret_crit_fixup) /* * Paranoia: Make sure we're really coming from kernel space. * One could imagine a case where userspace jumps into the @@ -204,4 +205,4 @@ ENTRY(xen_iret_crit_fixup) lea 4(%edi), %esp /* point esp to new frame */ 2: jmp xen_do_upcall - +SYM_CODE_END(xen_iret_crit_fixup) From mboxrd@z Thu Jan 1 00:00:00 1970 From: "tip-bot2 for Jiri Slaby" Subject: [tip: x86/asm] x86/asm/32: Add ENDs to some functions and relabel with SYM_CODE_* Date: Fri, 18 Oct 2019 16:30:26 -0000 Message-ID: <157141622620.29376.2511020268312198257.tip-bot2@tip-bot2> References: <20191011115108.12392-26-jslaby@suse.cz> Reply-To: linux-kernel@vger.kernel.org Mime-Version: 1.0 Content-Type: text/plain; charset="utf-8" Content-Transfer-Encoding: 7bit Return-path: In-Reply-To: <20191011115108.12392-26-jslaby@suse.cz> Sender: linux-kernel-owner@vger.kernel.org To: linux-tip-commits@vger.kernel.org Cc: Jiri Slaby , Borislav Petkov , Boris Ostrovsky , "Rafael J. Wysocki" , Andy Lutomirski , "H. Peter Anvin" , Ingo Molnar , Josh Poimboeuf , Juergen Gross , Len Brown , linux-arch@vger.kernel.org, linux-pm@vger.kernel.org, Pavel Machek , Peter Zijlstra , Pingfan Liu , Stefano Stabellini , "Steven Rostedt (VMware)" , Thomas Gleixner , x86-ml , xen-devel@lists.xenproject.org, Ingo Molnar , Borislav Petkov List-Id: linux-arch.vger.kernel.org The following commit has been merged into the x86/asm branch of tip: Commit-ID: 78762b0e79bc1dd01347be061abdf505202152c9 Gitweb: https://git.kernel.org/tip/78762b0e79bc1dd01347be061abdf505202152c9 Author: Jiri Slaby AuthorDate: Fri, 11 Oct 2019 13:51:05 +02:00 Committer: Borislav Petkov CommitterDate: Fri, 18 Oct 2019 11:58:33 +02:00 x86/asm/32: Add ENDs to some functions and relabel with SYM_CODE_* All these are functions which are invoked from elsewhere but they are not typical C functions. So annotate them using the new SYM_CODE_START. All these were not balanced with any END, so mark their ends by SYM_CODE_END, appropriately. Signed-off-by: Jiri Slaby Signed-off-by: Borislav Petkov Reviewed-by: Boris Ostrovsky [xen bits] Reviewed-by: Rafael J. Wysocki [hibernate] Cc: Andy Lutomirski Cc: "H. Peter Anvin" Cc: Ingo Molnar Cc: Josh Poimboeuf Cc: Juergen Gross Cc: Len Brown Cc: linux-arch@vger.kernel.org Cc: linux-pm@vger.kernel.org Cc: Pavel Machek Cc: Peter Zijlstra Cc: Pingfan Liu Cc: Stefano Stabellini Cc: "Steven Rostedt (VMware)" Cc: Thomas Gleixner Cc: x86-ml Cc: xen-devel@lists.xenproject.org Link: https://lkml.kernel.org/r/20191011115108.12392-26-jslaby@suse.cz --- arch/x86/entry/entry_32.S | 3 ++- arch/x86/kernel/acpi/wakeup_32.S | 7 ++++--- arch/x86/kernel/ftrace_32.S | 3 ++- arch/x86/kernel/head_32.S | 3 ++- arch/x86/power/hibernate_asm_32.S | 6 ++++-- arch/x86/realmode/rm/trampoline_32.S | 6 ++++-- arch/x86/xen/xen-asm_32.S | 7 ++++--- 7 files changed, 22 insertions(+), 13 deletions(-) diff --git a/arch/x86/entry/entry_32.S b/arch/x86/entry/entry_32.S index 4900a6a..64fe7aa 100644 --- a/arch/x86/entry/entry_32.S +++ b/arch/x86/entry/entry_32.S @@ -847,9 +847,10 @@ SYM_ENTRY(__begin_SYSENTER_singlestep_region, SYM_L_GLOBAL, SYM_A_NONE) * Xen doesn't set %esp to be precisely what the normal SYSENTER * entry point expects, so fix it up before using the normal path. */ -ENTRY(xen_sysenter_target) +SYM_CODE_START(xen_sysenter_target) addl $5*4, %esp /* remove xen-provided frame */ jmp .Lsysenter_past_esp +SYM_CODE_END(xen_sysenter_target) #endif /* diff --git a/arch/x86/kernel/acpi/wakeup_32.S b/arch/x86/kernel/acpi/wakeup_32.S index 4272492..daf88f8 100644 --- a/arch/x86/kernel/acpi/wakeup_32.S +++ b/arch/x86/kernel/acpi/wakeup_32.S @@ -9,8 +9,7 @@ .code32 ALIGN -ENTRY(wakeup_pmode_return) -wakeup_pmode_return: +SYM_CODE_START(wakeup_pmode_return) movw $__KERNEL_DS, %ax movw %ax, %ss movw %ax, %fs @@ -39,6 +38,7 @@ wakeup_pmode_return: # jump to place where we left off movl saved_eip, %eax jmp *%eax +SYM_CODE_END(wakeup_pmode_return) bogus_magic: jmp bogus_magic @@ -72,7 +72,7 @@ restore_registers: popfl ret -ENTRY(do_suspend_lowlevel) +SYM_CODE_START(do_suspend_lowlevel) call save_processor_state call save_registers pushl $3 @@ -87,6 +87,7 @@ ret_point: call restore_registers call restore_processor_state ret +SYM_CODE_END(do_suspend_lowlevel) .data ALIGN diff --git a/arch/x86/kernel/ftrace_32.S b/arch/x86/kernel/ftrace_32.S index 219be13..a43ed4c 100644 --- a/arch/x86/kernel/ftrace_32.S +++ b/arch/x86/kernel/ftrace_32.S @@ -89,7 +89,7 @@ WEAK(ftrace_stub) ret END(ftrace_caller) -ENTRY(ftrace_regs_caller) +SYM_CODE_START(ftrace_regs_caller) /* * We're here from an mcount/fentry CALL, and the stack frame looks like: * @@ -163,6 +163,7 @@ SYM_INNER_LABEL(ftrace_regs_call, SYM_L_GLOBAL) popl %eax jmp .Lftrace_ret +SYM_CODE_END(ftrace_regs_caller) #ifdef CONFIG_FUNCTION_GRAPH_TRACER ENTRY(ftrace_graph_caller) diff --git a/arch/x86/kernel/head_32.S b/arch/x86/kernel/head_32.S index e2b3e6c..7029bba 100644 --- a/arch/x86/kernel/head_32.S +++ b/arch/x86/kernel/head_32.S @@ -64,7 +64,7 @@ RESERVE_BRK(pagetables, INIT_MAP_SIZE) * can. */ __HEAD -ENTRY(startup_32) +SYM_CODE_START(startup_32) movl pa(initial_stack),%ecx /* test KEEP_SEGMENTS flag to see if the bootloader is asking @@ -172,6 +172,7 @@ num_subarch_entries = (. - subarch_entries) / 4 #else jmp .Ldefault_entry #endif /* CONFIG_PARAVIRT */ +SYM_CODE_END(startup_32) #ifdef CONFIG_HOTPLUG_CPU /* diff --git a/arch/x86/power/hibernate_asm_32.S b/arch/x86/power/hibernate_asm_32.S index 6fe3830..a19ed3d 100644 --- a/arch/x86/power/hibernate_asm_32.S +++ b/arch/x86/power/hibernate_asm_32.S @@ -35,7 +35,7 @@ ENTRY(swsusp_arch_suspend) ret ENDPROC(swsusp_arch_suspend) -ENTRY(restore_image) +SYM_CODE_START(restore_image) /* prepare to jump to the image kernel */ movl restore_jump_address, %ebx movl restore_cr3, %ebp @@ -45,9 +45,10 @@ ENTRY(restore_image) /* jump to relocated restore code */ movl relocated_restore_code, %eax jmpl *%eax +SYM_CODE_END(restore_image) /* code below has been relocated to a safe page */ -ENTRY(core_restore_code) +SYM_CODE_START(core_restore_code) movl temp_pgt, %eax movl %eax, %cr3 @@ -77,6 +78,7 @@ copy_loop: done: jmpl *%ebx +SYM_CODE_END(core_restore_code) /* code below belongs to the image kernel */ .align PAGE_SIZE diff --git a/arch/x86/realmode/rm/trampoline_32.S b/arch/x86/realmode/rm/trampoline_32.S index ff00594..3fad907 100644 --- a/arch/x86/realmode/rm/trampoline_32.S +++ b/arch/x86/realmode/rm/trampoline_32.S @@ -29,7 +29,7 @@ .code16 .balign PAGE_SIZE -ENTRY(trampoline_start) +SYM_CODE_START(trampoline_start) wbinvd # Needed for NUMA-Q should be harmless for others LJMPW_RM(1f) @@ -54,11 +54,13 @@ ENTRY(trampoline_start) lmsw %dx # into protected mode ljmpl $__BOOT_CS, $pa_startup_32 +SYM_CODE_END(trampoline_start) .section ".text32","ax" .code32 -ENTRY(startup_32) # note: also used from wakeup_asm.S +SYM_CODE_START(startup_32) # note: also used from wakeup_asm.S jmp *%eax +SYM_CODE_END(startup_32) .bss .balign 8 diff --git a/arch/x86/xen/xen-asm_32.S b/arch/x86/xen/xen-asm_32.S index c15db06..8b8f835 100644 --- a/arch/x86/xen/xen-asm_32.S +++ b/arch/x86/xen/xen-asm_32.S @@ -56,7 +56,7 @@ _ASM_EXTABLE(1b,2b) .endm -ENTRY(xen_iret) +SYM_CODE_START(xen_iret) /* test eflags for special cases */ testl $(X86_EFLAGS_VM | XEN_EFLAGS_NMI), 8(%esp) jnz hyper_iret @@ -122,6 +122,7 @@ xen_iret_end_crit: hyper_iret: /* put this out of line since its very rarely used */ jmp hypercall_page + __HYPERVISOR_iret * 32 +SYM_CODE_END(xen_iret) .globl xen_iret_start_crit, xen_iret_end_crit @@ -165,7 +166,7 @@ hyper_iret: * SAVE_ALL state before going on, since it's usermode state which we * eventually need to restore. */ -ENTRY(xen_iret_crit_fixup) +SYM_CODE_START(xen_iret_crit_fixup) /* * Paranoia: Make sure we're really coming from kernel space. * One could imagine a case where userspace jumps into the @@ -204,4 +205,4 @@ ENTRY(xen_iret_crit_fixup) lea 4(%edi), %esp /* point esp to new frame */ 2: jmp xen_do_upcall - +SYM_CODE_END(xen_iret_crit_fixup) From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-11.8 required=3.0 tests=HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_PATCH,MAILING_LIST_MULTI, MENTIONS_GIT_HOSTING,SIGNED_OFF_BY,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 77EF5CA9EA0 for ; Fri, 18 Oct 2019 16:31:28 +0000 (UTC) Received: from lists.xenproject.org (lists.xenproject.org [192.237.175.120]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 510EF21835 for ; Fri, 18 Oct 2019 16:31:28 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 510EF21835 Authentication-Results: mail.kernel.org; dmarc=none (p=none dis=none) header.from=linutronix.de Authentication-Results: mail.kernel.org; spf=none smtp.mailfrom=xen-devel-bounces@lists.xenproject.org Received: from localhost ([127.0.0.1] helo=lists.xenproject.org) by lists.xenproject.org with esmtp (Exim 4.89) (envelope-from ) id 1iLV9d-0006Xp-7B; Fri, 18 Oct 2019 16:31:05 +0000 Received: from us1-rack-iad1.inumbo.com ([172.99.69.81]) by lists.xenproject.org with esmtp (Exim 4.89) (envelope-from ) id 1iLV9c-0006XD-5K for xen-devel@lists.xenproject.org; Fri, 18 Oct 2019 16:31:04 +0000 X-Inumbo-ID: a56e55f4-f1c4-11e9-a531-bc764e2007e4 Received: from Galois.linutronix.de (unknown [2a0a:51c0:0:12e:550::1]) by us1-rack-iad1.inumbo.com (Halon) with ESMTPS id a56e55f4-f1c4-11e9-a531-bc764e2007e4; Fri, 18 Oct 2019 16:30:50 +0000 (UTC) Received: from [5.158.153.53] (helo=tip-bot2.lab.linutronix.de) by Galois.linutronix.de with esmtpsa (TLS1.2:DHE_RSA_AES_256_CBC_SHA256:256) (Exim 4.80) (envelope-from ) id 1iLV90-0002j8-Qm; Fri, 18 Oct 2019 18:30:27 +0200 Received: from [127.0.1.1] (localhost [IPv6:::1]) by tip-bot2.lab.linutronix.de (Postfix) with ESMTP id 54A741C048C; Fri, 18 Oct 2019 18:30:26 +0200 (CEST) Date: Fri, 18 Oct 2019 16:30:26 -0000 From: "tip-bot2 for Jiri Slaby" To: linux-tip-commits@vger.kernel.org In-Reply-To: <20191011115108.12392-26-jslaby@suse.cz> References: <20191011115108.12392-26-jslaby@suse.cz> MIME-Version: 1.0 Message-ID: <157141622620.29376.2511020268312198257.tip-bot2@tip-bot2> X-Mailer: tip-git-log-daemon Robot-ID: Robot-Unsubscribe: Contact to get blacklisted from these emails Precedence: bulk X-Linutronix-Spam-Score: -1.0 X-Linutronix-Spam-Level: - X-Linutronix-Spam-Status: No , -1.0 points, 5.0 required, ALL_TRUSTED=-1, SHORTCIRCUIT=-0.0001 Subject: [Xen-devel] [tip: x86/asm] x86/asm/32: Add ENDs to some functions and relabel with SYM_CODE_* X-BeenThere: xen-devel@lists.xenproject.org X-Mailman-Version: 2.1.23 List-Id: Xen developer discussion List-Unsubscribe: , List-Post: List-Help: List-Subscribe: , Reply-To: linux-kernel@vger.kernel.org Cc: x86-ml , Peter Zijlstra , Pingfan Liu , Pavel Machek , "H. Peter Anvin" , Jiri Slaby , Ingo Molnar , linux-arch@vger.kernel.org, Stefano Stabellini , "Rafael J. Wysocki" , Ingo Molnar , xen-devel@lists.xenproject.org, Borislav Petkov , Len Brown , linux-pm@vger.kernel.org, "Steven Rostedt \(VMware\)" , Borislav Petkov , Andy Lutomirski , Josh Poimboeuf , Boris Ostrovsky , Juergen Gross , linux-kernel@vger.kernel.org, Thomas Gleixner Content-Type: text/plain; charset="utf-8" Content-Transfer-Encoding: base64 Errors-To: xen-devel-bounces@lists.xenproject.org Sender: "Xen-devel" VGhlIGZvbGxvd2luZyBjb21taXQgaGFzIGJlZW4gbWVyZ2VkIGludG8gdGhlIHg4Ni9hc20gYnJh bmNoIG9mIHRpcDoKCkNvbW1pdC1JRDogICAgIDc4NzYyYjBlNzliYzFkZDAxMzQ3YmUwNjFhYmRm NTA1MjAyMTUyYzkKR2l0d2ViOiAgICAgICAgaHR0cHM6Ly9naXQua2VybmVsLm9yZy90aXAvNzg3 NjJiMGU3OWJjMWRkMDEzNDdiZTA2MWFiZGY1MDUyMDIxNTJjOQpBdXRob3I6ICAgICAgICBKaXJp IFNsYWJ5IDxqc2xhYnlAc3VzZS5jej4KQXV0aG9yRGF0ZTogICAgRnJpLCAxMSBPY3QgMjAxOSAx Mzo1MTowNSArMDI6MDAKQ29tbWl0dGVyOiAgICAgQm9yaXNsYXYgUGV0a292IDxicEBzdXNlLmRl PgpDb21taXR0ZXJEYXRlOiBGcmksIDE4IE9jdCAyMDE5IDExOjU4OjMzICswMjowMAoKeDg2L2Fz bS8zMjogQWRkIEVORHMgdG8gc29tZSBmdW5jdGlvbnMgYW5kIHJlbGFiZWwgd2l0aCBTWU1fQ09E RV8qCgpBbGwgdGhlc2UgYXJlIGZ1bmN0aW9ucyB3aGljaCBhcmUgaW52b2tlZCBmcm9tIGVsc2V3 aGVyZSBidXQgdGhleSBhcmUKbm90IHR5cGljYWwgQyBmdW5jdGlvbnMuIFNvIGFubm90YXRlIHRo ZW0gdXNpbmcgdGhlIG5ldyBTWU1fQ09ERV9TVEFSVC4KQWxsIHRoZXNlIHdlcmUgbm90IGJhbGFu Y2VkIHdpdGggYW55IEVORCwgc28gbWFyayB0aGVpciBlbmRzIGJ5ClNZTV9DT0RFX0VORCwgYXBw cm9wcmlhdGVseS4KClNpZ25lZC1vZmYtYnk6IEppcmkgU2xhYnkgPGpzbGFieUBzdXNlLmN6PgpT aWduZWQtb2ZmLWJ5OiBCb3Jpc2xhdiBQZXRrb3YgPGJwQHN1c2UuZGU+ClJldmlld2VkLWJ5OiBC b3JpcyBPc3Ryb3Zza3kgPGJvcmlzLm9zdHJvdnNreUBvcmFjbGUuY29tPiBbeGVuIGJpdHNdClJl dmlld2VkLWJ5OiBSYWZhZWwgSi4gV3lzb2NraSA8cmFmYWVsLmoud3lzb2NraUBpbnRlbC5jb20+ IFtoaWJlcm5hdGVdCkNjOiBBbmR5IEx1dG9taXJza2kgPGx1dG9Aa2VybmVsLm9yZz4KQ2M6ICJI LiBQZXRlciBBbnZpbiIgPGhwYUB6eXRvci5jb20+CkNjOiBJbmdvIE1vbG5hciA8bWluZ29AcmVk aGF0LmNvbT4KQ2M6IEpvc2ggUG9pbWJvZXVmIDxqcG9pbWJvZUByZWRoYXQuY29tPgpDYzogSnVl cmdlbiBHcm9zcyA8amdyb3NzQHN1c2UuY29tPgpDYzogTGVuIEJyb3duIDxsZW4uYnJvd25AaW50 ZWwuY29tPgpDYzogbGludXgtYXJjaEB2Z2VyLmtlcm5lbC5vcmcKQ2M6IGxpbnV4LXBtQHZnZXIu a2VybmVsLm9yZwpDYzogUGF2ZWwgTWFjaGVrIDxwYXZlbEB1Y3cuY3o+CkNjOiBQZXRlciBaaWps c3RyYSA8cGV0ZXJ6QGluZnJhZGVhZC5vcmc+CkNjOiBQaW5nZmFuIExpdSA8a2VybmVsZmFuc0Bn bWFpbC5jb20+CkNjOiBTdGVmYW5vIFN0YWJlbGxpbmkgPHNzdGFiZWxsaW5pQGtlcm5lbC5vcmc+ CkNjOiAiU3RldmVuIFJvc3RlZHQgKFZNd2FyZSkiIDxyb3N0ZWR0QGdvb2RtaXMub3JnPgpDYzog VGhvbWFzIEdsZWl4bmVyIDx0Z2x4QGxpbnV0cm9uaXguZGU+CkNjOiB4ODYtbWwgPHg4NkBrZXJu ZWwub3JnPgpDYzogeGVuLWRldmVsQGxpc3RzLnhlbnByb2plY3Qub3JnCkxpbms6IGh0dHBzOi8v bGttbC5rZXJuZWwub3JnL3IvMjAxOTEwMTExMTUxMDguMTIzOTItMjYtanNsYWJ5QHN1c2UuY3oK LS0tCiBhcmNoL3g4Ni9lbnRyeS9lbnRyeV8zMi5TICAgICAgICAgICAgfCAzICsrLQogYXJjaC94 ODYva2VybmVsL2FjcGkvd2FrZXVwXzMyLlMgICAgIHwgNyArKysrLS0tCiBhcmNoL3g4Ni9rZXJu ZWwvZnRyYWNlXzMyLlMgICAgICAgICAgfCAzICsrLQogYXJjaC94ODYva2VybmVsL2hlYWRfMzIu UyAgICAgICAgICAgIHwgMyArKy0KIGFyY2gveDg2L3Bvd2VyL2hpYmVybmF0ZV9hc21fMzIuUyAg ICB8IDYgKysrKy0tCiBhcmNoL3g4Ni9yZWFsbW9kZS9ybS90cmFtcG9saW5lXzMyLlMgfCA2ICsr KystLQogYXJjaC94ODYveGVuL3hlbi1hc21fMzIuUyAgICAgICAgICAgIHwgNyArKysrLS0tCiA3 IGZpbGVzIGNoYW5nZWQsIDIyIGluc2VydGlvbnMoKyksIDEzIGRlbGV0aW9ucygtKQoKZGlmZiAt LWdpdCBhL2FyY2gveDg2L2VudHJ5L2VudHJ5XzMyLlMgYi9hcmNoL3g4Ni9lbnRyeS9lbnRyeV8z Mi5TCmluZGV4IDQ5MDBhNmEuLjY0ZmU3YWEgMTAwNjQ0Ci0tLSBhL2FyY2gveDg2L2VudHJ5L2Vu dHJ5XzMyLlMKKysrIGIvYXJjaC94ODYvZW50cnkvZW50cnlfMzIuUwpAQCAtODQ3LDkgKzg0Nywx MCBAQCBTWU1fRU5UUlkoX19iZWdpbl9TWVNFTlRFUl9zaW5nbGVzdGVwX3JlZ2lvbiwgU1lNX0xf R0xPQkFMLCBTWU1fQV9OT05FKQogICogWGVuIGRvZXNuJ3Qgc2V0ICVlc3AgdG8gYmUgcHJlY2lz ZWx5IHdoYXQgdGhlIG5vcm1hbCBTWVNFTlRFUgogICogZW50cnkgcG9pbnQgZXhwZWN0cywgc28g Zml4IGl0IHVwIGJlZm9yZSB1c2luZyB0aGUgbm9ybWFsIHBhdGguCiAgKi8KLUVOVFJZKHhlbl9z eXNlbnRlcl90YXJnZXQpCitTWU1fQ09ERV9TVEFSVCh4ZW5fc3lzZW50ZXJfdGFyZ2V0KQogCWFk ZGwJJDUqNCwgJWVzcAkJCS8qIHJlbW92ZSB4ZW4tcHJvdmlkZWQgZnJhbWUgKi8KIAlqbXAJLkxz eXNlbnRlcl9wYXN0X2VzcAorU1lNX0NPREVfRU5EKHhlbl9zeXNlbnRlcl90YXJnZXQpCiAjZW5k aWYKIAogLyoKZGlmZiAtLWdpdCBhL2FyY2gveDg2L2tlcm5lbC9hY3BpL3dha2V1cF8zMi5TIGIv YXJjaC94ODYva2VybmVsL2FjcGkvd2FrZXVwXzMyLlMKaW5kZXggNDI3MjQ5Mi4uZGFmODhmOCAx MDA2NDQKLS0tIGEvYXJjaC94ODYva2VybmVsL2FjcGkvd2FrZXVwXzMyLlMKKysrIGIvYXJjaC94 ODYva2VybmVsL2FjcGkvd2FrZXVwXzMyLlMKQEAgLTksOCArOSw3IEBACiAJLmNvZGUzMgogCUFM SUdOCiAKLUVOVFJZKHdha2V1cF9wbW9kZV9yZXR1cm4pCi13YWtldXBfcG1vZGVfcmV0dXJuOgor U1lNX0NPREVfU1RBUlQod2FrZXVwX3Btb2RlX3JldHVybikKIAltb3Z3CSRfX0tFUk5FTF9EUywg JWF4CiAJbW92dwklYXgsICVzcwogCW1vdncJJWF4LCAlZnMKQEAgLTM5LDYgKzM4LDcgQEAgd2Fr ZXVwX3Btb2RlX3JldHVybjoKIAkjIGp1bXAgdG8gcGxhY2Ugd2hlcmUgd2UgbGVmdCBvZmYKIAlt b3ZsCXNhdmVkX2VpcCwgJWVheAogCWptcAkqJWVheAorU1lNX0NPREVfRU5EKHdha2V1cF9wbW9k ZV9yZXR1cm4pCiAKIGJvZ3VzX21hZ2ljOgogCWptcAlib2d1c19tYWdpYwpAQCAtNzIsNyArNzIs NyBAQCByZXN0b3JlX3JlZ2lzdGVyczoKIAlwb3BmbAogCXJldAogCi1FTlRSWShkb19zdXNwZW5k X2xvd2xldmVsKQorU1lNX0NPREVfU1RBUlQoZG9fc3VzcGVuZF9sb3dsZXZlbCkKIAljYWxsCXNh dmVfcHJvY2Vzc29yX3N0YXRlCiAJY2FsbAlzYXZlX3JlZ2lzdGVycwogCXB1c2hsCSQzCkBAIC04 Nyw2ICs4Nyw3IEBAIHJldF9wb2ludDoKIAljYWxsCXJlc3RvcmVfcmVnaXN0ZXJzCiAJY2FsbAly ZXN0b3JlX3Byb2Nlc3Nvcl9zdGF0ZQogCXJldAorU1lNX0NPREVfRU5EKGRvX3N1c3BlbmRfbG93 bGV2ZWwpCiAKIC5kYXRhCiBBTElHTgpkaWZmIC0tZ2l0IGEvYXJjaC94ODYva2VybmVsL2Z0cmFj ZV8zMi5TIGIvYXJjaC94ODYva2VybmVsL2Z0cmFjZV8zMi5TCmluZGV4IDIxOWJlMTMuLmE0M2Vk NGMgMTAwNjQ0Ci0tLSBhL2FyY2gveDg2L2tlcm5lbC9mdHJhY2VfMzIuUworKysgYi9hcmNoL3g4 Ni9rZXJuZWwvZnRyYWNlXzMyLlMKQEAgLTg5LDcgKzg5LDcgQEAgV0VBSyhmdHJhY2Vfc3R1YikK IAlyZXQKIEVORChmdHJhY2VfY2FsbGVyKQogCi1FTlRSWShmdHJhY2VfcmVnc19jYWxsZXIpCitT WU1fQ09ERV9TVEFSVChmdHJhY2VfcmVnc19jYWxsZXIpCiAJLyoKIAkgKiBXZSdyZSBoZXJlIGZy b20gYW4gbWNvdW50L2ZlbnRyeSBDQUxMLCBhbmQgdGhlIHN0YWNrIGZyYW1lIGxvb2tzIGxpa2U6 CiAJICoKQEAgLTE2Myw2ICsxNjMsNyBAQCBTWU1fSU5ORVJfTEFCRUwoZnRyYWNlX3JlZ3NfY2Fs bCwgU1lNX0xfR0xPQkFMKQogCXBvcGwJJWVheAogCiAJam1wCS5MZnRyYWNlX3JldAorU1lNX0NP REVfRU5EKGZ0cmFjZV9yZWdzX2NhbGxlcikKIAogI2lmZGVmIENPTkZJR19GVU5DVElPTl9HUkFQ SF9UUkFDRVIKIEVOVFJZKGZ0cmFjZV9ncmFwaF9jYWxsZXIpCmRpZmYgLS1naXQgYS9hcmNoL3g4 Ni9rZXJuZWwvaGVhZF8zMi5TIGIvYXJjaC94ODYva2VybmVsL2hlYWRfMzIuUwppbmRleCBlMmIz ZTZjLi43MDI5YmJhIDEwMDY0NAotLS0gYS9hcmNoL3g4Ni9rZXJuZWwvaGVhZF8zMi5TCisrKyBi L2FyY2gveDg2L2tlcm5lbC9oZWFkXzMyLlMKQEAgLTY0LDcgKzY0LDcgQEAgUkVTRVJWRV9CUkso cGFnZXRhYmxlcywgSU5JVF9NQVBfU0laRSkKICAqIGNhbi4KICAqLwogX19IRUFECi1FTlRSWShz dGFydHVwXzMyKQorU1lNX0NPREVfU1RBUlQoc3RhcnR1cF8zMikKIAltb3ZsIHBhKGluaXRpYWxf c3RhY2spLCVlY3gKIAkKIAkvKiB0ZXN0IEtFRVBfU0VHTUVOVFMgZmxhZyB0byBzZWUgaWYgdGhl IGJvb3Rsb2FkZXIgaXMgYXNraW5nCkBAIC0xNzIsNiArMTcyLDcgQEAgbnVtX3N1YmFyY2hfZW50 cmllcyA9ICguIC0gc3ViYXJjaF9lbnRyaWVzKSAvIDQKICNlbHNlCiAJam1wIC5MZGVmYXVsdF9l bnRyeQogI2VuZGlmIC8qIENPTkZJR19QQVJBVklSVCAqLworU1lNX0NPREVfRU5EKHN0YXJ0dXBf MzIpCiAKICNpZmRlZiBDT05GSUdfSE9UUExVR19DUFUKIC8qCmRpZmYgLS1naXQgYS9hcmNoL3g4 Ni9wb3dlci9oaWJlcm5hdGVfYXNtXzMyLlMgYi9hcmNoL3g4Ni9wb3dlci9oaWJlcm5hdGVfYXNt XzMyLlMKaW5kZXggNmZlMzgzMC4uYTE5ZWQzZCAxMDA2NDQKLS0tIGEvYXJjaC94ODYvcG93ZXIv aGliZXJuYXRlX2FzbV8zMi5TCisrKyBiL2FyY2gveDg2L3Bvd2VyL2hpYmVybmF0ZV9hc21fMzIu UwpAQCAtMzUsNyArMzUsNyBAQCBFTlRSWShzd3N1c3BfYXJjaF9zdXNwZW5kKQogCXJldAogRU5E UFJPQyhzd3N1c3BfYXJjaF9zdXNwZW5kKQogCi1FTlRSWShyZXN0b3JlX2ltYWdlKQorU1lNX0NP REVfU1RBUlQocmVzdG9yZV9pbWFnZSkKIAkvKiBwcmVwYXJlIHRvIGp1bXAgdG8gdGhlIGltYWdl IGtlcm5lbCAqLwogCW1vdmwJcmVzdG9yZV9qdW1wX2FkZHJlc3MsICVlYngKIAltb3ZsCXJlc3Rv cmVfY3IzLCAlZWJwCkBAIC00NSw5ICs0NSwxMCBAQCBFTlRSWShyZXN0b3JlX2ltYWdlKQogCS8q IGp1bXAgdG8gcmVsb2NhdGVkIHJlc3RvcmUgY29kZSAqLwogCW1vdmwJcmVsb2NhdGVkX3Jlc3Rv cmVfY29kZSwgJWVheAogCWptcGwJKiVlYXgKK1NZTV9DT0RFX0VORChyZXN0b3JlX2ltYWdlKQog CiAvKiBjb2RlIGJlbG93IGhhcyBiZWVuIHJlbG9jYXRlZCB0byBhIHNhZmUgcGFnZSAqLwotRU5U UlkoY29yZV9yZXN0b3JlX2NvZGUpCitTWU1fQ09ERV9TVEFSVChjb3JlX3Jlc3RvcmVfY29kZSkK IAltb3ZsCXRlbXBfcGd0LCAlZWF4CiAJbW92bAklZWF4LCAlY3IzCiAKQEAgLTc3LDYgKzc4LDcg QEAgY29weV9sb29wOgogCiBkb25lOgogCWptcGwJKiVlYngKK1NZTV9DT0RFX0VORChjb3JlX3Jl c3RvcmVfY29kZSkKIAogCS8qIGNvZGUgYmVsb3cgYmVsb25ncyB0byB0aGUgaW1hZ2Uga2VybmVs ICovCiAJLmFsaWduIFBBR0VfU0laRQpkaWZmIC0tZ2l0IGEvYXJjaC94ODYvcmVhbG1vZGUvcm0v dHJhbXBvbGluZV8zMi5TIGIvYXJjaC94ODYvcmVhbG1vZGUvcm0vdHJhbXBvbGluZV8zMi5TCmlu ZGV4IGZmMDA1OTQuLjNmYWQ5MDcgMTAwNjQ0Ci0tLSBhL2FyY2gveDg2L3JlYWxtb2RlL3JtL3Ry YW1wb2xpbmVfMzIuUworKysgYi9hcmNoL3g4Ni9yZWFsbW9kZS9ybS90cmFtcG9saW5lXzMyLlMK QEAgLTI5LDcgKzI5LDcgQEAKIAkuY29kZTE2CiAKIAkuYmFsaWduCVBBR0VfU0laRQotRU5UUlko dHJhbXBvbGluZV9zdGFydCkKK1NZTV9DT0RFX1NUQVJUKHRyYW1wb2xpbmVfc3RhcnQpCiAJd2Jp bnZkCQkJIyBOZWVkZWQgZm9yIE5VTUEtUSBzaG91bGQgYmUgaGFybWxlc3MgZm9yIG90aGVycwog CiAJTEpNUFdfUk0oMWYpCkBAIC01NCwxMSArNTQsMTMgQEAgRU5UUlkodHJhbXBvbGluZV9zdGFy dCkKIAlsbXN3CSVkeAkJCSMgaW50byBwcm90ZWN0ZWQgbW9kZQogCiAJbGptcGwJJF9fQk9PVF9D UywgJHBhX3N0YXJ0dXBfMzIKK1NZTV9DT0RFX0VORCh0cmFtcG9saW5lX3N0YXJ0KQogCiAJLnNl Y3Rpb24gIi50ZXh0MzIiLCJheCIKIAkuY29kZTMyCi1FTlRSWShzdGFydHVwXzMyKQkJCSMgbm90 ZTogYWxzbyB1c2VkIGZyb20gd2FrZXVwX2FzbS5TCitTWU1fQ09ERV9TVEFSVChzdGFydHVwXzMy KQkJCSMgbm90ZTogYWxzbyB1c2VkIGZyb20gd2FrZXVwX2FzbS5TCiAJam1wCSolZWF4CitTWU1f Q09ERV9FTkQoc3RhcnR1cF8zMikKIAogCS5ic3MKIAkuYmFsaWduIDgKZGlmZiAtLWdpdCBhL2Fy Y2gveDg2L3hlbi94ZW4tYXNtXzMyLlMgYi9hcmNoL3g4Ni94ZW4veGVuLWFzbV8zMi5TCmluZGV4 IGMxNWRiMDYuLjhiOGY4MzUgMTAwNjQ0Ci0tLSBhL2FyY2gveDg2L3hlbi94ZW4tYXNtXzMyLlMK KysrIGIvYXJjaC94ODYveGVuL3hlbi1hc21fMzIuUwpAQCAtNTYsNyArNTYsNyBAQAogCV9BU01f RVhUQUJMRSgxYiwyYikKIC5lbmRtCiAKLUVOVFJZKHhlbl9pcmV0KQorU1lNX0NPREVfU1RBUlQo eGVuX2lyZXQpCiAJLyogdGVzdCBlZmxhZ3MgZm9yIHNwZWNpYWwgY2FzZXMgKi8KIAl0ZXN0bCAk KFg4Nl9FRkxBR1NfVk0gfCBYRU5fRUZMQUdTX05NSSksIDgoJWVzcCkKIAlqbnogaHlwZXJfaXJl dApAQCAtMTIyLDYgKzEyMiw3IEBAIHhlbl9pcmV0X2VuZF9jcml0OgogaHlwZXJfaXJldDoKIAkv KiBwdXQgdGhpcyBvdXQgb2YgbGluZSBzaW5jZSBpdHMgdmVyeSByYXJlbHkgdXNlZCAqLwogCWpt cCBoeXBlcmNhbGxfcGFnZSArIF9fSFlQRVJWSVNPUl9pcmV0ICogMzIKK1NZTV9DT0RFX0VORCh4 ZW5faXJldCkKIAogCS5nbG9ibCB4ZW5faXJldF9zdGFydF9jcml0LCB4ZW5faXJldF9lbmRfY3Jp dAogCkBAIC0xNjUsNyArMTY2LDcgQEAgaHlwZXJfaXJldDoKICAqIFNBVkVfQUxMIHN0YXRlIGJl Zm9yZSBnb2luZyBvbiwgc2luY2UgaXQncyB1c2VybW9kZSBzdGF0ZSB3aGljaCB3ZQogICogZXZl bnR1YWxseSBuZWVkIHRvIHJlc3RvcmUuCiAgKi8KLUVOVFJZKHhlbl9pcmV0X2NyaXRfZml4dXAp CitTWU1fQ09ERV9TVEFSVCh4ZW5faXJldF9jcml0X2ZpeHVwKQogCS8qCiAJICogUGFyYW5vaWE6 IE1ha2Ugc3VyZSB3ZSdyZSByZWFsbHkgY29taW5nIGZyb20ga2VybmVsIHNwYWNlLgogCSAqIE9u ZSBjb3VsZCBpbWFnaW5lIGEgY2FzZSB3aGVyZSB1c2Vyc3BhY2UganVtcHMgaW50byB0aGUKQEAg LTIwNCw0ICsyMDUsNCBAQCBFTlRSWSh4ZW5faXJldF9jcml0X2ZpeHVwKQogCiAJbGVhIDQoJWVk aSksICVlc3AJCS8qIHBvaW50IGVzcCB0byBuZXcgZnJhbWUgKi8KIDI6CWptcCB4ZW5fZG9fdXBj YWxsCi0KK1NZTV9DT0RFX0VORCh4ZW5faXJldF9jcml0X2ZpeHVwKQoKX19fX19fX19fX19fX19f X19fX19fX19fX19fX19fX19fX19fX19fX19fX19fX18KWGVuLWRldmVsIG1haWxpbmcgbGlzdApY ZW4tZGV2ZWxAbGlzdHMueGVucHJvamVjdC5vcmcKaHR0cHM6Ly9saXN0cy54ZW5wcm9qZWN0Lm9y Zy9tYWlsbWFuL2xpc3RpbmZvL3hlbi1kZXZlbA==