From 6c7fb1458ddf7943dceac20440226d3d968e518d Mon Sep 17 00:00:00 2001 From: Roland McGrath Date: Fri, 18 May 2012 15:32:08 -0700 Subject: x32: Don't lose high bits of %rbp in setjmp/longjmp mangling/demangling. --- sysdeps/x86_64/__longjmp.S | 14 +++++++++++--- 1 file changed, 11 insertions(+), 3 deletions(-) (limited to 'sysdeps/x86_64/__longjmp.S') diff --git a/sysdeps/x86_64/__longjmp.S b/sysdeps/x86_64/__longjmp.S index 22beb888a1..c0c15dc810 100644 --- a/sysdeps/x86_64/__longjmp.S +++ b/sysdeps/x86_64/__longjmp.S @@ -1,4 +1,4 @@ -/* Copyright (C) 2001,2004,2005,2006,2009 Free Software Foundation, Inc. +/* Copyright (C) 2001-2012 Free Software Foundation, Inc. This file is part of the GNU C Library. The GNU C Library is free software; you can redistribute it and/or @@ -26,12 +26,20 @@ ENTRY(__longjmp) /* Restore registers. */ mov (JB_RSP*8)(%rdi),%R8_LP - movq (JB_RBP*8)(%rdi),%r9 + mov (JB_RBP*8)(%rdi),%R9_LP mov (JB_PC*8)(%rdi),%RDX_LP #ifdef PTR_DEMANGLE PTR_DEMANGLE (%R8_LP) - PTR_DEMANGLE (%r9) + PTR_DEMANGLE (%R9_LP) PTR_DEMANGLE (%RDX_LP) +# ifdef __ILP32__ + /* We ignored the high bits of the %rbp value because only the low + bits are mangled. But we cannot presume that %rbp is being used + as a pointer and truncate it, so recover the high bits. */ + movl (JB_RBP*8 + 4)(%rdi), %eax + shlq $32, %rax + orq %rax, %r9 +# endif #endif /* We add unwind information for the target here. */ cfi_def_cfa(%rdi, 0) -- cgit 1.4.1