summary refs log tree commit diff
path: root/sysdeps/sparc/sparc64/dl-machine.h
diff options
context:
space:
mode:
Diffstat (limited to 'sysdeps/sparc/sparc64/dl-machine.h')
-rw-r--r--sysdeps/sparc/sparc64/dl-machine.h330
1 files changed, 270 insertions, 60 deletions
diff --git a/sysdeps/sparc/sparc64/dl-machine.h b/sysdeps/sparc/sparc64/dl-machine.h
index 788e239779..33def6b455 100644
--- a/sysdeps/sparc/sparc64/dl-machine.h
+++ b/sysdeps/sparc/sparc64/dl-machine.h
@@ -21,19 +21,15 @@
 
 #include <assert.h>
 #include <string.h>
-#include <link.h>
 #include <sys/param.h>
+#include <elf/ldsodefs.h>
 #include <sysdep.h>
 
-
-/* Translate a processor-specific dynamic tag to the index into l_info.  */
-#define DT_SPARC(x)	(DT_SPARC_##x - DT_LOPROC + DT_NUM)
-
 /* Return nonzero iff E_MACHINE is compatible with the running host.  */
 static inline int
 elf_machine_matches_host (Elf64_Half e_machine)
 {
-  return e_machine == EM_SPARC64;
+  return e_machine == EM_SPARCV9;
 }
 
 /* Return the link-time address of _DYNAMIC.  Conveniently, this is the
@@ -42,12 +38,11 @@ elf_machine_matches_host (Elf64_Half e_machine)
 static inline Elf64_Addr
 elf_machine_dynamic (void)
 {
-  register Elf64_Addr elf_pic_register __asm__("%l7");
+  register Elf64_Addr *elf_pic_register __asm__("%l7");
 
-  return *(Elf64_Addr *)elf_pic_register;
+  return *elf_pic_register;
 }
 
-
 /* Return the run-time load address of the shared object.  */
 static inline Elf64_Addr
 elf_machine_load_address (void)
@@ -67,21 +62,92 @@ elf_machine_load_address (void)
   return pc - *(Elf64_Addr *)(elf_pic_register + la);
 }
 
+/* We have 3 cases to handle.  And we code different code sequences
+   for each one.  I love V9 code models...  */
 static inline void
 elf_machine_fixup_plt(struct link_map *map, const Elf64_Rela *reloc,
                       Elf64_Addr *reloc_addr, Elf64_Addr value)
 {
-  Elf64_Dyn *pltfmt = map->l_info[DT_SPARC(PLTFMT)];
-  switch (pltfmt ? pltfmt->d_un.d_val : 0)
+  unsigned int *insns = (unsigned int *) reloc_addr;
+  Elf64_Addr plt_vaddr = (Elf64_Addr) reloc_addr;
+
+  /* Now move plt_vaddr up to the call instruction.  */
+  plt_vaddr += (2 * 4);
+
+  /* 32-bit Sparc style, the target is in the lower 32-bits of
+     address space.  */
+  if ((value >> 32) == 0)
     {
-    case 1: /* .got.plt with absolute addresses */
-      *reloc_addr = value;
-      break;
-    case 2: /* .got.plt with got-relative addresses */
-      *reloc_addr = value - (map->l_info[DT_PLTGOT]->d_un.d_ptr + map->l_addr);
-      break;
-    default:
-      assert (! "unexpected .plt format type");
+      /* sethi	%hi(target), %g1
+	 jmpl	%g1 + %lo(target), %g0  */
+
+      insns[2] = 0x81c06000 | (value & 0x3ff);
+      __asm __volatile ("flush %0 + 8" : : "r" (insns));
+
+      insns[1] = 0x03000000 | ((unsigned int)(value >> 10));
+      __asm __volatile ("flush %0 + 4" : : "r" (insns));
+    }
+  /* We can also get somewhat simple sequences if the distance between
+     the target and the PLT entry is within +/- 2GB.  */
+  else if ((plt_vaddr > value
+	    && ((plt_vaddr - value) >> 32) == 0)
+	   || (value > plt_vaddr
+	       && ((value - plt_vaddr) >> 32) == 0))
+    {
+      unsigned int displacement;
+
+      if (plt_vaddr > value)
+	displacement = (0 - (plt_vaddr - value));
+      else
+	displacement = value - plt_vaddr;
+
+      /* mov	%o7, %g1
+	 call	displacement
+	  mov	%g1, %o7  */
+
+      insns[3] = 0x9e100001;
+      __asm __volatile ("flush %0 + 12" : : "r" (insns));
+
+      insns[2] = 0x40000000 | (displacement >> 2);
+      __asm __volatile ("flush %0 + 8" : : "r" (insns));
+
+      insns[1] = 0x8210000f;
+      __asm __volatile ("flush %0 + 4" : : "r" (insns));
+    }
+  /* Worst case, ho hum...  */
+  else
+    {
+      unsigned int high32 = (value >> 32);
+      unsigned int low32 = (unsigned int) value;
+
+      /* ??? Some tricks can be stolen from the sparc64 egcs backend
+	     constant formation code I wrote.  -DaveM  */
+
+      /* sethi	%hh(value), %g1
+	 sethi	%lm(value), %g2
+	 or	%g1, %hl(value), %g1
+	 or	%g2, %lo(value), %g2
+	 sllx	%g1, 32, %g1
+	 jmpl	%g1 + %g2, %g0
+	  nop  */
+
+      insns[6] = 0x81c04002;
+      __asm __volatile ("flush %0 + 24" : : "r" (insns));
+
+      insns[5] = 0x83287020;
+      __asm __volatile ("flush %0 + 20" : : "r" (insns));
+
+      insns[4] = 0x8410a000 | (low32 & 0x3ff);
+      __asm __volatile ("flush %0 + 16" : : "r" (insns));
+
+      insns[3] = 0x82106000 | (high32 & 0x3ff);
+      __asm __volatile ("flush %0 + 12" : : "r" (insns));
+
+      insns[2] = 0x05000000 | (low32 >> 10);
+      __asm __volatile ("flush %0 + 8" : : "r" (insns));
+
+      insns[1] = 0x03000000 | (high32 >> 10);
+      __asm __volatile ("flush %0 + 4" : : "r" (insns));
     }
 }
 
@@ -119,7 +185,7 @@ elf_machine_rela (struct link_map *map, const Elf64_Rela *reloc,
 #endif
 	*reloc_addr = map->l_addr + reloc->r_addend;
     }
-  else
+  else if (ELF64_R_TYPE (reloc->r_info) != R_SPARC_NONE) /* Who is Wilbur? */
     {
       const Elf64_Sym *const refsym = sym;
       Elf64_Addr value;
@@ -137,6 +203,10 @@ elf_machine_rela (struct link_map *map, const Elf64_Rela *reloc,
       switch (ELF64_R_TYPE (reloc->r_info))
 	{
 	case R_SPARC_COPY:
+	  if (sym == NULL)
+	    /* This can happen in trace mode if an object could not be
+	       found.  */
+	    break;
 	  if (sym->st_size > refsym->st_size
 	      || (_dl_verbose && sym->st_size < refsym->st_size))
 	    {
@@ -164,6 +234,9 @@ elf_machine_rela (struct link_map *map, const Elf64_Rela *reloc,
 	case R_SPARC_16:
 	  *(short *) reloc_addr = value;
 	  break;
+	case R_SPARC_32:
+	  *(unsigned int *) reloc_addr = value;
+	  break;
 	case R_SPARC_DISP8:
 	  *(char *) reloc_addr = (value - (Elf64_Addr) reloc_addr);
 	  break;
@@ -171,27 +244,64 @@ elf_machine_rela (struct link_map *map, const Elf64_Rela *reloc,
 	  *(short *) reloc_addr = (value - (Elf64_Addr) reloc_addr);
 	  break;
 	case R_SPARC_DISP32:
-	  *(unsigned int *)reloc_addr = (value - (Elf64_Addr) reloc_addr);
-	  break;
-	case R_SPARC_LO10:
-	  *(unsigned *)reloc_addr = (*(unsigned *)reloc_addr & ~0x3ff)
-				     | (value & 0x3ff);
+	  *(unsigned int *) reloc_addr = (value - (Elf64_Addr) reloc_addr);
 	  break;
 	case R_SPARC_WDISP30:
-	  *(unsigned *)reloc_addr = ((*(unsigned *)reloc_addr & 0xc0000000)
-			 | ((value - (Elf64_Addr) reloc_addr) >> 2));
+	  *(unsigned int *) reloc_addr =
+	    ((*(unsigned int *)reloc_addr & 0xc0000000) |
+	     ((value - (Elf64_Addr) reloc_addr) >> 2));
+	  break;
+
+	/* MEDLOW code model relocs */
+	case R_SPARC_LO10:
+	  *(unsigned int *) reloc_addr =
+	    ((*(unsigned int *)reloc_addr & ~0x3ff) |
+	     (value & 0x3ff));
 	  break;
 	case R_SPARC_HI22:
-	  *(unsigned *)reloc_addr = (*(unsigned *)reloc_addr & 0xffc00000)
-				     | (value >> 10);
+	  *(unsigned int *) reloc_addr =
+	    ((*(unsigned int *)reloc_addr & 0xffc00000) |
+	     (value >> 10));
+	  break;
+
+	/* MEDMID code model relocs */
+	case R_SPARC_H44:
+	  *(unsigned int *) reloc_addr =
+	    ((*(unsigned int *)reloc_addr & 0xffc00000) |
+	     (value >> 22));
+	  break;
+	case R_SPARC_M44:
+	  *(unsigned int *) reloc_addr =
+	    ((*(unsigned int *)reloc_addr & ~0x3ff) |
+	     ((value >> 12) & 0x3ff));
+	  break;
+	case R_SPARC_L44:
+	  *(unsigned int *) reloc_addr =
+	    ((*(unsigned int *)reloc_addr & ~0xfff) |
+	     (value & 0xfff));
+	  break;
+
+	/* MEDANY code model relocs */
+	case R_SPARC_HH22:
+	  *(unsigned int *) reloc_addr =
+	    ((*(unsigned int *)reloc_addr & 0xffc00000) |
+	     (value >> 42));
+	  break;
+	case R_SPARC_HM10:
+	  *(unsigned int *) reloc_addr =
+	    ((*(unsigned int *)reloc_addr & ~0x3ff) |
+	     ((value >> 32) & 0x3ff));
+	  break;
+	case R_SPARC_LM22:
+	  *(unsigned int *) reloc_addr =
+	    ((*(unsigned int *)reloc_addr & 0xffc00000) |
+	     ((value >> 10) & 0x003fffff));
 	  break;
 
 	case R_SPARC_JMP_SLOT:
 	  elf_machine_fixup_plt(map, reloc, reloc_addr, value);
 	  break;
 
-	case R_SPARC_NONE:		/* Alright, Wilbur.  */
-	  break;
 	default:
 	  assert (! "unexpected dynamic reloc type");
 	  break;
@@ -239,16 +349,63 @@ elf_machine_lazy_rel (Elf64_Addr l_addr, const Elf64_Rela *reloc)
 static inline int
 elf_machine_runtime_setup (struct link_map *l, int lazy, int profile)
 {
-  Elf64_Addr *got;
-  extern void _dl_runtime_resolve (void);
-
   if (l->l_info[DT_JMPREL] && lazy)
     {
-      got = (Elf64_Addr *) (l->l_addr + l->l_info[DT_PLTGOT]->d_un.d_ptr);
-      /* This function will get called to fix up the GOT entry indicated by
-         the offset on the stack, and then jump to the resolved address.  */
-      got[1] = (Elf64_Addr) &_dl_runtime_resolve;
-      got[2] = (Elf64_Addr) l;  /* Identify this shared object.  */
+      extern void _dl_runtime_resolve_0 (void);
+      extern void _dl_runtime_resolve_1 (void);
+      extern void _dl_runtime_profile_0 (void);
+      extern void _dl_runtime_profile_1 (void);
+      Elf64_Addr res0_addr, res1_addr;
+      unsigned int *plt = (unsigned int *)
+	(l->l_addr + l->l_info[DT_PLTGOT]->d_un.d_ptr);
+
+      if (! profile)
+	{
+	  res0_addr = (Elf64_Addr) &_dl_runtime_resolve_0;
+	  res1_addr = (Elf64_Addr) &_dl_runtime_resolve_1;
+	}
+      else
+	{
+	  res0_addr = (Elf64_Addr) &_dl_runtime_profile_0;
+	  res1_addr = (Elf64_Addr) &_dl_runtime_profile_1;
+	  if (_dl_name_match_p (_dl_profile, l))
+	    _dl_profile_map = l;
+	}
+
+      /* PLT0 looks like:
+
+	 save	%sp, -192, %sp
+	 sethi	%hh(_dl_runtime_{resolve,profile}_0), %g3
+	 sethi	%lm(_dl_runtime_{resolve,profile}_0), %g4
+	 or	%g3, %hm(_dl_runtime_{resolve,profile}_0), %g3
+	 or	%g4, %lo(_dl_runtime_{resolve,profile}_0), %g4
+	 sllx	%g3, 32, %g3
+	 jmpl	%g3 + %g4, %o0
+	  nop
+
+	 PLT1 is similar except we jump to _dl_runtime_{resolve,profile}_1.  */
+
+      plt[0] = 0x9de3bf40;
+      plt[1] = 0x07000000 | (res0_addr >> (64 - 22));
+      plt[2] = 0x09000000 | ((res0_addr >> 10) & 0x003fffff);
+      plt[3] = 0x8610e000 | ((res0_addr >> 32) & 0x3ff);
+      plt[4] = 0x88112000 | (res0_addr & 0x3ff);
+      plt[5] = 0x8728f020;
+      plt[6] = 0x91c0c004;
+      plt[7] = 0x01000000;
+
+      plt[8 + 0] = 0x9de3bf40;
+      plt[8 + 1] = 0x07000000 | (res1_addr >> (64 - 22));
+      plt[8 + 2] = 0x09000000 | ((res1_addr >> 10) & 0x003fffff);
+      plt[8 + 3] = 0x8610e000 | ((res1_addr >> 32) & 0x3ff);
+      plt[8 + 4] = 0x88112000 | (res1_addr & 0x3ff);
+      plt[8 + 5] = 0x8728f020;
+      plt[8 + 6] = 0x91c0c004;
+      plt[8 + 7] = 0x01000000;
+
+      /* Now put the magic cookie at the beginning of .PLT3
+	 Entry .PLT4 is unused by this implementation.  */
+      *((struct link_map **)(&plt[16 + 0])) = l;
     }
 
   return lazy;
@@ -256,23 +413,68 @@ elf_machine_runtime_setup (struct link_map *l, int lazy, int profile)
 
 /* This code is used in dl-runtime.c to call the `fixup' function
    and then redirect to the address it returns.  */
-#define ELF_MACHINE_RUNTIME_TRAMPOLINE asm ("\
-	.globl _dl_runtime_resolve
-	.type _dl_runtime_resolve, @function
-_dl_runtime_resolve:
-	save %sp, -160, %sp
-	mov %g1, %o0
-	call fixup
-	 mov %g2, %o1
-	jmp %o0
+#define TRAMPOLINE_TEMPLATE(tramp_name, fixup_name)	\
+  asm ("\
+	.text
+	.globl	" #tramp_name "_0
+	.type	" #tramp_name "_0, @function
+	.align	32
+" #tramp_name "_0:
+	ldx	[%o0 + 32 + 8], %l0
+	sethi	%hi(1048576), %g2
+	sub	%g1, %o0, %o0
+	xor	%g2, -20, %g2
+	sethi	%hi(5120), %g3
+	add	%o0, %g2, %o0
+	sethi	%hi(32768), %o2
+	udivx	%o0, %g3, %g3
+	sllx	%g3, 2, %g1
+	add	%g1, %g3, %g1
+	sllx	%g1, 10, %g2
+	sllx	%g1, 5, %g1
+	sub	%o0, %g2, %o0
+	udivx	%o0, 24, %o0
+	add	%o0, %o2, %o0
+	add	%g1, %o0, %g1
+	sllx	%g1, 1, %o1
+	mov	%l0, %o0
+	add	%o1, %g1, %o1
+	mov	%i7, %o2
+	call	" #fixup_name "
+	 sllx	%o1, 3, %o1
+	jmp	%o0
+	 restore
+	.size	" #tramp_name "_0, . - " #tramp_name "_0
+
+	.globl	" #tramp_name "_1
+	.type	" #tramp_name "_1, @function
+	.align	32
+" #tramp_name "_1:
+	srlx	%g1, 15, %o1
+	ldx	[%o0 + 8], %o0
+	sllx	%o1, 1, %o3
+	add	%o1, %o3, %o1
+	mov	%i7, %o2
+	call	" #fixup_name "
+	 sllx	%o1, 3, %o1
+	jmp	%o0
 	 restore
-	.size _dl_runtime_resolve, .-_dl_runtime_resolve
-");
+	.size	" #tramp_name "_1, . - " #tramp_name "_1
+	.previous");
+
+#ifndef PROF
+#define ELF_MACHINE_RUNTIME_TRAMPOLINE 			\
+  TRAMPOLINE_TEMPLATE (_dl_runtime_resolve, fixup);	\
+  TRAMPOLINE_TEMPLATE (_dl_runtime_profile, profile_fixup);
+#else
+#define ELF_MACHINE_RUNTIME_TRAMPOLINE			\
+  TRAMPOLINE_TEMPLATE (_dl_runtime_resolve, fixup);	\
+  TRAMPOLINE_TEMPLATE (_dl_runtime_profile, fixup);
+#endif
 
 /* The PLT uses Elf64_Rela relocs.  */
 #define elf_machine_relplt elf_machine_rela
 
-
 /* Initial entry point code for the dynamic linker.
    The C function `_dl_start' is the real entry point;
    its return value is the user program's entry point.  */
@@ -281,8 +483,10 @@ _dl_runtime_resolve:
 #define __S(x)	__S1(x)
 
 #define RTLD_START __asm__ ( "\
-	.global _start
-	.type _start, @function
+	.text
+	.global	_start
+	.type	_start, @function
+	.align	32
 _start:
    /* Make room for functions to drop their arguments on the stack.  */
 	sub	%sp, 6*8, %sp
@@ -292,8 +496,8 @@ _start:
 	/* FALLTHRU */
 	.size _start, .-_start
 
-	.global _dl_start_user
-	.type _dl_start_user, @function
+	.global	_dl_start_user
+	.type	_dl_start_user, @function
 _dl_start_user:
    /* Load the GOT register.  */
 1:	call	11f
@@ -302,6 +506,12 @@ _dl_start_user:
 	add	%l7,%o7,%l7
    /* Save the user entry point address in %l0.  */
 	mov	%o0,%l0
+  /* Store the highest stack address.  */
+	sethi	%hi(__libc_stack_end), %g2
+	or	%g2, %lo(__libc_stack_end), %g2
+	ldx	[%l7 + %g2], %l1
+	add	%sp, 6*8, %l2
+	stx	%l2, [%l1]
    /* See if we were run as a command with the executable file name as an
       extra leading argument.  If so, we must shift things around since we
       must keep the stack doubleword aligned.  */
@@ -338,7 +548,7 @@ _dl_start_user:
 	stx	%i4, [%i1+8]
 	brnz,pt	%i3, 13b
 	 add	%i1, 16, %i1
-   /* Load _dl_main_searchlist to pass to _dl_init_next.  */
+  /* Load searchlist of the main object to pass to _dl_init_next.  */
 2:	sethi	%hi(_dl_main_searchlist), %g2
 	or	%g2, %lo(_dl_main_searchlist), %g2
 	ldx	[%l7+%g2], %g2
@@ -349,8 +559,7 @@ _dl_start_user:
 	brz,pn	%o0, 4f
 	 nop
 	jmpl	%o0, %o7
-	 nop
-	ba,a	3b
+	 sub	%o7, 24, %o7
    /* Clear the startup flag.  */
 4:	sethi	%hi(_dl_starting_up), %g2
 	or	%g2, %lo(_dl_starting_up), %g2
@@ -360,7 +569,8 @@ _dl_start_user:
 	sethi	%hi(_dl_fini), %g1
 	or	%g1, %lo(_dl_fini), %g1
 	ldx	[%l7+%g1], %g1
-   /* Jump to the user's entry point & undo the allocation of the xtra regs.  */
+  /* Jump to the user's entry point and deallocate the extra stack we got.  */
 	jmp	%l0
 	 add	%sp, 6*8, %sp
-	.size _dl_start_user, .-_dl_start_user");
+	.size	_dl_start_user, . - _dl_start_user
+	.previous");